basedpyright 1.36.2 → 1.37.0-961d59bacff1958014fe4e9ed2ef3116fe2266e2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (732) hide show
  1. package/dist/pyright-langserver.js +1 -1
  2. package/dist/pyright-langserver.js.map +1 -1
  3. package/dist/pyright.js +1 -1
  4. package/dist/pyright.js.map +1 -1
  5. package/dist/typeshed-fallback/README.md +6 -6
  6. package/dist/typeshed-fallback/commit.txt +1 -1
  7. package/dist/typeshed-fallback/stdlib/_compression.pyi +13 -2
  8. package/dist/typeshed-fallback/stdlib/_ctypes.pyi +49 -16
  9. package/dist/typeshed-fallback/stdlib/_winapi.pyi +9 -0
  10. package/dist/typeshed-fallback/stdlib/_zstd.pyi +7 -2
  11. package/dist/typeshed-fallback/stdlib/argparse.pyi +39 -10
  12. package/dist/typeshed-fallback/stdlib/asyncio/coroutines.pyi +27 -8
  13. package/dist/typeshed-fallback/stdlib/asyncio/protocols.pyi +2 -2
  14. package/dist/typeshed-fallback/stdlib/asyncio/runners.pyi +5 -2
  15. package/dist/typeshed-fallback/stdlib/asyncio/trsock.pyi +1 -1
  16. package/dist/typeshed-fallback/stdlib/builtins.pyi +54 -31
  17. package/dist/typeshed-fallback/stdlib/calendar.pyi +3 -1
  18. package/dist/typeshed-fallback/stdlib/codecs.pyi +2 -1
  19. package/dist/typeshed-fallback/stdlib/compression/_common/_streams.pyi +13 -2
  20. package/dist/typeshed-fallback/stdlib/compression/zstd/__init__.pyi +7 -2
  21. package/dist/typeshed-fallback/stdlib/compression/zstd/_zstdfile.pyi +6 -6
  22. package/dist/typeshed-fallback/stdlib/contextlib.pyi +15 -6
  23. package/dist/typeshed-fallback/stdlib/ctypes/__init__.pyi +3 -0
  24. package/dist/typeshed-fallback/stdlib/enum.pyi +2 -0
  25. package/dist/typeshed-fallback/stdlib/heapq.pyi +5 -0
  26. package/dist/typeshed-fallback/stdlib/html/parser.pyi +2 -1
  27. package/dist/typeshed-fallback/stdlib/http/client.pyi +1 -0
  28. package/dist/typeshed-fallback/stdlib/imaplib.pyi +2 -1
  29. package/dist/typeshed-fallback/stdlib/importlib/readers.pyi +2 -2
  30. package/dist/typeshed-fallback/stdlib/importlib/resources/abc.pyi +3 -2
  31. package/dist/typeshed-fallback/stdlib/importlib/resources/simple.pyi +2 -1
  32. package/dist/typeshed-fallback/stdlib/importlib/util.pyi +15 -1
  33. package/dist/typeshed-fallback/stdlib/ipaddress.pyi +1 -1
  34. package/dist/typeshed-fallback/stdlib/locale.pyi +4 -0
  35. package/dist/typeshed-fallback/stdlib/mimetypes.pyi +6 -6
  36. package/dist/typeshed-fallback/stdlib/mmap.pyi +12 -12
  37. package/dist/typeshed-fallback/stdlib/multiprocessing/dummy/__init__.pyi +20 -8
  38. package/dist/typeshed-fallback/stdlib/multiprocessing/managers.pyi +4 -1
  39. package/dist/typeshed-fallback/stdlib/multiprocessing/process.pyi +4 -0
  40. package/dist/typeshed-fallback/stdlib/multiprocessing/synchronize.pyi +3 -0
  41. package/dist/typeshed-fallback/stdlib/optparse.pyi +1 -1
  42. package/dist/typeshed-fallback/stdlib/os/__init__.pyi +217 -127
  43. package/dist/typeshed-fallback/stdlib/parser.pyi +4 -4
  44. package/dist/typeshed-fallback/stdlib/pathlib/__init__.pyi +11 -0
  45. package/dist/typeshed-fallback/stdlib/pdb.pyi +14 -4
  46. package/dist/typeshed-fallback/stdlib/pyexpat/__init__.pyi +6 -0
  47. package/dist/typeshed-fallback/stdlib/select.pyi +8 -4
  48. package/dist/typeshed-fallback/stdlib/socket.pyi +15 -0
  49. package/dist/typeshed-fallback/stdlib/sqlite3/__init__.pyi +3 -2
  50. package/dist/typeshed-fallback/stdlib/ssl.pyi +3 -0
  51. package/dist/typeshed-fallback/stdlib/stat.pyi +108 -1
  52. package/dist/typeshed-fallback/stdlib/subprocess.pyi +60 -61
  53. package/dist/typeshed-fallback/stdlib/sys/__init__.pyi +41 -4
  54. package/dist/typeshed-fallback/stdlib/sys/_monitoring.pyi +8 -4
  55. package/dist/typeshed-fallback/stdlib/sysconfig.pyi +13 -4
  56. package/dist/typeshed-fallback/stdlib/tarfile.pyi +6 -6
  57. package/dist/typeshed-fallback/stdlib/threading.pyi +9 -4
  58. package/dist/typeshed-fallback/stdlib/tkinter/constants.pyi +6 -6
  59. package/dist/typeshed-fallback/stdlib/tkinter/ttk.pyi +14 -6
  60. package/dist/typeshed-fallback/stdlib/types.pyi +8 -11
  61. package/dist/typeshed-fallback/stdlib/typing.pyi +18 -12
  62. package/dist/typeshed-fallback/stdlib/typing_extensions.pyi +1 -0
  63. package/dist/typeshed-fallback/stdlib/unittest/mock.pyi +1 -0
  64. package/dist/typeshed-fallback/stdlib/unittest/util.pyi +19 -2
  65. package/dist/typeshed-fallback/stdlib/urllib/request.pyi +7 -0
  66. package/dist/typeshed-fallback/stdlib/uuid.pyi +6 -5
  67. package/dist/typeshed-fallback/stdlib/winreg.pyi +3 -3
  68. package/dist/typeshed-fallback/stdlib/xml/etree/ElementTree.pyi +1 -0
  69. package/dist/typeshed-fallback/stdlib/zipfile/__init__.pyi +2 -0
  70. package/dist/typeshed-fallback/stdlib/zlib.pyi +2 -2
  71. package/dist/typeshed-fallback/stubs/Authlib/METADATA.toml +1 -2
  72. package/dist/typeshed-fallback/stubs/Authlib/authlib/common/urls.pyi +1 -1
  73. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/base_client/async_app.pyi +4 -3
  74. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/base_client/async_openid.pyi +4 -2
  75. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/base_client/framework_integration.pyi +2 -2
  76. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/base_client/registry.pyi +1 -1
  77. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/base_client/sync_app.pyi +3 -3
  78. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/base_client/sync_openid.pyi +4 -2
  79. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/django_client/__init__.pyi +10 -0
  80. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/django_client/apps.pyi +14 -0
  81. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/django_client/integration.pyi +11 -0
  82. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/django_oauth1/__init__.pyi +4 -0
  83. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/django_oauth1/authorization_server.pyi +26 -0
  84. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/django_oauth1/nonce.pyi +1 -0
  85. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/django_oauth1/resource_protector.pyi +14 -0
  86. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/django_oauth2/__init__.pyi +8 -0
  87. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/django_oauth2/authorization_server.pyi +24 -0
  88. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/django_oauth2/endpoints.pyi +5 -0
  89. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/django_oauth2/requests.pyi +20 -0
  90. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/django_oauth2/resource_protector.pyi +15 -0
  91. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/django_oauth2/signals.pyi +6 -0
  92. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/flask_client/__init__.pyi +20 -0
  93. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/flask_client/apps.pyi +18 -0
  94. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/flask_client/integration.pyi +10 -0
  95. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/flask_oauth1/__init__.pyi +7 -0
  96. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/flask_oauth1/authorization_server.pyi +29 -0
  97. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/flask_oauth1/cache.pyi +8 -0
  98. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/flask_oauth1/resource_protector.pyi +18 -0
  99. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/flask_oauth2/__init__.pyi +7 -0
  100. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/flask_oauth2/authorization_server.pyi +23 -0
  101. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/flask_oauth2/errors.pyi +14 -0
  102. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/flask_oauth2/requests.pyi +27 -0
  103. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/flask_oauth2/resource_protector.pyi +15 -0
  104. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/flask_oauth2/signals.pyi +5 -0
  105. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/httpx_client/__init__.pyi +37 -0
  106. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/httpx_client/assertion_client.pyi +50 -0
  107. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/httpx_client/oauth1_client.pyi +56 -0
  108. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/httpx_client/oauth2_client.pyi +72 -0
  109. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/httpx_client/utils.pyi +7 -0
  110. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/requests_client/__init__.pyi +28 -0
  111. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/requests_client/assertion_session.pyi +31 -0
  112. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/requests_client/oauth1_session.pyi +29 -0
  113. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/requests_client/oauth2_session.pyi +43 -0
  114. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/requests_client/utils.pyi +6 -0
  115. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/sqla_oauth2/__init__.pyi +20 -0
  116. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/sqla_oauth2/client_mixin.pyi +55 -0
  117. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/sqla_oauth2/functions.pyi +21 -0
  118. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/sqla_oauth2/tokens_mixins.pyi +39 -0
  119. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/starlette_client/__init__.pyi +14 -0
  120. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/starlette_client/apps.pyi +16 -0
  121. package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/starlette_client/integration.pyi +13 -0
  122. package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/drafts/_jwe_algorithms.pyi +12 -10
  123. package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/drafts/_jwe_enc_cryptography.pyi +2 -2
  124. package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/jwk.pyi +6 -1
  125. package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7515/jws.pyi +6 -4
  126. package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7515/models.pyi +3 -2
  127. package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7516/jwe.pyi +9 -8
  128. package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7516/models.pyi +12 -11
  129. package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7517/_cryptography_key.pyi +35 -1
  130. package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7517/asymmetric_key.pyi +6 -5
  131. package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7517/base_key.pyi +8 -8
  132. package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7517/jwk.pyi +1 -1
  133. package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7517/key_set.pyi +3 -2
  134. package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7518/jwe_algs.pyi +28 -23
  135. package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7518/jwe_encs.pyi +6 -5
  136. package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7518/jws_algs.pyi +1 -1
  137. package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7518/oct_key.pyi +5 -4
  138. package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7518/rsa_key.pyi +2 -1
  139. package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7518/util.pyi +4 -2
  140. package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7519/jwt.pyi +2 -2
  141. package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc8037/jws_eddsa.pyi +4 -2
  142. package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc8037/okp_key.pyi +2 -2
  143. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth1/client.pyi +2 -2
  144. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth1/rfc5849/authorization_server.pyi +7 -7
  145. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth1/rfc5849/base_server.pyi +2 -2
  146. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth1/rfc5849/client_auth.pyi +6 -6
  147. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth1/rfc5849/errors.pyi +1 -1
  148. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth1/rfc5849/models.pyi +8 -8
  149. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth1/rfc5849/parameters.pyi +1 -1
  150. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth1/rfc5849/resource_protector.pyi +4 -2
  151. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth1/rfc5849/rsa.pyi +1 -1
  152. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth1/rfc5849/signature.pyi +20 -18
  153. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth1/rfc5849/util.pyi +2 -2
  154. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/base.pyi +1 -1
  155. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/client.pyi +1 -1
  156. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc6749/authorization_server.pyi +1 -1
  157. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc6749/grants/authorization_code.pyi +2 -2
  158. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc6749/grants/base.pyi +1 -1
  159. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc6749/grants/resource_owner_password_credentials.pyi +1 -1
  160. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc6749/parameters.pyi +5 -3
  161. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc6749/resource_protector.pyi +1 -1
  162. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc6749/token_endpoint.pyi +2 -2
  163. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc6749/wrappers.pyi +1 -1
  164. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc6750/errors.pyi +1 -1
  165. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc6750/parameters.pyi +1 -1
  166. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc6750/validator.pyi +1 -1
  167. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc7009/revocation.pyi +2 -2
  168. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc7523/assertion.pyi +3 -3
  169. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc7523/auth.pyi +2 -2
  170. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc7523/client.pyi +5 -3
  171. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc7523/jwt_bearer.pyi +6 -5
  172. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc7591/endpoint.pyi +4 -4
  173. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc7592/endpoint.pyi +8 -8
  174. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc7662/introspection.pyi +3 -3
  175. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc7662/token_validator.pyi +1 -1
  176. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc8414/well_known.pyi +1 -1
  177. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc8628/device_code.pyi +3 -3
  178. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc8628/endpoint.pyi +2 -2
  179. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc8628/models.pyi +4 -4
  180. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc9068/introspection.pyi +1 -1
  181. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc9068/revocation.pyi +3 -2
  182. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc9068/token.pyi +1 -1
  183. package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc9068/token_validator.pyi +1 -1
  184. package/dist/typeshed-fallback/stubs/Authlib/authlib/oidc/core/claims.pyi +6 -8
  185. package/dist/typeshed-fallback/stubs/Authlib/authlib/oidc/core/grants/code.pyi +2 -1
  186. package/dist/typeshed-fallback/stubs/Authlib/authlib/oidc/core/grants/hybrid.pyi +4 -4
  187. package/dist/typeshed-fallback/stubs/Authlib/authlib/oidc/core/grants/implicit.pyi +6 -4
  188. package/dist/typeshed-fallback/stubs/Authlib/authlib/oidc/core/userinfo.pyi +1 -1
  189. package/dist/typeshed-fallback/stubs/Authlib/authlib/oidc/core/util.pyi +7 -1
  190. package/dist/typeshed-fallback/stubs/Authlib/authlib/oidc/discovery/models.pyi +1 -3
  191. package/dist/typeshed-fallback/stubs/Authlib/authlib/oidc/discovery/well_known.pyi +1 -1
  192. package/dist/typeshed-fallback/stubs/Deprecated/METADATA.toml +1 -1
  193. package/dist/typeshed-fallback/stubs/Deprecated/deprecated/__init__.pyi +7 -2
  194. package/dist/typeshed-fallback/stubs/Deprecated/deprecated/params.pyi +20 -0
  195. package/dist/typeshed-fallback/stubs/Flask-SocketIO/METADATA.toml +1 -1
  196. package/dist/typeshed-fallback/stubs/Markdown/METADATA.toml +1 -1
  197. package/dist/typeshed-fallback/stubs/Markdown/markdown/htmlparser.pyi +2 -0
  198. package/dist/typeshed-fallback/stubs/PyMySQL/pymysql/cursors.pyi +1 -0
  199. package/dist/typeshed-fallback/stubs/Pygments/pygments/console.pyi +4 -4
  200. package/dist/typeshed-fallback/stubs/Pygments/pygments/filter.pyi +3 -3
  201. package/dist/typeshed-fallback/stubs/Pygments/pygments/filters/__init__.pyi +14 -14
  202. package/dist/typeshed-fallback/stubs/Pygments/pygments/formatter.pyi +10 -9
  203. package/dist/typeshed-fallback/stubs/Pygments/pygments/formatters/__init__.pyi +2 -2
  204. package/dist/typeshed-fallback/stubs/Pygments/pygments/formatters/_mapping.pyi +2 -2
  205. package/dist/typeshed-fallback/stubs/Pygments/pygments/formatters/bbcode.pyi +5 -4
  206. package/dist/typeshed-fallback/stubs/Pygments/pygments/formatters/html.pyi +27 -26
  207. package/dist/typeshed-fallback/stubs/Pygments/pygments/formatters/img.pyi +33 -32
  208. package/dist/typeshed-fallback/stubs/Pygments/pygments/formatters/irc.pyi +7 -6
  209. package/dist/typeshed-fallback/stubs/Pygments/pygments/formatters/latex.pyi +21 -20
  210. package/dist/typeshed-fallback/stubs/Pygments/pygments/formatters/other.pyi +9 -8
  211. package/dist/typeshed-fallback/stubs/Pygments/pygments/formatters/pangomarkup.pyi +5 -4
  212. package/dist/typeshed-fallback/stubs/Pygments/pygments/formatters/rtf.pyi +6 -5
  213. package/dist/typeshed-fallback/stubs/Pygments/pygments/formatters/svg.pyi +15 -14
  214. package/dist/typeshed-fallback/stubs/Pygments/pygments/formatters/terminal.pyi +7 -6
  215. package/dist/typeshed-fallback/stubs/Pygments/pygments/formatters/terminal256.pyi +18 -17
  216. package/dist/typeshed-fallback/stubs/Pygments/pygments/lexers/__init__.pyi +6 -7
  217. package/dist/typeshed-fallback/stubs/Pygments/pygments/plugin.pyi +5 -5
  218. package/dist/typeshed-fallback/stubs/Pygments/pygments/regexopt.pyi +3 -3
  219. package/dist/typeshed-fallback/stubs/Pygments/pygments/scanner.pyi +6 -6
  220. package/dist/typeshed-fallback/stubs/Pygments/pygments/unistring.pyi +2 -2
  221. package/dist/typeshed-fallback/stubs/Pygments/pygments/util.pyi +5 -5
  222. package/dist/typeshed-fallback/stubs/antlr4-python3-runtime/antlr4/BufferedTokenStream.pyi +14 -15
  223. package/dist/typeshed-fallback/stubs/antlr4-python3-runtime/antlr4/CommonTokenFactory.pyi +15 -7
  224. package/dist/typeshed-fallback/stubs/antlr4-python3-runtime/antlr4/CommonTokenStream.pyi +1 -3
  225. package/dist/typeshed-fallback/stubs/antlr4-python3-runtime/antlr4/FileStream.pyi +2 -4
  226. package/dist/typeshed-fallback/stubs/antlr4-python3-runtime/antlr4/InputStream.pyi +13 -11
  227. package/dist/typeshed-fallback/stubs/antlr4-python3-runtime/antlr4/Lexer.pyi +43 -29
  228. package/dist/typeshed-fallback/stubs/antlr4-python3-runtime/antlr4/Parser.pyi +47 -34
  229. package/dist/typeshed-fallback/stubs/antlr4-python3-runtime/antlr4/ParserInterpreter.pyi +13 -13
  230. package/dist/typeshed-fallback/stubs/antlr4-python3-runtime/antlr4/ParserRuleContext.pyi +30 -23
  231. package/dist/typeshed-fallback/stubs/antlr4-python3-runtime/antlr4/Recognizer.pyi +26 -17
  232. package/dist/typeshed-fallback/stubs/antlr4-python3-runtime/antlr4/RuleContext.pyi +22 -20
  233. package/dist/typeshed-fallback/stubs/antlr4-python3-runtime/antlr4/Token.pyi +26 -25
  234. package/dist/typeshed-fallback/stubs/antlr4-python3-runtime/antlr4/tree/Tree.pyi +30 -26
  235. package/dist/typeshed-fallback/stubs/assertpy/assertpy/assertpy.pyi +13 -10
  236. package/dist/typeshed-fallback/stubs/assertpy/assertpy/base.pyi +4 -5
  237. package/dist/typeshed-fallback/stubs/assertpy/assertpy/collection.pyi +12 -4
  238. package/dist/typeshed-fallback/stubs/assertpy/assertpy/contains.pyi +10 -8
  239. package/dist/typeshed-fallback/stubs/assertpy/assertpy/dict.pyi +8 -6
  240. package/dist/typeshed-fallback/stubs/assertpy/assertpy/dynamic.pyi +1 -1
  241. package/dist/typeshed-fallback/stubs/assertpy/assertpy/extracting.pyi +4 -2
  242. package/dist/typeshed-fallback/stubs/assertpy/assertpy/file.pyi +2 -3
  243. package/dist/typeshed-fallback/stubs/atheris/METADATA.toml +6 -0
  244. package/dist/typeshed-fallback/stubs/atheris/atheris/__init__.pyi +9 -0
  245. package/dist/typeshed-fallback/stubs/atheris/atheris/function_hooks.pyi +14 -0
  246. package/dist/typeshed-fallback/stubs/atheris/atheris/import_hook.pyi +36 -0
  247. package/dist/typeshed-fallback/stubs/atheris/atheris/instrument_bytecode.pyi +7 -0
  248. package/dist/typeshed-fallback/stubs/atheris/atheris/utils.pyi +18 -0
  249. package/dist/typeshed-fallback/stubs/atheris/atheris/version_dependent.pyi +27 -0
  250. package/dist/typeshed-fallback/stubs/aws-xray-sdk/METADATA.toml +1 -1
  251. package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/__init__.pyi +2 -1
  252. package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/async_context.pyi +15 -5
  253. package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/async_recorder.pyi +28 -9
  254. package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/context.pyi +7 -7
  255. package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/daemon_config.pyi +9 -7
  256. package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/emitters/udp_emitter.pyi +2 -2
  257. package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/lambda_launcher.pyi +8 -12
  258. package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/models/entity.pyi +1 -1
  259. package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/models/facade_segment.pyi +1 -1
  260. package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/models/http.pyi +9 -9
  261. package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/models/segment.pyi +1 -1
  262. package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/models/subsegment.pyi +3 -2
  263. package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/models/throwable.pyi +11 -3
  264. package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/patcher.pyi +3 -2
  265. package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/plugins/ec2_plugin.pyi +3 -3
  266. package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/plugins/ecs_plugin.pyi +2 -2
  267. package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/plugins/elasticbeanstalk_plugin.pyi +3 -3
  268. package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/recorder.pyi +19 -15
  269. package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/sampling/rule_cache.pyi +1 -1
  270. package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/sampling/sampler.pyi +3 -1
  271. package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/sampling/sampling_rule.pyi +23 -13
  272. package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/sampling/target_poller.pyi +5 -1
  273. package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/ext/psycopg/__init__.pyi +3 -0
  274. package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/ext/psycopg/patch.pyi +1 -0
  275. package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/ext/util.pyi +1 -1
  276. package/dist/typeshed-fallback/stubs/bleach/METADATA.toml +1 -1
  277. package/dist/typeshed-fallback/stubs/bleach/bleach/html5lib_shim.pyi +1 -1
  278. package/dist/typeshed-fallback/stubs/bleach/bleach/linkifier.pyi +1 -1
  279. package/dist/typeshed-fallback/stubs/bleach/bleach/sanitizer.pyi +2 -2
  280. package/dist/typeshed-fallback/stubs/braintree/METADATA.toml +1 -1
  281. package/dist/typeshed-fallback/stubs/braintree/braintree/address.pyi +8 -4
  282. package/dist/typeshed-fallback/stubs/braintree/braintree/address_gateway.pyi +3 -2
  283. package/dist/typeshed-fallback/stubs/braintree/braintree/client_token.pyi +5 -1
  284. package/dist/typeshed-fallback/stubs/braintree/braintree/client_token_gateway.pyi +5 -3
  285. package/dist/typeshed-fallback/stubs/braintree/braintree/error_codes.pyi +11 -2
  286. package/dist/typeshed-fallback/stubs/braintree/braintree/payment_method.pyi +30 -1
  287. package/dist/typeshed-fallback/stubs/braintree/braintree/payment_method_gateway.pyi +30 -2
  288. package/dist/typeshed-fallback/stubs/braintree/braintree/subscription.pyi +10 -5
  289. package/dist/typeshed-fallback/stubs/braintree/braintree/subscription_details.pyi +5 -1
  290. package/dist/typeshed-fallback/stubs/braintree/braintree/subscription_gateway.pyi +12 -5
  291. package/dist/typeshed-fallback/stubs/braintree/braintree/transaction.pyi +6 -2
  292. package/dist/typeshed-fallback/stubs/braintree/braintree/transaction_gateway.pyi +5 -2
  293. package/dist/typeshed-fallback/stubs/colorful/METADATA.toml +2 -0
  294. package/dist/typeshed-fallback/stubs/colorful/colorful/ansi.pyi +14 -0
  295. package/dist/typeshed-fallback/stubs/colorful/colorful/colors.pyi +6 -0
  296. package/dist/typeshed-fallback/stubs/colorful/colorful/core.pyi +102 -0
  297. package/dist/typeshed-fallback/stubs/colorful/colorful/styles.pyi +4 -0
  298. package/dist/typeshed-fallback/stubs/colorful/colorful/terminal.pyi +16 -0
  299. package/dist/typeshed-fallback/stubs/colorful/colorful/utils.pyi +2 -0
  300. package/dist/typeshed-fallback/stubs/decorator/decorator.pyi +2 -2
  301. package/dist/typeshed-fallback/stubs/django-filter/django_filters/filterset.pyi +3 -1
  302. package/dist/typeshed-fallback/stubs/docker/docker/api/container.pyi +6 -1
  303. package/dist/typeshed-fallback/stubs/docker/docker/api/daemon.pyi +12 -4
  304. package/dist/typeshed-fallback/stubs/docker/docker/api/exec_api.pyi +123 -7
  305. package/dist/typeshed-fallback/stubs/docker/docker/client.pyi +12 -8
  306. package/dist/typeshed-fallback/stubs/docker/docker/models/containers.pyi +27 -17
  307. package/dist/typeshed-fallback/stubs/docker/docker/models/images.pyi +1 -1
  308. package/dist/typeshed-fallback/stubs/docker/docker/models/resource.pyi +5 -5
  309. package/dist/typeshed-fallback/stubs/docutils/METADATA.toml +1 -1
  310. package/dist/typeshed-fallback/stubs/docutils/docutils/parsers/rst/directives/body.pyi +10 -0
  311. package/dist/typeshed-fallback/stubs/et_xmlfile/METADATA.toml +2 -0
  312. package/dist/typeshed-fallback/stubs/et_xmlfile/et_xmlfile/__init__.pyi +9 -0
  313. package/dist/typeshed-fallback/stubs/et_xmlfile/et_xmlfile/incremental_tree.pyi +170 -0
  314. package/dist/typeshed-fallback/stubs/et_xmlfile/et_xmlfile/xmlfile.pyi +37 -0
  315. package/dist/typeshed-fallback/stubs/fanstatic/METADATA.toml +1 -1
  316. package/dist/typeshed-fallback/stubs/flake8-bugbear/METADATA.toml +1 -1
  317. package/dist/typeshed-fallback/stubs/flake8-bugbear/bugbear.pyi +37 -65
  318. package/dist/typeshed-fallback/stubs/flake8-builtins/METADATA.toml +1 -1
  319. package/dist/typeshed-fallback/stubs/flake8-simplify/METADATA.toml +1 -1
  320. package/dist/typeshed-fallback/stubs/flake8-simplify/flake8_simplify/__init__.pyi +1 -0
  321. package/dist/typeshed-fallback/stubs/flake8-simplify/flake8_simplify/constants.pyi +6 -3
  322. package/dist/typeshed-fallback/stubs/flake8-simplify/flake8_simplify/utils.pyi +0 -1
  323. package/dist/typeshed-fallback/stubs/fpdf2/fpdf/drawing.pyi +145 -81
  324. package/dist/typeshed-fallback/stubs/fpdf2/fpdf/text_region.pyi +1 -1
  325. package/dist/typeshed-fallback/stubs/geopandas/METADATA.toml +1 -1
  326. package/dist/typeshed-fallback/stubs/geopandas/geopandas/array.pyi +7 -14
  327. package/dist/typeshed-fallback/stubs/geopandas/geopandas/geodataframe.pyi +1 -0
  328. package/dist/typeshed-fallback/stubs/gevent/METADATA.toml +2 -2
  329. package/dist/typeshed-fallback/stubs/gevent/gevent/events.pyi +1 -12
  330. package/dist/typeshed-fallback/stubs/gevent/gevent/libuv/watcher.pyi +1 -0
  331. package/dist/typeshed-fallback/stubs/gevent/gevent/os.pyi +1 -0
  332. package/dist/typeshed-fallback/stubs/gevent/gevent/signal.pyi +2 -1
  333. package/dist/typeshed-fallback/stubs/google-cloud-ndb/METADATA.toml +1 -1
  334. package/dist/typeshed-fallback/stubs/google-cloud-ndb/google/cloud/ndb/_cache.pyi +9 -10
  335. package/dist/typeshed-fallback/stubs/google-cloud-ndb/google/cloud/ndb/_datastore_query.pyi +1 -2
  336. package/dist/typeshed-fallback/stubs/google-cloud-ndb/google/cloud/ndb/_eventloop.pyi +11 -10
  337. package/dist/typeshed-fallback/stubs/google-cloud-ndb/google/cloud/ndb/blobstore.pyi +6 -6
  338. package/dist/typeshed-fallback/stubs/google-cloud-ndb/google/cloud/ndb/context.pyi +21 -21
  339. package/dist/typeshed-fallback/stubs/google-cloud-ndb/google/cloud/ndb/exceptions.pyi +2 -2
  340. package/dist/typeshed-fallback/stubs/google-cloud-ndb/google/cloud/ndb/global_cache.pyi +13 -14
  341. package/dist/typeshed-fallback/stubs/google-cloud-ndb/google/cloud/ndb/key.pyi +2 -3
  342. package/dist/typeshed-fallback/stubs/google-cloud-ndb/google/cloud/ndb/metadata.pyi +1 -2
  343. package/dist/typeshed-fallback/stubs/google-cloud-ndb/google/cloud/ndb/polymodel.pyi +2 -2
  344. package/dist/typeshed-fallback/stubs/google-cloud-ndb/google/cloud/ndb/query.pyi +26 -27
  345. package/dist/typeshed-fallback/stubs/google-cloud-ndb/google/cloud/ndb/stats.pyi +37 -37
  346. package/dist/typeshed-fallback/stubs/google-cloud-ndb/google/cloud/ndb/tasklets.pyi +7 -7
  347. package/dist/typeshed-fallback/stubs/google-cloud-ndb/google/cloud/ndb/utils.pyi +4 -4
  348. package/dist/typeshed-fallback/stubs/greenlet/METADATA.toml +1 -1
  349. package/dist/typeshed-fallback/stubs/gunicorn/gunicorn/arbiter.pyi +2 -2
  350. package/dist/typeshed-fallback/stubs/gunicorn/gunicorn/sock.pyi +2 -0
  351. package/dist/typeshed-fallback/stubs/html5lib/METADATA.toml +1 -0
  352. package/dist/typeshed-fallback/stubs/html5lib/html5lib/_ihatexml.pyi +21 -23
  353. package/dist/typeshed-fallback/stubs/html5lib/html5lib/_inputstream.pyi +79 -71
  354. package/dist/typeshed-fallback/stubs/html5lib/html5lib/_tokenizer.pyi +87 -51
  355. package/dist/typeshed-fallback/stubs/html5lib/html5lib/_utils.pyi +18 -11
  356. package/dist/typeshed-fallback/stubs/html5lib/html5lib/constants.pyi +1 -3
  357. package/dist/typeshed-fallback/stubs/html5lib/html5lib/filters/alphabeticalattributes.pyi +3 -2
  358. package/dist/typeshed-fallback/stubs/html5lib/html5lib/filters/base.pyi +9 -6
  359. package/dist/typeshed-fallback/stubs/html5lib/html5lib/filters/inject_meta_charset.pyi +5 -3
  360. package/dist/typeshed-fallback/stubs/html5lib/html5lib/filters/lint.pyi +5 -3
  361. package/dist/typeshed-fallback/stubs/html5lib/html5lib/filters/optionaltags.pyi +1 -2
  362. package/dist/typeshed-fallback/stubs/html5lib/html5lib/filters/sanitizer.pyi +6 -6
  363. package/dist/typeshed-fallback/stubs/html5lib/html5lib/filters/whitespace.pyi +2 -2
  364. package/dist/typeshed-fallback/stubs/html5lib/html5lib/html5parser.pyi +2 -2
  365. package/dist/typeshed-fallback/stubs/html5lib/html5lib/serializer.pyi +75 -19
  366. package/dist/typeshed-fallback/stubs/html5lib/html5lib/treeadapters/sax.pyi +0 -3
  367. package/dist/typeshed-fallback/stubs/hvac/METADATA.toml +1 -1
  368. package/dist/typeshed-fallback/stubs/hvac/hvac/api/auth_methods/__init__.pyi +3 -4
  369. package/dist/typeshed-fallback/stubs/hvac/hvac/api/auth_methods/aws.pyi +2 -2
  370. package/dist/typeshed-fallback/stubs/hvac/hvac/api/auth_methods/azure.pyi +3 -2
  371. package/dist/typeshed-fallback/stubs/hvac/hvac/api/auth_methods/gcp.pyi +3 -2
  372. package/dist/typeshed-fallback/stubs/hvac/hvac/api/auth_methods/kubernetes.pyi +1 -0
  373. package/dist/typeshed-fallback/stubs/hvac/hvac/api/auth_methods/legacy_mfa.pyi +2 -4
  374. package/dist/typeshed-fallback/stubs/hvac/hvac/api/secrets_engines/__init__.pyi +3 -4
  375. package/dist/typeshed-fallback/stubs/hvac/hvac/api/secrets_engines/identity.pyi +2 -2
  376. package/dist/typeshed-fallback/stubs/hvac/hvac/api/secrets_engines/kv.pyi +3 -3
  377. package/dist/typeshed-fallback/stubs/hvac/hvac/api/system_backend/__init__.pyi +3 -4
  378. package/dist/typeshed-fallback/stubs/hvac/hvac/api/system_backend/system_backend_mixin.pyi +2 -2
  379. package/dist/typeshed-fallback/stubs/hvac/hvac/api/system_backend/wrapping.pyi +2 -3
  380. package/dist/typeshed-fallback/stubs/hvac/hvac/api/vault_api_category.pyi +2 -3
  381. package/dist/typeshed-fallback/stubs/icalendar/METADATA.toml +1 -1
  382. package/dist/typeshed-fallback/stubs/icalendar/icalendar/__init__.pyi +0 -1
  383. package/dist/typeshed-fallback/stubs/icalendar/icalendar/alarms.pyi +2 -2
  384. package/dist/typeshed-fallback/stubs/icalendar/icalendar/cal.pyi +16 -9
  385. package/dist/typeshed-fallback/stubs/icalendar/icalendar/parser.pyi +2 -1
  386. package/dist/typeshed-fallback/stubs/icalendar/icalendar/prop.pyi +3 -3
  387. package/dist/typeshed-fallback/stubs/icalendar/icalendar/timezone/zoneinfo.pyi +1 -5
  388. package/dist/typeshed-fallback/stubs/ldap3/ldap3/__init__.pyi +5 -5
  389. package/dist/typeshed-fallback/stubs/ldap3/ldap3/abstract/__init__.pyi +3 -3
  390. package/dist/typeshed-fallback/stubs/ldap3/ldap3/abstract/attrDef.pyi +13 -13
  391. package/dist/typeshed-fallback/stubs/ldap3/ldap3/abstract/attribute.pyi +9 -9
  392. package/dist/typeshed-fallback/stubs/ldap3/ldap3/abstract/cursor.pyi +28 -27
  393. package/dist/typeshed-fallback/stubs/ldap3/ldap3/abstract/entry.pyi +11 -11
  394. package/dist/typeshed-fallback/stubs/ldap3/ldap3/core/exceptions.pyi +8 -8
  395. package/dist/typeshed-fallback/stubs/ldap3/ldap3/core/pooling.pyi +15 -15
  396. package/dist/typeshed-fallback/stubs/ldap3/ldap3/core/rdns.pyi +7 -7
  397. package/dist/typeshed-fallback/stubs/ldap3/ldap3/core/results.pyi +3 -3
  398. package/dist/typeshed-fallback/stubs/ldap3/ldap3/core/server.pyi +15 -14
  399. package/dist/typeshed-fallback/stubs/ldap3/ldap3/core/timezone.pyi +3 -3
  400. package/dist/typeshed-fallback/stubs/ldap3/ldap3/core/tls.pyi +13 -13
  401. package/dist/typeshed-fallback/stubs/ldap3/ldap3/core/usage.pyi +6 -6
  402. package/dist/typeshed-fallback/stubs/ldap3/ldap3/extend/__init__.pyi +4 -4
  403. package/dist/typeshed-fallback/stubs/ldap3/ldap3/extend/microsoft/dirSync.pyi +12 -12
  404. package/dist/typeshed-fallback/stubs/ldap3/ldap3/extend/microsoft/persistentSearch.pyi +7 -7
  405. package/dist/typeshed-fallback/stubs/ldap3/ldap3/extend/standard/PersistentSearch.pyi +13 -13
  406. package/dist/typeshed-fallback/stubs/ldap3/ldap3/operation/modify.pyi +2 -2
  407. package/dist/typeshed-fallback/stubs/ldap3/ldap3/operation/search.pyi +5 -5
  408. package/dist/typeshed-fallback/stubs/ldap3/ldap3/protocol/formatters/formatters.pyi +2 -2
  409. package/dist/typeshed-fallback/stubs/ldap3/ldap3/protocol/formatters/standard.pyi +2 -2
  410. package/dist/typeshed-fallback/stubs/ldap3/ldap3/protocol/oid.pyi +2 -2
  411. package/dist/typeshed-fallback/stubs/ldap3/ldap3/protocol/rfc2849.pyi +2 -2
  412. package/dist/typeshed-fallback/stubs/ldap3/ldap3/protocol/rfc4512.pyi +59 -59
  413. package/dist/typeshed-fallback/stubs/ldap3/ldap3/strategy/asyncStream.pyi +7 -7
  414. package/dist/typeshed-fallback/stubs/ldap3/ldap3/strategy/asynchronous.pyi +6 -6
  415. package/dist/typeshed-fallback/stubs/ldap3/ldap3/strategy/base.pyi +7 -7
  416. package/dist/typeshed-fallback/stubs/ldap3/ldap3/strategy/ldifProducer.pyi +4 -4
  417. package/dist/typeshed-fallback/stubs/ldap3/ldap3/strategy/mockAsync.pyi +2 -2
  418. package/dist/typeshed-fallback/stubs/ldap3/ldap3/strategy/mockBase.pyi +9 -9
  419. package/dist/typeshed-fallback/stubs/ldap3/ldap3/strategy/mockSync.pyi +2 -2
  420. package/dist/typeshed-fallback/stubs/ldap3/ldap3/strategy/restartable.pyi +4 -4
  421. package/dist/typeshed-fallback/stubs/ldap3/ldap3/strategy/reusable.pyi +20 -20
  422. package/dist/typeshed-fallback/stubs/ldap3/ldap3/strategy/sync.pyi +3 -3
  423. package/dist/typeshed-fallback/stubs/ldap3/ldap3/utils/config.pyi +2 -2
  424. package/dist/typeshed-fallback/stubs/ldap3/ldap3/utils/hashed.pyi +3 -3
  425. package/dist/typeshed-fallback/stubs/ldap3/ldap3/utils/log.pyi +3 -3
  426. package/dist/typeshed-fallback/stubs/ldap3/ldap3/utils/ntlm.pyi +41 -41
  427. package/dist/typeshed-fallback/stubs/ldap3/ldap3/utils/repr.pyi +2 -2
  428. package/dist/typeshed-fallback/stubs/lupa/METADATA.toml +1 -1
  429. package/dist/typeshed-fallback/stubs/netaddr/netaddr/ip/__init__.pyi +4 -1
  430. package/dist/typeshed-fallback/stubs/netaddr/netaddr/ip/iana.pyi +8 -8
  431. package/dist/typeshed-fallback/stubs/networkx/METADATA.toml +1 -1
  432. package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/__init__.pyi +1 -0
  433. package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/approximation/steinertree.pyi +4 -0
  434. package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/approximation/treewidth.pyi +11 -7
  435. package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/assortativity/mixing.pyi +4 -4
  436. package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/bipartite/matching.pyi +3 -3
  437. package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/bipartite/matrix.pyi +8 -1
  438. package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/centrality/closeness.pyi +2 -2
  439. package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/centrality/current_flow_betweenness.pyi +2 -0
  440. package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/centrality/eigenvector.pyi +3 -2
  441. package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/centrality/katz.pyi +5 -4
  442. package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/centrality/percolation.pyi +3 -2
  443. package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/centrality/reaching.pyi +3 -2
  444. package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/centrality/subgraph_alg.pyi +2 -2
  445. package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/cluster.pyi +13 -3
  446. package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/community/__init__.pyi +1 -1
  447. package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/community/bipartitions.pyi +22 -0
  448. package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/core.pyi +6 -5
  449. package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/dag.pyi +0 -3
  450. package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/distance_regular.pyi +2 -2
  451. package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/isomorphism/ismags.pyi +17 -4
  452. package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/non_randomness.pyi +1 -1
  453. package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/perfect_graph.pyi +7 -0
  454. package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/planarity.pyi +13 -4
  455. package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/similarity.pyi +19 -3
  456. package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/threshold.pyi +0 -1
  457. package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/tree/__init__.pyi +1 -0
  458. package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/tree/distance_measures.pyi +10 -0
  459. package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/wiener.pyi +3 -1
  460. package/dist/typeshed-fallback/stubs/networkx/networkx/classes/digraph.pyi +13 -7
  461. package/dist/typeshed-fallback/stubs/networkx/networkx/classes/function.pyi +3 -1
  462. package/dist/typeshed-fallback/stubs/networkx/networkx/classes/graph.pyi +12 -6
  463. package/dist/typeshed-fallback/stubs/networkx/networkx/classes/multidigraph.pyi +10 -9
  464. package/dist/typeshed-fallback/stubs/networkx/networkx/classes/multigraph.pyi +23 -13
  465. package/dist/typeshed-fallback/stubs/networkx/networkx/drawing/layout.pyi +0 -1
  466. package/dist/typeshed-fallback/stubs/networkx/networkx/drawing/nx_pylab.pyi +20 -0
  467. package/dist/typeshed-fallback/stubs/networkx/networkx/generators/expanders.pyi +9 -1
  468. package/dist/typeshed-fallback/stubs/networkx/networkx/generators/random_graphs.pyi +21 -16
  469. package/dist/typeshed-fallback/stubs/networkx/networkx/generators/small.pyi +3 -0
  470. package/dist/typeshed-fallback/stubs/networkx/networkx/readwrite/json_graph/node_link.pyi +2 -24
  471. package/dist/typeshed-fallback/stubs/networkx/networkx/utils/misc.pyi +1 -1
  472. package/dist/typeshed-fallback/stubs/networkx/networkx/utils/random_sequence.pyi +5 -0
  473. package/dist/typeshed-fallback/stubs/networkx/networkx/utils/union_find.pyi +12 -10
  474. package/dist/typeshed-fallback/stubs/parsimonious/METADATA.toml +1 -1
  475. package/dist/typeshed-fallback/stubs/parsimonious/parsimonious/exceptions.pyi +5 -3
  476. package/dist/typeshed-fallback/stubs/peewee/METADATA.toml +1 -1
  477. package/dist/typeshed-fallback/stubs/protobuf/google/protobuf/internal/containers.pyi +8 -3
  478. package/dist/typeshed-fallback/stubs/protobuf/google/protobuf/internal/well_known_types.pyi +2 -2
  479. package/dist/typeshed-fallback/stubs/protobuf/google/protobuf/message.pyi +1 -1
  480. package/dist/typeshed-fallback/stubs/psutil/METADATA.toml +1 -1
  481. package/dist/typeshed-fallback/stubs/psutil/psutil/__init__.pyi +57 -77
  482. package/dist/typeshed-fallback/stubs/psutil/psutil/_common.pyi +135 -262
  483. package/dist/typeshed-fallback/stubs/psutil/psutil/_ntuples.pyi +384 -0
  484. package/dist/typeshed-fallback/stubs/psutil/psutil/_psaix.pyi +100 -86
  485. package/dist/typeshed-fallback/stubs/psutil/psutil/_psbsd.pyi +213 -168
  486. package/dist/typeshed-fallback/stubs/psutil/psutil/_pslinux.pyi +181 -237
  487. package/dist/typeshed-fallback/stubs/psutil/psutil/_psosx.pyi +81 -100
  488. package/dist/typeshed-fallback/stubs/psutil/psutil/_psposix.pyi +78 -4
  489. package/dist/typeshed-fallback/stubs/psutil/psutil/_pssunos.pyi +141 -119
  490. package/dist/typeshed-fallback/stubs/psutil/psutil/_psutil_aix.pyi +58 -0
  491. package/dist/typeshed-fallback/stubs/psutil/psutil/_psutil_bsd.pyi +141 -0
  492. package/dist/typeshed-fallback/stubs/psutil/psutil/_psutil_linux.pyi +47 -15
  493. package/dist/typeshed-fallback/stubs/psutil/psutil/_psutil_osx.pyi +77 -49
  494. package/dist/typeshed-fallback/stubs/psutil/psutil/_psutil_sunos.pyi +63 -0
  495. package/dist/typeshed-fallback/stubs/psutil/psutil/_psutil_windows.pyi +102 -91
  496. package/dist/typeshed-fallback/stubs/psutil/psutil/_pswindows.pyi +175 -200
  497. package/dist/typeshed-fallback/stubs/pycurl/pycurl.pyi +5 -3
  498. package/dist/typeshed-fallback/stubs/pyinstaller/METADATA.toml +1 -1
  499. package/dist/typeshed-fallback/stubs/python-crontab/crontab.pyi +4 -4
  500. package/dist/typeshed-fallback/stubs/python-dateutil/dateutil/parser/_parser.pyi +6 -7
  501. package/dist/typeshed-fallback/stubs/python-dateutil/dateutil/rrule.pyi +28 -25
  502. package/dist/typeshed-fallback/stubs/python-dateutil/dateutil/tz/_common.pyi +8 -3
  503. package/dist/typeshed-fallback/stubs/python-dateutil/dateutil/tz/tz.pyi +21 -19
  504. package/dist/typeshed-fallback/stubs/python-dateutil/dateutil/tz/win.pyi +5 -4
  505. package/dist/typeshed-fallback/stubs/python-dateutil/dateutil/zoneinfo/__init__.pyi +2 -1
  506. package/dist/typeshed-fallback/stubs/python-dateutil/dateutil/zoneinfo/rebuild.pyi +5 -1
  507. package/dist/typeshed-fallback/stubs/python-jenkins/jenkins/__init__.pyi +13 -6
  508. package/dist/typeshed-fallback/stubs/pytz/pytz/lazy.pyi +2 -2
  509. package/dist/typeshed-fallback/stubs/pyxdg/xdg/RecentFiles.pyi +5 -7
  510. package/dist/typeshed-fallback/stubs/qrbill/METADATA.toml +1 -1
  511. package/dist/typeshed-fallback/stubs/qrbill/qrbill/bill.pyi +10 -0
  512. package/dist/typeshed-fallback/stubs/regex/METADATA.toml +1 -1
  513. package/dist/typeshed-fallback/stubs/regex/regex/__init__.pyi +64 -1
  514. package/dist/typeshed-fallback/stubs/regex/regex/{regex.pyi → _main.pyi} +66 -0
  515. package/dist/typeshed-fallback/stubs/regex/regex/_regex.pyi +1 -1
  516. package/dist/typeshed-fallback/stubs/regex/regex/_regex_core.pyi +59 -19
  517. package/dist/typeshed-fallback/stubs/reportlab/METADATA.toml +1 -1
  518. package/dist/typeshed-fallback/stubs/reportlab/reportlab/lib/testutils.pyi +27 -25
  519. package/dist/typeshed-fallback/stubs/reportlab/reportlab/platypus/flowables.pyi +1 -1
  520. package/dist/typeshed-fallback/stubs/requests/requests/exceptions.pyi +2 -1
  521. package/dist/typeshed-fallback/stubs/seaborn/seaborn/_core/data.pyi +2 -3
  522. package/dist/typeshed-fallback/stubs/seaborn/seaborn/_core/typing.pyi +0 -2
  523. package/dist/typeshed-fallback/stubs/seaborn/seaborn/categorical.pyi +1 -1
  524. package/dist/typeshed-fallback/stubs/setuptools/setuptools/__init__.pyi +102 -45
  525. package/dist/typeshed-fallback/stubs/setuptools/setuptools/_distutils/cmd.pyi +6 -3
  526. package/dist/typeshed-fallback/stubs/setuptools/setuptools/_distutils/compilers/C/base.pyi +8 -2
  527. package/dist/typeshed-fallback/stubs/setuptools/setuptools/_distutils/compilers/C/msvc.pyi +3 -0
  528. package/dist/typeshed-fallback/stubs/setuptools/setuptools/_distutils/spawn.pyi +14 -4
  529. package/dist/typeshed-fallback/stubs/simple-websocket/METADATA.toml +3 -0
  530. package/dist/typeshed-fallback/stubs/simple-websocket/simple_websocket/__init__.pyi +3 -0
  531. package/dist/typeshed-fallback/stubs/simple-websocket/simple_websocket/aiows.pyi +130 -0
  532. package/dist/typeshed-fallback/stubs/simple-websocket/simple_websocket/asgi.pyi +44 -0
  533. package/dist/typeshed-fallback/stubs/simple-websocket/simple_websocket/errors.pyi +12 -0
  534. package/dist/typeshed-fallback/stubs/simple-websocket/simple_websocket/ws.pyi +136 -0
  535. package/dist/typeshed-fallback/stubs/tensorflow/METADATA.toml +16 -0
  536. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/__init__.pyi +437 -0
  537. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/_aliases.pyi +70 -0
  538. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/audio.pyi +7 -0
  539. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/autodiff.pyi +63 -0
  540. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/autograph/__init__.pyi +17 -0
  541. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/autograph/experimental.pyi +30 -0
  542. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/bitwise.pyi +37 -0
  543. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/compiler/xla/service/hlo_pb2.pyi +2113 -0
  544. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/compiler/xla/service/hlo_profile_printer_data_pb2.pyi +187 -0
  545. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/compiler/xla/service/metrics_pb2.pyi +283 -0
  546. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/compiler/xla/service/test_compilation_environment_pb2.pyi +59 -0
  547. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/compiler/xla/service/xla_compile_result_pb2.pyi +167 -0
  548. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/compiler/xla/tsl/protobuf/bfc_memory_map_pb2.pyi +218 -0
  549. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/compiler/xla/tsl/protobuf/test_log_pb2.pyi +707 -0
  550. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/compiler/xla/xla_data_pb2.pyi +2681 -0
  551. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/compiler/xla/xla_pb2.pyi +2558 -0
  552. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/config/__init__.pyi +12 -0
  553. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/config/experimental.pyi +16 -0
  554. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/example/example_parser_configuration_pb2.pyi +153 -0
  555. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/example/example_pb2.pyi +330 -0
  556. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/example/feature_pb2.pyi +222 -0
  557. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/allocation_description_pb2.pyi +64 -0
  558. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/api_def_pb2.pyi +312 -0
  559. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/attr_value_pb2.pyi +274 -0
  560. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/cost_graph_pb2.pyi +229 -0
  561. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/cpp_shape_inference_pb2.pyi +110 -0
  562. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/dataset_metadata_pb2.pyi +25 -0
  563. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/dataset_options_pb2.pyi +720 -0
  564. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/dataset_pb2.pyi +116 -0
  565. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/device_attributes_pb2.pyi +140 -0
  566. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/full_type_pb2.pyi +617 -0
  567. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/function_pb2.pyi +309 -0
  568. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/graph_debug_info_pb2.pyi +210 -0
  569. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/graph_pb2.pyi +100 -0
  570. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/graph_transfer_info_pb2.pyi +290 -0
  571. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/kernel_def_pb2.pyi +116 -0
  572. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/log_memory_pb2.pyi +218 -0
  573. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/model_pb2.pyi +368 -0
  574. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/node_def_pb2.pyi +192 -0
  575. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/op_def_pb2.pyi +391 -0
  576. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/optimized_function_graph_pb2.pyi +166 -0
  577. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/reader_base_pb2.pyi +52 -0
  578. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/resource_handle_pb2.pyi +104 -0
  579. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/step_stats_pb2.pyi +332 -0
  580. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/summary_pb2.pyi +368 -0
  581. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/tensor_description_pb2.pyi +49 -0
  582. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/tensor_pb2.pyi +235 -0
  583. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/tensor_shape_pb2.pyi +74 -0
  584. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/tensor_slice_pb2.pyi +56 -0
  585. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/types_pb2.pyi +220 -0
  586. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/variable_pb2.pyi +229 -0
  587. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/versions_pb2.pyi +57 -0
  588. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/__init__.pyi +0 -0
  589. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/bfc_memory_map_pb2.pyi +15 -0
  590. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/cluster_pb2.pyi +126 -0
  591. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/composite_tensor_variant_pb2.pyi +33 -0
  592. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/config_pb2.pyi +1867 -0
  593. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/control_flow_pb2.pyi +243 -0
  594. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/core_platform_payloads_pb2.pyi +66 -0
  595. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/data_service_pb2.pyi +243 -0
  596. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/debug_event_pb2.pyi +746 -0
  597. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/debug_pb2.pyi +199 -0
  598. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/device_filters_pb2.pyi +124 -0
  599. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/device_properties_pb2.pyi +154 -0
  600. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/error_codes_pb2.pyi +33 -0
  601. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/fingerprint_pb2.pyi +76 -0
  602. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/meta_graph_pb2.pyi +735 -0
  603. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/named_tensor_pb2.pyi +41 -0
  604. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/queue_runner_pb2.pyi +73 -0
  605. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/remote_tensor_handle_pb2.pyi +96 -0
  606. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/rewriter_config_pb2.pyi +588 -0
  607. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/rpc_options_pb2.pyi +9 -0
  608. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/saved_model_pb2.pyi +51 -0
  609. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/saved_object_graph_pb2.pyi +715 -0
  610. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/saver_pb2.pyi +113 -0
  611. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/service_config_pb2.pyi +275 -0
  612. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/snapshot_pb2.pyi +164 -0
  613. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/status_pb2.pyi +13 -0
  614. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/struct_pb2.pyi +536 -0
  615. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/tensor_bundle_pb2.pyi +160 -0
  616. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/tensorflow_server_pb2.pyi +125 -0
  617. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/tpu/compilation_result_pb2.pyi +85 -0
  618. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/tpu/dynamic_padding_pb2.pyi +47 -0
  619. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/tpu/optimization_parameters_pb2.pyi +1326 -0
  620. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/tpu/topology_pb2.pyi +149 -0
  621. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/tpu/tpu_embedding_configuration_pb2.pyi +326 -0
  622. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/trackable_object_graph_pb2.pyi +213 -0
  623. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/transport_options_pb2.pyi +28 -0
  624. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/verifier_config_pb2.pyi +65 -0
  625. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/util/event_pb2.pyi +419 -0
  626. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/util/memmapped_file_system_pb2.pyi +65 -0
  627. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/util/saved_tensor_slice_pb2.pyi +171 -0
  628. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/util/test_log_pb2.pyi +24 -0
  629. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/data/__init__.pyi +272 -0
  630. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/data/experimental.pyi +32 -0
  631. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/distribute/__init__.pyi +3 -0
  632. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/distribute/coordinator.pyi +3 -0
  633. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/distribute/experimental/coordinator.pyi +11 -0
  634. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/dtypes.pyi +57 -0
  635. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/experimental/__init__.pyi +10 -0
  636. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/experimental/dtensor.pyi +19 -0
  637. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/feature_column/__init__.pyi +95 -0
  638. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/initializers.pyi +1 -0
  639. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/io/__init__.pyi +104 -0
  640. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/io/gfile.pyi +11 -0
  641. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/keras/__init__.pyi +15 -0
  642. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/keras/activations.pyi +35 -0
  643. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/keras/callbacks.pyi +170 -0
  644. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/keras/constraints.pyi +16 -0
  645. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/keras/initializers.pyi +50 -0
  646. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/keras/layers/__init__.pyi +446 -0
  647. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/keras/losses.pyi +177 -0
  648. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/keras/metrics.pyi +117 -0
  649. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/keras/models.pyi +167 -0
  650. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/keras/optimizers/__init__.pyi +7 -0
  651. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/keras/optimizers/legacy/__init__.pyi +61 -0
  652. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/keras/optimizers/schedules.pyi +103 -0
  653. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/keras/regularizers.pyi +21 -0
  654. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/linalg.pyi +55 -0
  655. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/math.pyi +298 -0
  656. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/nn.pyi +194 -0
  657. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/__init__.pyi +1 -0
  658. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/distribute/distribute_lib.pyi +5 -0
  659. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/feature_column/__init__.pyi +0 -0
  660. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/feature_column/feature_column_v2.pyi +273 -0
  661. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/feature_column/sequence_feature_column.pyi +30 -0
  662. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/framework/dtypes.pyi +7 -0
  663. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/keras/__init__.pyi +1 -0
  664. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/keras/protobuf/projector_config_pb2.pyi +129 -0
  665. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/keras/protobuf/saved_metadata_pb2.pyi +89 -0
  666. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/keras/protobuf/versions_pb2.pyi +63 -0
  667. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/trackable/__init__.pyi +0 -0
  668. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/trackable/autotrackable.pyi +3 -0
  669. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/trackable/base.pyi +5 -0
  670. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/trackable/resource.pyi +9 -0
  671. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/trackable/ressource.pyi +7 -0
  672. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/training/tracking/autotrackable.pyi +3 -0
  673. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/random.pyi +231 -0
  674. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/raw_ops.pyi +44 -0
  675. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/saved_model/__init__.pyi +125 -0
  676. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/saved_model/experimental.pyi +39 -0
  677. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/signal.pyi +6 -0
  678. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/sparse.pyi +31 -0
  679. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/strings.pyi +241 -0
  680. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/summary.pyi +58 -0
  681. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/train/__init__.pyi +74 -0
  682. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/train/experimental.pyi +12 -0
  683. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/tsl/protobuf/coordination_config_pb2.pyi +156 -0
  684. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/tsl/protobuf/coordination_service_pb2.pyi +666 -0
  685. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/tsl/protobuf/distributed_runtime_payloads_pb2.pyi +69 -0
  686. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/tsl/protobuf/dnn_pb2.pyi +619 -0
  687. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/tsl/protobuf/error_codes_pb2.pyi +291 -0
  688. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/tsl/protobuf/histogram_pb2.pyi +78 -0
  689. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/tsl/protobuf/rpc_options_pb2.pyi +85 -0
  690. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/tsl/protobuf/status_pb2.pyi +34 -0
  691. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/types/__init__.pyi +1 -0
  692. package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/types/experimental.pyi +30 -0
  693. package/dist/typeshed-fallback/stubs/ttkthemes/METADATA.toml +1 -1
  694. package/dist/typeshed-fallback/stubs/ttkthemes/ttkthemes/themed_style.pyi +1 -1
  695. package/dist/typeshed-fallback/stubs/ttkthemes/ttkthemes/themed_tk.pyi +11 -12
  696. package/dist/typeshed-fallback/stubs/tzdata/METADATA.toml +2 -0
  697. package/dist/typeshed-fallback/stubs/tzdata/tzdata/__init__.pyi +4 -0
  698. package/dist/typeshed-fallback/stubs/webencodings/METADATA.toml +2 -0
  699. package/dist/typeshed-fallback/stubs/webencodings/webencodings/__init__.pyi +33 -0
  700. package/dist/typeshed-fallback/stubs/webencodings/webencodings/labels.pyi +3 -0
  701. package/dist/typeshed-fallback/stubs/webencodings/webencodings/mklabels.pyi +5 -0
  702. package/dist/typeshed-fallback/stubs/webencodings/webencodings/x_user_defined.pyi +21 -0
  703. package/dist/typeshed-fallback/stubs/yt-dlp/METADATA.toml +1 -1
  704. package/dist/typeshed-fallback/stubs/yt-dlp/yt_dlp/__init__.pyi +2 -0
  705. package/dist/typeshed-fallback/stubs/yt-dlp/yt_dlp/extractor/common.pyi +1 -0
  706. package/dist/typeshed-fallback/stubs/yt-dlp/yt_dlp/globals.pyi +12 -1
  707. package/dist/typeshed-fallback/stubs/yt-dlp/yt_dlp/utils/_jsruntime.pyi +38 -0
  708. package/dist/typeshed-fallback/stubs/yt-dlp/yt_dlp/utils/_legacy.pyi +2 -68
  709. package/dist/typeshed-fallback/stubs/yt-dlp/yt_dlp/utils/_utils.pyi +1 -1
  710. package/package.json +7 -7
  711. package/dist/typeshed-fallback/stdlib/_typeshed/README.md +0 -34
  712. package/dist/typeshed-fallback/stubs/ExifRead/METADATA.toml +0 -3
  713. package/dist/typeshed-fallback/stubs/ExifRead/exifread/__init__.pyi +0 -17
  714. package/dist/typeshed-fallback/stubs/ExifRead/exifread/_types.pyi +0 -15
  715. package/dist/typeshed-fallback/stubs/ExifRead/exifread/classes.pyi +0 -47
  716. package/dist/typeshed-fallback/stubs/ExifRead/exifread/exceptions.pyi +0 -2
  717. package/dist/typeshed-fallback/stubs/ExifRead/exifread/exif_log.pyi +0 -24
  718. package/dist/typeshed-fallback/stubs/ExifRead/exifread/heic.pyi +0 -56
  719. package/dist/typeshed-fallback/stubs/ExifRead/exifread/jpeg.pyi +0 -7
  720. package/dist/typeshed-fallback/stubs/ExifRead/exifread/tags/__init__.pyi +0 -15
  721. package/dist/typeshed-fallback/stubs/ExifRead/exifread/tags/exif.pyi +0 -7
  722. package/dist/typeshed-fallback/stubs/ExifRead/exifread/tags/makernote/apple.pyi +0 -3
  723. package/dist/typeshed-fallback/stubs/ExifRead/exifread/tags/makernote/canon.pyi +0 -26
  724. package/dist/typeshed-fallback/stubs/ExifRead/exifread/tags/makernote/casio.pyi +0 -3
  725. package/dist/typeshed-fallback/stubs/ExifRead/exifread/tags/makernote/fujifilm.pyi +0 -3
  726. package/dist/typeshed-fallback/stubs/ExifRead/exifread/tags/makernote/nikon.pyi +0 -6
  727. package/dist/typeshed-fallback/stubs/ExifRead/exifread/tags/makernote/olympus.pyi +0 -6
  728. package/dist/typeshed-fallback/stubs/ExifRead/exifread/utils.pyi +0 -22
  729. package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/community/kernighan_lin.pyi +0 -16
  730. package/dist/typeshed-fallback/stubs/psutil/psutil/_psutil_posix.pyi +0 -34
  731. package/dist/typeshed-fallback/stubs/vobject/@tests/stubtest_allowlist.txt +0 -22
  732. /package/dist/typeshed-fallback/stubs/{ExifRead/exifread/tags/makernote → colorful/colorful}/__init__.pyi +0 -0
@@ -0,0 +1,2558 @@
1
+ """
2
+ @generated by mypy-protobuf. Do not edit manually!
3
+ isort:skip_file
4
+ Copyright 2017 The OpenXLA Authors.
5
+
6
+ Licensed under the Apache License, Version 2.0 (the "License");
7
+ you may not use this file except in compliance with the License.
8
+ You may obtain a copy of the License at
9
+
10
+ http://www.apache.org/licenses/LICENSE-2.0
11
+
12
+ Unless required by applicable law or agreed to in writing, software
13
+ distributed under the License is distributed on an "AS IS" BASIS,
14
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
15
+ See the License for the specific language governing permissions and
16
+ limitations under the License.
17
+ ==============================================================================
18
+ """
19
+
20
+ import builtins
21
+ import collections.abc
22
+ import sys
23
+ import typing
24
+
25
+ import google.protobuf.any_pb2
26
+ import google.protobuf.descriptor
27
+ import google.protobuf.internal.containers
28
+ import google.protobuf.internal.enum_type_wrapper
29
+ import google.protobuf.message
30
+ import tensorflow.compiler.xla.service.hlo_pb2
31
+ import tensorflow.compiler.xla.xla_data_pb2
32
+
33
+ if sys.version_info >= (3, 10):
34
+ import typing as typing_extensions
35
+ else:
36
+ import typing_extensions
37
+
38
+ DESCRIPTOR: google.protobuf.descriptor.FileDescriptor
39
+
40
+ @typing.final
41
+ class CompilationEnvironmentsProto(google.protobuf.message.Message):
42
+ """Proto version of `xla::CompilationEnvironments`."""
43
+
44
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
45
+
46
+ ENVIRONMENTS_FIELD_NUMBER: builtins.int
47
+ @property
48
+ def environments(
49
+ self,
50
+ ) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[google.protobuf.any_pb2.Any]: ...
51
+ def __init__(self, *, environments: collections.abc.Iterable[google.protobuf.any_pb2.Any] | None = ...) -> None: ...
52
+ def ClearField(self, field_name: typing.Literal["environments", b"environments"]) -> None: ...
53
+
54
+ global___CompilationEnvironmentsProto = CompilationEnvironmentsProto
55
+
56
+ @typing.final
57
+ class DebugOptions(google.protobuf.message.Message):
58
+ """Debugging options for XLA. These options may change at any time - there are
59
+ no guarantees about backward or forward compatibility for these fields.
60
+
61
+ Debug options naming and organization:
62
+
63
+ 1. Backend-agnostic options: `xla_$flag_name` - go first, and sorted
64
+ alphabetically by the flag name.
65
+
66
+ 2. Backend-specific options: `xla_$backend_$flag_name` - must be in the
67
+ corresponding backend section, and sorted alphabetically by the flag name.
68
+ --------------------------------------------------------------------------//
69
+ XLA backend-agnostic options.
70
+ --------------------------------------------------------------------------//
71
+ go/keep-sorted start
72
+ """
73
+
74
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
75
+
76
+ class _ShapeChecks:
77
+ ValueType = typing.NewType("ValueType", builtins.int)
78
+ V: typing_extensions.TypeAlias = ValueType
79
+
80
+ class _ShapeChecksEnumTypeWrapper(
81
+ google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[DebugOptions._ShapeChecks.ValueType], builtins.type
82
+ ):
83
+ DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
84
+ IGNORE: DebugOptions._ShapeChecks.ValueType # 0
85
+ """Do not insert any shape checks for dynamically shaped operations; output
86
+ buffers might contain garbage data if shapes don't match.
87
+ """
88
+ RUNTIME: DebugOptions._ShapeChecks.ValueType # 1
89
+ """Check shapes at runtime, will insert an extra synchronization if shapes
90
+ cannot be proven correct at compile time.
91
+ """
92
+ COMPILE_TIME: DebugOptions._ShapeChecks.ValueType # 2
93
+ """Will refuse to compile any program where shape correctness can not be
94
+ established at compile time.
95
+ """
96
+
97
+ class ShapeChecks(_ShapeChecks, metaclass=_ShapeChecksEnumTypeWrapper): ...
98
+ IGNORE: DebugOptions.ShapeChecks.ValueType # 0
99
+ """Do not insert any shape checks for dynamically shaped operations; output
100
+ buffers might contain garbage data if shapes don't match.
101
+ """
102
+ RUNTIME: DebugOptions.ShapeChecks.ValueType # 1
103
+ """Check shapes at runtime, will insert an extra synchronization if shapes
104
+ cannot be proven correct at compile time.
105
+ """
106
+ COMPILE_TIME: DebugOptions.ShapeChecks.ValueType # 2
107
+ """Will refuse to compile any program where shape correctness can not be
108
+ established at compile time.
109
+ """
110
+
111
+ class _StepMarkerLocation:
112
+ ValueType = typing.NewType("ValueType", builtins.int)
113
+ V: typing_extensions.TypeAlias = ValueType
114
+
115
+ class _StepMarkerLocationEnumTypeWrapper(
116
+ google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[DebugOptions._StepMarkerLocation.ValueType], builtins.type
117
+ ):
118
+ DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
119
+ STEP_MARK_AT_ENTRY: DebugOptions._StepMarkerLocation.ValueType # 0
120
+ """Generate a step marker at the program entry. This handles the case where
121
+ each step is done by one or multiple program execution(s). Only the first
122
+ program will be tagged for generating a step marker at the program entry.
123
+ This is the default.
124
+ """
125
+ STEP_MARK_AT_TOP_LEVEL_WHILE_LOOP: DebugOptions._StepMarkerLocation.ValueType # 1
126
+ """Generate a step marker at each iteration of the top level while loop,
127
+ which is assumed to be a training loop.
128
+ """
129
+ STEP_MARK_AT_SECOND_LEVEL_WHILE_LOOP: DebugOptions._StepMarkerLocation.ValueType # 3
130
+ """Generate a step marker at each iteration of the second level while loops,
131
+ which is assumed to be a training or eval loop.
132
+ """
133
+ STEP_MARK_NONE: DebugOptions._StepMarkerLocation.ValueType # 2
134
+ """No step marker generated."""
135
+
136
+ class StepMarkerLocation(_StepMarkerLocation, metaclass=_StepMarkerLocationEnumTypeWrapper): ...
137
+ STEP_MARK_AT_ENTRY: DebugOptions.StepMarkerLocation.ValueType # 0
138
+ """Generate a step marker at the program entry. This handles the case where
139
+ each step is done by one or multiple program execution(s). Only the first
140
+ program will be tagged for generating a step marker at the program entry.
141
+ This is the default.
142
+ """
143
+ STEP_MARK_AT_TOP_LEVEL_WHILE_LOOP: DebugOptions.StepMarkerLocation.ValueType # 1
144
+ """Generate a step marker at each iteration of the top level while loop,
145
+ which is assumed to be a training loop.
146
+ """
147
+ STEP_MARK_AT_SECOND_LEVEL_WHILE_LOOP: DebugOptions.StepMarkerLocation.ValueType # 3
148
+ """Generate a step marker at each iteration of the second level while loops,
149
+ which is assumed to be a training or eval loop.
150
+ """
151
+ STEP_MARK_NONE: DebugOptions.StepMarkerLocation.ValueType # 2
152
+ """No step marker generated."""
153
+
154
+ class _CollectiveOpType:
155
+ ValueType = typing.NewType("ValueType", builtins.int)
156
+ V: typing_extensions.TypeAlias = ValueType
157
+
158
+ class _CollectiveOpTypeEnumTypeWrapper(
159
+ google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[DebugOptions._CollectiveOpType.ValueType], builtins.type
160
+ ):
161
+ DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
162
+ NOOP: DebugOptions._CollectiveOpType.ValueType # 0
163
+ ALLREDUCE: DebugOptions._CollectiveOpType.ValueType # 1
164
+ ALLGATHER: DebugOptions._CollectiveOpType.ValueType # 2
165
+ REDUCESCATTER: DebugOptions._CollectiveOpType.ValueType # 3
166
+ COLLECTIVEBROADCAST: DebugOptions._CollectiveOpType.ValueType # 4
167
+ ALLTOALL: DebugOptions._CollectiveOpType.ValueType # 5
168
+ COLLECTIVEPERMUTE: DebugOptions._CollectiveOpType.ValueType # 6
169
+
170
+ class CollectiveOpType(_CollectiveOpType, metaclass=_CollectiveOpTypeEnumTypeWrapper):
171
+ """Enum to define all collective ops
172
+ that xla supports.
173
+ """
174
+
175
+ NOOP: DebugOptions.CollectiveOpType.ValueType # 0
176
+ ALLREDUCE: DebugOptions.CollectiveOpType.ValueType # 1
177
+ ALLGATHER: DebugOptions.CollectiveOpType.ValueType # 2
178
+ REDUCESCATTER: DebugOptions.CollectiveOpType.ValueType # 3
179
+ COLLECTIVEBROADCAST: DebugOptions.CollectiveOpType.ValueType # 4
180
+ ALLTOALL: DebugOptions.CollectiveOpType.ValueType # 5
181
+ COLLECTIVEPERMUTE: DebugOptions.CollectiveOpType.ValueType # 6
182
+
183
+ class _CommandBufferCmdType:
184
+ ValueType = typing.NewType("ValueType", builtins.int)
185
+ V: typing_extensions.TypeAlias = ValueType
186
+
187
+ class _CommandBufferCmdTypeEnumTypeWrapper(
188
+ google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[DebugOptions._CommandBufferCmdType.ValueType], builtins.type
189
+ ):
190
+ DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
191
+ INVALID: DebugOptions._CommandBufferCmdType.ValueType # 0
192
+ FUSION: DebugOptions._CommandBufferCmdType.ValueType # 1
193
+ CUBLAS: DebugOptions._CommandBufferCmdType.ValueType # 2
194
+ CUDNN: DebugOptions._CommandBufferCmdType.ValueType # 3
195
+ COLLECTIVES: DebugOptions._CommandBufferCmdType.ValueType # 4
196
+ CONDITIONALS: DebugOptions._CommandBufferCmdType.ValueType # 5
197
+ CUSTOM_CALL: DebugOptions._CommandBufferCmdType.ValueType # 6
198
+ CUBLASLT: DebugOptions._CommandBufferCmdType.ValueType # 7
199
+
200
+ class CommandBufferCmdType(_CommandBufferCmdType, metaclass=_CommandBufferCmdTypeEnumTypeWrapper):
201
+ """Commands are categorized into 5 types:
202
+ FUSION represents regular fusion kernels.
203
+ CUBLAS/CUBLASLT, CUDNN, and COLLECTIVES represent library calls.
204
+ CONDITIONALS represents control flow.
205
+ """
206
+
207
+ INVALID: DebugOptions.CommandBufferCmdType.ValueType # 0
208
+ FUSION: DebugOptions.CommandBufferCmdType.ValueType # 1
209
+ CUBLAS: DebugOptions.CommandBufferCmdType.ValueType # 2
210
+ CUDNN: DebugOptions.CommandBufferCmdType.ValueType # 3
211
+ COLLECTIVES: DebugOptions.CommandBufferCmdType.ValueType # 4
212
+ CONDITIONALS: DebugOptions.CommandBufferCmdType.ValueType # 5
213
+ CUSTOM_CALL: DebugOptions.CommandBufferCmdType.ValueType # 6
214
+ CUBLASLT: DebugOptions.CommandBufferCmdType.ValueType # 7
215
+
216
+ class _PartitioningAlgorithm:
217
+ ValueType = typing.NewType("ValueType", builtins.int)
218
+ V: typing_extensions.TypeAlias = ValueType
219
+
220
+ class _PartitioningAlgorithmEnumTypeWrapper(
221
+ google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[DebugOptions._PartitioningAlgorithm.ValueType], builtins.type
222
+ ):
223
+ DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
224
+ PARTITIONING_ALGORITHM_NOOP: DebugOptions._PartitioningAlgorithm.ValueType # 0
225
+ PARTITIONING_ALGORITHM_EXP0: DebugOptions._PartitioningAlgorithm.ValueType # 1
226
+ PARTITIONING_ALGORITHM_EXP1: DebugOptions._PartitioningAlgorithm.ValueType # 2
227
+ PARTITIONING_ALGORITHM_EXP2: DebugOptions._PartitioningAlgorithm.ValueType # 3
228
+
229
+ class PartitioningAlgorithm(_PartitioningAlgorithm, metaclass=_PartitioningAlgorithmEnumTypeWrapper): ...
230
+ PARTITIONING_ALGORITHM_NOOP: DebugOptions.PartitioningAlgorithm.ValueType # 0
231
+ PARTITIONING_ALGORITHM_EXP0: DebugOptions.PartitioningAlgorithm.ValueType # 1
232
+ PARTITIONING_ALGORITHM_EXP1: DebugOptions.PartitioningAlgorithm.ValueType # 2
233
+ PARTITIONING_ALGORITHM_EXP2: DebugOptions.PartitioningAlgorithm.ValueType # 3
234
+
235
+ class _WhileLoopUnrolling:
236
+ ValueType = typing.NewType("ValueType", builtins.int)
237
+ V: typing_extensions.TypeAlias = ValueType
238
+
239
+ class _WhileLoopUnrollingEnumTypeWrapper(
240
+ google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[DebugOptions._WhileLoopUnrolling.ValueType], builtins.type
241
+ ):
242
+ DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
243
+ WHILE_LOOP_UNROLLING_NO_UNROLL: DebugOptions._WhileLoopUnrolling.ValueType # 0
244
+ WHILE_LOOP_UNROLLING_DOUBLE_BUFFER: DebugOptions._WhileLoopUnrolling.ValueType # 1
245
+ """Has the same effect as setting
246
+ `xla_gpu_enable_while_loop_double_buffering`.
247
+ """
248
+ WHILE_LOOP_UNROLLING_FULL_UNROLL: DebugOptions._WhileLoopUnrolling.ValueType # 2
249
+ """Enables full loop unrolling using the same strategy as `DOUBLE_BUFFER`."""
250
+
251
+ class WhileLoopUnrolling(_WhileLoopUnrolling, metaclass=_WhileLoopUnrollingEnumTypeWrapper): ...
252
+ WHILE_LOOP_UNROLLING_NO_UNROLL: DebugOptions.WhileLoopUnrolling.ValueType # 0
253
+ WHILE_LOOP_UNROLLING_DOUBLE_BUFFER: DebugOptions.WhileLoopUnrolling.ValueType # 1
254
+ """Has the same effect as setting
255
+ `xla_gpu_enable_while_loop_double_buffering`.
256
+ """
257
+ WHILE_LOOP_UNROLLING_FULL_UNROLL: DebugOptions.WhileLoopUnrolling.ValueType # 2
258
+ """Enables full loop unrolling using the same strategy as `DOUBLE_BUFFER`."""
259
+
260
+ class _AutotuneCacheMode:
261
+ ValueType = typing.NewType("ValueType", builtins.int)
262
+ V: typing_extensions.TypeAlias = ValueType
263
+
264
+ class _AutotuneCacheModeEnumTypeWrapper(
265
+ google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[DebugOptions._AutotuneCacheMode.ValueType], builtins.type
266
+ ):
267
+ DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
268
+ AUTOTUNE_CACHE_MODE_UNSPECIFIED: DebugOptions._AutotuneCacheMode.ValueType # 0
269
+ AUTOTUNE_CACHE_MODE_UPDATE: DebugOptions._AutotuneCacheMode.ValueType # 1
270
+ """If the cache exists per fusion autotuner loads it and terminates,
271
+ otherwise runs autotuner and dumps the result.
272
+ """
273
+ AUTOTUNE_CACHE_MODE_READ: DebugOptions._AutotuneCacheMode.ValueType # 2
274
+ """Sets readonly access to the cache for the per fusion autotuner. Same as
275
+ above, but doesn't dump anything.
276
+ """
277
+
278
+ class AutotuneCacheMode(_AutotuneCacheMode, metaclass=_AutotuneCacheModeEnumTypeWrapper): ...
279
+ AUTOTUNE_CACHE_MODE_UNSPECIFIED: DebugOptions.AutotuneCacheMode.ValueType # 0
280
+ AUTOTUNE_CACHE_MODE_UPDATE: DebugOptions.AutotuneCacheMode.ValueType # 1
281
+ """If the cache exists per fusion autotuner loads it and terminates,
282
+ otherwise runs autotuner and dumps the result.
283
+ """
284
+ AUTOTUNE_CACHE_MODE_READ: DebugOptions.AutotuneCacheMode.ValueType # 2
285
+ """Sets readonly access to the cache for the per fusion autotuner. Same as
286
+ above, but doesn't dump anything.
287
+ """
288
+
289
+ @typing.final
290
+ class XlaBackendExtraOptionsEntry(google.protobuf.message.Message):
291
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
292
+
293
+ KEY_FIELD_NUMBER: builtins.int
294
+ VALUE_FIELD_NUMBER: builtins.int
295
+ key: builtins.str
296
+ value: builtins.str
297
+ def __init__(self, *, key: builtins.str | None = ..., value: builtins.str | None = ...) -> None: ...
298
+ def ClearField(self, field_name: typing.Literal["key", b"key", "value", b"value"]) -> None: ...
299
+
300
+ XLA_CPU_ENABLE_CONCURRENCY_OPTIMIZED_SCHEDULER_FIELD_NUMBER: builtins.int
301
+ XLA_CPU_ENABLE_FAST_MATH_FIELD_NUMBER: builtins.int
302
+ XLA_CPU_ENABLE_FAST_MIN_MAX_FIELD_NUMBER: builtins.int
303
+ XLA_CPU_FAST_MATH_HONOR_DIVISION_FIELD_NUMBER: builtins.int
304
+ XLA_CPU_FAST_MATH_HONOR_FUNCTIONS_FIELD_NUMBER: builtins.int
305
+ XLA_CPU_FAST_MATH_HONOR_INFS_FIELD_NUMBER: builtins.int
306
+ XLA_CPU_FAST_MATH_HONOR_NANS_FIELD_NUMBER: builtins.int
307
+ XLA_CPU_USE_THUNK_RUNTIME_FIELD_NUMBER: builtins.int
308
+ XLA_CPU_PARALLEL_CODEGEN_SPLIT_COUNT_FIELD_NUMBER: builtins.int
309
+ XLA_CPU_PREFER_VECTOR_WIDTH_FIELD_NUMBER: builtins.int
310
+ XLA_GPU_EXPERIMENTAL_AUTOTUNE_CACHE_MODE_FIELD_NUMBER: builtins.int
311
+ XLA_GPU_EXPERIMENTAL_DISABLE_BINARY_LIBRARIES_FIELD_NUMBER: builtins.int
312
+ XLA_GPU_EXPERIMENTAL_ENABLE_TRITON_SOFTMAX_PRIORITY_FUSION_FIELD_NUMBER: builtins.int
313
+ XLA_GPU_UNSUPPORTED_ENABLE_TRITON_GEMM_FIELD_NUMBER: builtins.int
314
+ XLA_HLO_GRAPH_ADDRESSES_FIELD_NUMBER: builtins.int
315
+ XLA_HLO_PROFILE_FIELD_NUMBER: builtins.int
316
+ XLA_DISABLE_HLO_PASSES_FIELD_NUMBER: builtins.int
317
+ XLA_ENABLE_HLO_PASSES_ONLY_FIELD_NUMBER: builtins.int
318
+ XLA_DISABLE_ALL_HLO_PASSES_FIELD_NUMBER: builtins.int
319
+ XLA_BACKEND_OPTIMIZATION_LEVEL_FIELD_NUMBER: builtins.int
320
+ XLA_EMBED_IR_IN_EXECUTABLE_FIELD_NUMBER: builtins.int
321
+ XLA_ELIMINATE_HLO_IMPLICIT_BROADCAST_FIELD_NUMBER: builtins.int
322
+ XLA_CPU_MULTI_THREAD_EIGEN_FIELD_NUMBER: builtins.int
323
+ XLA_GPU_CUDA_DATA_DIR_FIELD_NUMBER: builtins.int
324
+ XLA_GPU_FTZ_FIELD_NUMBER: builtins.int
325
+ XLA_LLVM_ENABLE_ALIAS_SCOPE_METADATA_FIELD_NUMBER: builtins.int
326
+ XLA_LLVM_ENABLE_NOALIAS_METADATA_FIELD_NUMBER: builtins.int
327
+ XLA_LLVM_ENABLE_INVARIANT_LOAD_METADATA_FIELD_NUMBER: builtins.int
328
+ XLA_LLVM_DISABLE_EXPENSIVE_PASSES_FIELD_NUMBER: builtins.int
329
+ XLA_TEST_ALL_OUTPUT_LAYOUTS_FIELD_NUMBER: builtins.int
330
+ XLA_TEST_ALL_INPUT_LAYOUTS_FIELD_NUMBER: builtins.int
331
+ XLA_HLO_GRAPH_SHARDING_COLOR_FIELD_NUMBER: builtins.int
332
+ XLA_CPU_USE_MKL_DNN_FIELD_NUMBER: builtins.int
333
+ XLA_GPU_ENABLE_FAST_MIN_MAX_FIELD_NUMBER: builtins.int
334
+ XLA_ALLOW_EXCESS_PRECISION_FIELD_NUMBER: builtins.int
335
+ XLA_GPU_CRASH_ON_VERIFICATION_FAILURES_FIELD_NUMBER: builtins.int
336
+ XLA_GPU_AUTOTUNE_LEVEL_FIELD_NUMBER: builtins.int
337
+ XLA_FORCE_HOST_PLATFORM_DEVICE_COUNT_FIELD_NUMBER: builtins.int
338
+ XLA_GPU_DISABLE_GPUASM_OPTIMIZATIONS_FIELD_NUMBER: builtins.int
339
+ XLA_GPU_SHAPE_CHECKS_FIELD_NUMBER: builtins.int
340
+ XLA_HLO_EVALUATOR_USE_FAST_PATH_FIELD_NUMBER: builtins.int
341
+ XLA_ALLOW_SCALAR_INDEX_DYNAMIC_OPS_FIELD_NUMBER: builtins.int
342
+ XLA_STEP_MARKER_LOCATION_FIELD_NUMBER: builtins.int
343
+ XLA_DUMP_TO_FIELD_NUMBER: builtins.int
344
+ XLA_DUMP_HLO_MODULE_RE_FIELD_NUMBER: builtins.int
345
+ XLA_DUMP_HLO_PASS_RE_FIELD_NUMBER: builtins.int
346
+ XLA_DUMP_HLO_AS_TEXT_FIELD_NUMBER: builtins.int
347
+ XLA_DUMP_HLO_AS_PROTO_FIELD_NUMBER: builtins.int
348
+ XLA_DUMP_HLO_AS_DOT_FIELD_NUMBER: builtins.int
349
+ XLA_DUMP_HLO_AS_URL_FIELD_NUMBER: builtins.int
350
+ XLA_DUMP_HLO_AS_HTML_FIELD_NUMBER: builtins.int
351
+ XLA_DUMP_FUSION_VISUALIZATION_FIELD_NUMBER: builtins.int
352
+ XLA_DUMP_HLO_SNAPSHOTS_FIELD_NUMBER: builtins.int
353
+ XLA_DUMP_INCLUDE_TIMESTAMP_FIELD_NUMBER: builtins.int
354
+ XLA_DUMP_MAX_HLO_MODULES_FIELD_NUMBER: builtins.int
355
+ XLA_DUMP_MODULE_METADATA_FIELD_NUMBER: builtins.int
356
+ XLA_DUMP_COMPRESS_PROTOS_FIELD_NUMBER: builtins.int
357
+ XLA_DUMP_HLO_AS_LONG_TEXT_FIELD_NUMBER: builtins.int
358
+ XLA_GPU_FORCE_CONV_NCHW_FIELD_NUMBER: builtins.int
359
+ XLA_GPU_FORCE_CONV_NHWC_FIELD_NUMBER: builtins.int
360
+ XLA_GPU_PTX_FILE_FIELD_NUMBER: builtins.int
361
+ XLA_GPU_DUMP_LLVMIR_FIELD_NUMBER: builtins.int
362
+ XLA_DUMP_ENABLE_MLIR_PRETTY_FORM_FIELD_NUMBER: builtins.int
363
+ XLA_GPU_ALGORITHM_DENYLIST_PATH_FIELD_NUMBER: builtins.int
364
+ XLA_TPU_DETECT_NAN_FIELD_NUMBER: builtins.int
365
+ XLA_TPU_DETECT_INF_FIELD_NUMBER: builtins.int
366
+ XLA_CPU_ENABLE_XPROF_TRACEME_FIELD_NUMBER: builtins.int
367
+ XLA_GPU_UNSAFE_FALLBACK_TO_DRIVER_ON_PTXAS_NOT_FOUND_FIELD_NUMBER: builtins.int
368
+ XLA_GPU_ASM_EXTRA_FLAGS_FIELD_NUMBER: builtins.int
369
+ XLA_MULTIHEAP_SIZE_CONSTRAINT_PER_HEAP_FIELD_NUMBER: builtins.int
370
+ XLA_DETAILED_LOGGING_FIELD_NUMBER: builtins.int
371
+ XLA_ENABLE_DUMPING_FIELD_NUMBER: builtins.int
372
+ XLA_GPU_FORCE_COMPILATION_PARALLELISM_FIELD_NUMBER: builtins.int
373
+ XLA_GPU_ENABLE_LLVM_MODULE_COMPILATION_PARALLELISM_FIELD_NUMBER: builtins.int
374
+ XLA_GPU_DETERMINISTIC_OPS_FIELD_NUMBER: builtins.int
375
+ XLA_GPU_LLVM_IR_FILE_FIELD_NUMBER: builtins.int
376
+ XLA_GPU_DISABLE_ASYNC_COLLECTIVES_FIELD_NUMBER: builtins.int
377
+ XLA_GPU_ALL_REDUCE_COMBINE_THRESHOLD_BYTES_FIELD_NUMBER: builtins.int
378
+ XLA_GPU_ALL_GATHER_COMBINE_THRESHOLD_BYTES_FIELD_NUMBER: builtins.int
379
+ XLA_GPU_REDUCE_SCATTER_COMBINE_THRESHOLD_BYTES_FIELD_NUMBER: builtins.int
380
+ XLA_GPU_ENABLE_ALL_GATHER_COMBINE_BY_DIM_FIELD_NUMBER: builtins.int
381
+ XLA_GPU_ENABLE_REDUCE_SCATTER_COMBINE_BY_DIM_FIELD_NUMBER: builtins.int
382
+ XLA_GPU_ENABLE_REASSOCIATION_FOR_CONVERTED_AR_FIELD_NUMBER: builtins.int
383
+ XLA_GPU_ALL_REDUCE_BLUECONNECT_NUM_DEVICES_PER_HOST_FIELD_NUMBER: builtins.int
384
+ XLA_GPU_ENABLE_WHILE_LOOP_REDUCE_SCATTER_CODE_MOTION_FIELD_NUMBER: builtins.int
385
+ XLA_GPU_COLLECTIVE_INFLATION_FACTOR_FIELD_NUMBER: builtins.int
386
+ XLA_LLVM_FORCE_INLINE_BEFORE_SPLIT_FIELD_NUMBER: builtins.int
387
+ XLA_GPU_ENABLE_CUDNN_FRONTEND_FIELD_NUMBER: builtins.int
388
+ XLA_GPU_ENABLE_CUDNN_FMHA_FIELD_NUMBER: builtins.int
389
+ XLA_GPU_FUSED_ATTENTION_USE_CUDNN_RNG_FIELD_NUMBER: builtins.int
390
+ XLA_GPU_ENABLE_CUDNN_LAYER_NORM_FIELD_NUMBER: builtins.int
391
+ XLA_DUMP_DISABLE_METADATA_FIELD_NUMBER: builtins.int
392
+ XLA_DUMP_HLO_PIPELINE_RE_FIELD_NUMBER: builtins.int
393
+ XLA_GPU_STRICT_CONV_ALGORITHM_PICKER_FIELD_NUMBER: builtins.int
394
+ XLA_GPU_ENABLE_CUSTOM_FUSIONS_FIELD_NUMBER: builtins.int
395
+ XLA_GPU_ENABLE_CUSTOM_FUSIONS_RE_FIELD_NUMBER: builtins.int
396
+ XLA_GPU_ENABLE_DYNAMIC_SLICE_FUSION_FIELD_NUMBER: builtins.int
397
+ XLA_GPU_NCCL_TERMINATION_TIMEOUT_SECONDS_FIELD_NUMBER: builtins.int
398
+ XLA_GPU_ENABLE_SHARED_CONSTANTS_FIELD_NUMBER: builtins.int
399
+ XLA_GPU_ENABLE_CUBLASLT_FIELD_NUMBER: builtins.int
400
+ XLA_GPU_ENABLE_COMMAND_BUFFER_FIELD_NUMBER: builtins.int
401
+ XLA_GPU_GRAPH_MIN_GRAPH_SIZE_FIELD_NUMBER: builtins.int
402
+ XLA_GPU_GRAPH_ENABLE_CONCURRENT_REGION_FIELD_NUMBER: builtins.int
403
+ XLA_GPU_REDZONE_SCRATCH_MAX_MEGABYTES_FIELD_NUMBER: builtins.int
404
+ XLA_GPU_REDZONE_PADDING_BYTES_FIELD_NUMBER: builtins.int
405
+ XLA_CPU_USE_ACL_FIELD_NUMBER: builtins.int
406
+ XLA_CPU_STRICT_DOT_CONV_MATH_FIELD_NUMBER: builtins.int
407
+ XLA_GPU_USE_RUNTIME_FUSION_FIELD_NUMBER: builtins.int
408
+ XLA_DUMP_LATENCY_HIDING_SCHEDULE_FIELD_NUMBER: builtins.int
409
+ XLA_CPU_ENABLE_MLIR_TILING_AND_FUSION_FIELD_NUMBER: builtins.int
410
+ XLA_CPU_ENABLE_CUSTOM_MATMUL_TILING_FIELD_NUMBER: builtins.int
411
+ XLA_CPU_MATMUL_TILING_M_DIM_FIELD_NUMBER: builtins.int
412
+ XLA_CPU_MATMUL_TILING_N_DIM_FIELD_NUMBER: builtins.int
413
+ XLA_CPU_MATMUL_TILING_K_DIM_FIELD_NUMBER: builtins.int
414
+ XLA_CPU_ENABLE_MLIR_FUSION_OUTLINING_FIELD_NUMBER: builtins.int
415
+ XLA_CPU_ENABLE_EXPERIMENTAL_DEALLOCATION_FIELD_NUMBER: builtins.int
416
+ XLA_GPU_ENABLE_LATENCY_HIDING_SCHEDULER_FIELD_NUMBER: builtins.int
417
+ XLA_GPU_ENABLE_HIGHEST_PRIORITY_ASYNC_STREAM_FIELD_NUMBER: builtins.int
418
+ XLA_GPU_ENABLE_ANALYTICAL_LATENCY_ESTIMATOR_FIELD_NUMBER: builtins.int
419
+ XLA_GPU_LHS_ENABLE_GPU_ASYNC_TRACKER_FIELD_NUMBER: builtins.int
420
+ XLA_GPU_PGLE_PROFILE_FILE_OR_DIRECTORY_PATH_FIELD_NUMBER: builtins.int
421
+ XLA_GPU_MEMORY_LIMIT_SLOP_FACTOR_FIELD_NUMBER: builtins.int
422
+ XLA_GPU_ENABLE_PIPELINED_COLLECTIVES_FIELD_NUMBER: builtins.int
423
+ XLA_GPU_ENABLE_PIPELINED_ALL_REDUCE_FIELD_NUMBER: builtins.int
424
+ XLA_GPU_ENABLE_PIPELINED_ALL_GATHER_FIELD_NUMBER: builtins.int
425
+ XLA_GPU_ENABLE_PIPELINED_REDUCE_SCATTER_FIELD_NUMBER: builtins.int
426
+ XLA_GPU_ENABLE_PIPELINED_P2P_FIELD_NUMBER: builtins.int
427
+ XLA_GPU_RUN_POST_LAYOUT_COLLECTIVE_PIPELINER_FIELD_NUMBER: builtins.int
428
+ XLA_GPU_COLLECTIVE_PERMUTE_DECOMPOSER_THRESHOLD_FIELD_NUMBER: builtins.int
429
+ XLA_PARTITIONING_ALGORITHM_FIELD_NUMBER: builtins.int
430
+ XLA_GPU_ENABLE_TRITON_GEMM_FIELD_NUMBER: builtins.int
431
+ XLA_GPU_ENABLE_CUDNN_INT8X32_CONVOLUTION_REORDERING_FIELD_NUMBER: builtins.int
432
+ XLA_GPU_TRITON_GEMM_ANY_FIELD_NUMBER: builtins.int
433
+ XLA_GPU_EXHAUSTIVE_TILING_SEARCH_FIELD_NUMBER: builtins.int
434
+ XLA_GPU_ENABLE_PRIORITY_FUSION_FIELD_NUMBER: builtins.int
435
+ XLA_GPU_DUMP_AUTOTUNE_RESULTS_TO_FIELD_NUMBER: builtins.int
436
+ XLA_GPU_LOAD_AUTOTUNE_RESULTS_FROM_FIELD_NUMBER: builtins.int
437
+ XLA_GPU_TARGET_CONFIG_FILENAME_FIELD_NUMBER: builtins.int
438
+ XLA_GPU_AUTO_SPMD_PARTITIONING_MEMORY_BUDGET_GB_FIELD_NUMBER: builtins.int
439
+ XLA_GPU_AUTO_SPMD_PARTITIONING_MEMORY_BUDGET_RATIO_FIELD_NUMBER: builtins.int
440
+ XLA_GPU_TRITON_GEMM_DISABLE_REDUCED_PRECISION_REDUCTION_FIELD_NUMBER: builtins.int
441
+ XLA_GPU_TRITON_FUSION_LEVEL_FIELD_NUMBER: builtins.int
442
+ XLA_GPU_DUMP_AUTOTUNED_GEMM_FUSIONS_FIELD_NUMBER: builtins.int
443
+ XLA_GPU_OVERRIDE_GEMM_AUTOTUNER_FIELD_NUMBER: builtins.int
444
+ XLA_GPU_COPY_INSERTION_USE_REGION_ANALYSIS_FIELD_NUMBER: builtins.int
445
+ XLA_GPU_COLLECT_COST_MODEL_STATS_FIELD_NUMBER: builtins.int
446
+ XLA_GPU_ENABLE_SPLIT_K_AUTOTUNING_FIELD_NUMBER: builtins.int
447
+ XLA_GPU_ENABLE_REDUCTION_EPILOGUE_FUSION_FIELD_NUMBER: builtins.int
448
+ XLA_GPU_ENABLE_NCCL_CLIQUE_OPTIMIZATION_FIELD_NUMBER: builtins.int
449
+ XLA_GPU_MOCK_CUSTOM_CALLS_FIELD_NUMBER: builtins.int
450
+ XLA_GPU_CUBLAS_FALLBACK_FIELD_NUMBER: builtins.int
451
+ XLA_GPU_ENABLE_WHILE_LOOP_DOUBLE_BUFFERING_FIELD_NUMBER: builtins.int
452
+ XLA_GPU_ENABLE_WHILE_LOOP_UNROLLING_FIELD_NUMBER: builtins.int
453
+ XLA_GPU_ENSURE_MINOR_DOT_CONTRACTION_DIMS_FIELD_NUMBER: builtins.int
454
+ XLA_GPU_FILTER_KERNELS_SPILLING_REGISTERS_ON_AUTOTUNING_FIELD_NUMBER: builtins.int
455
+ XLA_DEBUG_BUFFER_ASSIGNMENT_SHOW_MAX_FIELD_NUMBER: builtins.int
456
+ XLA_GPU_LLVM_VERIFICATION_LEVEL_FIELD_NUMBER: builtins.int
457
+ XLA_GPU_ENABLE_CUB_RADIX_SORT_FIELD_NUMBER: builtins.int
458
+ XLA_GPU_THRESHOLD_FOR_WINDOWED_EINSUM_MIB_FIELD_NUMBER: builtins.int
459
+ XLA_GPU_ENABLE_TRITON_HOPPER_FIELD_NUMBER: builtins.int
460
+ XLA_GPU_ENABLE_NCCL_USER_BUFFERS_FIELD_NUMBER: builtins.int
461
+ XLA_GPU_ENABLE_NCCL_COMM_SPLITTING_FIELD_NUMBER: builtins.int
462
+ XLA_GPU_ENABLE_NCCL_PER_STREAM_COMMS_FIELD_NUMBER: builtins.int
463
+ XLA_GPU_ENABLE_LIBNVPTXCOMPILER_FIELD_NUMBER: builtins.int
464
+ XLA_GPU_ENABLE_DOT_STRENGTH_REDUCTION_FIELD_NUMBER: builtins.int
465
+ XLA_GPU_MULTI_STREAMED_WINDOWED_EINSUM_FIELD_NUMBER: builtins.int
466
+ XLA_GPU_ENABLE_BF16_6WAY_GEMM_FIELD_NUMBER: builtins.int
467
+ XLA_GPU_ENABLE_BF16_3WAY_GEMM_FIELD_NUMBER: builtins.int
468
+ XLA_GPU_NCCL_COLLECTIVE_MAX_NCHANNELS_FIELD_NUMBER: builtins.int
469
+ XLA_GPU_NCCL_P2P_MAX_NCHANNELS_FIELD_NUMBER: builtins.int
470
+ XLA_GPU_MLIR_EMITTER_LEVEL_FIELD_NUMBER: builtins.int
471
+ XLA_GPU_GEMM_REWRITE_SIZE_THRESHOLD_FIELD_NUMBER: builtins.int
472
+ XLA_GPU_REQUIRE_COMPLETE_AOT_AUTOTUNE_RESULTS_FIELD_NUMBER: builtins.int
473
+ XLA_GPU_CUDNN_GEMM_FUSION_LEVEL_FIELD_NUMBER: builtins.int
474
+ XLA_GPU_USE_MEMCPY_LOCAL_P2P_FIELD_NUMBER: builtins.int
475
+ XLA_GPU_AUTOTUNE_MAX_SOLUTIONS_FIELD_NUMBER: builtins.int
476
+ XLA_DUMP_LARGE_CONSTANTS_FIELD_NUMBER: builtins.int
477
+ XLA_GPU_VERIFY_TRITON_FUSION_NUMERICS_FIELD_NUMBER: builtins.int
478
+ XLA_GPU_DUMP_AUTOTUNE_LOGS_TO_FIELD_NUMBER: builtins.int
479
+ XLA_REDUCE_WINDOW_REWRITE_BASE_LENGTH_FIELD_NUMBER: builtins.int
480
+ XLA_GPU_ENABLE_HOST_MEMORY_OFFLOADING_FIELD_NUMBER: builtins.int
481
+ XLA_GPU_EXCLUDE_NONDETERMINISTIC_OPS_FIELD_NUMBER: builtins.int
482
+ XLA_GPU_NCCL_TERMINATE_ON_ERROR_FIELD_NUMBER: builtins.int
483
+ XLA_GPU_SHARD_AUTOTUNING_FIELD_NUMBER: builtins.int
484
+ XLA_GPU_ENABLE_APPROX_COSTLY_COLLECTIVES_FIELD_NUMBER: builtins.int
485
+ XLA_GPU_KERNEL_CACHE_FILE_FIELD_NUMBER: builtins.int
486
+ XLA_GPU_UNSAFE_PIPELINED_LOOP_ANNOTATOR_FIELD_NUMBER: builtins.int
487
+ XLA_GPU_PER_FUSION_AUTOTUNE_CACHE_DIR_FIELD_NUMBER: builtins.int
488
+ XLA_CMD_BUFFER_TRACE_CACHE_SIZE_FIELD_NUMBER: builtins.int
489
+ XLA_GPU_TEMP_BUFFER_USE_SEPARATE_COLOR_FIELD_NUMBER: builtins.int
490
+ LEGACY_COMMAND_BUFFER_CUSTOM_CALL_TARGETS_FIELD_NUMBER: builtins.int
491
+ XLA_SYNTAX_SUGAR_ASYNC_OPS_FIELD_NUMBER: builtins.int
492
+ XLA_GPU_AUTOTUNE_GEMM_RTOL_FIELD_NUMBER: builtins.int
493
+ XLA_ENABLE_COMMAND_BUFFERS_DURING_PROFILING_FIELD_NUMBER: builtins.int
494
+ XLA_GPU_CUDNN_GEMM_MAX_PLANS_FIELD_NUMBER: builtins.int
495
+ XLA_GPU_ENABLE_LIBNVJITLINK_FIELD_NUMBER: builtins.int
496
+ XLA_GPU_ENABLE_TRITON_GEMM_INT4_FIELD_NUMBER: builtins.int
497
+ XLA_GPU_ASYNC_DOT_FIELD_NUMBER: builtins.int
498
+ XLA_GPU_ENABLE_PGLE_ACCURACY_CHECKER_FIELD_NUMBER: builtins.int
499
+ XLA_GPU_EXECUTABLE_WARN_STUCK_TIMEOUT_SECONDS_FIELD_NUMBER: builtins.int
500
+ XLA_GPU_EXECUTABLE_TERMINATE_TIMEOUT_SECONDS_FIELD_NUMBER: builtins.int
501
+ XLA_EXPERIMENTAL_IGNORE_CHANNEL_ID_FIELD_NUMBER: builtins.int
502
+ XLA_BACKEND_EXTRA_OPTIONS_FIELD_NUMBER: builtins.int
503
+ xla_cpu_enable_concurrency_optimized_scheduler: builtins.bool
504
+ """--------------------------------------------------------------------------//
505
+ XLA:CPU options.
506
+ --------------------------------------------------------------------------//
507
+
508
+ go/keep-sorted start newline_separated=yes
509
+
510
+ When true, XLA:CPU uses HLO module scheduler that is optimized for
511
+ extracting concurrency at the cost of extra memory: we extend the live
512
+ ranges of temporaries to allow XLA runtime to schedule independent
513
+ operations in parallel on separate threads.
514
+ """
515
+ xla_cpu_enable_fast_math: builtins.bool
516
+ """When true, "unsafe" mathematical optimizations are enabled. These
517
+ transformations include but are not limited to:
518
+
519
+ - Reducing the precision of operations (e.g. using an approximate sin
520
+ function, or transforming x/y into x * (1/y)).
521
+ - Assuming that operations never produce or consume NaN or +/- Inf (this
522
+ behavior can be adjusted using xla_cpu_fast_math_allow_{nans|infs}).
523
+ - Assuming that +0 and -0 are indistinguishable.
524
+ """
525
+ xla_cpu_enable_fast_min_max: builtins.bool
526
+ """When false we lower the Minimum and Maximum hlos in the CPU backend such
527
+ that Min(NotNaN, NaN) = Min(NaN, NotNaN) = NaN. In other words, if flag
528
+ this is false we always propagate NaNs through Min and Max.
529
+
530
+ Note, this does not correspond to the exact same behavior as the gpu flag
531
+ below!
532
+ """
533
+ xla_cpu_fast_math_honor_division: builtins.bool
534
+ """When xla_cpu_enable_fast_math is true then this controls whether we forbid
535
+ to use the reciprocal of an argument instead of division. Ignored when
536
+ xla_cpu_enable_fast_math is false.
537
+ """
538
+ xla_cpu_fast_math_honor_functions: builtins.bool
539
+ """When xla_cpu_enable_fast_math is true then this controls whether we forbid
540
+ to approximate calculations for functions. Ignored when
541
+ xla_cpu_enable_fast_math is false.
542
+ """
543
+ xla_cpu_fast_math_honor_infs: builtins.bool
544
+ """When xla_cpu_enable_fast_math is true then this controls whether we allow
545
+ operations to produce infinites. Ignored when xla_cpu_enable_fast_math is
546
+ false.
547
+ """
548
+ xla_cpu_fast_math_honor_nans: builtins.bool
549
+ """When xla_cpu_enable_fast_math is true then this controls whether we allow
550
+ operations to produce NaNs. Ignored when xla_cpu_enable_fast_math is
551
+ false.
552
+ """
553
+ xla_cpu_use_thunk_runtime: builtins.bool
554
+ """When true, XLA:CPU uses the thunk runtime to execute compiled program."""
555
+ xla_cpu_parallel_codegen_split_count: builtins.int
556
+ """The number of parts to split the LLVM module into before codegen. This
557
+ allows XLA to compile all parts in parallel, and resolve kernel symbols
558
+ from different dynamic libraries.
559
+ """
560
+ xla_cpu_prefer_vector_width: builtins.int
561
+ """A `prefer-vector-width` value that is passed to the LLVM backend. Default
562
+ value is `256` (AVX2 on x86 platforms).
563
+ """
564
+ xla_gpu_experimental_autotune_cache_mode: global___DebugOptions.AutotuneCacheMode.ValueType
565
+ """--------------------------------------------------------------------------//
566
+ XLA:GPU options.
567
+ --------------------------------------------------------------------------//
568
+ go/keep-sorted start newline_separated=yes skip_lines=1
569
+
570
+ Specifies the behavior of per kernel autotuning cache.
571
+ """
572
+ xla_gpu_experimental_disable_binary_libraries: builtins.bool
573
+ """Experimentally disables binary libraries in GPU compiler passes."""
574
+ xla_gpu_experimental_enable_triton_softmax_priority_fusion: builtins.bool
575
+ """Gates the experimental feature coupling the Triton Softmax pattern matcher
576
+ with priority fusion.
577
+ """
578
+ xla_gpu_unsupported_enable_triton_gemm: builtins.bool
579
+ """Internal debug/testing flag to switch Triton GEMM fusions on or off."""
580
+ xla_hlo_graph_addresses: builtins.bool
581
+ """--------------------------------------------------------------------------//
582
+ A bag of XLA options that have to be categorized.
583
+ --------------------------------------------------------------------------//
584
+
585
+ Show addresses of HLO ops in graph dump.
586
+ """
587
+ xla_hlo_profile: builtins.bool
588
+ """Instrument the computation to collect per-HLO cycle counts."""
589
+ xla_disable_all_hlo_passes: builtins.bool
590
+ """Disables all HLO passes. Notes that some passes are necessary for
591
+ correctness and the invariants that must be satisfied by "fully optimized"
592
+ HLO are different for different devices and may change over time. The only
593
+ "guarantee", such as it is, is that if you compile XLA and dump the
594
+ optimized HLO for some graph, you should be able to run it again on the
595
+ same device with the same build of XLA.
596
+ """
597
+ xla_backend_optimization_level: builtins.int
598
+ """Numerical optimization level for the XLA compiler backend; the specific
599
+ interpretation of this value is left to the backends.
600
+ """
601
+ xla_embed_ir_in_executable: builtins.bool
602
+ """Embed the compiler IR as a string in the executable."""
603
+ xla_eliminate_hlo_implicit_broadcast: builtins.bool
604
+ """Eliminate implicit broadcasts when lowering user computations to HLO
605
+ instructions; use explicit broadcast instead.
606
+ """
607
+ xla_cpu_multi_thread_eigen: builtins.bool
608
+ """When generating calls to Eigen in the CPU backend, use multi-threaded Eigen
609
+ mode.
610
+ """
611
+ xla_gpu_cuda_data_dir: builtins.str
612
+ """Path to directory with cuda/ptx tools and libraries."""
613
+ xla_gpu_ftz: builtins.bool
614
+ """Enable flush-to-zero semantics in the GPU backend."""
615
+ xla_llvm_enable_alias_scope_metadata: builtins.bool
616
+ """If true, in LLVM-based backends, emit !alias.scope metadata in
617
+ generated IR.
618
+ """
619
+ xla_llvm_enable_noalias_metadata: builtins.bool
620
+ """If true, in LLVM-based backends, emit !noalias metadata in the
621
+ generated IR.
622
+ """
623
+ xla_llvm_enable_invariant_load_metadata: builtins.bool
624
+ """If true, in LLVM-based backends, emit !invariant.load metadata in
625
+ the generated IR.
626
+ """
627
+ xla_llvm_disable_expensive_passes: builtins.bool
628
+ """If true, a set of expensive LLVM optimization passes will not be run."""
629
+ xla_test_all_output_layouts: builtins.bool
630
+ """This is used by ClientLibraryTestBase::ComputeAndCompare*. If true, the
631
+ computation will run n! times with all permunations of layouts for the
632
+ output shape in rank n. For example, with a 3D shape, all permutations of
633
+ the set {0, 1, 2} are tried.
634
+ """
635
+ xla_test_all_input_layouts: builtins.bool
636
+ """This is used by ClientLibraryTestBase::ComputeAndCompare*. If true, the
637
+ computation will run for all permunations of layouts of all input
638
+ arguments. For example, with 2 input arguments in 2D and 4D shapes, the
639
+ computation will run 2! * 4! times.
640
+ """
641
+ xla_hlo_graph_sharding_color: builtins.bool
642
+ """Assign colors based on sharding information when generating the Graphviz
643
+ HLO graph.
644
+ """
645
+ xla_cpu_use_mkl_dnn: builtins.bool
646
+ """Generate calls to MKL-DNN in the CPU backend."""
647
+ xla_gpu_enable_fast_min_max: builtins.bool
648
+ """When true we lower the Minimum and Maximum hlos in the GPU backend such
649
+ that Min(NotNaN, NaN) = Min(NaN, NotNaN) = NotNaN. In other words, if flag
650
+ this is true we don't propagate NaNs through Min and Max.
651
+
652
+ Note, this does not correspond to the exact same behavior as the cpu flag
653
+ above!
654
+ """
655
+ xla_allow_excess_precision: builtins.bool
656
+ """Allows xla to increase the output precision of floating point operations
657
+ and all floating-point conversions to be simplified, including those
658
+ that affect the numerics. The `FloatNormalization` pass inserts many
659
+ `f32 -> bf16 -> f32` conversion pairs. These are not removed by the
660
+ `AlgebraicSimplifier`, as that will only simplify conversions that are
661
+ no-ops, e.g. `bf16 -> f32 -> bf16`. Removing these improves accuracy.
662
+ """
663
+ xla_gpu_crash_on_verification_failures: builtins.bool
664
+ """Crashes the program when any kind of verification fails, instead of just
665
+ logging the failures. One example is cross checking of convolution results
666
+ among different algorithms.
667
+ """
668
+ xla_gpu_autotune_level: builtins.int
669
+ """0: Disable gemm and convolution autotuning.
670
+ 1: Enable autotuning, but disable correctness checking.
671
+ 2: Also set output buffers to random numbers during autotuning.
672
+ 3: Also reset output buffers to random numbers after autotuning each
673
+ algorithm.
674
+ 4+: Also check for correct outputs and for out-of-bounds reads/writes.
675
+
676
+ Default: 4.
677
+ """
678
+ xla_force_host_platform_device_count: builtins.int
679
+ """Force the host platform to pretend that there are these many host
680
+ "devices". All these devices are backed by the same threadpool. Defaults
681
+ to 1.
682
+
683
+ Setting this to anything other than 1 can increase overhead from context
684
+ switching but we let the user override this behavior to help run tests on
685
+ the host that run models in parallel across multiple devices.
686
+ """
687
+ xla_gpu_disable_gpuasm_optimizations: builtins.bool
688
+ """If set to true XLA:GPU invokes `ptxas` with -O0 (default is -O3)."""
689
+ xla_gpu_shape_checks: global___DebugOptions.ShapeChecks.ValueType
690
+ xla_hlo_evaluator_use_fast_path: builtins.bool
691
+ """Enable fast math with eigen in the HLO evaluator."""
692
+ xla_allow_scalar_index_dynamic_ops: builtins.bool
693
+ """Temporary option to allow support for both the R1 and the scalar index
694
+ versions of DynamicSlice and DynamicUpdateSlice. Only used for testing.
695
+ """
696
+ xla_step_marker_location: global___DebugOptions.StepMarkerLocation.ValueType
697
+ """Option to emit a target-specific marker to indicate the start of a training
698
+ step. The location of the marker (if any) is determined by the option
699
+ value.
700
+ """
701
+ xla_dump_to: builtins.str
702
+ """
703
+ BEGIN flags controlling dumping HLO modules for debugging.
704
+
705
+ When dumping is enabled, HLO modules dumped at the very beginning and end
706
+ of compilation, and optionally also during the pass pipeline.
707
+
708
+ In general, if you set one of these flags, we will try to infer reasonable
709
+ defaults for the others. For example:
710
+
711
+ * Setting --xla_dump_to=/tmp/foo without specifying a format
712
+ with --xla_dump_hlo_as_* will turn on --xla_dump_hlo_as_text.
713
+
714
+ * Setting --xla_dump_hlo_as_text without specifying --xla_dump_to will
715
+ dump to stdout.
716
+
717
+ Directory to dump into.
718
+ """
719
+ xla_dump_hlo_module_re: builtins.str
720
+ """If specified, will only dump modules which match this regexp."""
721
+ xla_dump_hlo_pass_re: builtins.str
722
+ """If this flag is specified, will also dump HLO before and after passes that
723
+ match this regular expression. Set to .* to dump before/after all passes.
724
+ """
725
+ xla_dump_hlo_as_text: builtins.bool
726
+ """Specifies the format that HLO is dumped in. Multiple of these may be
727
+ specified.
728
+ """
729
+ xla_dump_hlo_as_proto: builtins.bool
730
+ xla_dump_hlo_as_dot: builtins.bool
731
+ xla_dump_hlo_as_url: builtins.bool
732
+ xla_dump_hlo_as_html: builtins.bool
733
+ """Dump HLO graphs as an HTML (DOT -> SVG inlined in HTML)"""
734
+ xla_dump_fusion_visualization: builtins.bool
735
+ """Dump the visualization of the fusion progress."""
736
+ xla_dump_hlo_snapshots: builtins.bool
737
+ """If true, every time an HLO module is run, we will dump an HloSnapshot
738
+ (essentially, a serialized module plus its inputs) to the --xla_dump_to
739
+ directory.
740
+ """
741
+ xla_dump_include_timestamp: builtins.bool
742
+ """Include a timestamp in the dumped filenames."""
743
+ xla_dump_max_hlo_modules: builtins.int
744
+ """Max number of hlo module dumps in a directory. Set to < 0 for unbounded."""
745
+ xla_dump_module_metadata: builtins.bool
746
+ """Dump HloModuleMetadata as a text proto for each HLO module."""
747
+ xla_dump_compress_protos: builtins.bool
748
+ """GZip-compress protos dumped via --xla_dump_hlo_as_proto."""
749
+ xla_dump_hlo_as_long_text: builtins.bool
750
+ """Dump HLO in long text format. Ignored unless xla_dump_hlo_as_text is true."""
751
+ xla_gpu_force_conv_nchw: builtins.bool
752
+ """
753
+ END flags controlling dumping HLO modules.
754
+
755
+ Overrides for XLA GPU's convolution layout heuristic.
756
+ """
757
+ xla_gpu_force_conv_nhwc: builtins.bool
758
+ xla_gpu_dump_llvmir: builtins.bool
759
+ """Whether to dump llvm ir when compiling to ptx."""
760
+ xla_dump_enable_mlir_pretty_form: builtins.bool
761
+ """Whether to dump mlir using pretty print form."""
762
+ xla_gpu_algorithm_denylist_path: builtins.str
763
+ """Denylist for cuDNN convolutions."""
764
+ xla_tpu_detect_nan: builtins.bool
765
+ """Debug options that trigger execution errors when NaN or Inf are detected."""
766
+ xla_tpu_detect_inf: builtins.bool
767
+ xla_cpu_enable_xprof_traceme: builtins.bool
768
+ """True if TraceMe annotations are enabled for XLA:CPU."""
769
+ xla_gpu_unsafe_fallback_to_driver_on_ptxas_not_found: builtins.bool
770
+ """It is usually preferable to not fallback to the driver; it can consume more
771
+ memory, or have bugs.
772
+ """
773
+ xla_gpu_asm_extra_flags: builtins.str
774
+ """Extra parameters to pass the GPU assembler."""
775
+ xla_multiheap_size_constraint_per_heap: builtins.int
776
+ """Per-heap size constraint. New heaps will be created if per-heap max size is
777
+ reached.
778
+ """
779
+ xla_detailed_logging: builtins.bool
780
+ """Enable detailed logging into vlog. If this is disabled, no
781
+ compilation summary will be printed in the end of computation.
782
+ """
783
+ xla_enable_dumping: builtins.bool
784
+ """Enable HLO dumping. If this is disabled, no HLO modules will be dumped."""
785
+ xla_gpu_force_compilation_parallelism: builtins.int
786
+ """Overrides normal multi-threaded compilation setting to use this many
787
+ threads. Setting to 0 (the default value) means no enforcement.
788
+ """
789
+ xla_gpu_enable_llvm_module_compilation_parallelism: builtins.bool
790
+ xla_gpu_deterministic_ops: builtins.bool
791
+ """Guarantees run-to-run determinism.
792
+ This flag implies --xla_gpu_exclude_nondeterministic_ops and in addition
793
+ disables autotuning.
794
+ """
795
+ xla_gpu_all_reduce_combine_threshold_bytes: builtins.int
796
+ """Size threshold (in bytes) for the GPU collective combiners."""
797
+ xla_gpu_all_gather_combine_threshold_bytes: builtins.int
798
+ xla_gpu_reduce_scatter_combine_threshold_bytes: builtins.int
799
+ xla_gpu_enable_all_gather_combine_by_dim: builtins.bool
800
+ """Combine all-gather/scatter-reduce ops with the same dimension or
801
+ irrespective of their dimension.
802
+ """
803
+ xla_gpu_enable_reduce_scatter_combine_by_dim: builtins.bool
804
+ xla_gpu_enable_reassociation_for_converted_ar: builtins.bool
805
+ """Enable allreduce reassociation on allreduces that are converted to a wider
806
+ type. The resulting allreduce will be promoted to a wider-typed allreduce.
807
+ """
808
+ xla_gpu_all_reduce_blueconnect_num_devices_per_host: builtins.int
809
+ """Number of devices per host for first stage of BlueConnect decomposition
810
+ pass. The pass will attempt to decompose all-reduces ops into a
811
+ ReduceScatter-AllReduce-AllGather sequence, with the initial ReduceScatter
812
+ being performed over all of the devices in the same host. Set to < 1 to
813
+ disable all-reduce decomposition.
814
+ """
815
+ xla_gpu_enable_while_loop_reduce_scatter_code_motion: builtins.bool
816
+ """Enable hoisting of reduce-scatter out of while loops."""
817
+ xla_gpu_collective_inflation_factor: builtins.int
818
+ """Inflate collective cost by running each collective multiple times."""
819
+ xla_llvm_force_inline_before_split: builtins.bool
820
+ """Whether to force inline before llvm module split to get a more balanced
821
+ splits for parallel compilation.
822
+ """
823
+ xla_gpu_enable_cudnn_frontend: builtins.bool
824
+ """Whether to use the cuDNN frontend API for convolutions when possible."""
825
+ xla_gpu_enable_cudnn_fmha: builtins.bool
826
+ xla_gpu_fused_attention_use_cudnn_rng: builtins.bool
827
+ xla_gpu_enable_cudnn_layer_norm: builtins.bool
828
+ """Rewrite layer norm patterns into cuDNN library calls."""
829
+ xla_dump_disable_metadata: builtins.bool
830
+ """Disable dumping metadata in HLO dumps."""
831
+ xla_dump_hlo_pipeline_re: builtins.str
832
+ """If this flag is specified, will only dump HLO before and after passes in
833
+ the pass pipeline that matches this regular expression. Default empty value
834
+ enables dumping in all pipelines.
835
+ """
836
+ xla_gpu_strict_conv_algorithm_picker: builtins.bool
837
+ """If true, abort immediately when conv algorithm picker fails, rather than
838
+ logging a warning and proceeding with fallback.
839
+ """
840
+ xla_gpu_enable_custom_fusions: builtins.bool
841
+ """If true, XLA will try to pattern match subgraphs of HLO operations into
842
+ custom fusions registered in the current process (pre-compiled hand written
843
+ kernels, e.g. various GEMM fusions writtent in CUTLASS).
844
+ """
845
+ xla_gpu_enable_custom_fusions_re: builtins.str
846
+ """A regular expression enabling only a subset of custom fusions. Enabled only
847
+ if `xla_gpu_enable_custom_fusion` set to true.
848
+ """
849
+ xla_gpu_enable_dynamic_slice_fusion: builtins.bool
850
+ """Enables address computation fusion to optimize dynamic-slice and
851
+ dynamic-update-slice operations around library calls.
852
+ """
853
+ xla_gpu_nccl_termination_timeout_seconds: builtins.int
854
+ """Timeout in seconds before terminating jobs that are stuck in a NCCL
855
+ Rendezvous. Negative value disables the timeout and will not terminate.
856
+ """
857
+ xla_gpu_enable_shared_constants: builtins.bool
858
+ """Enables shared constants for XLA/GPU. This allows large constants to be
859
+ shared among multiple GPU executables.
860
+ """
861
+ xla_gpu_enable_cublaslt: builtins.bool
862
+ """Whether to use cuBLASLt for GEMMs on GPUs."""
863
+ xla_gpu_graph_min_graph_size: builtins.int
864
+ """This number determines how many moved instructions like fusion kernels are
865
+ required for a region to be captured as a function to be launched as a GPU
866
+ graph.
867
+ """
868
+ xla_gpu_graph_enable_concurrent_region: builtins.bool
869
+ """Identify concurrent regions in GPU graphs and execute them concurrently."""
870
+ xla_gpu_redzone_scratch_max_megabytes: builtins.int
871
+ """Size threshold (in megabytes) for the GPU redzone scratch allocator."""
872
+ xla_gpu_redzone_padding_bytes: builtins.int
873
+ """Amount of padding the redzone allocator will put on one side of each buffer
874
+ it allocates. (So the buffer's total size will be increased by 2x this
875
+ value.)
876
+
877
+ Higher values make it more likely that we'll catch an out-of-bounds read or
878
+ write. Smaller values consume less memory during autotuning. Note that a
879
+ fused cudnn conv has up to 6 total buffers (4 inputs, 1 output, and 1
880
+ scratch), so this can be multiplied by quite a lot.
881
+ """
882
+ xla_cpu_use_acl: builtins.bool
883
+ """Generate calls to Arm Compute Library in the CPU backend."""
884
+ xla_cpu_strict_dot_conv_math: builtins.bool
885
+ """By default, XLA:CPU will run fp16 dot/conv as fp32, as this is generally
886
+ (much) faster on our hardware. Set this flag to disable this behavior.
887
+ """
888
+ xla_gpu_use_runtime_fusion: builtins.bool
889
+ """An option to enable using cuDNN runtime compiled fusion kernels which is
890
+ available and recommended for Ampere+ GPUs.
891
+ """
892
+ xla_dump_latency_hiding_schedule: builtins.bool
893
+ xla_cpu_enable_mlir_tiling_and_fusion: builtins.bool
894
+ """By default, MLIR lowering will use Linalg elementwise fusion. If this flag
895
+ is enabled, the pipeline will use tiling, fusion, peeling, vectorization
896
+ instead.
897
+ """
898
+ xla_cpu_enable_custom_matmul_tiling: builtins.bool
899
+ """XLA:CPU-Next tiling parameters for matmul."""
900
+ xla_cpu_matmul_tiling_m_dim: builtins.int
901
+ xla_cpu_matmul_tiling_n_dim: builtins.int
902
+ xla_cpu_matmul_tiling_k_dim: builtins.int
903
+ xla_cpu_enable_mlir_fusion_outlining: builtins.bool
904
+ xla_cpu_enable_experimental_deallocation: builtins.bool
905
+ """If set, use the experimental deallocation pass from mlir-hlo."""
906
+ xla_gpu_enable_latency_hiding_scheduler: builtins.bool
907
+ xla_gpu_enable_highest_priority_async_stream: builtins.bool
908
+ xla_gpu_enable_analytical_latency_estimator: builtins.bool
909
+ xla_gpu_lhs_enable_gpu_async_tracker: builtins.bool
910
+ xla_gpu_pgle_profile_file_or_directory_path: builtins.str
911
+ xla_gpu_memory_limit_slop_factor: builtins.int
912
+ xla_gpu_enable_pipelined_collectives: builtins.bool
913
+ xla_gpu_enable_pipelined_all_reduce: builtins.bool
914
+ xla_gpu_enable_pipelined_all_gather: builtins.bool
915
+ xla_gpu_enable_pipelined_reduce_scatter: builtins.bool
916
+ xla_gpu_enable_pipelined_p2p: builtins.bool
917
+ xla_gpu_run_post_layout_collective_pipeliner: builtins.bool
918
+ xla_gpu_collective_permute_decomposer_threshold: builtins.int
919
+ """The minimum data size in bytes to trigger collective-permute-decomposer
920
+ transformation.
921
+ """
922
+ xla_partitioning_algorithm: global___DebugOptions.PartitioningAlgorithm.ValueType
923
+ """The partitioning algorithm to be used in the PartitionAssignment pass."""
924
+ xla_gpu_enable_triton_gemm: builtins.bool
925
+ xla_gpu_enable_cudnn_int8x32_convolution_reordering: builtins.bool
926
+ xla_gpu_triton_gemm_any: builtins.bool
927
+ """Creates triton fusion for all supported gemms.
928
+ To make sure only triton gemm is chosen by the autotuner run with
929
+ `xla_gpu_cublas_fallback` set to false.
930
+ """
931
+ xla_gpu_exhaustive_tiling_search: builtins.bool
932
+ xla_gpu_enable_priority_fusion: builtins.bool
933
+ xla_gpu_dump_autotune_results_to: builtins.str
934
+ """File to write autotune results to. It will be a binary file unless the name
935
+ ends with .txt or .textproto. Warning: The results are written at every
936
+ compilation, possibly multiple times per process. This only works on CUDA.
937
+ """
938
+ xla_gpu_load_autotune_results_from: builtins.str
939
+ """File to load autotune results from. It will be considered a binary file
940
+ unless the name ends with .txt or .textproto. At most one loading will
941
+ happen during the lifetime of one process, even if the first one is
942
+ unsuccessful or different file paths are passed here. This only works on
943
+ CUDA.
944
+ """
945
+ xla_gpu_target_config_filename: builtins.str
946
+ """Description of the target platform in GpuTargetConfigProto format; if
947
+ provided, deviceless compilation is assumed, and the current device is
948
+ ignored.
949
+ """
950
+ xla_gpu_auto_spmd_partitioning_memory_budget_gb: builtins.int
951
+ """Memory budget in GB per device for AutoSharding."""
952
+ xla_gpu_auto_spmd_partitioning_memory_budget_ratio: builtins.float
953
+ """See the definition of the
954
+ xla_gpu_auto_spmd_partitioning_memory_budget_ratio flag for the meaning of
955
+ this field.
956
+ """
957
+ xla_gpu_triton_gemm_disable_reduced_precision_reduction: builtins.bool
958
+ xla_gpu_triton_fusion_level: builtins.int
959
+ xla_gpu_dump_autotuned_gemm_fusions: builtins.bool
960
+ xla_gpu_override_gemm_autotuner: builtins.str
961
+ xla_gpu_copy_insertion_use_region_analysis: builtins.bool
962
+ xla_gpu_collect_cost_model_stats: builtins.bool
963
+ """If true, each fusion instruction will have a cost model runtime estimate in
964
+ backend config after compilation.
965
+ """
966
+ xla_gpu_enable_split_k_autotuning: builtins.bool
967
+ xla_gpu_enable_reduction_epilogue_fusion: builtins.bool
968
+ """Whether reduction epilogue fusion is enabled in fusion passes."""
969
+ xla_gpu_enable_nccl_clique_optimization: builtins.bool
970
+ """Allow early return when acquiring NCCL cliques."""
971
+ xla_gpu_mock_custom_calls: builtins.bool
972
+ """Replace custom calls with noop operations."""
973
+ xla_gpu_cublas_fallback: builtins.bool
974
+ """Allow Triton GEMM autotuning to fall back to cuBLAS when that is
975
+ faster.
976
+ """
977
+ xla_gpu_enable_while_loop_double_buffering: builtins.bool
978
+ """Enable double buffering for loops."""
979
+ xla_gpu_enable_while_loop_unrolling: global___DebugOptions.WhileLoopUnrolling.ValueType
980
+ """Determine the while loop unrolling scheme."""
981
+ xla_gpu_ensure_minor_dot_contraction_dims: builtins.bool
982
+ """Change the layout of the second triton dot operand to be column major.
983
+ Only works for (bf16 x bf16) -> bf16.
984
+ """
985
+ xla_gpu_filter_kernels_spilling_registers_on_autotuning: builtins.bool
986
+ """Filter out kernels that spill registers during autotuning."""
987
+ xla_debug_buffer_assignment_show_max: builtins.int
988
+ """Maximum number of buffers to print when debugging buffer assignment."""
989
+ xla_gpu_llvm_verification_level: builtins.int
990
+ xla_gpu_enable_cub_radix_sort: builtins.bool
991
+ """Enable radix sort using CUB."""
992
+ xla_gpu_threshold_for_windowed_einsum_mib: builtins.int
993
+ """Threshold to enable windowed einsum (collective matmul) in MB."""
994
+ xla_gpu_enable_triton_hopper: builtins.bool
995
+ """Enables currently disabled features within Triton for Hopper."""
996
+ xla_gpu_enable_nccl_user_buffers: builtins.bool
997
+ """Enable NCCL user buffers."""
998
+ xla_gpu_enable_nccl_comm_splitting: builtins.bool
999
+ """Enable NCCL communicator splitting."""
1000
+ xla_gpu_enable_nccl_per_stream_comms: builtins.bool
1001
+ """Enable NCCL per stream communicators."""
1002
+ xla_gpu_enable_libnvptxcompiler: builtins.bool
1003
+ """If enabled, uses the libnvptxcompiler library to compile PTX to cuBIN."""
1004
+ xla_gpu_enable_dot_strength_reduction: builtins.bool
1005
+ xla_gpu_multi_streamed_windowed_einsum: builtins.bool
1006
+ """Whether to use multiple compute streams to run windowed einsum."""
1007
+ xla_gpu_enable_bf16_6way_gemm: builtins.bool
1008
+ """If enabled, uses bf16_6way gemm to compute F32 gemm."""
1009
+ xla_gpu_enable_bf16_3way_gemm: builtins.bool
1010
+ """If enabled, uses bf16_3way gemm to compute F32 gemm."""
1011
+ xla_gpu_nccl_collective_max_nchannels: builtins.int
1012
+ """Specify the maximum number of channels(SMs) NCCL
1013
+ will use for collective operations.
1014
+ """
1015
+ xla_gpu_nccl_p2p_max_nchannels: builtins.int
1016
+ """Specify the maximum number of channels(SMs) NCCL
1017
+ will use for p2p operations.
1018
+ """
1019
+ xla_gpu_mlir_emitter_level: builtins.int
1020
+ """Choose the level of mlir emitters that are enabled.
1021
+ Current levels:
1022
+ 0: Disabled.
1023
+ 1: Loop emitter
1024
+ 2: + Loop-like emitters
1025
+ 3: + Transpose
1026
+ 4: + Reduce
1027
+ """
1028
+ xla_gpu_gemm_rewrite_size_threshold: builtins.int
1029
+ """Threshold to rewrite matmul to cuBLAS or Triton (minimum combined number of
1030
+ elements of both matrices in non-batch dimensions to be considered for a
1031
+ rewrite).
1032
+ """
1033
+ xla_gpu_require_complete_aot_autotune_results: builtins.bool
1034
+ """If true, will require complete AOT autotuning results; in the case of
1035
+ missing AOT result, the model will not be compiled or executed, a
1036
+ `NotFound` error will be returned.
1037
+ """
1038
+ xla_gpu_cudnn_gemm_fusion_level: builtins.int
1039
+ """Let GEMM fusion autotuning probe cuDNN as a backend.
1040
+ Current levels:
1041
+ 0: Disabled.
1042
+ 1: Fusions of GEMM, elementwise, transpose/reshape operations.
1043
+ 2: + Broadcasts, slicing.
1044
+ 3: + Nontrivial noncontracting dimension reshapes/transposes.
1045
+ """
1046
+ xla_gpu_use_memcpy_local_p2p: builtins.bool
1047
+ """This instructs the runtime whether to use
1048
+ memcpy for p2p communication when source and
1049
+ target are located within a node(nvlink).
1050
+ """
1051
+ xla_gpu_autotune_max_solutions: builtins.int
1052
+ """If non-zero, limits the number of solutions to be used by GEMM autotuner.
1053
+ This might be useful if underlying math library returns too many GEMM
1054
+ solutions.
1055
+ """
1056
+ xla_dump_large_constants: builtins.bool
1057
+ """If true, large constants will be printed out when dumping HLOs."""
1058
+ xla_gpu_verify_triton_fusion_numerics: builtins.bool
1059
+ """If true, will verify that the numerical results of Triton fusions match
1060
+ the results of regular emitters.
1061
+ """
1062
+ xla_gpu_dump_autotune_logs_to: builtins.str
1063
+ """File to write autotune logs to. It will stored in txt format."""
1064
+ xla_reduce_window_rewrite_base_length: builtins.int
1065
+ """Base length to rewrite the reduce window to, no rewrite if set to 0."""
1066
+ xla_gpu_enable_host_memory_offloading: builtins.bool
1067
+ """If true, will enable host memory offloading on a device."""
1068
+ xla_gpu_exclude_nondeterministic_ops: builtins.bool
1069
+ """Excludes non-deterministic ops from compiled executables.
1070
+ Unlike --xla_gpu_deterministic_ops does not disable autotuning - the
1071
+ compilation itself can be non-deterministic.
1072
+ At present, the HLO op SelectAndScatter does not have a
1073
+ deterministic XLA:GPU implementation.
1074
+ Compilation errors out if SelectAndScatter is encountered.
1075
+ Scatter ops can non-deterministic by default; these get converted to
1076
+ a deterministic implementation.
1077
+ """
1078
+ xla_gpu_nccl_terminate_on_error: builtins.bool
1079
+ """If true, Nccl errors will terminate the process."""
1080
+ xla_gpu_shard_autotuning: builtins.bool
1081
+ xla_gpu_enable_approx_costly_collectives: builtins.bool
1082
+ xla_gpu_kernel_cache_file: builtins.str
1083
+ xla_gpu_unsafe_pipelined_loop_annotator: builtins.bool
1084
+ """Recognises rotate-right patterns (slice, slice, concat) within a while
1085
+ loop and labels the while loop as a pipelined while loop. This is an
1086
+ unsafe flag.
1087
+ """
1088
+ xla_gpu_per_fusion_autotune_cache_dir: builtins.str
1089
+ xla_cmd_buffer_trace_cache_size: builtins.int
1090
+ """The command buffer trace cache size, increasing the cache size may
1091
+ sometimes reduces the chances of doing command buffer tracing for
1092
+ updating command buffer instance.
1093
+ """
1094
+ xla_gpu_temp_buffer_use_separate_color: builtins.bool
1095
+ """Enable this flag will use a separate memory space color for
1096
+ temp buffer, and then will use separate memory allocator to allocate it,
1097
+ as there is no other memory allocation interference,
1098
+ it will allocate temp buffer to some fix address on every iteration,
1099
+ which is good for cuda-graph perf.
1100
+ """
1101
+ xla_syntax_sugar_async_ops: builtins.bool
1102
+ """This flag is used for controlling HLO dumping and NVTX marker. If turned
1103
+ on, both HLO dumping and NVTX marker will use syntactic sugar wrappers
1104
+ as op names, while the actual op names will be shown if turned off.
1105
+
1106
+ Here is an example HLO excerpt with the flag off:
1107
+
1108
+ async_computation {
1109
+ param_0 = f32[1,4,8]{1,0,2} parameter(0)
1110
+ ROOT all-to-all.3.1 = f32[1,4,8]{1,0,2} all-to-all(param_0),
1111
+ replica_groups={{0,1,2,3,4,5,6,7}}, dimensions={2}
1112
+ }
1113
+ ...
1114
+
1115
+ all-to-all-start =
1116
+ ((f32[1,4,8]{1,0,2}), f32[1,4,8]{1,0,2}) async-start(bitcast.24.0),
1117
+ calls=async_computation, backend_config={...}
1118
+ all-to-all-done = f32[1,4,8]{1,0,2} async-done(all-to-all-start)
1119
+
1120
+ and with the flag on:
1121
+
1122
+ all-to-all-start = ((f32[1,4,8]{1,0,2}), f32[1,4,8]{1,0,2})
1123
+ all-to-all-start(bitcast.24.0),
1124
+ replica_groups={{0,1,2,3,4,5,6,7}}, dimensions={2},
1125
+ backend_config={...}
1126
+ all-to-all-done = f32[1,4,8]{1,0,2} all-to-all-done(all-to-all-start)
1127
+ """
1128
+ xla_gpu_autotune_gemm_rtol: builtins.float
1129
+ """Relative precision for comparing different GEMM solutions"""
1130
+ xla_enable_command_buffers_during_profiling: builtins.bool
1131
+ """Allow launching command buffers while profiling active.
1132
+ When disabled, execute in op-by-op mode.
1133
+ TODO(b/355487968): Remove this option when validation complete.
1134
+ """
1135
+ xla_gpu_cudnn_gemm_max_plans: builtins.int
1136
+ """Limit for the number of kernel configurations (plans) to use during
1137
+ autotuning of cuDNN GEMM fusions. The more - the slower the autotuning
1138
+ but potentially higher the performance.
1139
+ """
1140
+ xla_gpu_enable_libnvjitlink: builtins.bool
1141
+ """If enabled, uses the libnvjitlink library for PTX compilation and linking"""
1142
+ xla_gpu_enable_triton_gemm_int4: builtins.bool
1143
+ """If enabled, generates triton gemm kernels for int4 inputs."""
1144
+ xla_gpu_async_dot: builtins.bool
1145
+ """If true, XLA will wrap `dot` operations into async computations in an
1146
+ effort to parallelize matrix operations.
1147
+ """
1148
+ xla_gpu_enable_pgle_accuracy_checker: builtins.bool
1149
+ """Enables strict PGLE checking. If an FDO profile is specified and latency
1150
+ hiding scheduler encounters missing instructions in the profile
1151
+ compilation will halt.
1152
+ """
1153
+ xla_gpu_executable_warn_stuck_timeout_seconds: builtins.int
1154
+ """Timeouts for RendezvousSingle stuck warning and termination."""
1155
+ xla_gpu_executable_terminate_timeout_seconds: builtins.int
1156
+ xla_experimental_ignore_channel_id: builtins.bool
1157
+ """Whether to ignore channel ids(including verifier channel id checks)
1158
+ for collectives in the given HLO.
1159
+ """
1160
+ @property
1161
+ def xla_disable_hlo_passes(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
1162
+ """List of HLO passes to disable/enable. These names must exactly match the
1163
+ pass names as specified by the HloPassInterface::name() method.
1164
+
1165
+ At least one of xla_disable_hlo_passes and xla_enable_hlo_passes_only must
1166
+ be empty.
1167
+ """
1168
+
1169
+ @property
1170
+ def xla_enable_hlo_passes_only(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]: ...
1171
+ @property
1172
+ def xla_gpu_ptx_file(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
1173
+ """Paths to files with ptx code."""
1174
+
1175
+ @property
1176
+ def xla_gpu_llvm_ir_file(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
1177
+ """Paths to files with LLVM code."""
1178
+
1179
+ @property
1180
+ def xla_gpu_disable_async_collectives(
1181
+ self,
1182
+ ) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[global___DebugOptions.CollectiveOpType.ValueType]: ...
1183
+ @property
1184
+ def xla_gpu_enable_command_buffer(
1185
+ self,
1186
+ ) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[global___DebugOptions.CommandBufferCmdType.ValueType]:
1187
+ """Determine the types of commands that are recorded into command buffers."""
1188
+
1189
+ @property
1190
+ def legacy_command_buffer_custom_call_targets(
1191
+ self,
1192
+ ) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
1193
+ """Custom call targets with legacy registry API (non FFI API),
1194
+ that support recording to command buffer custom command,
1195
+ i.e., custom call target supports cuda-graph capturing for CUDA devices.
1196
+ This flag is read if CUSTOM_CALL command type is recorded into
1197
+ command buffer.
1198
+ """
1199
+
1200
+ @property
1201
+ def xla_backend_extra_options(self) -> google.protobuf.internal.containers.ScalarMap[builtins.str, builtins.str]:
1202
+ """Next id: 331
1203
+
1204
+ Extra options to pass to the compilation backend (e.g. LLVM); specific
1205
+ interpretation of these values is left to the backend.
1206
+ """
1207
+
1208
+ def __init__(
1209
+ self,
1210
+ *,
1211
+ xla_cpu_enable_concurrency_optimized_scheduler: builtins.bool | None = ...,
1212
+ xla_cpu_enable_fast_math: builtins.bool | None = ...,
1213
+ xla_cpu_enable_fast_min_max: builtins.bool | None = ...,
1214
+ xla_cpu_fast_math_honor_division: builtins.bool | None = ...,
1215
+ xla_cpu_fast_math_honor_functions: builtins.bool | None = ...,
1216
+ xla_cpu_fast_math_honor_infs: builtins.bool | None = ...,
1217
+ xla_cpu_fast_math_honor_nans: builtins.bool | None = ...,
1218
+ xla_cpu_use_thunk_runtime: builtins.bool | None = ...,
1219
+ xla_cpu_parallel_codegen_split_count: builtins.int | None = ...,
1220
+ xla_cpu_prefer_vector_width: builtins.int | None = ...,
1221
+ xla_gpu_experimental_autotune_cache_mode: global___DebugOptions.AutotuneCacheMode.ValueType | None = ...,
1222
+ xla_gpu_experimental_disable_binary_libraries: builtins.bool | None = ...,
1223
+ xla_gpu_experimental_enable_triton_softmax_priority_fusion: builtins.bool | None = ...,
1224
+ xla_gpu_unsupported_enable_triton_gemm: builtins.bool | None = ...,
1225
+ xla_hlo_graph_addresses: builtins.bool | None = ...,
1226
+ xla_hlo_profile: builtins.bool | None = ...,
1227
+ xla_disable_hlo_passes: collections.abc.Iterable[builtins.str] | None = ...,
1228
+ xla_enable_hlo_passes_only: collections.abc.Iterable[builtins.str] | None = ...,
1229
+ xla_disable_all_hlo_passes: builtins.bool | None = ...,
1230
+ xla_backend_optimization_level: builtins.int | None = ...,
1231
+ xla_embed_ir_in_executable: builtins.bool | None = ...,
1232
+ xla_eliminate_hlo_implicit_broadcast: builtins.bool | None = ...,
1233
+ xla_cpu_multi_thread_eigen: builtins.bool | None = ...,
1234
+ xla_gpu_cuda_data_dir: builtins.str | None = ...,
1235
+ xla_gpu_ftz: builtins.bool | None = ...,
1236
+ xla_llvm_enable_alias_scope_metadata: builtins.bool | None = ...,
1237
+ xla_llvm_enable_noalias_metadata: builtins.bool | None = ...,
1238
+ xla_llvm_enable_invariant_load_metadata: builtins.bool | None = ...,
1239
+ xla_llvm_disable_expensive_passes: builtins.bool | None = ...,
1240
+ xla_test_all_output_layouts: builtins.bool | None = ...,
1241
+ xla_test_all_input_layouts: builtins.bool | None = ...,
1242
+ xla_hlo_graph_sharding_color: builtins.bool | None = ...,
1243
+ xla_cpu_use_mkl_dnn: builtins.bool | None = ...,
1244
+ xla_gpu_enable_fast_min_max: builtins.bool | None = ...,
1245
+ xla_allow_excess_precision: builtins.bool | None = ...,
1246
+ xla_gpu_crash_on_verification_failures: builtins.bool | None = ...,
1247
+ xla_gpu_autotune_level: builtins.int | None = ...,
1248
+ xla_force_host_platform_device_count: builtins.int | None = ...,
1249
+ xla_gpu_disable_gpuasm_optimizations: builtins.bool | None = ...,
1250
+ xla_gpu_shape_checks: global___DebugOptions.ShapeChecks.ValueType | None = ...,
1251
+ xla_hlo_evaluator_use_fast_path: builtins.bool | None = ...,
1252
+ xla_allow_scalar_index_dynamic_ops: builtins.bool | None = ...,
1253
+ xla_step_marker_location: global___DebugOptions.StepMarkerLocation.ValueType | None = ...,
1254
+ xla_dump_to: builtins.str | None = ...,
1255
+ xla_dump_hlo_module_re: builtins.str | None = ...,
1256
+ xla_dump_hlo_pass_re: builtins.str | None = ...,
1257
+ xla_dump_hlo_as_text: builtins.bool | None = ...,
1258
+ xla_dump_hlo_as_proto: builtins.bool | None = ...,
1259
+ xla_dump_hlo_as_dot: builtins.bool | None = ...,
1260
+ xla_dump_hlo_as_url: builtins.bool | None = ...,
1261
+ xla_dump_hlo_as_html: builtins.bool | None = ...,
1262
+ xla_dump_fusion_visualization: builtins.bool | None = ...,
1263
+ xla_dump_hlo_snapshots: builtins.bool | None = ...,
1264
+ xla_dump_include_timestamp: builtins.bool | None = ...,
1265
+ xla_dump_max_hlo_modules: builtins.int | None = ...,
1266
+ xla_dump_module_metadata: builtins.bool | None = ...,
1267
+ xla_dump_compress_protos: builtins.bool | None = ...,
1268
+ xla_dump_hlo_as_long_text: builtins.bool | None = ...,
1269
+ xla_gpu_force_conv_nchw: builtins.bool | None = ...,
1270
+ xla_gpu_force_conv_nhwc: builtins.bool | None = ...,
1271
+ xla_gpu_ptx_file: collections.abc.Iterable[builtins.str] | None = ...,
1272
+ xla_gpu_dump_llvmir: builtins.bool | None = ...,
1273
+ xla_dump_enable_mlir_pretty_form: builtins.bool | None = ...,
1274
+ xla_gpu_algorithm_denylist_path: builtins.str | None = ...,
1275
+ xla_tpu_detect_nan: builtins.bool | None = ...,
1276
+ xla_tpu_detect_inf: builtins.bool | None = ...,
1277
+ xla_cpu_enable_xprof_traceme: builtins.bool | None = ...,
1278
+ xla_gpu_unsafe_fallback_to_driver_on_ptxas_not_found: builtins.bool | None = ...,
1279
+ xla_gpu_asm_extra_flags: builtins.str | None = ...,
1280
+ xla_multiheap_size_constraint_per_heap: builtins.int | None = ...,
1281
+ xla_detailed_logging: builtins.bool | None = ...,
1282
+ xla_enable_dumping: builtins.bool | None = ...,
1283
+ xla_gpu_force_compilation_parallelism: builtins.int | None = ...,
1284
+ xla_gpu_enable_llvm_module_compilation_parallelism: builtins.bool | None = ...,
1285
+ xla_gpu_deterministic_ops: builtins.bool | None = ...,
1286
+ xla_gpu_llvm_ir_file: collections.abc.Iterable[builtins.str] | None = ...,
1287
+ xla_gpu_disable_async_collectives: (
1288
+ collections.abc.Iterable[global___DebugOptions.CollectiveOpType.ValueType] | None
1289
+ ) = ...,
1290
+ xla_gpu_all_reduce_combine_threshold_bytes: builtins.int | None = ...,
1291
+ xla_gpu_all_gather_combine_threshold_bytes: builtins.int | None = ...,
1292
+ xla_gpu_reduce_scatter_combine_threshold_bytes: builtins.int | None = ...,
1293
+ xla_gpu_enable_all_gather_combine_by_dim: builtins.bool | None = ...,
1294
+ xla_gpu_enable_reduce_scatter_combine_by_dim: builtins.bool | None = ...,
1295
+ xla_gpu_enable_reassociation_for_converted_ar: builtins.bool | None = ...,
1296
+ xla_gpu_all_reduce_blueconnect_num_devices_per_host: builtins.int | None = ...,
1297
+ xla_gpu_enable_while_loop_reduce_scatter_code_motion: builtins.bool | None = ...,
1298
+ xla_gpu_collective_inflation_factor: builtins.int | None = ...,
1299
+ xla_llvm_force_inline_before_split: builtins.bool | None = ...,
1300
+ xla_gpu_enable_cudnn_frontend: builtins.bool | None = ...,
1301
+ xla_gpu_enable_cudnn_fmha: builtins.bool | None = ...,
1302
+ xla_gpu_fused_attention_use_cudnn_rng: builtins.bool | None = ...,
1303
+ xla_gpu_enable_cudnn_layer_norm: builtins.bool | None = ...,
1304
+ xla_dump_disable_metadata: builtins.bool | None = ...,
1305
+ xla_dump_hlo_pipeline_re: builtins.str | None = ...,
1306
+ xla_gpu_strict_conv_algorithm_picker: builtins.bool | None = ...,
1307
+ xla_gpu_enable_custom_fusions: builtins.bool | None = ...,
1308
+ xla_gpu_enable_custom_fusions_re: builtins.str | None = ...,
1309
+ xla_gpu_enable_dynamic_slice_fusion: builtins.bool | None = ...,
1310
+ xla_gpu_nccl_termination_timeout_seconds: builtins.int | None = ...,
1311
+ xla_gpu_enable_shared_constants: builtins.bool | None = ...,
1312
+ xla_gpu_enable_cublaslt: builtins.bool | None = ...,
1313
+ xla_gpu_enable_command_buffer: (
1314
+ collections.abc.Iterable[global___DebugOptions.CommandBufferCmdType.ValueType] | None
1315
+ ) = ...,
1316
+ xla_gpu_graph_min_graph_size: builtins.int | None = ...,
1317
+ xla_gpu_graph_enable_concurrent_region: builtins.bool | None = ...,
1318
+ xla_gpu_redzone_scratch_max_megabytes: builtins.int | None = ...,
1319
+ xla_gpu_redzone_padding_bytes: builtins.int | None = ...,
1320
+ xla_cpu_use_acl: builtins.bool | None = ...,
1321
+ xla_cpu_strict_dot_conv_math: builtins.bool | None = ...,
1322
+ xla_gpu_use_runtime_fusion: builtins.bool | None = ...,
1323
+ xla_dump_latency_hiding_schedule: builtins.bool | None = ...,
1324
+ xla_cpu_enable_mlir_tiling_and_fusion: builtins.bool | None = ...,
1325
+ xla_cpu_enable_custom_matmul_tiling: builtins.bool | None = ...,
1326
+ xla_cpu_matmul_tiling_m_dim: builtins.int | None = ...,
1327
+ xla_cpu_matmul_tiling_n_dim: builtins.int | None = ...,
1328
+ xla_cpu_matmul_tiling_k_dim: builtins.int | None = ...,
1329
+ xla_cpu_enable_mlir_fusion_outlining: builtins.bool | None = ...,
1330
+ xla_cpu_enable_experimental_deallocation: builtins.bool | None = ...,
1331
+ xla_gpu_enable_latency_hiding_scheduler: builtins.bool | None = ...,
1332
+ xla_gpu_enable_highest_priority_async_stream: builtins.bool | None = ...,
1333
+ xla_gpu_enable_analytical_latency_estimator: builtins.bool | None = ...,
1334
+ xla_gpu_lhs_enable_gpu_async_tracker: builtins.bool | None = ...,
1335
+ xla_gpu_pgle_profile_file_or_directory_path: builtins.str | None = ...,
1336
+ xla_gpu_memory_limit_slop_factor: builtins.int | None = ...,
1337
+ xla_gpu_enable_pipelined_collectives: builtins.bool | None = ...,
1338
+ xla_gpu_enable_pipelined_all_reduce: builtins.bool | None = ...,
1339
+ xla_gpu_enable_pipelined_all_gather: builtins.bool | None = ...,
1340
+ xla_gpu_enable_pipelined_reduce_scatter: builtins.bool | None = ...,
1341
+ xla_gpu_enable_pipelined_p2p: builtins.bool | None = ...,
1342
+ xla_gpu_run_post_layout_collective_pipeliner: builtins.bool | None = ...,
1343
+ xla_gpu_collective_permute_decomposer_threshold: builtins.int | None = ...,
1344
+ xla_partitioning_algorithm: global___DebugOptions.PartitioningAlgorithm.ValueType | None = ...,
1345
+ xla_gpu_enable_triton_gemm: builtins.bool | None = ...,
1346
+ xla_gpu_enable_cudnn_int8x32_convolution_reordering: builtins.bool | None = ...,
1347
+ xla_gpu_triton_gemm_any: builtins.bool | None = ...,
1348
+ xla_gpu_exhaustive_tiling_search: builtins.bool | None = ...,
1349
+ xla_gpu_enable_priority_fusion: builtins.bool | None = ...,
1350
+ xla_gpu_dump_autotune_results_to: builtins.str | None = ...,
1351
+ xla_gpu_load_autotune_results_from: builtins.str | None = ...,
1352
+ xla_gpu_target_config_filename: builtins.str | None = ...,
1353
+ xla_gpu_auto_spmd_partitioning_memory_budget_gb: builtins.int | None = ...,
1354
+ xla_gpu_auto_spmd_partitioning_memory_budget_ratio: builtins.float | None = ...,
1355
+ xla_gpu_triton_gemm_disable_reduced_precision_reduction: builtins.bool | None = ...,
1356
+ xla_gpu_triton_fusion_level: builtins.int | None = ...,
1357
+ xla_gpu_dump_autotuned_gemm_fusions: builtins.bool | None = ...,
1358
+ xla_gpu_override_gemm_autotuner: builtins.str | None = ...,
1359
+ xla_gpu_copy_insertion_use_region_analysis: builtins.bool | None = ...,
1360
+ xla_gpu_collect_cost_model_stats: builtins.bool | None = ...,
1361
+ xla_gpu_enable_split_k_autotuning: builtins.bool | None = ...,
1362
+ xla_gpu_enable_reduction_epilogue_fusion: builtins.bool | None = ...,
1363
+ xla_gpu_enable_nccl_clique_optimization: builtins.bool | None = ...,
1364
+ xla_gpu_mock_custom_calls: builtins.bool | None = ...,
1365
+ xla_gpu_cublas_fallback: builtins.bool | None = ...,
1366
+ xla_gpu_enable_while_loop_double_buffering: builtins.bool | None = ...,
1367
+ xla_gpu_enable_while_loop_unrolling: global___DebugOptions.WhileLoopUnrolling.ValueType | None = ...,
1368
+ xla_gpu_ensure_minor_dot_contraction_dims: builtins.bool | None = ...,
1369
+ xla_gpu_filter_kernels_spilling_registers_on_autotuning: builtins.bool | None = ...,
1370
+ xla_debug_buffer_assignment_show_max: builtins.int | None = ...,
1371
+ xla_gpu_llvm_verification_level: builtins.int | None = ...,
1372
+ xla_gpu_enable_cub_radix_sort: builtins.bool | None = ...,
1373
+ xla_gpu_threshold_for_windowed_einsum_mib: builtins.int | None = ...,
1374
+ xla_gpu_enable_triton_hopper: builtins.bool | None = ...,
1375
+ xla_gpu_enable_nccl_user_buffers: builtins.bool | None = ...,
1376
+ xla_gpu_enable_nccl_comm_splitting: builtins.bool | None = ...,
1377
+ xla_gpu_enable_nccl_per_stream_comms: builtins.bool | None = ...,
1378
+ xla_gpu_enable_libnvptxcompiler: builtins.bool | None = ...,
1379
+ xla_gpu_enable_dot_strength_reduction: builtins.bool | None = ...,
1380
+ xla_gpu_multi_streamed_windowed_einsum: builtins.bool | None = ...,
1381
+ xla_gpu_enable_bf16_6way_gemm: builtins.bool | None = ...,
1382
+ xla_gpu_enable_bf16_3way_gemm: builtins.bool | None = ...,
1383
+ xla_gpu_nccl_collective_max_nchannels: builtins.int | None = ...,
1384
+ xla_gpu_nccl_p2p_max_nchannels: builtins.int | None = ...,
1385
+ xla_gpu_mlir_emitter_level: builtins.int | None = ...,
1386
+ xla_gpu_gemm_rewrite_size_threshold: builtins.int | None = ...,
1387
+ xla_gpu_require_complete_aot_autotune_results: builtins.bool | None = ...,
1388
+ xla_gpu_cudnn_gemm_fusion_level: builtins.int | None = ...,
1389
+ xla_gpu_use_memcpy_local_p2p: builtins.bool | None = ...,
1390
+ xla_gpu_autotune_max_solutions: builtins.int | None = ...,
1391
+ xla_dump_large_constants: builtins.bool | None = ...,
1392
+ xla_gpu_verify_triton_fusion_numerics: builtins.bool | None = ...,
1393
+ xla_gpu_dump_autotune_logs_to: builtins.str | None = ...,
1394
+ xla_reduce_window_rewrite_base_length: builtins.int | None = ...,
1395
+ xla_gpu_enable_host_memory_offloading: builtins.bool | None = ...,
1396
+ xla_gpu_exclude_nondeterministic_ops: builtins.bool | None = ...,
1397
+ xla_gpu_nccl_terminate_on_error: builtins.bool | None = ...,
1398
+ xla_gpu_shard_autotuning: builtins.bool | None = ...,
1399
+ xla_gpu_enable_approx_costly_collectives: builtins.bool | None = ...,
1400
+ xla_gpu_kernel_cache_file: builtins.str | None = ...,
1401
+ xla_gpu_unsafe_pipelined_loop_annotator: builtins.bool | None = ...,
1402
+ xla_gpu_per_fusion_autotune_cache_dir: builtins.str | None = ...,
1403
+ xla_cmd_buffer_trace_cache_size: builtins.int | None = ...,
1404
+ xla_gpu_temp_buffer_use_separate_color: builtins.bool | None = ...,
1405
+ legacy_command_buffer_custom_call_targets: collections.abc.Iterable[builtins.str] | None = ...,
1406
+ xla_syntax_sugar_async_ops: builtins.bool | None = ...,
1407
+ xla_gpu_autotune_gemm_rtol: builtins.float | None = ...,
1408
+ xla_enable_command_buffers_during_profiling: builtins.bool | None = ...,
1409
+ xla_gpu_cudnn_gemm_max_plans: builtins.int | None = ...,
1410
+ xla_gpu_enable_libnvjitlink: builtins.bool | None = ...,
1411
+ xla_gpu_enable_triton_gemm_int4: builtins.bool | None = ...,
1412
+ xla_gpu_async_dot: builtins.bool | None = ...,
1413
+ xla_gpu_enable_pgle_accuracy_checker: builtins.bool | None = ...,
1414
+ xla_gpu_executable_warn_stuck_timeout_seconds: builtins.int | None = ...,
1415
+ xla_gpu_executable_terminate_timeout_seconds: builtins.int | None = ...,
1416
+ xla_experimental_ignore_channel_id: builtins.bool | None = ...,
1417
+ xla_backend_extra_options: collections.abc.Mapping[builtins.str, builtins.str] | None = ...,
1418
+ ) -> None: ...
1419
+ def ClearField(
1420
+ self,
1421
+ field_name: typing.Literal[
1422
+ "legacy_command_buffer_custom_call_targets",
1423
+ b"legacy_command_buffer_custom_call_targets",
1424
+ "xla_allow_excess_precision",
1425
+ b"xla_allow_excess_precision",
1426
+ "xla_allow_scalar_index_dynamic_ops",
1427
+ b"xla_allow_scalar_index_dynamic_ops",
1428
+ "xla_backend_extra_options",
1429
+ b"xla_backend_extra_options",
1430
+ "xla_backend_optimization_level",
1431
+ b"xla_backend_optimization_level",
1432
+ "xla_cmd_buffer_trace_cache_size",
1433
+ b"xla_cmd_buffer_trace_cache_size",
1434
+ "xla_cpu_enable_concurrency_optimized_scheduler",
1435
+ b"xla_cpu_enable_concurrency_optimized_scheduler",
1436
+ "xla_cpu_enable_custom_matmul_tiling",
1437
+ b"xla_cpu_enable_custom_matmul_tiling",
1438
+ "xla_cpu_enable_experimental_deallocation",
1439
+ b"xla_cpu_enable_experimental_deallocation",
1440
+ "xla_cpu_enable_fast_math",
1441
+ b"xla_cpu_enable_fast_math",
1442
+ "xla_cpu_enable_fast_min_max",
1443
+ b"xla_cpu_enable_fast_min_max",
1444
+ "xla_cpu_enable_mlir_fusion_outlining",
1445
+ b"xla_cpu_enable_mlir_fusion_outlining",
1446
+ "xla_cpu_enable_mlir_tiling_and_fusion",
1447
+ b"xla_cpu_enable_mlir_tiling_and_fusion",
1448
+ "xla_cpu_enable_xprof_traceme",
1449
+ b"xla_cpu_enable_xprof_traceme",
1450
+ "xla_cpu_fast_math_honor_division",
1451
+ b"xla_cpu_fast_math_honor_division",
1452
+ "xla_cpu_fast_math_honor_functions",
1453
+ b"xla_cpu_fast_math_honor_functions",
1454
+ "xla_cpu_fast_math_honor_infs",
1455
+ b"xla_cpu_fast_math_honor_infs",
1456
+ "xla_cpu_fast_math_honor_nans",
1457
+ b"xla_cpu_fast_math_honor_nans",
1458
+ "xla_cpu_matmul_tiling_k_dim",
1459
+ b"xla_cpu_matmul_tiling_k_dim",
1460
+ "xla_cpu_matmul_tiling_m_dim",
1461
+ b"xla_cpu_matmul_tiling_m_dim",
1462
+ "xla_cpu_matmul_tiling_n_dim",
1463
+ b"xla_cpu_matmul_tiling_n_dim",
1464
+ "xla_cpu_multi_thread_eigen",
1465
+ b"xla_cpu_multi_thread_eigen",
1466
+ "xla_cpu_parallel_codegen_split_count",
1467
+ b"xla_cpu_parallel_codegen_split_count",
1468
+ "xla_cpu_prefer_vector_width",
1469
+ b"xla_cpu_prefer_vector_width",
1470
+ "xla_cpu_strict_dot_conv_math",
1471
+ b"xla_cpu_strict_dot_conv_math",
1472
+ "xla_cpu_use_acl",
1473
+ b"xla_cpu_use_acl",
1474
+ "xla_cpu_use_mkl_dnn",
1475
+ b"xla_cpu_use_mkl_dnn",
1476
+ "xla_cpu_use_thunk_runtime",
1477
+ b"xla_cpu_use_thunk_runtime",
1478
+ "xla_debug_buffer_assignment_show_max",
1479
+ b"xla_debug_buffer_assignment_show_max",
1480
+ "xla_detailed_logging",
1481
+ b"xla_detailed_logging",
1482
+ "xla_disable_all_hlo_passes",
1483
+ b"xla_disable_all_hlo_passes",
1484
+ "xla_disable_hlo_passes",
1485
+ b"xla_disable_hlo_passes",
1486
+ "xla_dump_compress_protos",
1487
+ b"xla_dump_compress_protos",
1488
+ "xla_dump_disable_metadata",
1489
+ b"xla_dump_disable_metadata",
1490
+ "xla_dump_enable_mlir_pretty_form",
1491
+ b"xla_dump_enable_mlir_pretty_form",
1492
+ "xla_dump_fusion_visualization",
1493
+ b"xla_dump_fusion_visualization",
1494
+ "xla_dump_hlo_as_dot",
1495
+ b"xla_dump_hlo_as_dot",
1496
+ "xla_dump_hlo_as_html",
1497
+ b"xla_dump_hlo_as_html",
1498
+ "xla_dump_hlo_as_long_text",
1499
+ b"xla_dump_hlo_as_long_text",
1500
+ "xla_dump_hlo_as_proto",
1501
+ b"xla_dump_hlo_as_proto",
1502
+ "xla_dump_hlo_as_text",
1503
+ b"xla_dump_hlo_as_text",
1504
+ "xla_dump_hlo_as_url",
1505
+ b"xla_dump_hlo_as_url",
1506
+ "xla_dump_hlo_module_re",
1507
+ b"xla_dump_hlo_module_re",
1508
+ "xla_dump_hlo_pass_re",
1509
+ b"xla_dump_hlo_pass_re",
1510
+ "xla_dump_hlo_pipeline_re",
1511
+ b"xla_dump_hlo_pipeline_re",
1512
+ "xla_dump_hlo_snapshots",
1513
+ b"xla_dump_hlo_snapshots",
1514
+ "xla_dump_include_timestamp",
1515
+ b"xla_dump_include_timestamp",
1516
+ "xla_dump_large_constants",
1517
+ b"xla_dump_large_constants",
1518
+ "xla_dump_latency_hiding_schedule",
1519
+ b"xla_dump_latency_hiding_schedule",
1520
+ "xla_dump_max_hlo_modules",
1521
+ b"xla_dump_max_hlo_modules",
1522
+ "xla_dump_module_metadata",
1523
+ b"xla_dump_module_metadata",
1524
+ "xla_dump_to",
1525
+ b"xla_dump_to",
1526
+ "xla_eliminate_hlo_implicit_broadcast",
1527
+ b"xla_eliminate_hlo_implicit_broadcast",
1528
+ "xla_embed_ir_in_executable",
1529
+ b"xla_embed_ir_in_executable",
1530
+ "xla_enable_command_buffers_during_profiling",
1531
+ b"xla_enable_command_buffers_during_profiling",
1532
+ "xla_enable_dumping",
1533
+ b"xla_enable_dumping",
1534
+ "xla_enable_hlo_passes_only",
1535
+ b"xla_enable_hlo_passes_only",
1536
+ "xla_experimental_ignore_channel_id",
1537
+ b"xla_experimental_ignore_channel_id",
1538
+ "xla_force_host_platform_device_count",
1539
+ b"xla_force_host_platform_device_count",
1540
+ "xla_gpu_algorithm_denylist_path",
1541
+ b"xla_gpu_algorithm_denylist_path",
1542
+ "xla_gpu_all_gather_combine_threshold_bytes",
1543
+ b"xla_gpu_all_gather_combine_threshold_bytes",
1544
+ "xla_gpu_all_reduce_blueconnect_num_devices_per_host",
1545
+ b"xla_gpu_all_reduce_blueconnect_num_devices_per_host",
1546
+ "xla_gpu_all_reduce_combine_threshold_bytes",
1547
+ b"xla_gpu_all_reduce_combine_threshold_bytes",
1548
+ "xla_gpu_asm_extra_flags",
1549
+ b"xla_gpu_asm_extra_flags",
1550
+ "xla_gpu_async_dot",
1551
+ b"xla_gpu_async_dot",
1552
+ "xla_gpu_auto_spmd_partitioning_memory_budget_gb",
1553
+ b"xla_gpu_auto_spmd_partitioning_memory_budget_gb",
1554
+ "xla_gpu_auto_spmd_partitioning_memory_budget_ratio",
1555
+ b"xla_gpu_auto_spmd_partitioning_memory_budget_ratio",
1556
+ "xla_gpu_autotune_gemm_rtol",
1557
+ b"xla_gpu_autotune_gemm_rtol",
1558
+ "xla_gpu_autotune_level",
1559
+ b"xla_gpu_autotune_level",
1560
+ "xla_gpu_autotune_max_solutions",
1561
+ b"xla_gpu_autotune_max_solutions",
1562
+ "xla_gpu_collect_cost_model_stats",
1563
+ b"xla_gpu_collect_cost_model_stats",
1564
+ "xla_gpu_collective_inflation_factor",
1565
+ b"xla_gpu_collective_inflation_factor",
1566
+ "xla_gpu_collective_permute_decomposer_threshold",
1567
+ b"xla_gpu_collective_permute_decomposer_threshold",
1568
+ "xla_gpu_copy_insertion_use_region_analysis",
1569
+ b"xla_gpu_copy_insertion_use_region_analysis",
1570
+ "xla_gpu_crash_on_verification_failures",
1571
+ b"xla_gpu_crash_on_verification_failures",
1572
+ "xla_gpu_cublas_fallback",
1573
+ b"xla_gpu_cublas_fallback",
1574
+ "xla_gpu_cuda_data_dir",
1575
+ b"xla_gpu_cuda_data_dir",
1576
+ "xla_gpu_cudnn_gemm_fusion_level",
1577
+ b"xla_gpu_cudnn_gemm_fusion_level",
1578
+ "xla_gpu_cudnn_gemm_max_plans",
1579
+ b"xla_gpu_cudnn_gemm_max_plans",
1580
+ "xla_gpu_deterministic_ops",
1581
+ b"xla_gpu_deterministic_ops",
1582
+ "xla_gpu_disable_async_collectives",
1583
+ b"xla_gpu_disable_async_collectives",
1584
+ "xla_gpu_disable_gpuasm_optimizations",
1585
+ b"xla_gpu_disable_gpuasm_optimizations",
1586
+ "xla_gpu_dump_autotune_logs_to",
1587
+ b"xla_gpu_dump_autotune_logs_to",
1588
+ "xla_gpu_dump_autotune_results_to",
1589
+ b"xla_gpu_dump_autotune_results_to",
1590
+ "xla_gpu_dump_autotuned_gemm_fusions",
1591
+ b"xla_gpu_dump_autotuned_gemm_fusions",
1592
+ "xla_gpu_dump_llvmir",
1593
+ b"xla_gpu_dump_llvmir",
1594
+ "xla_gpu_enable_all_gather_combine_by_dim",
1595
+ b"xla_gpu_enable_all_gather_combine_by_dim",
1596
+ "xla_gpu_enable_analytical_latency_estimator",
1597
+ b"xla_gpu_enable_analytical_latency_estimator",
1598
+ "xla_gpu_enable_approx_costly_collectives",
1599
+ b"xla_gpu_enable_approx_costly_collectives",
1600
+ "xla_gpu_enable_bf16_3way_gemm",
1601
+ b"xla_gpu_enable_bf16_3way_gemm",
1602
+ "xla_gpu_enable_bf16_6way_gemm",
1603
+ b"xla_gpu_enable_bf16_6way_gemm",
1604
+ "xla_gpu_enable_command_buffer",
1605
+ b"xla_gpu_enable_command_buffer",
1606
+ "xla_gpu_enable_cub_radix_sort",
1607
+ b"xla_gpu_enable_cub_radix_sort",
1608
+ "xla_gpu_enable_cublaslt",
1609
+ b"xla_gpu_enable_cublaslt",
1610
+ "xla_gpu_enable_cudnn_fmha",
1611
+ b"xla_gpu_enable_cudnn_fmha",
1612
+ "xla_gpu_enable_cudnn_frontend",
1613
+ b"xla_gpu_enable_cudnn_frontend",
1614
+ "xla_gpu_enable_cudnn_int8x32_convolution_reordering",
1615
+ b"xla_gpu_enable_cudnn_int8x32_convolution_reordering",
1616
+ "xla_gpu_enable_cudnn_layer_norm",
1617
+ b"xla_gpu_enable_cudnn_layer_norm",
1618
+ "xla_gpu_enable_custom_fusions",
1619
+ b"xla_gpu_enable_custom_fusions",
1620
+ "xla_gpu_enable_custom_fusions_re",
1621
+ b"xla_gpu_enable_custom_fusions_re",
1622
+ "xla_gpu_enable_dot_strength_reduction",
1623
+ b"xla_gpu_enable_dot_strength_reduction",
1624
+ "xla_gpu_enable_dynamic_slice_fusion",
1625
+ b"xla_gpu_enable_dynamic_slice_fusion",
1626
+ "xla_gpu_enable_fast_min_max",
1627
+ b"xla_gpu_enable_fast_min_max",
1628
+ "xla_gpu_enable_highest_priority_async_stream",
1629
+ b"xla_gpu_enable_highest_priority_async_stream",
1630
+ "xla_gpu_enable_host_memory_offloading",
1631
+ b"xla_gpu_enable_host_memory_offloading",
1632
+ "xla_gpu_enable_latency_hiding_scheduler",
1633
+ b"xla_gpu_enable_latency_hiding_scheduler",
1634
+ "xla_gpu_enable_libnvjitlink",
1635
+ b"xla_gpu_enable_libnvjitlink",
1636
+ "xla_gpu_enable_libnvptxcompiler",
1637
+ b"xla_gpu_enable_libnvptxcompiler",
1638
+ "xla_gpu_enable_llvm_module_compilation_parallelism",
1639
+ b"xla_gpu_enable_llvm_module_compilation_parallelism",
1640
+ "xla_gpu_enable_nccl_clique_optimization",
1641
+ b"xla_gpu_enable_nccl_clique_optimization",
1642
+ "xla_gpu_enable_nccl_comm_splitting",
1643
+ b"xla_gpu_enable_nccl_comm_splitting",
1644
+ "xla_gpu_enable_nccl_per_stream_comms",
1645
+ b"xla_gpu_enable_nccl_per_stream_comms",
1646
+ "xla_gpu_enable_nccl_user_buffers",
1647
+ b"xla_gpu_enable_nccl_user_buffers",
1648
+ "xla_gpu_enable_pgle_accuracy_checker",
1649
+ b"xla_gpu_enable_pgle_accuracy_checker",
1650
+ "xla_gpu_enable_pipelined_all_gather",
1651
+ b"xla_gpu_enable_pipelined_all_gather",
1652
+ "xla_gpu_enable_pipelined_all_reduce",
1653
+ b"xla_gpu_enable_pipelined_all_reduce",
1654
+ "xla_gpu_enable_pipelined_collectives",
1655
+ b"xla_gpu_enable_pipelined_collectives",
1656
+ "xla_gpu_enable_pipelined_p2p",
1657
+ b"xla_gpu_enable_pipelined_p2p",
1658
+ "xla_gpu_enable_pipelined_reduce_scatter",
1659
+ b"xla_gpu_enable_pipelined_reduce_scatter",
1660
+ "xla_gpu_enable_priority_fusion",
1661
+ b"xla_gpu_enable_priority_fusion",
1662
+ "xla_gpu_enable_reassociation_for_converted_ar",
1663
+ b"xla_gpu_enable_reassociation_for_converted_ar",
1664
+ "xla_gpu_enable_reduce_scatter_combine_by_dim",
1665
+ b"xla_gpu_enable_reduce_scatter_combine_by_dim",
1666
+ "xla_gpu_enable_reduction_epilogue_fusion",
1667
+ b"xla_gpu_enable_reduction_epilogue_fusion",
1668
+ "xla_gpu_enable_shared_constants",
1669
+ b"xla_gpu_enable_shared_constants",
1670
+ "xla_gpu_enable_split_k_autotuning",
1671
+ b"xla_gpu_enable_split_k_autotuning",
1672
+ "xla_gpu_enable_triton_gemm",
1673
+ b"xla_gpu_enable_triton_gemm",
1674
+ "xla_gpu_enable_triton_gemm_int4",
1675
+ b"xla_gpu_enable_triton_gemm_int4",
1676
+ "xla_gpu_enable_triton_hopper",
1677
+ b"xla_gpu_enable_triton_hopper",
1678
+ "xla_gpu_enable_while_loop_double_buffering",
1679
+ b"xla_gpu_enable_while_loop_double_buffering",
1680
+ "xla_gpu_enable_while_loop_reduce_scatter_code_motion",
1681
+ b"xla_gpu_enable_while_loop_reduce_scatter_code_motion",
1682
+ "xla_gpu_enable_while_loop_unrolling",
1683
+ b"xla_gpu_enable_while_loop_unrolling",
1684
+ "xla_gpu_ensure_minor_dot_contraction_dims",
1685
+ b"xla_gpu_ensure_minor_dot_contraction_dims",
1686
+ "xla_gpu_exclude_nondeterministic_ops",
1687
+ b"xla_gpu_exclude_nondeterministic_ops",
1688
+ "xla_gpu_executable_terminate_timeout_seconds",
1689
+ b"xla_gpu_executable_terminate_timeout_seconds",
1690
+ "xla_gpu_executable_warn_stuck_timeout_seconds",
1691
+ b"xla_gpu_executable_warn_stuck_timeout_seconds",
1692
+ "xla_gpu_exhaustive_tiling_search",
1693
+ b"xla_gpu_exhaustive_tiling_search",
1694
+ "xla_gpu_experimental_autotune_cache_mode",
1695
+ b"xla_gpu_experimental_autotune_cache_mode",
1696
+ "xla_gpu_experimental_disable_binary_libraries",
1697
+ b"xla_gpu_experimental_disable_binary_libraries",
1698
+ "xla_gpu_experimental_enable_triton_softmax_priority_fusion",
1699
+ b"xla_gpu_experimental_enable_triton_softmax_priority_fusion",
1700
+ "xla_gpu_filter_kernels_spilling_registers_on_autotuning",
1701
+ b"xla_gpu_filter_kernels_spilling_registers_on_autotuning",
1702
+ "xla_gpu_force_compilation_parallelism",
1703
+ b"xla_gpu_force_compilation_parallelism",
1704
+ "xla_gpu_force_conv_nchw",
1705
+ b"xla_gpu_force_conv_nchw",
1706
+ "xla_gpu_force_conv_nhwc",
1707
+ b"xla_gpu_force_conv_nhwc",
1708
+ "xla_gpu_ftz",
1709
+ b"xla_gpu_ftz",
1710
+ "xla_gpu_fused_attention_use_cudnn_rng",
1711
+ b"xla_gpu_fused_attention_use_cudnn_rng",
1712
+ "xla_gpu_gemm_rewrite_size_threshold",
1713
+ b"xla_gpu_gemm_rewrite_size_threshold",
1714
+ "xla_gpu_graph_enable_concurrent_region",
1715
+ b"xla_gpu_graph_enable_concurrent_region",
1716
+ "xla_gpu_graph_min_graph_size",
1717
+ b"xla_gpu_graph_min_graph_size",
1718
+ "xla_gpu_kernel_cache_file",
1719
+ b"xla_gpu_kernel_cache_file",
1720
+ "xla_gpu_lhs_enable_gpu_async_tracker",
1721
+ b"xla_gpu_lhs_enable_gpu_async_tracker",
1722
+ "xla_gpu_llvm_ir_file",
1723
+ b"xla_gpu_llvm_ir_file",
1724
+ "xla_gpu_llvm_verification_level",
1725
+ b"xla_gpu_llvm_verification_level",
1726
+ "xla_gpu_load_autotune_results_from",
1727
+ b"xla_gpu_load_autotune_results_from",
1728
+ "xla_gpu_memory_limit_slop_factor",
1729
+ b"xla_gpu_memory_limit_slop_factor",
1730
+ "xla_gpu_mlir_emitter_level",
1731
+ b"xla_gpu_mlir_emitter_level",
1732
+ "xla_gpu_mock_custom_calls",
1733
+ b"xla_gpu_mock_custom_calls",
1734
+ "xla_gpu_multi_streamed_windowed_einsum",
1735
+ b"xla_gpu_multi_streamed_windowed_einsum",
1736
+ "xla_gpu_nccl_collective_max_nchannels",
1737
+ b"xla_gpu_nccl_collective_max_nchannels",
1738
+ "xla_gpu_nccl_p2p_max_nchannels",
1739
+ b"xla_gpu_nccl_p2p_max_nchannels",
1740
+ "xla_gpu_nccl_terminate_on_error",
1741
+ b"xla_gpu_nccl_terminate_on_error",
1742
+ "xla_gpu_nccl_termination_timeout_seconds",
1743
+ b"xla_gpu_nccl_termination_timeout_seconds",
1744
+ "xla_gpu_override_gemm_autotuner",
1745
+ b"xla_gpu_override_gemm_autotuner",
1746
+ "xla_gpu_per_fusion_autotune_cache_dir",
1747
+ b"xla_gpu_per_fusion_autotune_cache_dir",
1748
+ "xla_gpu_pgle_profile_file_or_directory_path",
1749
+ b"xla_gpu_pgle_profile_file_or_directory_path",
1750
+ "xla_gpu_ptx_file",
1751
+ b"xla_gpu_ptx_file",
1752
+ "xla_gpu_reduce_scatter_combine_threshold_bytes",
1753
+ b"xla_gpu_reduce_scatter_combine_threshold_bytes",
1754
+ "xla_gpu_redzone_padding_bytes",
1755
+ b"xla_gpu_redzone_padding_bytes",
1756
+ "xla_gpu_redzone_scratch_max_megabytes",
1757
+ b"xla_gpu_redzone_scratch_max_megabytes",
1758
+ "xla_gpu_require_complete_aot_autotune_results",
1759
+ b"xla_gpu_require_complete_aot_autotune_results",
1760
+ "xla_gpu_run_post_layout_collective_pipeliner",
1761
+ b"xla_gpu_run_post_layout_collective_pipeliner",
1762
+ "xla_gpu_shape_checks",
1763
+ b"xla_gpu_shape_checks",
1764
+ "xla_gpu_shard_autotuning",
1765
+ b"xla_gpu_shard_autotuning",
1766
+ "xla_gpu_strict_conv_algorithm_picker",
1767
+ b"xla_gpu_strict_conv_algorithm_picker",
1768
+ "xla_gpu_target_config_filename",
1769
+ b"xla_gpu_target_config_filename",
1770
+ "xla_gpu_temp_buffer_use_separate_color",
1771
+ b"xla_gpu_temp_buffer_use_separate_color",
1772
+ "xla_gpu_threshold_for_windowed_einsum_mib",
1773
+ b"xla_gpu_threshold_for_windowed_einsum_mib",
1774
+ "xla_gpu_triton_fusion_level",
1775
+ b"xla_gpu_triton_fusion_level",
1776
+ "xla_gpu_triton_gemm_any",
1777
+ b"xla_gpu_triton_gemm_any",
1778
+ "xla_gpu_triton_gemm_disable_reduced_precision_reduction",
1779
+ b"xla_gpu_triton_gemm_disable_reduced_precision_reduction",
1780
+ "xla_gpu_unsafe_fallback_to_driver_on_ptxas_not_found",
1781
+ b"xla_gpu_unsafe_fallback_to_driver_on_ptxas_not_found",
1782
+ "xla_gpu_unsafe_pipelined_loop_annotator",
1783
+ b"xla_gpu_unsafe_pipelined_loop_annotator",
1784
+ "xla_gpu_unsupported_enable_triton_gemm",
1785
+ b"xla_gpu_unsupported_enable_triton_gemm",
1786
+ "xla_gpu_use_memcpy_local_p2p",
1787
+ b"xla_gpu_use_memcpy_local_p2p",
1788
+ "xla_gpu_use_runtime_fusion",
1789
+ b"xla_gpu_use_runtime_fusion",
1790
+ "xla_gpu_verify_triton_fusion_numerics",
1791
+ b"xla_gpu_verify_triton_fusion_numerics",
1792
+ "xla_hlo_evaluator_use_fast_path",
1793
+ b"xla_hlo_evaluator_use_fast_path",
1794
+ "xla_hlo_graph_addresses",
1795
+ b"xla_hlo_graph_addresses",
1796
+ "xla_hlo_graph_sharding_color",
1797
+ b"xla_hlo_graph_sharding_color",
1798
+ "xla_hlo_profile",
1799
+ b"xla_hlo_profile",
1800
+ "xla_llvm_disable_expensive_passes",
1801
+ b"xla_llvm_disable_expensive_passes",
1802
+ "xla_llvm_enable_alias_scope_metadata",
1803
+ b"xla_llvm_enable_alias_scope_metadata",
1804
+ "xla_llvm_enable_invariant_load_metadata",
1805
+ b"xla_llvm_enable_invariant_load_metadata",
1806
+ "xla_llvm_enable_noalias_metadata",
1807
+ b"xla_llvm_enable_noalias_metadata",
1808
+ "xla_llvm_force_inline_before_split",
1809
+ b"xla_llvm_force_inline_before_split",
1810
+ "xla_multiheap_size_constraint_per_heap",
1811
+ b"xla_multiheap_size_constraint_per_heap",
1812
+ "xla_partitioning_algorithm",
1813
+ b"xla_partitioning_algorithm",
1814
+ "xla_reduce_window_rewrite_base_length",
1815
+ b"xla_reduce_window_rewrite_base_length",
1816
+ "xla_step_marker_location",
1817
+ b"xla_step_marker_location",
1818
+ "xla_syntax_sugar_async_ops",
1819
+ b"xla_syntax_sugar_async_ops",
1820
+ "xla_test_all_input_layouts",
1821
+ b"xla_test_all_input_layouts",
1822
+ "xla_test_all_output_layouts",
1823
+ b"xla_test_all_output_layouts",
1824
+ "xla_tpu_detect_inf",
1825
+ b"xla_tpu_detect_inf",
1826
+ "xla_tpu_detect_nan",
1827
+ b"xla_tpu_detect_nan",
1828
+ ],
1829
+ ) -> None: ...
1830
+
1831
+ global___DebugOptions = DebugOptions
1832
+
1833
+ @typing.final
1834
+ class GpuCompilationEnvironment(google.protobuf.message.Message):
1835
+ """Contains flags which affects the GPU compilation result.
1836
+ These flags are part of Debug Options as of now, and will be migrated to
1837
+ this proto.
1838
+ """
1839
+
1840
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
1841
+
1842
+ DUMMY_FLAG_FIELD_NUMBER: builtins.int
1843
+ dummy_flag: builtins.int
1844
+ """Temporary dummy flag is added to test the flow.
1845
+ To be removed when we add flags here.
1846
+ """
1847
+ def __init__(self, *, dummy_flag: builtins.int | None = ...) -> None: ...
1848
+ def ClearField(self, field_name: typing.Literal["dummy_flag", b"dummy_flag"]) -> None: ...
1849
+
1850
+ global___GpuCompilationEnvironment = GpuCompilationEnvironment
1851
+
1852
+ @typing.final
1853
+ class ShardableValueUpdatePairProto(google.protobuf.message.Message):
1854
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
1855
+
1856
+ INPUT_PARAMETER_NUMBER_FIELD_NUMBER: builtins.int
1857
+ PARAMETER_SHAPE_INDEX_FIELD_NUMBER: builtins.int
1858
+ OUTPUT_SHAPE_INDEX_FIELD_NUMBER: builtins.int
1859
+ input_parameter_number: builtins.int
1860
+ @property
1861
+ def parameter_shape_index(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]: ...
1862
+ @property
1863
+ def output_shape_index(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]: ...
1864
+ def __init__(
1865
+ self,
1866
+ *,
1867
+ input_parameter_number: builtins.int | None = ...,
1868
+ parameter_shape_index: collections.abc.Iterable[builtins.int] | None = ...,
1869
+ output_shape_index: collections.abc.Iterable[builtins.int] | None = ...,
1870
+ ) -> None: ...
1871
+ def ClearField(
1872
+ self,
1873
+ field_name: typing.Literal[
1874
+ "input_parameter_number",
1875
+ b"input_parameter_number",
1876
+ "output_shape_index",
1877
+ b"output_shape_index",
1878
+ "parameter_shape_index",
1879
+ b"parameter_shape_index",
1880
+ ],
1881
+ ) -> None: ...
1882
+
1883
+ global___ShardableValueUpdatePairProto = ShardableValueUpdatePairProto
1884
+
1885
+ @typing.final
1886
+ class ExecutionOptions(google.protobuf.message.Message):
1887
+ """These settings control how XLA compiles and/or runs code. Not all settings
1888
+ will have an effect on every platform.
1889
+
1890
+ When adding new fields, keep in mind that boolean fields default to false.
1891
+ Next id: 25.
1892
+ """
1893
+
1894
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
1895
+
1896
+ SHAPE_WITH_OUTPUT_LAYOUT_FIELD_NUMBER: builtins.int
1897
+ SEED_FIELD_NUMBER: builtins.int
1898
+ DEBUG_OPTIONS_FIELD_NUMBER: builtins.int
1899
+ DEVICE_HANDLES_FIELD_NUMBER: builtins.int
1900
+ NUM_REPLICAS_FIELD_NUMBER: builtins.int
1901
+ DEVICE_ASSIGNMENT_FIELD_NUMBER: builtins.int
1902
+ ALIAS_PASSTHROUGH_PARAMS_FIELD_NUMBER: builtins.int
1903
+ NUM_PARTITIONS_FIELD_NUMBER: builtins.int
1904
+ LAUNCH_ID_FIELD_NUMBER: builtins.int
1905
+ USE_SPMD_PARTITIONING_FIELD_NUMBER: builtins.int
1906
+ USE_AUTO_SPMD_PARTITIONING_FIELD_NUMBER: builtins.int
1907
+ AUTO_SPMD_PARTITIONING_MESH_SHAPE_FIELD_NUMBER: builtins.int
1908
+ AUTO_SPMD_PARTITIONING_MESH_IDS_FIELD_NUMBER: builtins.int
1909
+ DEDUPLICATE_HLO_FIELD_NUMBER: builtins.int
1910
+ ALLOW_SPMD_SHARDING_PROPAGATION_TO_PARAMETERS_FIELD_NUMBER: builtins.int
1911
+ ALLOW_SPMD_SHARDING_PROPAGATION_TO_OUTPUT_FIELD_NUMBER: builtins.int
1912
+ PARAM_REQUIRES_BROADCAST_VIA_COLLECTIVES_FIELD_NUMBER: builtins.int
1913
+ ALLOW_SEPARATE_SHARDING_PROGRAMS_FIELD_NUMBER: builtins.int
1914
+ SHARDABLE_VALUE_UPDATE_PAIRS_FIELD_NUMBER: builtins.int
1915
+ FDO_PROFILE_FIELD_NUMBER: builtins.int
1916
+ DEVICE_MEMORY_SIZE_FIELD_NUMBER: builtins.int
1917
+ USE_SHARDY_PARTITIONER_FIELD_NUMBER: builtins.int
1918
+ seed: builtins.int
1919
+ """Used to seed random-number generators used in this computation. If this is
1920
+ 0, we generate a seed ourselves.
1921
+
1922
+ TODO(b/32083678): Changing the seed unnecessarily forces a recompilation.
1923
+ """
1924
+ num_replicas: builtins.int
1925
+ """Number of replicas of the computation to run. If zero, uses the default
1926
+ number of replicas for the XLA service.
1927
+ """
1928
+ alias_passthrough_params: builtins.bool
1929
+ """Alias input and output buffers for parameters that are passed-through XLA
1930
+ modules without being changed.
1931
+ """
1932
+ num_partitions: builtins.int
1933
+ """Number of partitions of the computation to run (model parallelism).
1934
+ If zero, uses the default number of partitions for the XLA service.
1935
+ """
1936
+ launch_id: builtins.int
1937
+ """Used to identify a set of programs that should be launch together."""
1938
+ use_spmd_partitioning: builtins.bool
1939
+ """Indicates whether to use SPMD (true) or MPMD (false) partitioning when
1940
+ num_partitions > 1 and XLA is requested to partition the input program.
1941
+ """
1942
+ use_auto_spmd_partitioning: builtins.bool
1943
+ """Whether to automatically generate XLA shardings for SPMD partitioner."""
1944
+ deduplicate_hlo: builtins.bool
1945
+ """If set, deduplicate hlo into function calls to reduce binary size. Only
1946
+ works on TPU.
1947
+ """
1948
+ allow_separate_sharding_programs: builtins.bool
1949
+ """If enabled, the compiler may generate sharding and unsharding programs as
1950
+ separate HLO modules, and modify the main program's input and output to
1951
+ be sharded.
1952
+ """
1953
+ fdo_profile: builtins.bytes
1954
+ """Profiling data for feedback directed optimizations. Note that this is not
1955
+ the only way to feed FDO data into the compiler and individual backends
1956
+ may choose to get FDO data by other means.
1957
+ """
1958
+ device_memory_size: builtins.int
1959
+ """Amount of device memory available for the executable to use."""
1960
+ use_shardy_partitioner: builtins.bool
1961
+ """Use Shardy, a new partitioner, to replace the existing
1962
+ ShardingPropagation and SpmdPartitioner. See go/xla-sdy-pipeline for
1963
+ details.
1964
+ """
1965
+ @property
1966
+ def shape_with_output_layout(self) -> tensorflow.compiler.xla.xla_data_pb2.ShapeProto:
1967
+ """This optional field's layout is used as a hint when storing the output of
1968
+ this computation. Subsequent transfers of this output array to the client
1969
+ may be faster when using this layout.
1970
+
1971
+ We use a Shape here to accommodate computations that return a tuple.
1972
+ """
1973
+
1974
+ @property
1975
+ def debug_options(self) -> global___DebugOptions: ...
1976
+ @property
1977
+ def device_handles(
1978
+ self,
1979
+ ) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[tensorflow.compiler.xla.xla_data_pb2.DeviceHandle]:
1980
+ """This optional field specifies a particular set of devices to run the
1981
+ computation on. The computation will be partitioned across these devices.
1982
+ If not provided, the default device will be chosen.
1983
+ """
1984
+
1985
+ @property
1986
+ def device_assignment(self) -> tensorflow.compiler.xla.xla_data_pb2.DeviceAssignmentProto:
1987
+ """This optional field specifies the device assignment if known at compile
1988
+ time.
1989
+ """
1990
+
1991
+ @property
1992
+ def auto_spmd_partitioning_mesh_shape(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]:
1993
+ """Device mesh shape used to create the sharding search space when
1994
+ use_auto_spmd_partitioning=true.
1995
+ """
1996
+
1997
+ @property
1998
+ def auto_spmd_partitioning_mesh_ids(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]:
1999
+ """Device mesh ids compatible with the above mesh_shape used when
2000
+ use_auto_spmd_partitioning=true.
2001
+ """
2002
+
2003
+ @property
2004
+ def allow_spmd_sharding_propagation_to_parameters(
2005
+ self,
2006
+ ) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.bool]:
2007
+ """Allows sharding propagation to propagate to the parameters. This changes
2008
+ the input shape of the computation (which is undesirable), but it can be
2009
+ used to allow to run partial compilation to determine what would be the
2010
+ input sharding of a computation if XLA would be allowed to propagate the
2011
+ sharding which can be used by higher level framework as a way to query
2012
+ intermediate sharding of operations when multiple computation would be
2013
+ chained and merged together.
2014
+ This is a vector of bool, because the user can control which parameters can
2015
+ have the sharding substituted. If only one boolean value is passed in the
2016
+ vector that is interpreted as the value to be applied for every parameter.
2017
+ """
2018
+
2019
+ @property
2020
+ def allow_spmd_sharding_propagation_to_output(
2021
+ self,
2022
+ ) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.bool]:
2023
+ """Allows sharding propagation to propagate to the outputs. This changes the
2024
+ output shape of the computation (which is undesirable), but it can be used
2025
+ to allow to run partial compilation to determine what would be the output
2026
+ sharding of a computation if XLA would be allowed to propagate the sharding
2027
+ which can be used by higher level framework as a way to query intermediate
2028
+ sharding of operations when multiple computation would be chained and
2029
+ merged together.
2030
+ This is a vector of bool, because the user can control (if the output of
2031
+ the computation is a tuple) which elements of the tuple can have the
2032
+ sharding substituted and which don't. If only one boolean value is passed
2033
+ in the vector that's interpreted as the value to be applied for every
2034
+ single element of the output tuple. One value per element of the tuple
2035
+ means that each value is attached to one of the output elements.
2036
+ """
2037
+
2038
+ @property
2039
+ def param_requires_broadcast_via_collectives(
2040
+ self,
2041
+ ) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.bool]:
2042
+ """Whether to broadcast args across all replicas. One entry per arg."""
2043
+
2044
+ @property
2045
+ def shardable_value_update_pairs(
2046
+ self,
2047
+ ) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ShardableValueUpdatePairProto]:
2048
+ """The list of input/output pairs in the main program that could be sharded."""
2049
+
2050
+ def __init__(
2051
+ self,
2052
+ *,
2053
+ shape_with_output_layout: tensorflow.compiler.xla.xla_data_pb2.ShapeProto | None = ...,
2054
+ seed: builtins.int | None = ...,
2055
+ debug_options: global___DebugOptions | None = ...,
2056
+ device_handles: collections.abc.Iterable[tensorflow.compiler.xla.xla_data_pb2.DeviceHandle] | None = ...,
2057
+ num_replicas: builtins.int | None = ...,
2058
+ device_assignment: tensorflow.compiler.xla.xla_data_pb2.DeviceAssignmentProto | None = ...,
2059
+ alias_passthrough_params: builtins.bool | None = ...,
2060
+ num_partitions: builtins.int | None = ...,
2061
+ launch_id: builtins.int | None = ...,
2062
+ use_spmd_partitioning: builtins.bool | None = ...,
2063
+ use_auto_spmd_partitioning: builtins.bool | None = ...,
2064
+ auto_spmd_partitioning_mesh_shape: collections.abc.Iterable[builtins.int] | None = ...,
2065
+ auto_spmd_partitioning_mesh_ids: collections.abc.Iterable[builtins.int] | None = ...,
2066
+ deduplicate_hlo: builtins.bool | None = ...,
2067
+ allow_spmd_sharding_propagation_to_parameters: collections.abc.Iterable[builtins.bool] | None = ...,
2068
+ allow_spmd_sharding_propagation_to_output: collections.abc.Iterable[builtins.bool] | None = ...,
2069
+ param_requires_broadcast_via_collectives: collections.abc.Iterable[builtins.bool] | None = ...,
2070
+ allow_separate_sharding_programs: builtins.bool | None = ...,
2071
+ shardable_value_update_pairs: collections.abc.Iterable[global___ShardableValueUpdatePairProto] | None = ...,
2072
+ fdo_profile: builtins.bytes | None = ...,
2073
+ device_memory_size: builtins.int | None = ...,
2074
+ use_shardy_partitioner: builtins.bool | None = ...,
2075
+ ) -> None: ...
2076
+ def HasField(
2077
+ self,
2078
+ field_name: typing.Literal[
2079
+ "debug_options",
2080
+ b"debug_options",
2081
+ "device_assignment",
2082
+ b"device_assignment",
2083
+ "shape_with_output_layout",
2084
+ b"shape_with_output_layout",
2085
+ ],
2086
+ ) -> builtins.bool: ...
2087
+ def ClearField(
2088
+ self,
2089
+ field_name: typing.Literal[
2090
+ "alias_passthrough_params",
2091
+ b"alias_passthrough_params",
2092
+ "allow_separate_sharding_programs",
2093
+ b"allow_separate_sharding_programs",
2094
+ "allow_spmd_sharding_propagation_to_output",
2095
+ b"allow_spmd_sharding_propagation_to_output",
2096
+ "allow_spmd_sharding_propagation_to_parameters",
2097
+ b"allow_spmd_sharding_propagation_to_parameters",
2098
+ "auto_spmd_partitioning_mesh_ids",
2099
+ b"auto_spmd_partitioning_mesh_ids",
2100
+ "auto_spmd_partitioning_mesh_shape",
2101
+ b"auto_spmd_partitioning_mesh_shape",
2102
+ "debug_options",
2103
+ b"debug_options",
2104
+ "deduplicate_hlo",
2105
+ b"deduplicate_hlo",
2106
+ "device_assignment",
2107
+ b"device_assignment",
2108
+ "device_handles",
2109
+ b"device_handles",
2110
+ "device_memory_size",
2111
+ b"device_memory_size",
2112
+ "fdo_profile",
2113
+ b"fdo_profile",
2114
+ "launch_id",
2115
+ b"launch_id",
2116
+ "num_partitions",
2117
+ b"num_partitions",
2118
+ "num_replicas",
2119
+ b"num_replicas",
2120
+ "param_requires_broadcast_via_collectives",
2121
+ b"param_requires_broadcast_via_collectives",
2122
+ "seed",
2123
+ b"seed",
2124
+ "shape_with_output_layout",
2125
+ b"shape_with_output_layout",
2126
+ "shardable_value_update_pairs",
2127
+ b"shardable_value_update_pairs",
2128
+ "use_auto_spmd_partitioning",
2129
+ b"use_auto_spmd_partitioning",
2130
+ "use_shardy_partitioner",
2131
+ b"use_shardy_partitioner",
2132
+ "use_spmd_partitioning",
2133
+ b"use_spmd_partitioning",
2134
+ ],
2135
+ ) -> None: ...
2136
+
2137
+ global___ExecutionOptions = ExecutionOptions
2138
+
2139
+ @typing.final
2140
+ class HloModuleConfigProto(google.protobuf.message.Message):
2141
+ """Serialization of HloModuleConfig. See the C++ class definition for
2142
+ descriptions of each field.
2143
+ There are no guarantees of backwards or forwards compatibility.
2144
+ Next id: 36.
2145
+ """
2146
+
2147
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
2148
+
2149
+ class _FusionConfigCollection:
2150
+ ValueType = typing.NewType("ValueType", builtins.int)
2151
+ V: typing_extensions.TypeAlias = ValueType
2152
+
2153
+ class _FusionConfigCollectionEnumTypeWrapper(
2154
+ google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[HloModuleConfigProto._FusionConfigCollection.ValueType],
2155
+ builtins.type,
2156
+ ):
2157
+ DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
2158
+ OFF: HloModuleConfigProto._FusionConfigCollection.ValueType # 0
2159
+ """Do not collect configuration."""
2160
+ PER_EDGE: HloModuleConfigProto._FusionConfigCollection.ValueType # 1
2161
+ """Collect per-edge configuration."""
2162
+ PER_NODE: HloModuleConfigProto._FusionConfigCollection.ValueType # 2
2163
+ """Collect per-node configuration."""
2164
+
2165
+ class FusionConfigCollection(_FusionConfigCollection, metaclass=_FusionConfigCollectionEnumTypeWrapper): ...
2166
+ OFF: HloModuleConfigProto.FusionConfigCollection.ValueType # 0
2167
+ """Do not collect configuration."""
2168
+ PER_EDGE: HloModuleConfigProto.FusionConfigCollection.ValueType # 1
2169
+ """Collect per-edge configuration."""
2170
+ PER_NODE: HloModuleConfigProto.FusionConfigCollection.ValueType # 2
2171
+ """Collect per-node configuration."""
2172
+
2173
+ @typing.final
2174
+ class BoolList(google.protobuf.message.Message):
2175
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
2176
+
2177
+ VALS_FIELD_NUMBER: builtins.int
2178
+ @property
2179
+ def vals(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.bool]: ...
2180
+ def __init__(self, *, vals: collections.abc.Iterable[builtins.bool] | None = ...) -> None: ...
2181
+ def ClearField(self, field_name: typing.Literal["vals", b"vals"]) -> None: ...
2182
+
2183
+ @typing.final
2184
+ class Int64List(google.protobuf.message.Message):
2185
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
2186
+
2187
+ VALS_FIELD_NUMBER: builtins.int
2188
+ @property
2189
+ def vals(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]: ...
2190
+ def __init__(self, *, vals: collections.abc.Iterable[builtins.int] | None = ...) -> None: ...
2191
+ def ClearField(self, field_name: typing.Literal["vals", b"vals"]) -> None: ...
2192
+
2193
+ @typing.final
2194
+ class Int64ListList(google.protobuf.message.Message):
2195
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
2196
+
2197
+ LISTS_FIELD_NUMBER: builtins.int
2198
+ @property
2199
+ def lists(
2200
+ self,
2201
+ ) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___HloModuleConfigProto.Int64List]: ...
2202
+ def __init__(self, *, lists: collections.abc.Iterable[global___HloModuleConfigProto.Int64List] | None = ...) -> None: ...
2203
+ def ClearField(self, field_name: typing.Literal["lists", b"lists"]) -> None: ...
2204
+
2205
+ @typing.final
2206
+ class DotConfigEntry(google.protobuf.message.Message):
2207
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
2208
+
2209
+ KEY_FIELD_NUMBER: builtins.int
2210
+ VALUE_FIELD_NUMBER: builtins.int
2211
+ key: builtins.str
2212
+ @property
2213
+ def value(self) -> global___HloModuleConfigProto.Int64List: ...
2214
+ def __init__(
2215
+ self, *, key: builtins.str | None = ..., value: global___HloModuleConfigProto.Int64List | None = ...
2216
+ ) -> None: ...
2217
+ def HasField(self, field_name: typing.Literal["value", b"value"]) -> builtins.bool: ...
2218
+ def ClearField(self, field_name: typing.Literal["key", b"key", "value", b"value"]) -> None: ...
2219
+
2220
+ @typing.final
2221
+ class AnalysisAllowanceMapEntry(google.protobuf.message.Message):
2222
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
2223
+
2224
+ KEY_FIELD_NUMBER: builtins.int
2225
+ VALUE_FIELD_NUMBER: builtins.int
2226
+ key: builtins.str
2227
+ value: builtins.int
2228
+ def __init__(self, *, key: builtins.str | None = ..., value: builtins.int | None = ...) -> None: ...
2229
+ def ClearField(self, field_name: typing.Literal["key", b"key", "value", b"value"]) -> None: ...
2230
+
2231
+ ENTRY_COMPUTATION_LAYOUT_FIELD_NUMBER: builtins.int
2232
+ SEED_FIELD_NUMBER: builtins.int
2233
+ LAUNCH_ID_FIELD_NUMBER: builtins.int
2234
+ REPLICA_COUNT_FIELD_NUMBER: builtins.int
2235
+ NUM_PARTITIONS_FIELD_NUMBER: builtins.int
2236
+ PARAM_REQUIRES_BROADCAST_VIA_COLLECTIVES_FIELD_NUMBER: builtins.int
2237
+ USE_SPMD_PARTITIONING_FIELD_NUMBER: builtins.int
2238
+ USE_AUTO_SPMD_PARTITIONING_FIELD_NUMBER: builtins.int
2239
+ AUTO_SPMD_PARTITIONING_MESH_SHAPE_FIELD_NUMBER: builtins.int
2240
+ AUTO_SPMD_PARTITIONING_MESH_IDS_FIELD_NUMBER: builtins.int
2241
+ DEDUPLICATE_HLO_FIELD_NUMBER: builtins.int
2242
+ INTRA_OP_PARALLELISM_THREADS_FIELD_NUMBER: builtins.int
2243
+ DEVICE_TYPE_FIELD_NUMBER: builtins.int
2244
+ DEBUG_OPTIONS_FIELD_NUMBER: builtins.int
2245
+ STATIC_DEVICE_ASSIGNMENT_FIELD_NUMBER: builtins.int
2246
+ PRE_SIMULATION_DEVICE_ASSIGNMENT_FIELD_NUMBER: builtins.int
2247
+ ALLOW_SEPARATE_SHARDING_PROGRAMS_FIELD_NUMBER: builtins.int
2248
+ SHARDABLE_VALUE_UPDATE_PAIRS_FIELD_NUMBER: builtins.int
2249
+ ALIAS_PASSTHROUGH_PARAMS_FIELD_NUMBER: builtins.int
2250
+ CONTENT_AWARE_COMPUTATION_SORTING_FIELD_NUMBER: builtins.int
2251
+ FUSION_CONFIG_COLLECTION_FIELD_NUMBER: builtins.int
2252
+ FUSION_CONFIG_FIELD_NUMBER: builtins.int
2253
+ DOT_CONFIG_FIELD_NUMBER: builtins.int
2254
+ LAYOUT_CONFIG_FIELD_NUMBER: builtins.int
2255
+ MEMORY_SPACE_ASSIGNMENT_CONFIG_FIELD_NUMBER: builtins.int
2256
+ PHASE_ORDERING_CONFIG_FIELD_NUMBER: builtins.int
2257
+ PHASE_INDEX_FIELD_NUMBER: builtins.int
2258
+ ALLOW_SPMD_SHARDING_PROPAGATION_TO_PARAMETERS_FIELD_NUMBER: builtins.int
2259
+ ALLOW_SPMD_SHARDING_PROPAGATION_TO_OUTPUT_FIELD_NUMBER: builtins.int
2260
+ ANALYSIS_ALLOWANCE_MAP_FIELD_NUMBER: builtins.int
2261
+ MATRIX_UNIT_OPERAND_PRECISION_FIELD_NUMBER: builtins.int
2262
+ FDO_PROFILE_FIELD_NUMBER: builtins.int
2263
+ DEVICE_MEMORY_SIZE_FIELD_NUMBER: builtins.int
2264
+ USE_SHARDY_PARTITIONER_FIELD_NUMBER: builtins.int
2265
+ seed: builtins.int
2266
+ launch_id: builtins.int
2267
+ replica_count: builtins.int
2268
+ num_partitions: builtins.int
2269
+ use_spmd_partitioning: builtins.bool
2270
+ use_auto_spmd_partitioning: builtins.bool
2271
+ deduplicate_hlo: builtins.bool
2272
+ intra_op_parallelism_threads: builtins.int
2273
+ device_type: builtins.str
2274
+ allow_separate_sharding_programs: builtins.bool
2275
+ alias_passthrough_params: builtins.bool
2276
+ content_aware_computation_sorting: builtins.bool
2277
+ fusion_config_collection: global___HloModuleConfigProto.FusionConfigCollection.ValueType
2278
+ phase_index: builtins.int
2279
+ matrix_unit_operand_precision: tensorflow.compiler.xla.xla_data_pb2.PrecisionConfig.Precision.ValueType
2280
+ fdo_profile: builtins.bytes
2281
+ device_memory_size: builtins.int
2282
+ use_shardy_partitioner: builtins.bool
2283
+ @property
2284
+ def entry_computation_layout(self) -> tensorflow.compiler.xla.xla_data_pb2.ProgramShapeProto: ...
2285
+ @property
2286
+ def param_requires_broadcast_via_collectives(
2287
+ self,
2288
+ ) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.bool]: ...
2289
+ @property
2290
+ def auto_spmd_partitioning_mesh_shape(
2291
+ self,
2292
+ ) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]: ...
2293
+ @property
2294
+ def auto_spmd_partitioning_mesh_ids(
2295
+ self,
2296
+ ) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]: ...
2297
+ @property
2298
+ def debug_options(self) -> global___DebugOptions: ...
2299
+ @property
2300
+ def static_device_assignment(self) -> tensorflow.compiler.xla.xla_data_pb2.DeviceAssignmentProto: ...
2301
+ @property
2302
+ def pre_simulation_device_assignment(self) -> tensorflow.compiler.xla.xla_data_pb2.DeviceAssignmentProto:
2303
+ """The original device assignment before being changed by a simulator.
2304
+ Simulators, like HybridSim, may change the device assignment to a smaller
2305
+ topology, to make simulation easier.
2306
+ """
2307
+
2308
+ @property
2309
+ def shardable_value_update_pairs(
2310
+ self,
2311
+ ) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ShardableValueUpdatePairProto]: ...
2312
+ @property
2313
+ def fusion_config(
2314
+ self,
2315
+ ) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___HloModuleConfigProto.BoolList]: ...
2316
+ @property
2317
+ def dot_config(
2318
+ self,
2319
+ ) -> google.protobuf.internal.containers.MessageMap[builtins.str, global___HloModuleConfigProto.Int64List]: ...
2320
+ @property
2321
+ def layout_config(
2322
+ self,
2323
+ ) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___HloModuleConfigProto.Int64ListList]: ...
2324
+ @property
2325
+ def memory_space_assignment_config(
2326
+ self,
2327
+ ) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]: ...
2328
+ @property
2329
+ def phase_ordering_config(
2330
+ self,
2331
+ ) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___HloModuleConfigProto.BoolList]: ...
2332
+ @property
2333
+ def allow_spmd_sharding_propagation_to_parameters(
2334
+ self,
2335
+ ) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.bool]: ...
2336
+ @property
2337
+ def allow_spmd_sharding_propagation_to_output(
2338
+ self,
2339
+ ) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.bool]: ...
2340
+ @property
2341
+ def analysis_allowance_map(self) -> google.protobuf.internal.containers.ScalarMap[builtins.str, builtins.int]: ...
2342
+ def __init__(
2343
+ self,
2344
+ *,
2345
+ entry_computation_layout: tensorflow.compiler.xla.xla_data_pb2.ProgramShapeProto | None = ...,
2346
+ seed: builtins.int | None = ...,
2347
+ launch_id: builtins.int | None = ...,
2348
+ replica_count: builtins.int | None = ...,
2349
+ num_partitions: builtins.int | None = ...,
2350
+ param_requires_broadcast_via_collectives: collections.abc.Iterable[builtins.bool] | None = ...,
2351
+ use_spmd_partitioning: builtins.bool | None = ...,
2352
+ use_auto_spmd_partitioning: builtins.bool | None = ...,
2353
+ auto_spmd_partitioning_mesh_shape: collections.abc.Iterable[builtins.int] | None = ...,
2354
+ auto_spmd_partitioning_mesh_ids: collections.abc.Iterable[builtins.int] | None = ...,
2355
+ deduplicate_hlo: builtins.bool | None = ...,
2356
+ intra_op_parallelism_threads: builtins.int | None = ...,
2357
+ device_type: builtins.str | None = ...,
2358
+ debug_options: global___DebugOptions | None = ...,
2359
+ static_device_assignment: tensorflow.compiler.xla.xla_data_pb2.DeviceAssignmentProto | None = ...,
2360
+ pre_simulation_device_assignment: tensorflow.compiler.xla.xla_data_pb2.DeviceAssignmentProto | None = ...,
2361
+ allow_separate_sharding_programs: builtins.bool | None = ...,
2362
+ shardable_value_update_pairs: collections.abc.Iterable[global___ShardableValueUpdatePairProto] | None = ...,
2363
+ alias_passthrough_params: builtins.bool | None = ...,
2364
+ content_aware_computation_sorting: builtins.bool | None = ...,
2365
+ fusion_config_collection: global___HloModuleConfigProto.FusionConfigCollection.ValueType | None = ...,
2366
+ fusion_config: collections.abc.Iterable[global___HloModuleConfigProto.BoolList] | None = ...,
2367
+ dot_config: collections.abc.Mapping[builtins.str, global___HloModuleConfigProto.Int64List] | None = ...,
2368
+ layout_config: collections.abc.Iterable[global___HloModuleConfigProto.Int64ListList] | None = ...,
2369
+ memory_space_assignment_config: collections.abc.Iterable[builtins.int] | None = ...,
2370
+ phase_ordering_config: collections.abc.Iterable[global___HloModuleConfigProto.BoolList] | None = ...,
2371
+ phase_index: builtins.int | None = ...,
2372
+ allow_spmd_sharding_propagation_to_parameters: collections.abc.Iterable[builtins.bool] | None = ...,
2373
+ allow_spmd_sharding_propagation_to_output: collections.abc.Iterable[builtins.bool] | None = ...,
2374
+ analysis_allowance_map: collections.abc.Mapping[builtins.str, builtins.int] | None = ...,
2375
+ matrix_unit_operand_precision: tensorflow.compiler.xla.xla_data_pb2.PrecisionConfig.Precision.ValueType | None = ...,
2376
+ fdo_profile: builtins.bytes | None = ...,
2377
+ device_memory_size: builtins.int | None = ...,
2378
+ use_shardy_partitioner: builtins.bool | None = ...,
2379
+ ) -> None: ...
2380
+ def HasField(
2381
+ self,
2382
+ field_name: typing.Literal[
2383
+ "debug_options",
2384
+ b"debug_options",
2385
+ "entry_computation_layout",
2386
+ b"entry_computation_layout",
2387
+ "pre_simulation_device_assignment",
2388
+ b"pre_simulation_device_assignment",
2389
+ "static_device_assignment",
2390
+ b"static_device_assignment",
2391
+ ],
2392
+ ) -> builtins.bool: ...
2393
+ def ClearField(
2394
+ self,
2395
+ field_name: typing.Literal[
2396
+ "alias_passthrough_params",
2397
+ b"alias_passthrough_params",
2398
+ "allow_separate_sharding_programs",
2399
+ b"allow_separate_sharding_programs",
2400
+ "allow_spmd_sharding_propagation_to_output",
2401
+ b"allow_spmd_sharding_propagation_to_output",
2402
+ "allow_spmd_sharding_propagation_to_parameters",
2403
+ b"allow_spmd_sharding_propagation_to_parameters",
2404
+ "analysis_allowance_map",
2405
+ b"analysis_allowance_map",
2406
+ "auto_spmd_partitioning_mesh_ids",
2407
+ b"auto_spmd_partitioning_mesh_ids",
2408
+ "auto_spmd_partitioning_mesh_shape",
2409
+ b"auto_spmd_partitioning_mesh_shape",
2410
+ "content_aware_computation_sorting",
2411
+ b"content_aware_computation_sorting",
2412
+ "debug_options",
2413
+ b"debug_options",
2414
+ "deduplicate_hlo",
2415
+ b"deduplicate_hlo",
2416
+ "device_memory_size",
2417
+ b"device_memory_size",
2418
+ "device_type",
2419
+ b"device_type",
2420
+ "dot_config",
2421
+ b"dot_config",
2422
+ "entry_computation_layout",
2423
+ b"entry_computation_layout",
2424
+ "fdo_profile",
2425
+ b"fdo_profile",
2426
+ "fusion_config",
2427
+ b"fusion_config",
2428
+ "fusion_config_collection",
2429
+ b"fusion_config_collection",
2430
+ "intra_op_parallelism_threads",
2431
+ b"intra_op_parallelism_threads",
2432
+ "launch_id",
2433
+ b"launch_id",
2434
+ "layout_config",
2435
+ b"layout_config",
2436
+ "matrix_unit_operand_precision",
2437
+ b"matrix_unit_operand_precision",
2438
+ "memory_space_assignment_config",
2439
+ b"memory_space_assignment_config",
2440
+ "num_partitions",
2441
+ b"num_partitions",
2442
+ "param_requires_broadcast_via_collectives",
2443
+ b"param_requires_broadcast_via_collectives",
2444
+ "phase_index",
2445
+ b"phase_index",
2446
+ "phase_ordering_config",
2447
+ b"phase_ordering_config",
2448
+ "pre_simulation_device_assignment",
2449
+ b"pre_simulation_device_assignment",
2450
+ "replica_count",
2451
+ b"replica_count",
2452
+ "seed",
2453
+ b"seed",
2454
+ "shardable_value_update_pairs",
2455
+ b"shardable_value_update_pairs",
2456
+ "static_device_assignment",
2457
+ b"static_device_assignment",
2458
+ "use_auto_spmd_partitioning",
2459
+ b"use_auto_spmd_partitioning",
2460
+ "use_shardy_partitioner",
2461
+ b"use_shardy_partitioner",
2462
+ "use_spmd_partitioning",
2463
+ b"use_spmd_partitioning",
2464
+ ],
2465
+ ) -> None: ...
2466
+
2467
+ global___HloModuleConfigProto = HloModuleConfigProto
2468
+
2469
+ @typing.final
2470
+ class HloModuleProtoWithConfig(google.protobuf.message.Message):
2471
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
2472
+
2473
+ HLO_MODULE_FIELD_NUMBER: builtins.int
2474
+ CONFIG_FIELD_NUMBER: builtins.int
2475
+ @property
2476
+ def hlo_module(self) -> tensorflow.compiler.xla.service.hlo_pb2.HloModuleProto: ...
2477
+ @property
2478
+ def config(self) -> global___HloModuleConfigProto: ...
2479
+ def __init__(
2480
+ self,
2481
+ *,
2482
+ hlo_module: tensorflow.compiler.xla.service.hlo_pb2.HloModuleProto | None = ...,
2483
+ config: global___HloModuleConfigProto | None = ...,
2484
+ ) -> None: ...
2485
+ def HasField(self, field_name: typing.Literal["config", b"config", "hlo_module", b"hlo_module"]) -> builtins.bool: ...
2486
+ def ClearField(self, field_name: typing.Literal["config", b"config", "hlo_module", b"hlo_module"]) -> None: ...
2487
+
2488
+ global___HloModuleProtoWithConfig = HloModuleProtoWithConfig
2489
+
2490
+ @typing.final
2491
+ class ScheduleProto(google.protobuf.message.Message):
2492
+ """A trace estimated by the Latency Hiding Scheduler."""
2493
+
2494
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
2495
+
2496
+ @typing.final
2497
+ class Instruction(google.protobuf.message.Message):
2498
+ DESCRIPTOR: google.protobuf.descriptor.Descriptor
2499
+
2500
+ ID_FIELD_NUMBER: builtins.int
2501
+ START_TIMESTAMP_CYCLES_FIELD_NUMBER: builtins.int
2502
+ END_TIMESTAMP_CYCLES_FIELD_NUMBER: builtins.int
2503
+ id: builtins.int
2504
+ """Instruction id (matches the id in HloInstructionProto)."""
2505
+ start_timestamp_cycles: builtins.float
2506
+ """Start and end timestamps in cycles."""
2507
+ end_timestamp_cycles: builtins.float
2508
+ def __init__(
2509
+ self,
2510
+ *,
2511
+ id: builtins.int | None = ...,
2512
+ start_timestamp_cycles: builtins.float | None = ...,
2513
+ end_timestamp_cycles: builtins.float | None = ...,
2514
+ ) -> None: ...
2515
+ def ClearField(
2516
+ self,
2517
+ field_name: typing.Literal[
2518
+ "end_timestamp_cycles", b"end_timestamp_cycles", "id", b"id", "start_timestamp_cycles", b"start_timestamp_cycles"
2519
+ ],
2520
+ ) -> None: ...
2521
+
2522
+ INSTRUCTIONS_FIELD_NUMBER: builtins.int
2523
+ COMPUTATION_ID_FIELD_NUMBER: builtins.int
2524
+ HLO_MODULE_FIELD_NUMBER: builtins.int
2525
+ CYCLES_PER_MICROSECOND_FIELD_NUMBER: builtins.int
2526
+ computation_id: builtins.int
2527
+ """Computation id (matches the id in HloComputationProto)."""
2528
+ cycles_per_microsecond: builtins.int
2529
+ @property
2530
+ def instructions(
2531
+ self,
2532
+ ) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ScheduleProto.Instruction]: ...
2533
+ @property
2534
+ def hlo_module(self) -> tensorflow.compiler.xla.service.hlo_pb2.HloModuleProto: ...
2535
+ def __init__(
2536
+ self,
2537
+ *,
2538
+ instructions: collections.abc.Iterable[global___ScheduleProto.Instruction] | None = ...,
2539
+ computation_id: builtins.int | None = ...,
2540
+ hlo_module: tensorflow.compiler.xla.service.hlo_pb2.HloModuleProto | None = ...,
2541
+ cycles_per_microsecond: builtins.int | None = ...,
2542
+ ) -> None: ...
2543
+ def HasField(self, field_name: typing.Literal["hlo_module", b"hlo_module"]) -> builtins.bool: ...
2544
+ def ClearField(
2545
+ self,
2546
+ field_name: typing.Literal[
2547
+ "computation_id",
2548
+ b"computation_id",
2549
+ "cycles_per_microsecond",
2550
+ b"cycles_per_microsecond",
2551
+ "hlo_module",
2552
+ b"hlo_module",
2553
+ "instructions",
2554
+ b"instructions",
2555
+ ],
2556
+ ) -> None: ...
2557
+
2558
+ global___ScheduleProto = ScheduleProto