basedpyright 1.36.2 → 1.37.0-961d59bacff1958014fe4e9ed2ef3116fe2266e2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/pyright-langserver.js +1 -1
- package/dist/pyright-langserver.js.map +1 -1
- package/dist/pyright.js +1 -1
- package/dist/pyright.js.map +1 -1
- package/dist/typeshed-fallback/README.md +6 -6
- package/dist/typeshed-fallback/commit.txt +1 -1
- package/dist/typeshed-fallback/stdlib/_compression.pyi +13 -2
- package/dist/typeshed-fallback/stdlib/_ctypes.pyi +49 -16
- package/dist/typeshed-fallback/stdlib/_winapi.pyi +9 -0
- package/dist/typeshed-fallback/stdlib/_zstd.pyi +7 -2
- package/dist/typeshed-fallback/stdlib/argparse.pyi +39 -10
- package/dist/typeshed-fallback/stdlib/asyncio/coroutines.pyi +27 -8
- package/dist/typeshed-fallback/stdlib/asyncio/protocols.pyi +2 -2
- package/dist/typeshed-fallback/stdlib/asyncio/runners.pyi +5 -2
- package/dist/typeshed-fallback/stdlib/asyncio/trsock.pyi +1 -1
- package/dist/typeshed-fallback/stdlib/builtins.pyi +54 -31
- package/dist/typeshed-fallback/stdlib/calendar.pyi +3 -1
- package/dist/typeshed-fallback/stdlib/codecs.pyi +2 -1
- package/dist/typeshed-fallback/stdlib/compression/_common/_streams.pyi +13 -2
- package/dist/typeshed-fallback/stdlib/compression/zstd/__init__.pyi +7 -2
- package/dist/typeshed-fallback/stdlib/compression/zstd/_zstdfile.pyi +6 -6
- package/dist/typeshed-fallback/stdlib/contextlib.pyi +15 -6
- package/dist/typeshed-fallback/stdlib/ctypes/__init__.pyi +3 -0
- package/dist/typeshed-fallback/stdlib/enum.pyi +2 -0
- package/dist/typeshed-fallback/stdlib/heapq.pyi +5 -0
- package/dist/typeshed-fallback/stdlib/html/parser.pyi +2 -1
- package/dist/typeshed-fallback/stdlib/http/client.pyi +1 -0
- package/dist/typeshed-fallback/stdlib/imaplib.pyi +2 -1
- package/dist/typeshed-fallback/stdlib/importlib/readers.pyi +2 -2
- package/dist/typeshed-fallback/stdlib/importlib/resources/abc.pyi +3 -2
- package/dist/typeshed-fallback/stdlib/importlib/resources/simple.pyi +2 -1
- package/dist/typeshed-fallback/stdlib/importlib/util.pyi +15 -1
- package/dist/typeshed-fallback/stdlib/ipaddress.pyi +1 -1
- package/dist/typeshed-fallback/stdlib/locale.pyi +4 -0
- package/dist/typeshed-fallback/stdlib/mimetypes.pyi +6 -6
- package/dist/typeshed-fallback/stdlib/mmap.pyi +12 -12
- package/dist/typeshed-fallback/stdlib/multiprocessing/dummy/__init__.pyi +20 -8
- package/dist/typeshed-fallback/stdlib/multiprocessing/managers.pyi +4 -1
- package/dist/typeshed-fallback/stdlib/multiprocessing/process.pyi +4 -0
- package/dist/typeshed-fallback/stdlib/multiprocessing/synchronize.pyi +3 -0
- package/dist/typeshed-fallback/stdlib/optparse.pyi +1 -1
- package/dist/typeshed-fallback/stdlib/os/__init__.pyi +217 -127
- package/dist/typeshed-fallback/stdlib/parser.pyi +4 -4
- package/dist/typeshed-fallback/stdlib/pathlib/__init__.pyi +11 -0
- package/dist/typeshed-fallback/stdlib/pdb.pyi +14 -4
- package/dist/typeshed-fallback/stdlib/pyexpat/__init__.pyi +6 -0
- package/dist/typeshed-fallback/stdlib/select.pyi +8 -4
- package/dist/typeshed-fallback/stdlib/socket.pyi +15 -0
- package/dist/typeshed-fallback/stdlib/sqlite3/__init__.pyi +3 -2
- package/dist/typeshed-fallback/stdlib/ssl.pyi +3 -0
- package/dist/typeshed-fallback/stdlib/stat.pyi +108 -1
- package/dist/typeshed-fallback/stdlib/subprocess.pyi +60 -61
- package/dist/typeshed-fallback/stdlib/sys/__init__.pyi +41 -4
- package/dist/typeshed-fallback/stdlib/sys/_monitoring.pyi +8 -4
- package/dist/typeshed-fallback/stdlib/sysconfig.pyi +13 -4
- package/dist/typeshed-fallback/stdlib/tarfile.pyi +6 -6
- package/dist/typeshed-fallback/stdlib/threading.pyi +9 -4
- package/dist/typeshed-fallback/stdlib/tkinter/constants.pyi +6 -6
- package/dist/typeshed-fallback/stdlib/tkinter/ttk.pyi +14 -6
- package/dist/typeshed-fallback/stdlib/types.pyi +8 -11
- package/dist/typeshed-fallback/stdlib/typing.pyi +18 -12
- package/dist/typeshed-fallback/stdlib/typing_extensions.pyi +1 -0
- package/dist/typeshed-fallback/stdlib/unittest/mock.pyi +1 -0
- package/dist/typeshed-fallback/stdlib/unittest/util.pyi +19 -2
- package/dist/typeshed-fallback/stdlib/urllib/request.pyi +7 -0
- package/dist/typeshed-fallback/stdlib/uuid.pyi +6 -5
- package/dist/typeshed-fallback/stdlib/winreg.pyi +3 -3
- package/dist/typeshed-fallback/stdlib/xml/etree/ElementTree.pyi +1 -0
- package/dist/typeshed-fallback/stdlib/zipfile/__init__.pyi +2 -0
- package/dist/typeshed-fallback/stdlib/zlib.pyi +2 -2
- package/dist/typeshed-fallback/stubs/Authlib/METADATA.toml +1 -2
- package/dist/typeshed-fallback/stubs/Authlib/authlib/common/urls.pyi +1 -1
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/base_client/async_app.pyi +4 -3
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/base_client/async_openid.pyi +4 -2
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/base_client/framework_integration.pyi +2 -2
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/base_client/registry.pyi +1 -1
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/base_client/sync_app.pyi +3 -3
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/base_client/sync_openid.pyi +4 -2
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/django_client/__init__.pyi +10 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/django_client/apps.pyi +14 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/django_client/integration.pyi +11 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/django_oauth1/__init__.pyi +4 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/django_oauth1/authorization_server.pyi +26 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/django_oauth1/nonce.pyi +1 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/django_oauth1/resource_protector.pyi +14 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/django_oauth2/__init__.pyi +8 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/django_oauth2/authorization_server.pyi +24 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/django_oauth2/endpoints.pyi +5 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/django_oauth2/requests.pyi +20 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/django_oauth2/resource_protector.pyi +15 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/django_oauth2/signals.pyi +6 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/flask_client/__init__.pyi +20 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/flask_client/apps.pyi +18 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/flask_client/integration.pyi +10 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/flask_oauth1/__init__.pyi +7 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/flask_oauth1/authorization_server.pyi +29 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/flask_oauth1/cache.pyi +8 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/flask_oauth1/resource_protector.pyi +18 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/flask_oauth2/__init__.pyi +7 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/flask_oauth2/authorization_server.pyi +23 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/flask_oauth2/errors.pyi +14 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/flask_oauth2/requests.pyi +27 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/flask_oauth2/resource_protector.pyi +15 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/flask_oauth2/signals.pyi +5 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/httpx_client/__init__.pyi +37 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/httpx_client/assertion_client.pyi +50 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/httpx_client/oauth1_client.pyi +56 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/httpx_client/oauth2_client.pyi +72 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/httpx_client/utils.pyi +7 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/requests_client/__init__.pyi +28 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/requests_client/assertion_session.pyi +31 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/requests_client/oauth1_session.pyi +29 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/requests_client/oauth2_session.pyi +43 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/requests_client/utils.pyi +6 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/sqla_oauth2/__init__.pyi +20 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/sqla_oauth2/client_mixin.pyi +55 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/sqla_oauth2/functions.pyi +21 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/sqla_oauth2/tokens_mixins.pyi +39 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/starlette_client/__init__.pyi +14 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/starlette_client/apps.pyi +16 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/integrations/starlette_client/integration.pyi +13 -0
- package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/drafts/_jwe_algorithms.pyi +12 -10
- package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/drafts/_jwe_enc_cryptography.pyi +2 -2
- package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/jwk.pyi +6 -1
- package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7515/jws.pyi +6 -4
- package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7515/models.pyi +3 -2
- package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7516/jwe.pyi +9 -8
- package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7516/models.pyi +12 -11
- package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7517/_cryptography_key.pyi +35 -1
- package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7517/asymmetric_key.pyi +6 -5
- package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7517/base_key.pyi +8 -8
- package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7517/jwk.pyi +1 -1
- package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7517/key_set.pyi +3 -2
- package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7518/jwe_algs.pyi +28 -23
- package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7518/jwe_encs.pyi +6 -5
- package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7518/jws_algs.pyi +1 -1
- package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7518/oct_key.pyi +5 -4
- package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7518/rsa_key.pyi +2 -1
- package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7518/util.pyi +4 -2
- package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc7519/jwt.pyi +2 -2
- package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc8037/jws_eddsa.pyi +4 -2
- package/dist/typeshed-fallback/stubs/Authlib/authlib/jose/rfc8037/okp_key.pyi +2 -2
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth1/client.pyi +2 -2
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth1/rfc5849/authorization_server.pyi +7 -7
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth1/rfc5849/base_server.pyi +2 -2
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth1/rfc5849/client_auth.pyi +6 -6
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth1/rfc5849/errors.pyi +1 -1
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth1/rfc5849/models.pyi +8 -8
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth1/rfc5849/parameters.pyi +1 -1
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth1/rfc5849/resource_protector.pyi +4 -2
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth1/rfc5849/rsa.pyi +1 -1
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth1/rfc5849/signature.pyi +20 -18
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth1/rfc5849/util.pyi +2 -2
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/base.pyi +1 -1
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/client.pyi +1 -1
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc6749/authorization_server.pyi +1 -1
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc6749/grants/authorization_code.pyi +2 -2
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc6749/grants/base.pyi +1 -1
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc6749/grants/resource_owner_password_credentials.pyi +1 -1
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc6749/parameters.pyi +5 -3
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc6749/resource_protector.pyi +1 -1
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc6749/token_endpoint.pyi +2 -2
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc6749/wrappers.pyi +1 -1
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc6750/errors.pyi +1 -1
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc6750/parameters.pyi +1 -1
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc6750/validator.pyi +1 -1
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc7009/revocation.pyi +2 -2
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc7523/assertion.pyi +3 -3
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc7523/auth.pyi +2 -2
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc7523/client.pyi +5 -3
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc7523/jwt_bearer.pyi +6 -5
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc7591/endpoint.pyi +4 -4
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc7592/endpoint.pyi +8 -8
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc7662/introspection.pyi +3 -3
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc7662/token_validator.pyi +1 -1
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc8414/well_known.pyi +1 -1
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc8628/device_code.pyi +3 -3
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc8628/endpoint.pyi +2 -2
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc8628/models.pyi +4 -4
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc9068/introspection.pyi +1 -1
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc9068/revocation.pyi +3 -2
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc9068/token.pyi +1 -1
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oauth2/rfc9068/token_validator.pyi +1 -1
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oidc/core/claims.pyi +6 -8
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oidc/core/grants/code.pyi +2 -1
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oidc/core/grants/hybrid.pyi +4 -4
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oidc/core/grants/implicit.pyi +6 -4
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oidc/core/userinfo.pyi +1 -1
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oidc/core/util.pyi +7 -1
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oidc/discovery/models.pyi +1 -3
- package/dist/typeshed-fallback/stubs/Authlib/authlib/oidc/discovery/well_known.pyi +1 -1
- package/dist/typeshed-fallback/stubs/Deprecated/METADATA.toml +1 -1
- package/dist/typeshed-fallback/stubs/Deprecated/deprecated/__init__.pyi +7 -2
- package/dist/typeshed-fallback/stubs/Deprecated/deprecated/params.pyi +20 -0
- package/dist/typeshed-fallback/stubs/Flask-SocketIO/METADATA.toml +1 -1
- package/dist/typeshed-fallback/stubs/Markdown/METADATA.toml +1 -1
- package/dist/typeshed-fallback/stubs/Markdown/markdown/htmlparser.pyi +2 -0
- package/dist/typeshed-fallback/stubs/PyMySQL/pymysql/cursors.pyi +1 -0
- package/dist/typeshed-fallback/stubs/Pygments/pygments/console.pyi +4 -4
- package/dist/typeshed-fallback/stubs/Pygments/pygments/filter.pyi +3 -3
- package/dist/typeshed-fallback/stubs/Pygments/pygments/filters/__init__.pyi +14 -14
- package/dist/typeshed-fallback/stubs/Pygments/pygments/formatter.pyi +10 -9
- package/dist/typeshed-fallback/stubs/Pygments/pygments/formatters/__init__.pyi +2 -2
- package/dist/typeshed-fallback/stubs/Pygments/pygments/formatters/_mapping.pyi +2 -2
- package/dist/typeshed-fallback/stubs/Pygments/pygments/formatters/bbcode.pyi +5 -4
- package/dist/typeshed-fallback/stubs/Pygments/pygments/formatters/html.pyi +27 -26
- package/dist/typeshed-fallback/stubs/Pygments/pygments/formatters/img.pyi +33 -32
- package/dist/typeshed-fallback/stubs/Pygments/pygments/formatters/irc.pyi +7 -6
- package/dist/typeshed-fallback/stubs/Pygments/pygments/formatters/latex.pyi +21 -20
- package/dist/typeshed-fallback/stubs/Pygments/pygments/formatters/other.pyi +9 -8
- package/dist/typeshed-fallback/stubs/Pygments/pygments/formatters/pangomarkup.pyi +5 -4
- package/dist/typeshed-fallback/stubs/Pygments/pygments/formatters/rtf.pyi +6 -5
- package/dist/typeshed-fallback/stubs/Pygments/pygments/formatters/svg.pyi +15 -14
- package/dist/typeshed-fallback/stubs/Pygments/pygments/formatters/terminal.pyi +7 -6
- package/dist/typeshed-fallback/stubs/Pygments/pygments/formatters/terminal256.pyi +18 -17
- package/dist/typeshed-fallback/stubs/Pygments/pygments/lexers/__init__.pyi +6 -7
- package/dist/typeshed-fallback/stubs/Pygments/pygments/plugin.pyi +5 -5
- package/dist/typeshed-fallback/stubs/Pygments/pygments/regexopt.pyi +3 -3
- package/dist/typeshed-fallback/stubs/Pygments/pygments/scanner.pyi +6 -6
- package/dist/typeshed-fallback/stubs/Pygments/pygments/unistring.pyi +2 -2
- package/dist/typeshed-fallback/stubs/Pygments/pygments/util.pyi +5 -5
- package/dist/typeshed-fallback/stubs/antlr4-python3-runtime/antlr4/BufferedTokenStream.pyi +14 -15
- package/dist/typeshed-fallback/stubs/antlr4-python3-runtime/antlr4/CommonTokenFactory.pyi +15 -7
- package/dist/typeshed-fallback/stubs/antlr4-python3-runtime/antlr4/CommonTokenStream.pyi +1 -3
- package/dist/typeshed-fallback/stubs/antlr4-python3-runtime/antlr4/FileStream.pyi +2 -4
- package/dist/typeshed-fallback/stubs/antlr4-python3-runtime/antlr4/InputStream.pyi +13 -11
- package/dist/typeshed-fallback/stubs/antlr4-python3-runtime/antlr4/Lexer.pyi +43 -29
- package/dist/typeshed-fallback/stubs/antlr4-python3-runtime/antlr4/Parser.pyi +47 -34
- package/dist/typeshed-fallback/stubs/antlr4-python3-runtime/antlr4/ParserInterpreter.pyi +13 -13
- package/dist/typeshed-fallback/stubs/antlr4-python3-runtime/antlr4/ParserRuleContext.pyi +30 -23
- package/dist/typeshed-fallback/stubs/antlr4-python3-runtime/antlr4/Recognizer.pyi +26 -17
- package/dist/typeshed-fallback/stubs/antlr4-python3-runtime/antlr4/RuleContext.pyi +22 -20
- package/dist/typeshed-fallback/stubs/antlr4-python3-runtime/antlr4/Token.pyi +26 -25
- package/dist/typeshed-fallback/stubs/antlr4-python3-runtime/antlr4/tree/Tree.pyi +30 -26
- package/dist/typeshed-fallback/stubs/assertpy/assertpy/assertpy.pyi +13 -10
- package/dist/typeshed-fallback/stubs/assertpy/assertpy/base.pyi +4 -5
- package/dist/typeshed-fallback/stubs/assertpy/assertpy/collection.pyi +12 -4
- package/dist/typeshed-fallback/stubs/assertpy/assertpy/contains.pyi +10 -8
- package/dist/typeshed-fallback/stubs/assertpy/assertpy/dict.pyi +8 -6
- package/dist/typeshed-fallback/stubs/assertpy/assertpy/dynamic.pyi +1 -1
- package/dist/typeshed-fallback/stubs/assertpy/assertpy/extracting.pyi +4 -2
- package/dist/typeshed-fallback/stubs/assertpy/assertpy/file.pyi +2 -3
- package/dist/typeshed-fallback/stubs/atheris/METADATA.toml +6 -0
- package/dist/typeshed-fallback/stubs/atheris/atheris/__init__.pyi +9 -0
- package/dist/typeshed-fallback/stubs/atheris/atheris/function_hooks.pyi +14 -0
- package/dist/typeshed-fallback/stubs/atheris/atheris/import_hook.pyi +36 -0
- package/dist/typeshed-fallback/stubs/atheris/atheris/instrument_bytecode.pyi +7 -0
- package/dist/typeshed-fallback/stubs/atheris/atheris/utils.pyi +18 -0
- package/dist/typeshed-fallback/stubs/atheris/atheris/version_dependent.pyi +27 -0
- package/dist/typeshed-fallback/stubs/aws-xray-sdk/METADATA.toml +1 -1
- package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/__init__.pyi +2 -1
- package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/async_context.pyi +15 -5
- package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/async_recorder.pyi +28 -9
- package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/context.pyi +7 -7
- package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/daemon_config.pyi +9 -7
- package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/emitters/udp_emitter.pyi +2 -2
- package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/lambda_launcher.pyi +8 -12
- package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/models/entity.pyi +1 -1
- package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/models/facade_segment.pyi +1 -1
- package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/models/http.pyi +9 -9
- package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/models/segment.pyi +1 -1
- package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/models/subsegment.pyi +3 -2
- package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/models/throwable.pyi +11 -3
- package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/patcher.pyi +3 -2
- package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/plugins/ec2_plugin.pyi +3 -3
- package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/plugins/ecs_plugin.pyi +2 -2
- package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/plugins/elasticbeanstalk_plugin.pyi +3 -3
- package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/recorder.pyi +19 -15
- package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/sampling/rule_cache.pyi +1 -1
- package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/sampling/sampler.pyi +3 -1
- package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/sampling/sampling_rule.pyi +23 -13
- package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/core/sampling/target_poller.pyi +5 -1
- package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/ext/psycopg/__init__.pyi +3 -0
- package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/ext/psycopg/patch.pyi +1 -0
- package/dist/typeshed-fallback/stubs/aws-xray-sdk/aws_xray_sdk/ext/util.pyi +1 -1
- package/dist/typeshed-fallback/stubs/bleach/METADATA.toml +1 -1
- package/dist/typeshed-fallback/stubs/bleach/bleach/html5lib_shim.pyi +1 -1
- package/dist/typeshed-fallback/stubs/bleach/bleach/linkifier.pyi +1 -1
- package/dist/typeshed-fallback/stubs/bleach/bleach/sanitizer.pyi +2 -2
- package/dist/typeshed-fallback/stubs/braintree/METADATA.toml +1 -1
- package/dist/typeshed-fallback/stubs/braintree/braintree/address.pyi +8 -4
- package/dist/typeshed-fallback/stubs/braintree/braintree/address_gateway.pyi +3 -2
- package/dist/typeshed-fallback/stubs/braintree/braintree/client_token.pyi +5 -1
- package/dist/typeshed-fallback/stubs/braintree/braintree/client_token_gateway.pyi +5 -3
- package/dist/typeshed-fallback/stubs/braintree/braintree/error_codes.pyi +11 -2
- package/dist/typeshed-fallback/stubs/braintree/braintree/payment_method.pyi +30 -1
- package/dist/typeshed-fallback/stubs/braintree/braintree/payment_method_gateway.pyi +30 -2
- package/dist/typeshed-fallback/stubs/braintree/braintree/subscription.pyi +10 -5
- package/dist/typeshed-fallback/stubs/braintree/braintree/subscription_details.pyi +5 -1
- package/dist/typeshed-fallback/stubs/braintree/braintree/subscription_gateway.pyi +12 -5
- package/dist/typeshed-fallback/stubs/braintree/braintree/transaction.pyi +6 -2
- package/dist/typeshed-fallback/stubs/braintree/braintree/transaction_gateway.pyi +5 -2
- package/dist/typeshed-fallback/stubs/colorful/METADATA.toml +2 -0
- package/dist/typeshed-fallback/stubs/colorful/colorful/ansi.pyi +14 -0
- package/dist/typeshed-fallback/stubs/colorful/colorful/colors.pyi +6 -0
- package/dist/typeshed-fallback/stubs/colorful/colorful/core.pyi +102 -0
- package/dist/typeshed-fallback/stubs/colorful/colorful/styles.pyi +4 -0
- package/dist/typeshed-fallback/stubs/colorful/colorful/terminal.pyi +16 -0
- package/dist/typeshed-fallback/stubs/colorful/colorful/utils.pyi +2 -0
- package/dist/typeshed-fallback/stubs/decorator/decorator.pyi +2 -2
- package/dist/typeshed-fallback/stubs/django-filter/django_filters/filterset.pyi +3 -1
- package/dist/typeshed-fallback/stubs/docker/docker/api/container.pyi +6 -1
- package/dist/typeshed-fallback/stubs/docker/docker/api/daemon.pyi +12 -4
- package/dist/typeshed-fallback/stubs/docker/docker/api/exec_api.pyi +123 -7
- package/dist/typeshed-fallback/stubs/docker/docker/client.pyi +12 -8
- package/dist/typeshed-fallback/stubs/docker/docker/models/containers.pyi +27 -17
- package/dist/typeshed-fallback/stubs/docker/docker/models/images.pyi +1 -1
- package/dist/typeshed-fallback/stubs/docker/docker/models/resource.pyi +5 -5
- package/dist/typeshed-fallback/stubs/docutils/METADATA.toml +1 -1
- package/dist/typeshed-fallback/stubs/docutils/docutils/parsers/rst/directives/body.pyi +10 -0
- package/dist/typeshed-fallback/stubs/et_xmlfile/METADATA.toml +2 -0
- package/dist/typeshed-fallback/stubs/et_xmlfile/et_xmlfile/__init__.pyi +9 -0
- package/dist/typeshed-fallback/stubs/et_xmlfile/et_xmlfile/incremental_tree.pyi +170 -0
- package/dist/typeshed-fallback/stubs/et_xmlfile/et_xmlfile/xmlfile.pyi +37 -0
- package/dist/typeshed-fallback/stubs/fanstatic/METADATA.toml +1 -1
- package/dist/typeshed-fallback/stubs/flake8-bugbear/METADATA.toml +1 -1
- package/dist/typeshed-fallback/stubs/flake8-bugbear/bugbear.pyi +37 -65
- package/dist/typeshed-fallback/stubs/flake8-builtins/METADATA.toml +1 -1
- package/dist/typeshed-fallback/stubs/flake8-simplify/METADATA.toml +1 -1
- package/dist/typeshed-fallback/stubs/flake8-simplify/flake8_simplify/__init__.pyi +1 -0
- package/dist/typeshed-fallback/stubs/flake8-simplify/flake8_simplify/constants.pyi +6 -3
- package/dist/typeshed-fallback/stubs/flake8-simplify/flake8_simplify/utils.pyi +0 -1
- package/dist/typeshed-fallback/stubs/fpdf2/fpdf/drawing.pyi +145 -81
- package/dist/typeshed-fallback/stubs/fpdf2/fpdf/text_region.pyi +1 -1
- package/dist/typeshed-fallback/stubs/geopandas/METADATA.toml +1 -1
- package/dist/typeshed-fallback/stubs/geopandas/geopandas/array.pyi +7 -14
- package/dist/typeshed-fallback/stubs/geopandas/geopandas/geodataframe.pyi +1 -0
- package/dist/typeshed-fallback/stubs/gevent/METADATA.toml +2 -2
- package/dist/typeshed-fallback/stubs/gevent/gevent/events.pyi +1 -12
- package/dist/typeshed-fallback/stubs/gevent/gevent/libuv/watcher.pyi +1 -0
- package/dist/typeshed-fallback/stubs/gevent/gevent/os.pyi +1 -0
- package/dist/typeshed-fallback/stubs/gevent/gevent/signal.pyi +2 -1
- package/dist/typeshed-fallback/stubs/google-cloud-ndb/METADATA.toml +1 -1
- package/dist/typeshed-fallback/stubs/google-cloud-ndb/google/cloud/ndb/_cache.pyi +9 -10
- package/dist/typeshed-fallback/stubs/google-cloud-ndb/google/cloud/ndb/_datastore_query.pyi +1 -2
- package/dist/typeshed-fallback/stubs/google-cloud-ndb/google/cloud/ndb/_eventloop.pyi +11 -10
- package/dist/typeshed-fallback/stubs/google-cloud-ndb/google/cloud/ndb/blobstore.pyi +6 -6
- package/dist/typeshed-fallback/stubs/google-cloud-ndb/google/cloud/ndb/context.pyi +21 -21
- package/dist/typeshed-fallback/stubs/google-cloud-ndb/google/cloud/ndb/exceptions.pyi +2 -2
- package/dist/typeshed-fallback/stubs/google-cloud-ndb/google/cloud/ndb/global_cache.pyi +13 -14
- package/dist/typeshed-fallback/stubs/google-cloud-ndb/google/cloud/ndb/key.pyi +2 -3
- package/dist/typeshed-fallback/stubs/google-cloud-ndb/google/cloud/ndb/metadata.pyi +1 -2
- package/dist/typeshed-fallback/stubs/google-cloud-ndb/google/cloud/ndb/polymodel.pyi +2 -2
- package/dist/typeshed-fallback/stubs/google-cloud-ndb/google/cloud/ndb/query.pyi +26 -27
- package/dist/typeshed-fallback/stubs/google-cloud-ndb/google/cloud/ndb/stats.pyi +37 -37
- package/dist/typeshed-fallback/stubs/google-cloud-ndb/google/cloud/ndb/tasklets.pyi +7 -7
- package/dist/typeshed-fallback/stubs/google-cloud-ndb/google/cloud/ndb/utils.pyi +4 -4
- package/dist/typeshed-fallback/stubs/greenlet/METADATA.toml +1 -1
- package/dist/typeshed-fallback/stubs/gunicorn/gunicorn/arbiter.pyi +2 -2
- package/dist/typeshed-fallback/stubs/gunicorn/gunicorn/sock.pyi +2 -0
- package/dist/typeshed-fallback/stubs/html5lib/METADATA.toml +1 -0
- package/dist/typeshed-fallback/stubs/html5lib/html5lib/_ihatexml.pyi +21 -23
- package/dist/typeshed-fallback/stubs/html5lib/html5lib/_inputstream.pyi +79 -71
- package/dist/typeshed-fallback/stubs/html5lib/html5lib/_tokenizer.pyi +87 -51
- package/dist/typeshed-fallback/stubs/html5lib/html5lib/_utils.pyi +18 -11
- package/dist/typeshed-fallback/stubs/html5lib/html5lib/constants.pyi +1 -3
- package/dist/typeshed-fallback/stubs/html5lib/html5lib/filters/alphabeticalattributes.pyi +3 -2
- package/dist/typeshed-fallback/stubs/html5lib/html5lib/filters/base.pyi +9 -6
- package/dist/typeshed-fallback/stubs/html5lib/html5lib/filters/inject_meta_charset.pyi +5 -3
- package/dist/typeshed-fallback/stubs/html5lib/html5lib/filters/lint.pyi +5 -3
- package/dist/typeshed-fallback/stubs/html5lib/html5lib/filters/optionaltags.pyi +1 -2
- package/dist/typeshed-fallback/stubs/html5lib/html5lib/filters/sanitizer.pyi +6 -6
- package/dist/typeshed-fallback/stubs/html5lib/html5lib/filters/whitespace.pyi +2 -2
- package/dist/typeshed-fallback/stubs/html5lib/html5lib/html5parser.pyi +2 -2
- package/dist/typeshed-fallback/stubs/html5lib/html5lib/serializer.pyi +75 -19
- package/dist/typeshed-fallback/stubs/html5lib/html5lib/treeadapters/sax.pyi +0 -3
- package/dist/typeshed-fallback/stubs/hvac/METADATA.toml +1 -1
- package/dist/typeshed-fallback/stubs/hvac/hvac/api/auth_methods/__init__.pyi +3 -4
- package/dist/typeshed-fallback/stubs/hvac/hvac/api/auth_methods/aws.pyi +2 -2
- package/dist/typeshed-fallback/stubs/hvac/hvac/api/auth_methods/azure.pyi +3 -2
- package/dist/typeshed-fallback/stubs/hvac/hvac/api/auth_methods/gcp.pyi +3 -2
- package/dist/typeshed-fallback/stubs/hvac/hvac/api/auth_methods/kubernetes.pyi +1 -0
- package/dist/typeshed-fallback/stubs/hvac/hvac/api/auth_methods/legacy_mfa.pyi +2 -4
- package/dist/typeshed-fallback/stubs/hvac/hvac/api/secrets_engines/__init__.pyi +3 -4
- package/dist/typeshed-fallback/stubs/hvac/hvac/api/secrets_engines/identity.pyi +2 -2
- package/dist/typeshed-fallback/stubs/hvac/hvac/api/secrets_engines/kv.pyi +3 -3
- package/dist/typeshed-fallback/stubs/hvac/hvac/api/system_backend/__init__.pyi +3 -4
- package/dist/typeshed-fallback/stubs/hvac/hvac/api/system_backend/system_backend_mixin.pyi +2 -2
- package/dist/typeshed-fallback/stubs/hvac/hvac/api/system_backend/wrapping.pyi +2 -3
- package/dist/typeshed-fallback/stubs/hvac/hvac/api/vault_api_category.pyi +2 -3
- package/dist/typeshed-fallback/stubs/icalendar/METADATA.toml +1 -1
- package/dist/typeshed-fallback/stubs/icalendar/icalendar/__init__.pyi +0 -1
- package/dist/typeshed-fallback/stubs/icalendar/icalendar/alarms.pyi +2 -2
- package/dist/typeshed-fallback/stubs/icalendar/icalendar/cal.pyi +16 -9
- package/dist/typeshed-fallback/stubs/icalendar/icalendar/parser.pyi +2 -1
- package/dist/typeshed-fallback/stubs/icalendar/icalendar/prop.pyi +3 -3
- package/dist/typeshed-fallback/stubs/icalendar/icalendar/timezone/zoneinfo.pyi +1 -5
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/__init__.pyi +5 -5
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/abstract/__init__.pyi +3 -3
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/abstract/attrDef.pyi +13 -13
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/abstract/attribute.pyi +9 -9
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/abstract/cursor.pyi +28 -27
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/abstract/entry.pyi +11 -11
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/core/exceptions.pyi +8 -8
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/core/pooling.pyi +15 -15
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/core/rdns.pyi +7 -7
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/core/results.pyi +3 -3
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/core/server.pyi +15 -14
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/core/timezone.pyi +3 -3
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/core/tls.pyi +13 -13
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/core/usage.pyi +6 -6
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/extend/__init__.pyi +4 -4
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/extend/microsoft/dirSync.pyi +12 -12
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/extend/microsoft/persistentSearch.pyi +7 -7
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/extend/standard/PersistentSearch.pyi +13 -13
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/operation/modify.pyi +2 -2
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/operation/search.pyi +5 -5
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/protocol/formatters/formatters.pyi +2 -2
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/protocol/formatters/standard.pyi +2 -2
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/protocol/oid.pyi +2 -2
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/protocol/rfc2849.pyi +2 -2
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/protocol/rfc4512.pyi +59 -59
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/strategy/asyncStream.pyi +7 -7
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/strategy/asynchronous.pyi +6 -6
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/strategy/base.pyi +7 -7
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/strategy/ldifProducer.pyi +4 -4
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/strategy/mockAsync.pyi +2 -2
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/strategy/mockBase.pyi +9 -9
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/strategy/mockSync.pyi +2 -2
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/strategy/restartable.pyi +4 -4
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/strategy/reusable.pyi +20 -20
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/strategy/sync.pyi +3 -3
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/utils/config.pyi +2 -2
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/utils/hashed.pyi +3 -3
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/utils/log.pyi +3 -3
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/utils/ntlm.pyi +41 -41
- package/dist/typeshed-fallback/stubs/ldap3/ldap3/utils/repr.pyi +2 -2
- package/dist/typeshed-fallback/stubs/lupa/METADATA.toml +1 -1
- package/dist/typeshed-fallback/stubs/netaddr/netaddr/ip/__init__.pyi +4 -1
- package/dist/typeshed-fallback/stubs/netaddr/netaddr/ip/iana.pyi +8 -8
- package/dist/typeshed-fallback/stubs/networkx/METADATA.toml +1 -1
- package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/__init__.pyi +1 -0
- package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/approximation/steinertree.pyi +4 -0
- package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/approximation/treewidth.pyi +11 -7
- package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/assortativity/mixing.pyi +4 -4
- package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/bipartite/matching.pyi +3 -3
- package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/bipartite/matrix.pyi +8 -1
- package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/centrality/closeness.pyi +2 -2
- package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/centrality/current_flow_betweenness.pyi +2 -0
- package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/centrality/eigenvector.pyi +3 -2
- package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/centrality/katz.pyi +5 -4
- package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/centrality/percolation.pyi +3 -2
- package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/centrality/reaching.pyi +3 -2
- package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/centrality/subgraph_alg.pyi +2 -2
- package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/cluster.pyi +13 -3
- package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/community/__init__.pyi +1 -1
- package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/community/bipartitions.pyi +22 -0
- package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/core.pyi +6 -5
- package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/dag.pyi +0 -3
- package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/distance_regular.pyi +2 -2
- package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/isomorphism/ismags.pyi +17 -4
- package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/non_randomness.pyi +1 -1
- package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/perfect_graph.pyi +7 -0
- package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/planarity.pyi +13 -4
- package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/similarity.pyi +19 -3
- package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/threshold.pyi +0 -1
- package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/tree/__init__.pyi +1 -0
- package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/tree/distance_measures.pyi +10 -0
- package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/wiener.pyi +3 -1
- package/dist/typeshed-fallback/stubs/networkx/networkx/classes/digraph.pyi +13 -7
- package/dist/typeshed-fallback/stubs/networkx/networkx/classes/function.pyi +3 -1
- package/dist/typeshed-fallback/stubs/networkx/networkx/classes/graph.pyi +12 -6
- package/dist/typeshed-fallback/stubs/networkx/networkx/classes/multidigraph.pyi +10 -9
- package/dist/typeshed-fallback/stubs/networkx/networkx/classes/multigraph.pyi +23 -13
- package/dist/typeshed-fallback/stubs/networkx/networkx/drawing/layout.pyi +0 -1
- package/dist/typeshed-fallback/stubs/networkx/networkx/drawing/nx_pylab.pyi +20 -0
- package/dist/typeshed-fallback/stubs/networkx/networkx/generators/expanders.pyi +9 -1
- package/dist/typeshed-fallback/stubs/networkx/networkx/generators/random_graphs.pyi +21 -16
- package/dist/typeshed-fallback/stubs/networkx/networkx/generators/small.pyi +3 -0
- package/dist/typeshed-fallback/stubs/networkx/networkx/readwrite/json_graph/node_link.pyi +2 -24
- package/dist/typeshed-fallback/stubs/networkx/networkx/utils/misc.pyi +1 -1
- package/dist/typeshed-fallback/stubs/networkx/networkx/utils/random_sequence.pyi +5 -0
- package/dist/typeshed-fallback/stubs/networkx/networkx/utils/union_find.pyi +12 -10
- package/dist/typeshed-fallback/stubs/parsimonious/METADATA.toml +1 -1
- package/dist/typeshed-fallback/stubs/parsimonious/parsimonious/exceptions.pyi +5 -3
- package/dist/typeshed-fallback/stubs/peewee/METADATA.toml +1 -1
- package/dist/typeshed-fallback/stubs/protobuf/google/protobuf/internal/containers.pyi +8 -3
- package/dist/typeshed-fallback/stubs/protobuf/google/protobuf/internal/well_known_types.pyi +2 -2
- package/dist/typeshed-fallback/stubs/protobuf/google/protobuf/message.pyi +1 -1
- package/dist/typeshed-fallback/stubs/psutil/METADATA.toml +1 -1
- package/dist/typeshed-fallback/stubs/psutil/psutil/__init__.pyi +57 -77
- package/dist/typeshed-fallback/stubs/psutil/psutil/_common.pyi +135 -262
- package/dist/typeshed-fallback/stubs/psutil/psutil/_ntuples.pyi +384 -0
- package/dist/typeshed-fallback/stubs/psutil/psutil/_psaix.pyi +100 -86
- package/dist/typeshed-fallback/stubs/psutil/psutil/_psbsd.pyi +213 -168
- package/dist/typeshed-fallback/stubs/psutil/psutil/_pslinux.pyi +181 -237
- package/dist/typeshed-fallback/stubs/psutil/psutil/_psosx.pyi +81 -100
- package/dist/typeshed-fallback/stubs/psutil/psutil/_psposix.pyi +78 -4
- package/dist/typeshed-fallback/stubs/psutil/psutil/_pssunos.pyi +141 -119
- package/dist/typeshed-fallback/stubs/psutil/psutil/_psutil_aix.pyi +58 -0
- package/dist/typeshed-fallback/stubs/psutil/psutil/_psutil_bsd.pyi +141 -0
- package/dist/typeshed-fallback/stubs/psutil/psutil/_psutil_linux.pyi +47 -15
- package/dist/typeshed-fallback/stubs/psutil/psutil/_psutil_osx.pyi +77 -49
- package/dist/typeshed-fallback/stubs/psutil/psutil/_psutil_sunos.pyi +63 -0
- package/dist/typeshed-fallback/stubs/psutil/psutil/_psutil_windows.pyi +102 -91
- package/dist/typeshed-fallback/stubs/psutil/psutil/_pswindows.pyi +175 -200
- package/dist/typeshed-fallback/stubs/pycurl/pycurl.pyi +5 -3
- package/dist/typeshed-fallback/stubs/pyinstaller/METADATA.toml +1 -1
- package/dist/typeshed-fallback/stubs/python-crontab/crontab.pyi +4 -4
- package/dist/typeshed-fallback/stubs/python-dateutil/dateutil/parser/_parser.pyi +6 -7
- package/dist/typeshed-fallback/stubs/python-dateutil/dateutil/rrule.pyi +28 -25
- package/dist/typeshed-fallback/stubs/python-dateutil/dateutil/tz/_common.pyi +8 -3
- package/dist/typeshed-fallback/stubs/python-dateutil/dateutil/tz/tz.pyi +21 -19
- package/dist/typeshed-fallback/stubs/python-dateutil/dateutil/tz/win.pyi +5 -4
- package/dist/typeshed-fallback/stubs/python-dateutil/dateutil/zoneinfo/__init__.pyi +2 -1
- package/dist/typeshed-fallback/stubs/python-dateutil/dateutil/zoneinfo/rebuild.pyi +5 -1
- package/dist/typeshed-fallback/stubs/python-jenkins/jenkins/__init__.pyi +13 -6
- package/dist/typeshed-fallback/stubs/pytz/pytz/lazy.pyi +2 -2
- package/dist/typeshed-fallback/stubs/pyxdg/xdg/RecentFiles.pyi +5 -7
- package/dist/typeshed-fallback/stubs/qrbill/METADATA.toml +1 -1
- package/dist/typeshed-fallback/stubs/qrbill/qrbill/bill.pyi +10 -0
- package/dist/typeshed-fallback/stubs/regex/METADATA.toml +1 -1
- package/dist/typeshed-fallback/stubs/regex/regex/__init__.pyi +64 -1
- package/dist/typeshed-fallback/stubs/regex/regex/{regex.pyi → _main.pyi} +66 -0
- package/dist/typeshed-fallback/stubs/regex/regex/_regex.pyi +1 -1
- package/dist/typeshed-fallback/stubs/regex/regex/_regex_core.pyi +59 -19
- package/dist/typeshed-fallback/stubs/reportlab/METADATA.toml +1 -1
- package/dist/typeshed-fallback/stubs/reportlab/reportlab/lib/testutils.pyi +27 -25
- package/dist/typeshed-fallback/stubs/reportlab/reportlab/platypus/flowables.pyi +1 -1
- package/dist/typeshed-fallback/stubs/requests/requests/exceptions.pyi +2 -1
- package/dist/typeshed-fallback/stubs/seaborn/seaborn/_core/data.pyi +2 -3
- package/dist/typeshed-fallback/stubs/seaborn/seaborn/_core/typing.pyi +0 -2
- package/dist/typeshed-fallback/stubs/seaborn/seaborn/categorical.pyi +1 -1
- package/dist/typeshed-fallback/stubs/setuptools/setuptools/__init__.pyi +102 -45
- package/dist/typeshed-fallback/stubs/setuptools/setuptools/_distutils/cmd.pyi +6 -3
- package/dist/typeshed-fallback/stubs/setuptools/setuptools/_distutils/compilers/C/base.pyi +8 -2
- package/dist/typeshed-fallback/stubs/setuptools/setuptools/_distutils/compilers/C/msvc.pyi +3 -0
- package/dist/typeshed-fallback/stubs/setuptools/setuptools/_distutils/spawn.pyi +14 -4
- package/dist/typeshed-fallback/stubs/simple-websocket/METADATA.toml +3 -0
- package/dist/typeshed-fallback/stubs/simple-websocket/simple_websocket/__init__.pyi +3 -0
- package/dist/typeshed-fallback/stubs/simple-websocket/simple_websocket/aiows.pyi +130 -0
- package/dist/typeshed-fallback/stubs/simple-websocket/simple_websocket/asgi.pyi +44 -0
- package/dist/typeshed-fallback/stubs/simple-websocket/simple_websocket/errors.pyi +12 -0
- package/dist/typeshed-fallback/stubs/simple-websocket/simple_websocket/ws.pyi +136 -0
- package/dist/typeshed-fallback/stubs/tensorflow/METADATA.toml +16 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/__init__.pyi +437 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/_aliases.pyi +70 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/audio.pyi +7 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/autodiff.pyi +63 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/autograph/__init__.pyi +17 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/autograph/experimental.pyi +30 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/bitwise.pyi +37 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/compiler/xla/service/hlo_pb2.pyi +2113 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/compiler/xla/service/hlo_profile_printer_data_pb2.pyi +187 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/compiler/xla/service/metrics_pb2.pyi +283 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/compiler/xla/service/test_compilation_environment_pb2.pyi +59 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/compiler/xla/service/xla_compile_result_pb2.pyi +167 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/compiler/xla/tsl/protobuf/bfc_memory_map_pb2.pyi +218 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/compiler/xla/tsl/protobuf/test_log_pb2.pyi +707 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/compiler/xla/xla_data_pb2.pyi +2681 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/compiler/xla/xla_pb2.pyi +2558 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/config/__init__.pyi +12 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/config/experimental.pyi +16 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/example/example_parser_configuration_pb2.pyi +153 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/example/example_pb2.pyi +330 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/example/feature_pb2.pyi +222 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/allocation_description_pb2.pyi +64 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/api_def_pb2.pyi +312 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/attr_value_pb2.pyi +274 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/cost_graph_pb2.pyi +229 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/cpp_shape_inference_pb2.pyi +110 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/dataset_metadata_pb2.pyi +25 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/dataset_options_pb2.pyi +720 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/dataset_pb2.pyi +116 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/device_attributes_pb2.pyi +140 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/full_type_pb2.pyi +617 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/function_pb2.pyi +309 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/graph_debug_info_pb2.pyi +210 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/graph_pb2.pyi +100 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/graph_transfer_info_pb2.pyi +290 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/kernel_def_pb2.pyi +116 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/log_memory_pb2.pyi +218 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/model_pb2.pyi +368 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/node_def_pb2.pyi +192 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/op_def_pb2.pyi +391 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/optimized_function_graph_pb2.pyi +166 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/reader_base_pb2.pyi +52 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/resource_handle_pb2.pyi +104 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/step_stats_pb2.pyi +332 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/summary_pb2.pyi +368 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/tensor_description_pb2.pyi +49 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/tensor_pb2.pyi +235 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/tensor_shape_pb2.pyi +74 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/tensor_slice_pb2.pyi +56 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/types_pb2.pyi +220 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/variable_pb2.pyi +229 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/framework/versions_pb2.pyi +57 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/__init__.pyi +0 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/bfc_memory_map_pb2.pyi +15 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/cluster_pb2.pyi +126 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/composite_tensor_variant_pb2.pyi +33 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/config_pb2.pyi +1867 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/control_flow_pb2.pyi +243 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/core_platform_payloads_pb2.pyi +66 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/data_service_pb2.pyi +243 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/debug_event_pb2.pyi +746 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/debug_pb2.pyi +199 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/device_filters_pb2.pyi +124 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/device_properties_pb2.pyi +154 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/error_codes_pb2.pyi +33 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/fingerprint_pb2.pyi +76 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/meta_graph_pb2.pyi +735 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/named_tensor_pb2.pyi +41 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/queue_runner_pb2.pyi +73 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/remote_tensor_handle_pb2.pyi +96 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/rewriter_config_pb2.pyi +588 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/rpc_options_pb2.pyi +9 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/saved_model_pb2.pyi +51 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/saved_object_graph_pb2.pyi +715 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/saver_pb2.pyi +113 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/service_config_pb2.pyi +275 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/snapshot_pb2.pyi +164 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/status_pb2.pyi +13 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/struct_pb2.pyi +536 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/tensor_bundle_pb2.pyi +160 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/tensorflow_server_pb2.pyi +125 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/tpu/compilation_result_pb2.pyi +85 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/tpu/dynamic_padding_pb2.pyi +47 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/tpu/optimization_parameters_pb2.pyi +1326 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/tpu/topology_pb2.pyi +149 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/tpu/tpu_embedding_configuration_pb2.pyi +326 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/trackable_object_graph_pb2.pyi +213 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/transport_options_pb2.pyi +28 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/protobuf/verifier_config_pb2.pyi +65 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/util/event_pb2.pyi +419 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/util/memmapped_file_system_pb2.pyi +65 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/util/saved_tensor_slice_pb2.pyi +171 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/core/util/test_log_pb2.pyi +24 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/data/__init__.pyi +272 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/data/experimental.pyi +32 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/distribute/__init__.pyi +3 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/distribute/coordinator.pyi +3 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/distribute/experimental/coordinator.pyi +11 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/dtypes.pyi +57 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/experimental/__init__.pyi +10 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/experimental/dtensor.pyi +19 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/feature_column/__init__.pyi +95 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/initializers.pyi +1 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/io/__init__.pyi +104 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/io/gfile.pyi +11 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/keras/__init__.pyi +15 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/keras/activations.pyi +35 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/keras/callbacks.pyi +170 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/keras/constraints.pyi +16 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/keras/initializers.pyi +50 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/keras/layers/__init__.pyi +446 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/keras/losses.pyi +177 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/keras/metrics.pyi +117 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/keras/models.pyi +167 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/keras/optimizers/__init__.pyi +7 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/keras/optimizers/legacy/__init__.pyi +61 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/keras/optimizers/schedules.pyi +103 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/keras/regularizers.pyi +21 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/linalg.pyi +55 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/math.pyi +298 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/nn.pyi +194 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/__init__.pyi +1 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/distribute/distribute_lib.pyi +5 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/feature_column/__init__.pyi +0 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/feature_column/feature_column_v2.pyi +273 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/feature_column/sequence_feature_column.pyi +30 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/framework/dtypes.pyi +7 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/keras/__init__.pyi +1 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/keras/protobuf/projector_config_pb2.pyi +129 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/keras/protobuf/saved_metadata_pb2.pyi +89 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/keras/protobuf/versions_pb2.pyi +63 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/trackable/__init__.pyi +0 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/trackable/autotrackable.pyi +3 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/trackable/base.pyi +5 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/trackable/resource.pyi +9 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/trackable/ressource.pyi +7 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/python/training/tracking/autotrackable.pyi +3 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/random.pyi +231 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/raw_ops.pyi +44 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/saved_model/__init__.pyi +125 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/saved_model/experimental.pyi +39 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/signal.pyi +6 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/sparse.pyi +31 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/strings.pyi +241 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/summary.pyi +58 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/train/__init__.pyi +74 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/train/experimental.pyi +12 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/tsl/protobuf/coordination_config_pb2.pyi +156 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/tsl/protobuf/coordination_service_pb2.pyi +666 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/tsl/protobuf/distributed_runtime_payloads_pb2.pyi +69 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/tsl/protobuf/dnn_pb2.pyi +619 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/tsl/protobuf/error_codes_pb2.pyi +291 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/tsl/protobuf/histogram_pb2.pyi +78 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/tsl/protobuf/rpc_options_pb2.pyi +85 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/tsl/protobuf/status_pb2.pyi +34 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/types/__init__.pyi +1 -0
- package/dist/typeshed-fallback/stubs/tensorflow/tensorflow/types/experimental.pyi +30 -0
- package/dist/typeshed-fallback/stubs/ttkthemes/METADATA.toml +1 -1
- package/dist/typeshed-fallback/stubs/ttkthemes/ttkthemes/themed_style.pyi +1 -1
- package/dist/typeshed-fallback/stubs/ttkthemes/ttkthemes/themed_tk.pyi +11 -12
- package/dist/typeshed-fallback/stubs/tzdata/METADATA.toml +2 -0
- package/dist/typeshed-fallback/stubs/tzdata/tzdata/__init__.pyi +4 -0
- package/dist/typeshed-fallback/stubs/webencodings/METADATA.toml +2 -0
- package/dist/typeshed-fallback/stubs/webencodings/webencodings/__init__.pyi +33 -0
- package/dist/typeshed-fallback/stubs/webencodings/webencodings/labels.pyi +3 -0
- package/dist/typeshed-fallback/stubs/webencodings/webencodings/mklabels.pyi +5 -0
- package/dist/typeshed-fallback/stubs/webencodings/webencodings/x_user_defined.pyi +21 -0
- package/dist/typeshed-fallback/stubs/yt-dlp/METADATA.toml +1 -1
- package/dist/typeshed-fallback/stubs/yt-dlp/yt_dlp/__init__.pyi +2 -0
- package/dist/typeshed-fallback/stubs/yt-dlp/yt_dlp/extractor/common.pyi +1 -0
- package/dist/typeshed-fallback/stubs/yt-dlp/yt_dlp/globals.pyi +12 -1
- package/dist/typeshed-fallback/stubs/yt-dlp/yt_dlp/utils/_jsruntime.pyi +38 -0
- package/dist/typeshed-fallback/stubs/yt-dlp/yt_dlp/utils/_legacy.pyi +2 -68
- package/dist/typeshed-fallback/stubs/yt-dlp/yt_dlp/utils/_utils.pyi +1 -1
- package/package.json +7 -7
- package/dist/typeshed-fallback/stdlib/_typeshed/README.md +0 -34
- package/dist/typeshed-fallback/stubs/ExifRead/METADATA.toml +0 -3
- package/dist/typeshed-fallback/stubs/ExifRead/exifread/__init__.pyi +0 -17
- package/dist/typeshed-fallback/stubs/ExifRead/exifread/_types.pyi +0 -15
- package/dist/typeshed-fallback/stubs/ExifRead/exifread/classes.pyi +0 -47
- package/dist/typeshed-fallback/stubs/ExifRead/exifread/exceptions.pyi +0 -2
- package/dist/typeshed-fallback/stubs/ExifRead/exifread/exif_log.pyi +0 -24
- package/dist/typeshed-fallback/stubs/ExifRead/exifread/heic.pyi +0 -56
- package/dist/typeshed-fallback/stubs/ExifRead/exifread/jpeg.pyi +0 -7
- package/dist/typeshed-fallback/stubs/ExifRead/exifread/tags/__init__.pyi +0 -15
- package/dist/typeshed-fallback/stubs/ExifRead/exifread/tags/exif.pyi +0 -7
- package/dist/typeshed-fallback/stubs/ExifRead/exifread/tags/makernote/apple.pyi +0 -3
- package/dist/typeshed-fallback/stubs/ExifRead/exifread/tags/makernote/canon.pyi +0 -26
- package/dist/typeshed-fallback/stubs/ExifRead/exifread/tags/makernote/casio.pyi +0 -3
- package/dist/typeshed-fallback/stubs/ExifRead/exifread/tags/makernote/fujifilm.pyi +0 -3
- package/dist/typeshed-fallback/stubs/ExifRead/exifread/tags/makernote/nikon.pyi +0 -6
- package/dist/typeshed-fallback/stubs/ExifRead/exifread/tags/makernote/olympus.pyi +0 -6
- package/dist/typeshed-fallback/stubs/ExifRead/exifread/utils.pyi +0 -22
- package/dist/typeshed-fallback/stubs/networkx/networkx/algorithms/community/kernighan_lin.pyi +0 -16
- package/dist/typeshed-fallback/stubs/psutil/psutil/_psutil_posix.pyi +0 -34
- package/dist/typeshed-fallback/stubs/vobject/@tests/stubtest_allowlist.txt +0 -22
- /package/dist/typeshed-fallback/stubs/{ExifRead/exifread/tags/makernote → colorful/colorful}/__init__.pyi +0 -0
|
@@ -0,0 +1,2558 @@
|
|
|
1
|
+
"""
|
|
2
|
+
@generated by mypy-protobuf. Do not edit manually!
|
|
3
|
+
isort:skip_file
|
|
4
|
+
Copyright 2017 The OpenXLA Authors.
|
|
5
|
+
|
|
6
|
+
Licensed under the Apache License, Version 2.0 (the "License");
|
|
7
|
+
you may not use this file except in compliance with the License.
|
|
8
|
+
You may obtain a copy of the License at
|
|
9
|
+
|
|
10
|
+
http://www.apache.org/licenses/LICENSE-2.0
|
|
11
|
+
|
|
12
|
+
Unless required by applicable law or agreed to in writing, software
|
|
13
|
+
distributed under the License is distributed on an "AS IS" BASIS,
|
|
14
|
+
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
15
|
+
See the License for the specific language governing permissions and
|
|
16
|
+
limitations under the License.
|
|
17
|
+
==============================================================================
|
|
18
|
+
"""
|
|
19
|
+
|
|
20
|
+
import builtins
|
|
21
|
+
import collections.abc
|
|
22
|
+
import sys
|
|
23
|
+
import typing
|
|
24
|
+
|
|
25
|
+
import google.protobuf.any_pb2
|
|
26
|
+
import google.protobuf.descriptor
|
|
27
|
+
import google.protobuf.internal.containers
|
|
28
|
+
import google.protobuf.internal.enum_type_wrapper
|
|
29
|
+
import google.protobuf.message
|
|
30
|
+
import tensorflow.compiler.xla.service.hlo_pb2
|
|
31
|
+
import tensorflow.compiler.xla.xla_data_pb2
|
|
32
|
+
|
|
33
|
+
if sys.version_info >= (3, 10):
|
|
34
|
+
import typing as typing_extensions
|
|
35
|
+
else:
|
|
36
|
+
import typing_extensions
|
|
37
|
+
|
|
38
|
+
DESCRIPTOR: google.protobuf.descriptor.FileDescriptor
|
|
39
|
+
|
|
40
|
+
@typing.final
|
|
41
|
+
class CompilationEnvironmentsProto(google.protobuf.message.Message):
|
|
42
|
+
"""Proto version of `xla::CompilationEnvironments`."""
|
|
43
|
+
|
|
44
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
45
|
+
|
|
46
|
+
ENVIRONMENTS_FIELD_NUMBER: builtins.int
|
|
47
|
+
@property
|
|
48
|
+
def environments(
|
|
49
|
+
self,
|
|
50
|
+
) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[google.protobuf.any_pb2.Any]: ...
|
|
51
|
+
def __init__(self, *, environments: collections.abc.Iterable[google.protobuf.any_pb2.Any] | None = ...) -> None: ...
|
|
52
|
+
def ClearField(self, field_name: typing.Literal["environments", b"environments"]) -> None: ...
|
|
53
|
+
|
|
54
|
+
global___CompilationEnvironmentsProto = CompilationEnvironmentsProto
|
|
55
|
+
|
|
56
|
+
@typing.final
|
|
57
|
+
class DebugOptions(google.protobuf.message.Message):
|
|
58
|
+
"""Debugging options for XLA. These options may change at any time - there are
|
|
59
|
+
no guarantees about backward or forward compatibility for these fields.
|
|
60
|
+
|
|
61
|
+
Debug options naming and organization:
|
|
62
|
+
|
|
63
|
+
1. Backend-agnostic options: `xla_$flag_name` - go first, and sorted
|
|
64
|
+
alphabetically by the flag name.
|
|
65
|
+
|
|
66
|
+
2. Backend-specific options: `xla_$backend_$flag_name` - must be in the
|
|
67
|
+
corresponding backend section, and sorted alphabetically by the flag name.
|
|
68
|
+
--------------------------------------------------------------------------//
|
|
69
|
+
XLA backend-agnostic options.
|
|
70
|
+
--------------------------------------------------------------------------//
|
|
71
|
+
go/keep-sorted start
|
|
72
|
+
"""
|
|
73
|
+
|
|
74
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
75
|
+
|
|
76
|
+
class _ShapeChecks:
|
|
77
|
+
ValueType = typing.NewType("ValueType", builtins.int)
|
|
78
|
+
V: typing_extensions.TypeAlias = ValueType
|
|
79
|
+
|
|
80
|
+
class _ShapeChecksEnumTypeWrapper(
|
|
81
|
+
google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[DebugOptions._ShapeChecks.ValueType], builtins.type
|
|
82
|
+
):
|
|
83
|
+
DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
|
|
84
|
+
IGNORE: DebugOptions._ShapeChecks.ValueType # 0
|
|
85
|
+
"""Do not insert any shape checks for dynamically shaped operations; output
|
|
86
|
+
buffers might contain garbage data if shapes don't match.
|
|
87
|
+
"""
|
|
88
|
+
RUNTIME: DebugOptions._ShapeChecks.ValueType # 1
|
|
89
|
+
"""Check shapes at runtime, will insert an extra synchronization if shapes
|
|
90
|
+
cannot be proven correct at compile time.
|
|
91
|
+
"""
|
|
92
|
+
COMPILE_TIME: DebugOptions._ShapeChecks.ValueType # 2
|
|
93
|
+
"""Will refuse to compile any program where shape correctness can not be
|
|
94
|
+
established at compile time.
|
|
95
|
+
"""
|
|
96
|
+
|
|
97
|
+
class ShapeChecks(_ShapeChecks, metaclass=_ShapeChecksEnumTypeWrapper): ...
|
|
98
|
+
IGNORE: DebugOptions.ShapeChecks.ValueType # 0
|
|
99
|
+
"""Do not insert any shape checks for dynamically shaped operations; output
|
|
100
|
+
buffers might contain garbage data if shapes don't match.
|
|
101
|
+
"""
|
|
102
|
+
RUNTIME: DebugOptions.ShapeChecks.ValueType # 1
|
|
103
|
+
"""Check shapes at runtime, will insert an extra synchronization if shapes
|
|
104
|
+
cannot be proven correct at compile time.
|
|
105
|
+
"""
|
|
106
|
+
COMPILE_TIME: DebugOptions.ShapeChecks.ValueType # 2
|
|
107
|
+
"""Will refuse to compile any program where shape correctness can not be
|
|
108
|
+
established at compile time.
|
|
109
|
+
"""
|
|
110
|
+
|
|
111
|
+
class _StepMarkerLocation:
|
|
112
|
+
ValueType = typing.NewType("ValueType", builtins.int)
|
|
113
|
+
V: typing_extensions.TypeAlias = ValueType
|
|
114
|
+
|
|
115
|
+
class _StepMarkerLocationEnumTypeWrapper(
|
|
116
|
+
google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[DebugOptions._StepMarkerLocation.ValueType], builtins.type
|
|
117
|
+
):
|
|
118
|
+
DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
|
|
119
|
+
STEP_MARK_AT_ENTRY: DebugOptions._StepMarkerLocation.ValueType # 0
|
|
120
|
+
"""Generate a step marker at the program entry. This handles the case where
|
|
121
|
+
each step is done by one or multiple program execution(s). Only the first
|
|
122
|
+
program will be tagged for generating a step marker at the program entry.
|
|
123
|
+
This is the default.
|
|
124
|
+
"""
|
|
125
|
+
STEP_MARK_AT_TOP_LEVEL_WHILE_LOOP: DebugOptions._StepMarkerLocation.ValueType # 1
|
|
126
|
+
"""Generate a step marker at each iteration of the top level while loop,
|
|
127
|
+
which is assumed to be a training loop.
|
|
128
|
+
"""
|
|
129
|
+
STEP_MARK_AT_SECOND_LEVEL_WHILE_LOOP: DebugOptions._StepMarkerLocation.ValueType # 3
|
|
130
|
+
"""Generate a step marker at each iteration of the second level while loops,
|
|
131
|
+
which is assumed to be a training or eval loop.
|
|
132
|
+
"""
|
|
133
|
+
STEP_MARK_NONE: DebugOptions._StepMarkerLocation.ValueType # 2
|
|
134
|
+
"""No step marker generated."""
|
|
135
|
+
|
|
136
|
+
class StepMarkerLocation(_StepMarkerLocation, metaclass=_StepMarkerLocationEnumTypeWrapper): ...
|
|
137
|
+
STEP_MARK_AT_ENTRY: DebugOptions.StepMarkerLocation.ValueType # 0
|
|
138
|
+
"""Generate a step marker at the program entry. This handles the case where
|
|
139
|
+
each step is done by one or multiple program execution(s). Only the first
|
|
140
|
+
program will be tagged for generating a step marker at the program entry.
|
|
141
|
+
This is the default.
|
|
142
|
+
"""
|
|
143
|
+
STEP_MARK_AT_TOP_LEVEL_WHILE_LOOP: DebugOptions.StepMarkerLocation.ValueType # 1
|
|
144
|
+
"""Generate a step marker at each iteration of the top level while loop,
|
|
145
|
+
which is assumed to be a training loop.
|
|
146
|
+
"""
|
|
147
|
+
STEP_MARK_AT_SECOND_LEVEL_WHILE_LOOP: DebugOptions.StepMarkerLocation.ValueType # 3
|
|
148
|
+
"""Generate a step marker at each iteration of the second level while loops,
|
|
149
|
+
which is assumed to be a training or eval loop.
|
|
150
|
+
"""
|
|
151
|
+
STEP_MARK_NONE: DebugOptions.StepMarkerLocation.ValueType # 2
|
|
152
|
+
"""No step marker generated."""
|
|
153
|
+
|
|
154
|
+
class _CollectiveOpType:
|
|
155
|
+
ValueType = typing.NewType("ValueType", builtins.int)
|
|
156
|
+
V: typing_extensions.TypeAlias = ValueType
|
|
157
|
+
|
|
158
|
+
class _CollectiveOpTypeEnumTypeWrapper(
|
|
159
|
+
google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[DebugOptions._CollectiveOpType.ValueType], builtins.type
|
|
160
|
+
):
|
|
161
|
+
DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
|
|
162
|
+
NOOP: DebugOptions._CollectiveOpType.ValueType # 0
|
|
163
|
+
ALLREDUCE: DebugOptions._CollectiveOpType.ValueType # 1
|
|
164
|
+
ALLGATHER: DebugOptions._CollectiveOpType.ValueType # 2
|
|
165
|
+
REDUCESCATTER: DebugOptions._CollectiveOpType.ValueType # 3
|
|
166
|
+
COLLECTIVEBROADCAST: DebugOptions._CollectiveOpType.ValueType # 4
|
|
167
|
+
ALLTOALL: DebugOptions._CollectiveOpType.ValueType # 5
|
|
168
|
+
COLLECTIVEPERMUTE: DebugOptions._CollectiveOpType.ValueType # 6
|
|
169
|
+
|
|
170
|
+
class CollectiveOpType(_CollectiveOpType, metaclass=_CollectiveOpTypeEnumTypeWrapper):
|
|
171
|
+
"""Enum to define all collective ops
|
|
172
|
+
that xla supports.
|
|
173
|
+
"""
|
|
174
|
+
|
|
175
|
+
NOOP: DebugOptions.CollectiveOpType.ValueType # 0
|
|
176
|
+
ALLREDUCE: DebugOptions.CollectiveOpType.ValueType # 1
|
|
177
|
+
ALLGATHER: DebugOptions.CollectiveOpType.ValueType # 2
|
|
178
|
+
REDUCESCATTER: DebugOptions.CollectiveOpType.ValueType # 3
|
|
179
|
+
COLLECTIVEBROADCAST: DebugOptions.CollectiveOpType.ValueType # 4
|
|
180
|
+
ALLTOALL: DebugOptions.CollectiveOpType.ValueType # 5
|
|
181
|
+
COLLECTIVEPERMUTE: DebugOptions.CollectiveOpType.ValueType # 6
|
|
182
|
+
|
|
183
|
+
class _CommandBufferCmdType:
|
|
184
|
+
ValueType = typing.NewType("ValueType", builtins.int)
|
|
185
|
+
V: typing_extensions.TypeAlias = ValueType
|
|
186
|
+
|
|
187
|
+
class _CommandBufferCmdTypeEnumTypeWrapper(
|
|
188
|
+
google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[DebugOptions._CommandBufferCmdType.ValueType], builtins.type
|
|
189
|
+
):
|
|
190
|
+
DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
|
|
191
|
+
INVALID: DebugOptions._CommandBufferCmdType.ValueType # 0
|
|
192
|
+
FUSION: DebugOptions._CommandBufferCmdType.ValueType # 1
|
|
193
|
+
CUBLAS: DebugOptions._CommandBufferCmdType.ValueType # 2
|
|
194
|
+
CUDNN: DebugOptions._CommandBufferCmdType.ValueType # 3
|
|
195
|
+
COLLECTIVES: DebugOptions._CommandBufferCmdType.ValueType # 4
|
|
196
|
+
CONDITIONALS: DebugOptions._CommandBufferCmdType.ValueType # 5
|
|
197
|
+
CUSTOM_CALL: DebugOptions._CommandBufferCmdType.ValueType # 6
|
|
198
|
+
CUBLASLT: DebugOptions._CommandBufferCmdType.ValueType # 7
|
|
199
|
+
|
|
200
|
+
class CommandBufferCmdType(_CommandBufferCmdType, metaclass=_CommandBufferCmdTypeEnumTypeWrapper):
|
|
201
|
+
"""Commands are categorized into 5 types:
|
|
202
|
+
FUSION represents regular fusion kernels.
|
|
203
|
+
CUBLAS/CUBLASLT, CUDNN, and COLLECTIVES represent library calls.
|
|
204
|
+
CONDITIONALS represents control flow.
|
|
205
|
+
"""
|
|
206
|
+
|
|
207
|
+
INVALID: DebugOptions.CommandBufferCmdType.ValueType # 0
|
|
208
|
+
FUSION: DebugOptions.CommandBufferCmdType.ValueType # 1
|
|
209
|
+
CUBLAS: DebugOptions.CommandBufferCmdType.ValueType # 2
|
|
210
|
+
CUDNN: DebugOptions.CommandBufferCmdType.ValueType # 3
|
|
211
|
+
COLLECTIVES: DebugOptions.CommandBufferCmdType.ValueType # 4
|
|
212
|
+
CONDITIONALS: DebugOptions.CommandBufferCmdType.ValueType # 5
|
|
213
|
+
CUSTOM_CALL: DebugOptions.CommandBufferCmdType.ValueType # 6
|
|
214
|
+
CUBLASLT: DebugOptions.CommandBufferCmdType.ValueType # 7
|
|
215
|
+
|
|
216
|
+
class _PartitioningAlgorithm:
|
|
217
|
+
ValueType = typing.NewType("ValueType", builtins.int)
|
|
218
|
+
V: typing_extensions.TypeAlias = ValueType
|
|
219
|
+
|
|
220
|
+
class _PartitioningAlgorithmEnumTypeWrapper(
|
|
221
|
+
google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[DebugOptions._PartitioningAlgorithm.ValueType], builtins.type
|
|
222
|
+
):
|
|
223
|
+
DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
|
|
224
|
+
PARTITIONING_ALGORITHM_NOOP: DebugOptions._PartitioningAlgorithm.ValueType # 0
|
|
225
|
+
PARTITIONING_ALGORITHM_EXP0: DebugOptions._PartitioningAlgorithm.ValueType # 1
|
|
226
|
+
PARTITIONING_ALGORITHM_EXP1: DebugOptions._PartitioningAlgorithm.ValueType # 2
|
|
227
|
+
PARTITIONING_ALGORITHM_EXP2: DebugOptions._PartitioningAlgorithm.ValueType # 3
|
|
228
|
+
|
|
229
|
+
class PartitioningAlgorithm(_PartitioningAlgorithm, metaclass=_PartitioningAlgorithmEnumTypeWrapper): ...
|
|
230
|
+
PARTITIONING_ALGORITHM_NOOP: DebugOptions.PartitioningAlgorithm.ValueType # 0
|
|
231
|
+
PARTITIONING_ALGORITHM_EXP0: DebugOptions.PartitioningAlgorithm.ValueType # 1
|
|
232
|
+
PARTITIONING_ALGORITHM_EXP1: DebugOptions.PartitioningAlgorithm.ValueType # 2
|
|
233
|
+
PARTITIONING_ALGORITHM_EXP2: DebugOptions.PartitioningAlgorithm.ValueType # 3
|
|
234
|
+
|
|
235
|
+
class _WhileLoopUnrolling:
|
|
236
|
+
ValueType = typing.NewType("ValueType", builtins.int)
|
|
237
|
+
V: typing_extensions.TypeAlias = ValueType
|
|
238
|
+
|
|
239
|
+
class _WhileLoopUnrollingEnumTypeWrapper(
|
|
240
|
+
google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[DebugOptions._WhileLoopUnrolling.ValueType], builtins.type
|
|
241
|
+
):
|
|
242
|
+
DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
|
|
243
|
+
WHILE_LOOP_UNROLLING_NO_UNROLL: DebugOptions._WhileLoopUnrolling.ValueType # 0
|
|
244
|
+
WHILE_LOOP_UNROLLING_DOUBLE_BUFFER: DebugOptions._WhileLoopUnrolling.ValueType # 1
|
|
245
|
+
"""Has the same effect as setting
|
|
246
|
+
`xla_gpu_enable_while_loop_double_buffering`.
|
|
247
|
+
"""
|
|
248
|
+
WHILE_LOOP_UNROLLING_FULL_UNROLL: DebugOptions._WhileLoopUnrolling.ValueType # 2
|
|
249
|
+
"""Enables full loop unrolling using the same strategy as `DOUBLE_BUFFER`."""
|
|
250
|
+
|
|
251
|
+
class WhileLoopUnrolling(_WhileLoopUnrolling, metaclass=_WhileLoopUnrollingEnumTypeWrapper): ...
|
|
252
|
+
WHILE_LOOP_UNROLLING_NO_UNROLL: DebugOptions.WhileLoopUnrolling.ValueType # 0
|
|
253
|
+
WHILE_LOOP_UNROLLING_DOUBLE_BUFFER: DebugOptions.WhileLoopUnrolling.ValueType # 1
|
|
254
|
+
"""Has the same effect as setting
|
|
255
|
+
`xla_gpu_enable_while_loop_double_buffering`.
|
|
256
|
+
"""
|
|
257
|
+
WHILE_LOOP_UNROLLING_FULL_UNROLL: DebugOptions.WhileLoopUnrolling.ValueType # 2
|
|
258
|
+
"""Enables full loop unrolling using the same strategy as `DOUBLE_BUFFER`."""
|
|
259
|
+
|
|
260
|
+
class _AutotuneCacheMode:
|
|
261
|
+
ValueType = typing.NewType("ValueType", builtins.int)
|
|
262
|
+
V: typing_extensions.TypeAlias = ValueType
|
|
263
|
+
|
|
264
|
+
class _AutotuneCacheModeEnumTypeWrapper(
|
|
265
|
+
google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[DebugOptions._AutotuneCacheMode.ValueType], builtins.type
|
|
266
|
+
):
|
|
267
|
+
DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
|
|
268
|
+
AUTOTUNE_CACHE_MODE_UNSPECIFIED: DebugOptions._AutotuneCacheMode.ValueType # 0
|
|
269
|
+
AUTOTUNE_CACHE_MODE_UPDATE: DebugOptions._AutotuneCacheMode.ValueType # 1
|
|
270
|
+
"""If the cache exists per fusion autotuner loads it and terminates,
|
|
271
|
+
otherwise runs autotuner and dumps the result.
|
|
272
|
+
"""
|
|
273
|
+
AUTOTUNE_CACHE_MODE_READ: DebugOptions._AutotuneCacheMode.ValueType # 2
|
|
274
|
+
"""Sets readonly access to the cache for the per fusion autotuner. Same as
|
|
275
|
+
above, but doesn't dump anything.
|
|
276
|
+
"""
|
|
277
|
+
|
|
278
|
+
class AutotuneCacheMode(_AutotuneCacheMode, metaclass=_AutotuneCacheModeEnumTypeWrapper): ...
|
|
279
|
+
AUTOTUNE_CACHE_MODE_UNSPECIFIED: DebugOptions.AutotuneCacheMode.ValueType # 0
|
|
280
|
+
AUTOTUNE_CACHE_MODE_UPDATE: DebugOptions.AutotuneCacheMode.ValueType # 1
|
|
281
|
+
"""If the cache exists per fusion autotuner loads it and terminates,
|
|
282
|
+
otherwise runs autotuner and dumps the result.
|
|
283
|
+
"""
|
|
284
|
+
AUTOTUNE_CACHE_MODE_READ: DebugOptions.AutotuneCacheMode.ValueType # 2
|
|
285
|
+
"""Sets readonly access to the cache for the per fusion autotuner. Same as
|
|
286
|
+
above, but doesn't dump anything.
|
|
287
|
+
"""
|
|
288
|
+
|
|
289
|
+
@typing.final
|
|
290
|
+
class XlaBackendExtraOptionsEntry(google.protobuf.message.Message):
|
|
291
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
292
|
+
|
|
293
|
+
KEY_FIELD_NUMBER: builtins.int
|
|
294
|
+
VALUE_FIELD_NUMBER: builtins.int
|
|
295
|
+
key: builtins.str
|
|
296
|
+
value: builtins.str
|
|
297
|
+
def __init__(self, *, key: builtins.str | None = ..., value: builtins.str | None = ...) -> None: ...
|
|
298
|
+
def ClearField(self, field_name: typing.Literal["key", b"key", "value", b"value"]) -> None: ...
|
|
299
|
+
|
|
300
|
+
XLA_CPU_ENABLE_CONCURRENCY_OPTIMIZED_SCHEDULER_FIELD_NUMBER: builtins.int
|
|
301
|
+
XLA_CPU_ENABLE_FAST_MATH_FIELD_NUMBER: builtins.int
|
|
302
|
+
XLA_CPU_ENABLE_FAST_MIN_MAX_FIELD_NUMBER: builtins.int
|
|
303
|
+
XLA_CPU_FAST_MATH_HONOR_DIVISION_FIELD_NUMBER: builtins.int
|
|
304
|
+
XLA_CPU_FAST_MATH_HONOR_FUNCTIONS_FIELD_NUMBER: builtins.int
|
|
305
|
+
XLA_CPU_FAST_MATH_HONOR_INFS_FIELD_NUMBER: builtins.int
|
|
306
|
+
XLA_CPU_FAST_MATH_HONOR_NANS_FIELD_NUMBER: builtins.int
|
|
307
|
+
XLA_CPU_USE_THUNK_RUNTIME_FIELD_NUMBER: builtins.int
|
|
308
|
+
XLA_CPU_PARALLEL_CODEGEN_SPLIT_COUNT_FIELD_NUMBER: builtins.int
|
|
309
|
+
XLA_CPU_PREFER_VECTOR_WIDTH_FIELD_NUMBER: builtins.int
|
|
310
|
+
XLA_GPU_EXPERIMENTAL_AUTOTUNE_CACHE_MODE_FIELD_NUMBER: builtins.int
|
|
311
|
+
XLA_GPU_EXPERIMENTAL_DISABLE_BINARY_LIBRARIES_FIELD_NUMBER: builtins.int
|
|
312
|
+
XLA_GPU_EXPERIMENTAL_ENABLE_TRITON_SOFTMAX_PRIORITY_FUSION_FIELD_NUMBER: builtins.int
|
|
313
|
+
XLA_GPU_UNSUPPORTED_ENABLE_TRITON_GEMM_FIELD_NUMBER: builtins.int
|
|
314
|
+
XLA_HLO_GRAPH_ADDRESSES_FIELD_NUMBER: builtins.int
|
|
315
|
+
XLA_HLO_PROFILE_FIELD_NUMBER: builtins.int
|
|
316
|
+
XLA_DISABLE_HLO_PASSES_FIELD_NUMBER: builtins.int
|
|
317
|
+
XLA_ENABLE_HLO_PASSES_ONLY_FIELD_NUMBER: builtins.int
|
|
318
|
+
XLA_DISABLE_ALL_HLO_PASSES_FIELD_NUMBER: builtins.int
|
|
319
|
+
XLA_BACKEND_OPTIMIZATION_LEVEL_FIELD_NUMBER: builtins.int
|
|
320
|
+
XLA_EMBED_IR_IN_EXECUTABLE_FIELD_NUMBER: builtins.int
|
|
321
|
+
XLA_ELIMINATE_HLO_IMPLICIT_BROADCAST_FIELD_NUMBER: builtins.int
|
|
322
|
+
XLA_CPU_MULTI_THREAD_EIGEN_FIELD_NUMBER: builtins.int
|
|
323
|
+
XLA_GPU_CUDA_DATA_DIR_FIELD_NUMBER: builtins.int
|
|
324
|
+
XLA_GPU_FTZ_FIELD_NUMBER: builtins.int
|
|
325
|
+
XLA_LLVM_ENABLE_ALIAS_SCOPE_METADATA_FIELD_NUMBER: builtins.int
|
|
326
|
+
XLA_LLVM_ENABLE_NOALIAS_METADATA_FIELD_NUMBER: builtins.int
|
|
327
|
+
XLA_LLVM_ENABLE_INVARIANT_LOAD_METADATA_FIELD_NUMBER: builtins.int
|
|
328
|
+
XLA_LLVM_DISABLE_EXPENSIVE_PASSES_FIELD_NUMBER: builtins.int
|
|
329
|
+
XLA_TEST_ALL_OUTPUT_LAYOUTS_FIELD_NUMBER: builtins.int
|
|
330
|
+
XLA_TEST_ALL_INPUT_LAYOUTS_FIELD_NUMBER: builtins.int
|
|
331
|
+
XLA_HLO_GRAPH_SHARDING_COLOR_FIELD_NUMBER: builtins.int
|
|
332
|
+
XLA_CPU_USE_MKL_DNN_FIELD_NUMBER: builtins.int
|
|
333
|
+
XLA_GPU_ENABLE_FAST_MIN_MAX_FIELD_NUMBER: builtins.int
|
|
334
|
+
XLA_ALLOW_EXCESS_PRECISION_FIELD_NUMBER: builtins.int
|
|
335
|
+
XLA_GPU_CRASH_ON_VERIFICATION_FAILURES_FIELD_NUMBER: builtins.int
|
|
336
|
+
XLA_GPU_AUTOTUNE_LEVEL_FIELD_NUMBER: builtins.int
|
|
337
|
+
XLA_FORCE_HOST_PLATFORM_DEVICE_COUNT_FIELD_NUMBER: builtins.int
|
|
338
|
+
XLA_GPU_DISABLE_GPUASM_OPTIMIZATIONS_FIELD_NUMBER: builtins.int
|
|
339
|
+
XLA_GPU_SHAPE_CHECKS_FIELD_NUMBER: builtins.int
|
|
340
|
+
XLA_HLO_EVALUATOR_USE_FAST_PATH_FIELD_NUMBER: builtins.int
|
|
341
|
+
XLA_ALLOW_SCALAR_INDEX_DYNAMIC_OPS_FIELD_NUMBER: builtins.int
|
|
342
|
+
XLA_STEP_MARKER_LOCATION_FIELD_NUMBER: builtins.int
|
|
343
|
+
XLA_DUMP_TO_FIELD_NUMBER: builtins.int
|
|
344
|
+
XLA_DUMP_HLO_MODULE_RE_FIELD_NUMBER: builtins.int
|
|
345
|
+
XLA_DUMP_HLO_PASS_RE_FIELD_NUMBER: builtins.int
|
|
346
|
+
XLA_DUMP_HLO_AS_TEXT_FIELD_NUMBER: builtins.int
|
|
347
|
+
XLA_DUMP_HLO_AS_PROTO_FIELD_NUMBER: builtins.int
|
|
348
|
+
XLA_DUMP_HLO_AS_DOT_FIELD_NUMBER: builtins.int
|
|
349
|
+
XLA_DUMP_HLO_AS_URL_FIELD_NUMBER: builtins.int
|
|
350
|
+
XLA_DUMP_HLO_AS_HTML_FIELD_NUMBER: builtins.int
|
|
351
|
+
XLA_DUMP_FUSION_VISUALIZATION_FIELD_NUMBER: builtins.int
|
|
352
|
+
XLA_DUMP_HLO_SNAPSHOTS_FIELD_NUMBER: builtins.int
|
|
353
|
+
XLA_DUMP_INCLUDE_TIMESTAMP_FIELD_NUMBER: builtins.int
|
|
354
|
+
XLA_DUMP_MAX_HLO_MODULES_FIELD_NUMBER: builtins.int
|
|
355
|
+
XLA_DUMP_MODULE_METADATA_FIELD_NUMBER: builtins.int
|
|
356
|
+
XLA_DUMP_COMPRESS_PROTOS_FIELD_NUMBER: builtins.int
|
|
357
|
+
XLA_DUMP_HLO_AS_LONG_TEXT_FIELD_NUMBER: builtins.int
|
|
358
|
+
XLA_GPU_FORCE_CONV_NCHW_FIELD_NUMBER: builtins.int
|
|
359
|
+
XLA_GPU_FORCE_CONV_NHWC_FIELD_NUMBER: builtins.int
|
|
360
|
+
XLA_GPU_PTX_FILE_FIELD_NUMBER: builtins.int
|
|
361
|
+
XLA_GPU_DUMP_LLVMIR_FIELD_NUMBER: builtins.int
|
|
362
|
+
XLA_DUMP_ENABLE_MLIR_PRETTY_FORM_FIELD_NUMBER: builtins.int
|
|
363
|
+
XLA_GPU_ALGORITHM_DENYLIST_PATH_FIELD_NUMBER: builtins.int
|
|
364
|
+
XLA_TPU_DETECT_NAN_FIELD_NUMBER: builtins.int
|
|
365
|
+
XLA_TPU_DETECT_INF_FIELD_NUMBER: builtins.int
|
|
366
|
+
XLA_CPU_ENABLE_XPROF_TRACEME_FIELD_NUMBER: builtins.int
|
|
367
|
+
XLA_GPU_UNSAFE_FALLBACK_TO_DRIVER_ON_PTXAS_NOT_FOUND_FIELD_NUMBER: builtins.int
|
|
368
|
+
XLA_GPU_ASM_EXTRA_FLAGS_FIELD_NUMBER: builtins.int
|
|
369
|
+
XLA_MULTIHEAP_SIZE_CONSTRAINT_PER_HEAP_FIELD_NUMBER: builtins.int
|
|
370
|
+
XLA_DETAILED_LOGGING_FIELD_NUMBER: builtins.int
|
|
371
|
+
XLA_ENABLE_DUMPING_FIELD_NUMBER: builtins.int
|
|
372
|
+
XLA_GPU_FORCE_COMPILATION_PARALLELISM_FIELD_NUMBER: builtins.int
|
|
373
|
+
XLA_GPU_ENABLE_LLVM_MODULE_COMPILATION_PARALLELISM_FIELD_NUMBER: builtins.int
|
|
374
|
+
XLA_GPU_DETERMINISTIC_OPS_FIELD_NUMBER: builtins.int
|
|
375
|
+
XLA_GPU_LLVM_IR_FILE_FIELD_NUMBER: builtins.int
|
|
376
|
+
XLA_GPU_DISABLE_ASYNC_COLLECTIVES_FIELD_NUMBER: builtins.int
|
|
377
|
+
XLA_GPU_ALL_REDUCE_COMBINE_THRESHOLD_BYTES_FIELD_NUMBER: builtins.int
|
|
378
|
+
XLA_GPU_ALL_GATHER_COMBINE_THRESHOLD_BYTES_FIELD_NUMBER: builtins.int
|
|
379
|
+
XLA_GPU_REDUCE_SCATTER_COMBINE_THRESHOLD_BYTES_FIELD_NUMBER: builtins.int
|
|
380
|
+
XLA_GPU_ENABLE_ALL_GATHER_COMBINE_BY_DIM_FIELD_NUMBER: builtins.int
|
|
381
|
+
XLA_GPU_ENABLE_REDUCE_SCATTER_COMBINE_BY_DIM_FIELD_NUMBER: builtins.int
|
|
382
|
+
XLA_GPU_ENABLE_REASSOCIATION_FOR_CONVERTED_AR_FIELD_NUMBER: builtins.int
|
|
383
|
+
XLA_GPU_ALL_REDUCE_BLUECONNECT_NUM_DEVICES_PER_HOST_FIELD_NUMBER: builtins.int
|
|
384
|
+
XLA_GPU_ENABLE_WHILE_LOOP_REDUCE_SCATTER_CODE_MOTION_FIELD_NUMBER: builtins.int
|
|
385
|
+
XLA_GPU_COLLECTIVE_INFLATION_FACTOR_FIELD_NUMBER: builtins.int
|
|
386
|
+
XLA_LLVM_FORCE_INLINE_BEFORE_SPLIT_FIELD_NUMBER: builtins.int
|
|
387
|
+
XLA_GPU_ENABLE_CUDNN_FRONTEND_FIELD_NUMBER: builtins.int
|
|
388
|
+
XLA_GPU_ENABLE_CUDNN_FMHA_FIELD_NUMBER: builtins.int
|
|
389
|
+
XLA_GPU_FUSED_ATTENTION_USE_CUDNN_RNG_FIELD_NUMBER: builtins.int
|
|
390
|
+
XLA_GPU_ENABLE_CUDNN_LAYER_NORM_FIELD_NUMBER: builtins.int
|
|
391
|
+
XLA_DUMP_DISABLE_METADATA_FIELD_NUMBER: builtins.int
|
|
392
|
+
XLA_DUMP_HLO_PIPELINE_RE_FIELD_NUMBER: builtins.int
|
|
393
|
+
XLA_GPU_STRICT_CONV_ALGORITHM_PICKER_FIELD_NUMBER: builtins.int
|
|
394
|
+
XLA_GPU_ENABLE_CUSTOM_FUSIONS_FIELD_NUMBER: builtins.int
|
|
395
|
+
XLA_GPU_ENABLE_CUSTOM_FUSIONS_RE_FIELD_NUMBER: builtins.int
|
|
396
|
+
XLA_GPU_ENABLE_DYNAMIC_SLICE_FUSION_FIELD_NUMBER: builtins.int
|
|
397
|
+
XLA_GPU_NCCL_TERMINATION_TIMEOUT_SECONDS_FIELD_NUMBER: builtins.int
|
|
398
|
+
XLA_GPU_ENABLE_SHARED_CONSTANTS_FIELD_NUMBER: builtins.int
|
|
399
|
+
XLA_GPU_ENABLE_CUBLASLT_FIELD_NUMBER: builtins.int
|
|
400
|
+
XLA_GPU_ENABLE_COMMAND_BUFFER_FIELD_NUMBER: builtins.int
|
|
401
|
+
XLA_GPU_GRAPH_MIN_GRAPH_SIZE_FIELD_NUMBER: builtins.int
|
|
402
|
+
XLA_GPU_GRAPH_ENABLE_CONCURRENT_REGION_FIELD_NUMBER: builtins.int
|
|
403
|
+
XLA_GPU_REDZONE_SCRATCH_MAX_MEGABYTES_FIELD_NUMBER: builtins.int
|
|
404
|
+
XLA_GPU_REDZONE_PADDING_BYTES_FIELD_NUMBER: builtins.int
|
|
405
|
+
XLA_CPU_USE_ACL_FIELD_NUMBER: builtins.int
|
|
406
|
+
XLA_CPU_STRICT_DOT_CONV_MATH_FIELD_NUMBER: builtins.int
|
|
407
|
+
XLA_GPU_USE_RUNTIME_FUSION_FIELD_NUMBER: builtins.int
|
|
408
|
+
XLA_DUMP_LATENCY_HIDING_SCHEDULE_FIELD_NUMBER: builtins.int
|
|
409
|
+
XLA_CPU_ENABLE_MLIR_TILING_AND_FUSION_FIELD_NUMBER: builtins.int
|
|
410
|
+
XLA_CPU_ENABLE_CUSTOM_MATMUL_TILING_FIELD_NUMBER: builtins.int
|
|
411
|
+
XLA_CPU_MATMUL_TILING_M_DIM_FIELD_NUMBER: builtins.int
|
|
412
|
+
XLA_CPU_MATMUL_TILING_N_DIM_FIELD_NUMBER: builtins.int
|
|
413
|
+
XLA_CPU_MATMUL_TILING_K_DIM_FIELD_NUMBER: builtins.int
|
|
414
|
+
XLA_CPU_ENABLE_MLIR_FUSION_OUTLINING_FIELD_NUMBER: builtins.int
|
|
415
|
+
XLA_CPU_ENABLE_EXPERIMENTAL_DEALLOCATION_FIELD_NUMBER: builtins.int
|
|
416
|
+
XLA_GPU_ENABLE_LATENCY_HIDING_SCHEDULER_FIELD_NUMBER: builtins.int
|
|
417
|
+
XLA_GPU_ENABLE_HIGHEST_PRIORITY_ASYNC_STREAM_FIELD_NUMBER: builtins.int
|
|
418
|
+
XLA_GPU_ENABLE_ANALYTICAL_LATENCY_ESTIMATOR_FIELD_NUMBER: builtins.int
|
|
419
|
+
XLA_GPU_LHS_ENABLE_GPU_ASYNC_TRACKER_FIELD_NUMBER: builtins.int
|
|
420
|
+
XLA_GPU_PGLE_PROFILE_FILE_OR_DIRECTORY_PATH_FIELD_NUMBER: builtins.int
|
|
421
|
+
XLA_GPU_MEMORY_LIMIT_SLOP_FACTOR_FIELD_NUMBER: builtins.int
|
|
422
|
+
XLA_GPU_ENABLE_PIPELINED_COLLECTIVES_FIELD_NUMBER: builtins.int
|
|
423
|
+
XLA_GPU_ENABLE_PIPELINED_ALL_REDUCE_FIELD_NUMBER: builtins.int
|
|
424
|
+
XLA_GPU_ENABLE_PIPELINED_ALL_GATHER_FIELD_NUMBER: builtins.int
|
|
425
|
+
XLA_GPU_ENABLE_PIPELINED_REDUCE_SCATTER_FIELD_NUMBER: builtins.int
|
|
426
|
+
XLA_GPU_ENABLE_PIPELINED_P2P_FIELD_NUMBER: builtins.int
|
|
427
|
+
XLA_GPU_RUN_POST_LAYOUT_COLLECTIVE_PIPELINER_FIELD_NUMBER: builtins.int
|
|
428
|
+
XLA_GPU_COLLECTIVE_PERMUTE_DECOMPOSER_THRESHOLD_FIELD_NUMBER: builtins.int
|
|
429
|
+
XLA_PARTITIONING_ALGORITHM_FIELD_NUMBER: builtins.int
|
|
430
|
+
XLA_GPU_ENABLE_TRITON_GEMM_FIELD_NUMBER: builtins.int
|
|
431
|
+
XLA_GPU_ENABLE_CUDNN_INT8X32_CONVOLUTION_REORDERING_FIELD_NUMBER: builtins.int
|
|
432
|
+
XLA_GPU_TRITON_GEMM_ANY_FIELD_NUMBER: builtins.int
|
|
433
|
+
XLA_GPU_EXHAUSTIVE_TILING_SEARCH_FIELD_NUMBER: builtins.int
|
|
434
|
+
XLA_GPU_ENABLE_PRIORITY_FUSION_FIELD_NUMBER: builtins.int
|
|
435
|
+
XLA_GPU_DUMP_AUTOTUNE_RESULTS_TO_FIELD_NUMBER: builtins.int
|
|
436
|
+
XLA_GPU_LOAD_AUTOTUNE_RESULTS_FROM_FIELD_NUMBER: builtins.int
|
|
437
|
+
XLA_GPU_TARGET_CONFIG_FILENAME_FIELD_NUMBER: builtins.int
|
|
438
|
+
XLA_GPU_AUTO_SPMD_PARTITIONING_MEMORY_BUDGET_GB_FIELD_NUMBER: builtins.int
|
|
439
|
+
XLA_GPU_AUTO_SPMD_PARTITIONING_MEMORY_BUDGET_RATIO_FIELD_NUMBER: builtins.int
|
|
440
|
+
XLA_GPU_TRITON_GEMM_DISABLE_REDUCED_PRECISION_REDUCTION_FIELD_NUMBER: builtins.int
|
|
441
|
+
XLA_GPU_TRITON_FUSION_LEVEL_FIELD_NUMBER: builtins.int
|
|
442
|
+
XLA_GPU_DUMP_AUTOTUNED_GEMM_FUSIONS_FIELD_NUMBER: builtins.int
|
|
443
|
+
XLA_GPU_OVERRIDE_GEMM_AUTOTUNER_FIELD_NUMBER: builtins.int
|
|
444
|
+
XLA_GPU_COPY_INSERTION_USE_REGION_ANALYSIS_FIELD_NUMBER: builtins.int
|
|
445
|
+
XLA_GPU_COLLECT_COST_MODEL_STATS_FIELD_NUMBER: builtins.int
|
|
446
|
+
XLA_GPU_ENABLE_SPLIT_K_AUTOTUNING_FIELD_NUMBER: builtins.int
|
|
447
|
+
XLA_GPU_ENABLE_REDUCTION_EPILOGUE_FUSION_FIELD_NUMBER: builtins.int
|
|
448
|
+
XLA_GPU_ENABLE_NCCL_CLIQUE_OPTIMIZATION_FIELD_NUMBER: builtins.int
|
|
449
|
+
XLA_GPU_MOCK_CUSTOM_CALLS_FIELD_NUMBER: builtins.int
|
|
450
|
+
XLA_GPU_CUBLAS_FALLBACK_FIELD_NUMBER: builtins.int
|
|
451
|
+
XLA_GPU_ENABLE_WHILE_LOOP_DOUBLE_BUFFERING_FIELD_NUMBER: builtins.int
|
|
452
|
+
XLA_GPU_ENABLE_WHILE_LOOP_UNROLLING_FIELD_NUMBER: builtins.int
|
|
453
|
+
XLA_GPU_ENSURE_MINOR_DOT_CONTRACTION_DIMS_FIELD_NUMBER: builtins.int
|
|
454
|
+
XLA_GPU_FILTER_KERNELS_SPILLING_REGISTERS_ON_AUTOTUNING_FIELD_NUMBER: builtins.int
|
|
455
|
+
XLA_DEBUG_BUFFER_ASSIGNMENT_SHOW_MAX_FIELD_NUMBER: builtins.int
|
|
456
|
+
XLA_GPU_LLVM_VERIFICATION_LEVEL_FIELD_NUMBER: builtins.int
|
|
457
|
+
XLA_GPU_ENABLE_CUB_RADIX_SORT_FIELD_NUMBER: builtins.int
|
|
458
|
+
XLA_GPU_THRESHOLD_FOR_WINDOWED_EINSUM_MIB_FIELD_NUMBER: builtins.int
|
|
459
|
+
XLA_GPU_ENABLE_TRITON_HOPPER_FIELD_NUMBER: builtins.int
|
|
460
|
+
XLA_GPU_ENABLE_NCCL_USER_BUFFERS_FIELD_NUMBER: builtins.int
|
|
461
|
+
XLA_GPU_ENABLE_NCCL_COMM_SPLITTING_FIELD_NUMBER: builtins.int
|
|
462
|
+
XLA_GPU_ENABLE_NCCL_PER_STREAM_COMMS_FIELD_NUMBER: builtins.int
|
|
463
|
+
XLA_GPU_ENABLE_LIBNVPTXCOMPILER_FIELD_NUMBER: builtins.int
|
|
464
|
+
XLA_GPU_ENABLE_DOT_STRENGTH_REDUCTION_FIELD_NUMBER: builtins.int
|
|
465
|
+
XLA_GPU_MULTI_STREAMED_WINDOWED_EINSUM_FIELD_NUMBER: builtins.int
|
|
466
|
+
XLA_GPU_ENABLE_BF16_6WAY_GEMM_FIELD_NUMBER: builtins.int
|
|
467
|
+
XLA_GPU_ENABLE_BF16_3WAY_GEMM_FIELD_NUMBER: builtins.int
|
|
468
|
+
XLA_GPU_NCCL_COLLECTIVE_MAX_NCHANNELS_FIELD_NUMBER: builtins.int
|
|
469
|
+
XLA_GPU_NCCL_P2P_MAX_NCHANNELS_FIELD_NUMBER: builtins.int
|
|
470
|
+
XLA_GPU_MLIR_EMITTER_LEVEL_FIELD_NUMBER: builtins.int
|
|
471
|
+
XLA_GPU_GEMM_REWRITE_SIZE_THRESHOLD_FIELD_NUMBER: builtins.int
|
|
472
|
+
XLA_GPU_REQUIRE_COMPLETE_AOT_AUTOTUNE_RESULTS_FIELD_NUMBER: builtins.int
|
|
473
|
+
XLA_GPU_CUDNN_GEMM_FUSION_LEVEL_FIELD_NUMBER: builtins.int
|
|
474
|
+
XLA_GPU_USE_MEMCPY_LOCAL_P2P_FIELD_NUMBER: builtins.int
|
|
475
|
+
XLA_GPU_AUTOTUNE_MAX_SOLUTIONS_FIELD_NUMBER: builtins.int
|
|
476
|
+
XLA_DUMP_LARGE_CONSTANTS_FIELD_NUMBER: builtins.int
|
|
477
|
+
XLA_GPU_VERIFY_TRITON_FUSION_NUMERICS_FIELD_NUMBER: builtins.int
|
|
478
|
+
XLA_GPU_DUMP_AUTOTUNE_LOGS_TO_FIELD_NUMBER: builtins.int
|
|
479
|
+
XLA_REDUCE_WINDOW_REWRITE_BASE_LENGTH_FIELD_NUMBER: builtins.int
|
|
480
|
+
XLA_GPU_ENABLE_HOST_MEMORY_OFFLOADING_FIELD_NUMBER: builtins.int
|
|
481
|
+
XLA_GPU_EXCLUDE_NONDETERMINISTIC_OPS_FIELD_NUMBER: builtins.int
|
|
482
|
+
XLA_GPU_NCCL_TERMINATE_ON_ERROR_FIELD_NUMBER: builtins.int
|
|
483
|
+
XLA_GPU_SHARD_AUTOTUNING_FIELD_NUMBER: builtins.int
|
|
484
|
+
XLA_GPU_ENABLE_APPROX_COSTLY_COLLECTIVES_FIELD_NUMBER: builtins.int
|
|
485
|
+
XLA_GPU_KERNEL_CACHE_FILE_FIELD_NUMBER: builtins.int
|
|
486
|
+
XLA_GPU_UNSAFE_PIPELINED_LOOP_ANNOTATOR_FIELD_NUMBER: builtins.int
|
|
487
|
+
XLA_GPU_PER_FUSION_AUTOTUNE_CACHE_DIR_FIELD_NUMBER: builtins.int
|
|
488
|
+
XLA_CMD_BUFFER_TRACE_CACHE_SIZE_FIELD_NUMBER: builtins.int
|
|
489
|
+
XLA_GPU_TEMP_BUFFER_USE_SEPARATE_COLOR_FIELD_NUMBER: builtins.int
|
|
490
|
+
LEGACY_COMMAND_BUFFER_CUSTOM_CALL_TARGETS_FIELD_NUMBER: builtins.int
|
|
491
|
+
XLA_SYNTAX_SUGAR_ASYNC_OPS_FIELD_NUMBER: builtins.int
|
|
492
|
+
XLA_GPU_AUTOTUNE_GEMM_RTOL_FIELD_NUMBER: builtins.int
|
|
493
|
+
XLA_ENABLE_COMMAND_BUFFERS_DURING_PROFILING_FIELD_NUMBER: builtins.int
|
|
494
|
+
XLA_GPU_CUDNN_GEMM_MAX_PLANS_FIELD_NUMBER: builtins.int
|
|
495
|
+
XLA_GPU_ENABLE_LIBNVJITLINK_FIELD_NUMBER: builtins.int
|
|
496
|
+
XLA_GPU_ENABLE_TRITON_GEMM_INT4_FIELD_NUMBER: builtins.int
|
|
497
|
+
XLA_GPU_ASYNC_DOT_FIELD_NUMBER: builtins.int
|
|
498
|
+
XLA_GPU_ENABLE_PGLE_ACCURACY_CHECKER_FIELD_NUMBER: builtins.int
|
|
499
|
+
XLA_GPU_EXECUTABLE_WARN_STUCK_TIMEOUT_SECONDS_FIELD_NUMBER: builtins.int
|
|
500
|
+
XLA_GPU_EXECUTABLE_TERMINATE_TIMEOUT_SECONDS_FIELD_NUMBER: builtins.int
|
|
501
|
+
XLA_EXPERIMENTAL_IGNORE_CHANNEL_ID_FIELD_NUMBER: builtins.int
|
|
502
|
+
XLA_BACKEND_EXTRA_OPTIONS_FIELD_NUMBER: builtins.int
|
|
503
|
+
xla_cpu_enable_concurrency_optimized_scheduler: builtins.bool
|
|
504
|
+
"""--------------------------------------------------------------------------//
|
|
505
|
+
XLA:CPU options.
|
|
506
|
+
--------------------------------------------------------------------------//
|
|
507
|
+
|
|
508
|
+
go/keep-sorted start newline_separated=yes
|
|
509
|
+
|
|
510
|
+
When true, XLA:CPU uses HLO module scheduler that is optimized for
|
|
511
|
+
extracting concurrency at the cost of extra memory: we extend the live
|
|
512
|
+
ranges of temporaries to allow XLA runtime to schedule independent
|
|
513
|
+
operations in parallel on separate threads.
|
|
514
|
+
"""
|
|
515
|
+
xla_cpu_enable_fast_math: builtins.bool
|
|
516
|
+
"""When true, "unsafe" mathematical optimizations are enabled. These
|
|
517
|
+
transformations include but are not limited to:
|
|
518
|
+
|
|
519
|
+
- Reducing the precision of operations (e.g. using an approximate sin
|
|
520
|
+
function, or transforming x/y into x * (1/y)).
|
|
521
|
+
- Assuming that operations never produce or consume NaN or +/- Inf (this
|
|
522
|
+
behavior can be adjusted using xla_cpu_fast_math_allow_{nans|infs}).
|
|
523
|
+
- Assuming that +0 and -0 are indistinguishable.
|
|
524
|
+
"""
|
|
525
|
+
xla_cpu_enable_fast_min_max: builtins.bool
|
|
526
|
+
"""When false we lower the Minimum and Maximum hlos in the CPU backend such
|
|
527
|
+
that Min(NotNaN, NaN) = Min(NaN, NotNaN) = NaN. In other words, if flag
|
|
528
|
+
this is false we always propagate NaNs through Min and Max.
|
|
529
|
+
|
|
530
|
+
Note, this does not correspond to the exact same behavior as the gpu flag
|
|
531
|
+
below!
|
|
532
|
+
"""
|
|
533
|
+
xla_cpu_fast_math_honor_division: builtins.bool
|
|
534
|
+
"""When xla_cpu_enable_fast_math is true then this controls whether we forbid
|
|
535
|
+
to use the reciprocal of an argument instead of division. Ignored when
|
|
536
|
+
xla_cpu_enable_fast_math is false.
|
|
537
|
+
"""
|
|
538
|
+
xla_cpu_fast_math_honor_functions: builtins.bool
|
|
539
|
+
"""When xla_cpu_enable_fast_math is true then this controls whether we forbid
|
|
540
|
+
to approximate calculations for functions. Ignored when
|
|
541
|
+
xla_cpu_enable_fast_math is false.
|
|
542
|
+
"""
|
|
543
|
+
xla_cpu_fast_math_honor_infs: builtins.bool
|
|
544
|
+
"""When xla_cpu_enable_fast_math is true then this controls whether we allow
|
|
545
|
+
operations to produce infinites. Ignored when xla_cpu_enable_fast_math is
|
|
546
|
+
false.
|
|
547
|
+
"""
|
|
548
|
+
xla_cpu_fast_math_honor_nans: builtins.bool
|
|
549
|
+
"""When xla_cpu_enable_fast_math is true then this controls whether we allow
|
|
550
|
+
operations to produce NaNs. Ignored when xla_cpu_enable_fast_math is
|
|
551
|
+
false.
|
|
552
|
+
"""
|
|
553
|
+
xla_cpu_use_thunk_runtime: builtins.bool
|
|
554
|
+
"""When true, XLA:CPU uses the thunk runtime to execute compiled program."""
|
|
555
|
+
xla_cpu_parallel_codegen_split_count: builtins.int
|
|
556
|
+
"""The number of parts to split the LLVM module into before codegen. This
|
|
557
|
+
allows XLA to compile all parts in parallel, and resolve kernel symbols
|
|
558
|
+
from different dynamic libraries.
|
|
559
|
+
"""
|
|
560
|
+
xla_cpu_prefer_vector_width: builtins.int
|
|
561
|
+
"""A `prefer-vector-width` value that is passed to the LLVM backend. Default
|
|
562
|
+
value is `256` (AVX2 on x86 platforms).
|
|
563
|
+
"""
|
|
564
|
+
xla_gpu_experimental_autotune_cache_mode: global___DebugOptions.AutotuneCacheMode.ValueType
|
|
565
|
+
"""--------------------------------------------------------------------------//
|
|
566
|
+
XLA:GPU options.
|
|
567
|
+
--------------------------------------------------------------------------//
|
|
568
|
+
go/keep-sorted start newline_separated=yes skip_lines=1
|
|
569
|
+
|
|
570
|
+
Specifies the behavior of per kernel autotuning cache.
|
|
571
|
+
"""
|
|
572
|
+
xla_gpu_experimental_disable_binary_libraries: builtins.bool
|
|
573
|
+
"""Experimentally disables binary libraries in GPU compiler passes."""
|
|
574
|
+
xla_gpu_experimental_enable_triton_softmax_priority_fusion: builtins.bool
|
|
575
|
+
"""Gates the experimental feature coupling the Triton Softmax pattern matcher
|
|
576
|
+
with priority fusion.
|
|
577
|
+
"""
|
|
578
|
+
xla_gpu_unsupported_enable_triton_gemm: builtins.bool
|
|
579
|
+
"""Internal debug/testing flag to switch Triton GEMM fusions on or off."""
|
|
580
|
+
xla_hlo_graph_addresses: builtins.bool
|
|
581
|
+
"""--------------------------------------------------------------------------//
|
|
582
|
+
A bag of XLA options that have to be categorized.
|
|
583
|
+
--------------------------------------------------------------------------//
|
|
584
|
+
|
|
585
|
+
Show addresses of HLO ops in graph dump.
|
|
586
|
+
"""
|
|
587
|
+
xla_hlo_profile: builtins.bool
|
|
588
|
+
"""Instrument the computation to collect per-HLO cycle counts."""
|
|
589
|
+
xla_disable_all_hlo_passes: builtins.bool
|
|
590
|
+
"""Disables all HLO passes. Notes that some passes are necessary for
|
|
591
|
+
correctness and the invariants that must be satisfied by "fully optimized"
|
|
592
|
+
HLO are different for different devices and may change over time. The only
|
|
593
|
+
"guarantee", such as it is, is that if you compile XLA and dump the
|
|
594
|
+
optimized HLO for some graph, you should be able to run it again on the
|
|
595
|
+
same device with the same build of XLA.
|
|
596
|
+
"""
|
|
597
|
+
xla_backend_optimization_level: builtins.int
|
|
598
|
+
"""Numerical optimization level for the XLA compiler backend; the specific
|
|
599
|
+
interpretation of this value is left to the backends.
|
|
600
|
+
"""
|
|
601
|
+
xla_embed_ir_in_executable: builtins.bool
|
|
602
|
+
"""Embed the compiler IR as a string in the executable."""
|
|
603
|
+
xla_eliminate_hlo_implicit_broadcast: builtins.bool
|
|
604
|
+
"""Eliminate implicit broadcasts when lowering user computations to HLO
|
|
605
|
+
instructions; use explicit broadcast instead.
|
|
606
|
+
"""
|
|
607
|
+
xla_cpu_multi_thread_eigen: builtins.bool
|
|
608
|
+
"""When generating calls to Eigen in the CPU backend, use multi-threaded Eigen
|
|
609
|
+
mode.
|
|
610
|
+
"""
|
|
611
|
+
xla_gpu_cuda_data_dir: builtins.str
|
|
612
|
+
"""Path to directory with cuda/ptx tools and libraries."""
|
|
613
|
+
xla_gpu_ftz: builtins.bool
|
|
614
|
+
"""Enable flush-to-zero semantics in the GPU backend."""
|
|
615
|
+
xla_llvm_enable_alias_scope_metadata: builtins.bool
|
|
616
|
+
"""If true, in LLVM-based backends, emit !alias.scope metadata in
|
|
617
|
+
generated IR.
|
|
618
|
+
"""
|
|
619
|
+
xla_llvm_enable_noalias_metadata: builtins.bool
|
|
620
|
+
"""If true, in LLVM-based backends, emit !noalias metadata in the
|
|
621
|
+
generated IR.
|
|
622
|
+
"""
|
|
623
|
+
xla_llvm_enable_invariant_load_metadata: builtins.bool
|
|
624
|
+
"""If true, in LLVM-based backends, emit !invariant.load metadata in
|
|
625
|
+
the generated IR.
|
|
626
|
+
"""
|
|
627
|
+
xla_llvm_disable_expensive_passes: builtins.bool
|
|
628
|
+
"""If true, a set of expensive LLVM optimization passes will not be run."""
|
|
629
|
+
xla_test_all_output_layouts: builtins.bool
|
|
630
|
+
"""This is used by ClientLibraryTestBase::ComputeAndCompare*. If true, the
|
|
631
|
+
computation will run n! times with all permunations of layouts for the
|
|
632
|
+
output shape in rank n. For example, with a 3D shape, all permutations of
|
|
633
|
+
the set {0, 1, 2} are tried.
|
|
634
|
+
"""
|
|
635
|
+
xla_test_all_input_layouts: builtins.bool
|
|
636
|
+
"""This is used by ClientLibraryTestBase::ComputeAndCompare*. If true, the
|
|
637
|
+
computation will run for all permunations of layouts of all input
|
|
638
|
+
arguments. For example, with 2 input arguments in 2D and 4D shapes, the
|
|
639
|
+
computation will run 2! * 4! times.
|
|
640
|
+
"""
|
|
641
|
+
xla_hlo_graph_sharding_color: builtins.bool
|
|
642
|
+
"""Assign colors based on sharding information when generating the Graphviz
|
|
643
|
+
HLO graph.
|
|
644
|
+
"""
|
|
645
|
+
xla_cpu_use_mkl_dnn: builtins.bool
|
|
646
|
+
"""Generate calls to MKL-DNN in the CPU backend."""
|
|
647
|
+
xla_gpu_enable_fast_min_max: builtins.bool
|
|
648
|
+
"""When true we lower the Minimum and Maximum hlos in the GPU backend such
|
|
649
|
+
that Min(NotNaN, NaN) = Min(NaN, NotNaN) = NotNaN. In other words, if flag
|
|
650
|
+
this is true we don't propagate NaNs through Min and Max.
|
|
651
|
+
|
|
652
|
+
Note, this does not correspond to the exact same behavior as the cpu flag
|
|
653
|
+
above!
|
|
654
|
+
"""
|
|
655
|
+
xla_allow_excess_precision: builtins.bool
|
|
656
|
+
"""Allows xla to increase the output precision of floating point operations
|
|
657
|
+
and all floating-point conversions to be simplified, including those
|
|
658
|
+
that affect the numerics. The `FloatNormalization` pass inserts many
|
|
659
|
+
`f32 -> bf16 -> f32` conversion pairs. These are not removed by the
|
|
660
|
+
`AlgebraicSimplifier`, as that will only simplify conversions that are
|
|
661
|
+
no-ops, e.g. `bf16 -> f32 -> bf16`. Removing these improves accuracy.
|
|
662
|
+
"""
|
|
663
|
+
xla_gpu_crash_on_verification_failures: builtins.bool
|
|
664
|
+
"""Crashes the program when any kind of verification fails, instead of just
|
|
665
|
+
logging the failures. One example is cross checking of convolution results
|
|
666
|
+
among different algorithms.
|
|
667
|
+
"""
|
|
668
|
+
xla_gpu_autotune_level: builtins.int
|
|
669
|
+
"""0: Disable gemm and convolution autotuning.
|
|
670
|
+
1: Enable autotuning, but disable correctness checking.
|
|
671
|
+
2: Also set output buffers to random numbers during autotuning.
|
|
672
|
+
3: Also reset output buffers to random numbers after autotuning each
|
|
673
|
+
algorithm.
|
|
674
|
+
4+: Also check for correct outputs and for out-of-bounds reads/writes.
|
|
675
|
+
|
|
676
|
+
Default: 4.
|
|
677
|
+
"""
|
|
678
|
+
xla_force_host_platform_device_count: builtins.int
|
|
679
|
+
"""Force the host platform to pretend that there are these many host
|
|
680
|
+
"devices". All these devices are backed by the same threadpool. Defaults
|
|
681
|
+
to 1.
|
|
682
|
+
|
|
683
|
+
Setting this to anything other than 1 can increase overhead from context
|
|
684
|
+
switching but we let the user override this behavior to help run tests on
|
|
685
|
+
the host that run models in parallel across multiple devices.
|
|
686
|
+
"""
|
|
687
|
+
xla_gpu_disable_gpuasm_optimizations: builtins.bool
|
|
688
|
+
"""If set to true XLA:GPU invokes `ptxas` with -O0 (default is -O3)."""
|
|
689
|
+
xla_gpu_shape_checks: global___DebugOptions.ShapeChecks.ValueType
|
|
690
|
+
xla_hlo_evaluator_use_fast_path: builtins.bool
|
|
691
|
+
"""Enable fast math with eigen in the HLO evaluator."""
|
|
692
|
+
xla_allow_scalar_index_dynamic_ops: builtins.bool
|
|
693
|
+
"""Temporary option to allow support for both the R1 and the scalar index
|
|
694
|
+
versions of DynamicSlice and DynamicUpdateSlice. Only used for testing.
|
|
695
|
+
"""
|
|
696
|
+
xla_step_marker_location: global___DebugOptions.StepMarkerLocation.ValueType
|
|
697
|
+
"""Option to emit a target-specific marker to indicate the start of a training
|
|
698
|
+
step. The location of the marker (if any) is determined by the option
|
|
699
|
+
value.
|
|
700
|
+
"""
|
|
701
|
+
xla_dump_to: builtins.str
|
|
702
|
+
"""
|
|
703
|
+
BEGIN flags controlling dumping HLO modules for debugging.
|
|
704
|
+
|
|
705
|
+
When dumping is enabled, HLO modules dumped at the very beginning and end
|
|
706
|
+
of compilation, and optionally also during the pass pipeline.
|
|
707
|
+
|
|
708
|
+
In general, if you set one of these flags, we will try to infer reasonable
|
|
709
|
+
defaults for the others. For example:
|
|
710
|
+
|
|
711
|
+
* Setting --xla_dump_to=/tmp/foo without specifying a format
|
|
712
|
+
with --xla_dump_hlo_as_* will turn on --xla_dump_hlo_as_text.
|
|
713
|
+
|
|
714
|
+
* Setting --xla_dump_hlo_as_text without specifying --xla_dump_to will
|
|
715
|
+
dump to stdout.
|
|
716
|
+
|
|
717
|
+
Directory to dump into.
|
|
718
|
+
"""
|
|
719
|
+
xla_dump_hlo_module_re: builtins.str
|
|
720
|
+
"""If specified, will only dump modules which match this regexp."""
|
|
721
|
+
xla_dump_hlo_pass_re: builtins.str
|
|
722
|
+
"""If this flag is specified, will also dump HLO before and after passes that
|
|
723
|
+
match this regular expression. Set to .* to dump before/after all passes.
|
|
724
|
+
"""
|
|
725
|
+
xla_dump_hlo_as_text: builtins.bool
|
|
726
|
+
"""Specifies the format that HLO is dumped in. Multiple of these may be
|
|
727
|
+
specified.
|
|
728
|
+
"""
|
|
729
|
+
xla_dump_hlo_as_proto: builtins.bool
|
|
730
|
+
xla_dump_hlo_as_dot: builtins.bool
|
|
731
|
+
xla_dump_hlo_as_url: builtins.bool
|
|
732
|
+
xla_dump_hlo_as_html: builtins.bool
|
|
733
|
+
"""Dump HLO graphs as an HTML (DOT -> SVG inlined in HTML)"""
|
|
734
|
+
xla_dump_fusion_visualization: builtins.bool
|
|
735
|
+
"""Dump the visualization of the fusion progress."""
|
|
736
|
+
xla_dump_hlo_snapshots: builtins.bool
|
|
737
|
+
"""If true, every time an HLO module is run, we will dump an HloSnapshot
|
|
738
|
+
(essentially, a serialized module plus its inputs) to the --xla_dump_to
|
|
739
|
+
directory.
|
|
740
|
+
"""
|
|
741
|
+
xla_dump_include_timestamp: builtins.bool
|
|
742
|
+
"""Include a timestamp in the dumped filenames."""
|
|
743
|
+
xla_dump_max_hlo_modules: builtins.int
|
|
744
|
+
"""Max number of hlo module dumps in a directory. Set to < 0 for unbounded."""
|
|
745
|
+
xla_dump_module_metadata: builtins.bool
|
|
746
|
+
"""Dump HloModuleMetadata as a text proto for each HLO module."""
|
|
747
|
+
xla_dump_compress_protos: builtins.bool
|
|
748
|
+
"""GZip-compress protos dumped via --xla_dump_hlo_as_proto."""
|
|
749
|
+
xla_dump_hlo_as_long_text: builtins.bool
|
|
750
|
+
"""Dump HLO in long text format. Ignored unless xla_dump_hlo_as_text is true."""
|
|
751
|
+
xla_gpu_force_conv_nchw: builtins.bool
|
|
752
|
+
"""
|
|
753
|
+
END flags controlling dumping HLO modules.
|
|
754
|
+
|
|
755
|
+
Overrides for XLA GPU's convolution layout heuristic.
|
|
756
|
+
"""
|
|
757
|
+
xla_gpu_force_conv_nhwc: builtins.bool
|
|
758
|
+
xla_gpu_dump_llvmir: builtins.bool
|
|
759
|
+
"""Whether to dump llvm ir when compiling to ptx."""
|
|
760
|
+
xla_dump_enable_mlir_pretty_form: builtins.bool
|
|
761
|
+
"""Whether to dump mlir using pretty print form."""
|
|
762
|
+
xla_gpu_algorithm_denylist_path: builtins.str
|
|
763
|
+
"""Denylist for cuDNN convolutions."""
|
|
764
|
+
xla_tpu_detect_nan: builtins.bool
|
|
765
|
+
"""Debug options that trigger execution errors when NaN or Inf are detected."""
|
|
766
|
+
xla_tpu_detect_inf: builtins.bool
|
|
767
|
+
xla_cpu_enable_xprof_traceme: builtins.bool
|
|
768
|
+
"""True if TraceMe annotations are enabled for XLA:CPU."""
|
|
769
|
+
xla_gpu_unsafe_fallback_to_driver_on_ptxas_not_found: builtins.bool
|
|
770
|
+
"""It is usually preferable to not fallback to the driver; it can consume more
|
|
771
|
+
memory, or have bugs.
|
|
772
|
+
"""
|
|
773
|
+
xla_gpu_asm_extra_flags: builtins.str
|
|
774
|
+
"""Extra parameters to pass the GPU assembler."""
|
|
775
|
+
xla_multiheap_size_constraint_per_heap: builtins.int
|
|
776
|
+
"""Per-heap size constraint. New heaps will be created if per-heap max size is
|
|
777
|
+
reached.
|
|
778
|
+
"""
|
|
779
|
+
xla_detailed_logging: builtins.bool
|
|
780
|
+
"""Enable detailed logging into vlog. If this is disabled, no
|
|
781
|
+
compilation summary will be printed in the end of computation.
|
|
782
|
+
"""
|
|
783
|
+
xla_enable_dumping: builtins.bool
|
|
784
|
+
"""Enable HLO dumping. If this is disabled, no HLO modules will be dumped."""
|
|
785
|
+
xla_gpu_force_compilation_parallelism: builtins.int
|
|
786
|
+
"""Overrides normal multi-threaded compilation setting to use this many
|
|
787
|
+
threads. Setting to 0 (the default value) means no enforcement.
|
|
788
|
+
"""
|
|
789
|
+
xla_gpu_enable_llvm_module_compilation_parallelism: builtins.bool
|
|
790
|
+
xla_gpu_deterministic_ops: builtins.bool
|
|
791
|
+
"""Guarantees run-to-run determinism.
|
|
792
|
+
This flag implies --xla_gpu_exclude_nondeterministic_ops and in addition
|
|
793
|
+
disables autotuning.
|
|
794
|
+
"""
|
|
795
|
+
xla_gpu_all_reduce_combine_threshold_bytes: builtins.int
|
|
796
|
+
"""Size threshold (in bytes) for the GPU collective combiners."""
|
|
797
|
+
xla_gpu_all_gather_combine_threshold_bytes: builtins.int
|
|
798
|
+
xla_gpu_reduce_scatter_combine_threshold_bytes: builtins.int
|
|
799
|
+
xla_gpu_enable_all_gather_combine_by_dim: builtins.bool
|
|
800
|
+
"""Combine all-gather/scatter-reduce ops with the same dimension or
|
|
801
|
+
irrespective of their dimension.
|
|
802
|
+
"""
|
|
803
|
+
xla_gpu_enable_reduce_scatter_combine_by_dim: builtins.bool
|
|
804
|
+
xla_gpu_enable_reassociation_for_converted_ar: builtins.bool
|
|
805
|
+
"""Enable allreduce reassociation on allreduces that are converted to a wider
|
|
806
|
+
type. The resulting allreduce will be promoted to a wider-typed allreduce.
|
|
807
|
+
"""
|
|
808
|
+
xla_gpu_all_reduce_blueconnect_num_devices_per_host: builtins.int
|
|
809
|
+
"""Number of devices per host for first stage of BlueConnect decomposition
|
|
810
|
+
pass. The pass will attempt to decompose all-reduces ops into a
|
|
811
|
+
ReduceScatter-AllReduce-AllGather sequence, with the initial ReduceScatter
|
|
812
|
+
being performed over all of the devices in the same host. Set to < 1 to
|
|
813
|
+
disable all-reduce decomposition.
|
|
814
|
+
"""
|
|
815
|
+
xla_gpu_enable_while_loop_reduce_scatter_code_motion: builtins.bool
|
|
816
|
+
"""Enable hoisting of reduce-scatter out of while loops."""
|
|
817
|
+
xla_gpu_collective_inflation_factor: builtins.int
|
|
818
|
+
"""Inflate collective cost by running each collective multiple times."""
|
|
819
|
+
xla_llvm_force_inline_before_split: builtins.bool
|
|
820
|
+
"""Whether to force inline before llvm module split to get a more balanced
|
|
821
|
+
splits for parallel compilation.
|
|
822
|
+
"""
|
|
823
|
+
xla_gpu_enable_cudnn_frontend: builtins.bool
|
|
824
|
+
"""Whether to use the cuDNN frontend API for convolutions when possible."""
|
|
825
|
+
xla_gpu_enable_cudnn_fmha: builtins.bool
|
|
826
|
+
xla_gpu_fused_attention_use_cudnn_rng: builtins.bool
|
|
827
|
+
xla_gpu_enable_cudnn_layer_norm: builtins.bool
|
|
828
|
+
"""Rewrite layer norm patterns into cuDNN library calls."""
|
|
829
|
+
xla_dump_disable_metadata: builtins.bool
|
|
830
|
+
"""Disable dumping metadata in HLO dumps."""
|
|
831
|
+
xla_dump_hlo_pipeline_re: builtins.str
|
|
832
|
+
"""If this flag is specified, will only dump HLO before and after passes in
|
|
833
|
+
the pass pipeline that matches this regular expression. Default empty value
|
|
834
|
+
enables dumping in all pipelines.
|
|
835
|
+
"""
|
|
836
|
+
xla_gpu_strict_conv_algorithm_picker: builtins.bool
|
|
837
|
+
"""If true, abort immediately when conv algorithm picker fails, rather than
|
|
838
|
+
logging a warning and proceeding with fallback.
|
|
839
|
+
"""
|
|
840
|
+
xla_gpu_enable_custom_fusions: builtins.bool
|
|
841
|
+
"""If true, XLA will try to pattern match subgraphs of HLO operations into
|
|
842
|
+
custom fusions registered in the current process (pre-compiled hand written
|
|
843
|
+
kernels, e.g. various GEMM fusions writtent in CUTLASS).
|
|
844
|
+
"""
|
|
845
|
+
xla_gpu_enable_custom_fusions_re: builtins.str
|
|
846
|
+
"""A regular expression enabling only a subset of custom fusions. Enabled only
|
|
847
|
+
if `xla_gpu_enable_custom_fusion` set to true.
|
|
848
|
+
"""
|
|
849
|
+
xla_gpu_enable_dynamic_slice_fusion: builtins.bool
|
|
850
|
+
"""Enables address computation fusion to optimize dynamic-slice and
|
|
851
|
+
dynamic-update-slice operations around library calls.
|
|
852
|
+
"""
|
|
853
|
+
xla_gpu_nccl_termination_timeout_seconds: builtins.int
|
|
854
|
+
"""Timeout in seconds before terminating jobs that are stuck in a NCCL
|
|
855
|
+
Rendezvous. Negative value disables the timeout and will not terminate.
|
|
856
|
+
"""
|
|
857
|
+
xla_gpu_enable_shared_constants: builtins.bool
|
|
858
|
+
"""Enables shared constants for XLA/GPU. This allows large constants to be
|
|
859
|
+
shared among multiple GPU executables.
|
|
860
|
+
"""
|
|
861
|
+
xla_gpu_enable_cublaslt: builtins.bool
|
|
862
|
+
"""Whether to use cuBLASLt for GEMMs on GPUs."""
|
|
863
|
+
xla_gpu_graph_min_graph_size: builtins.int
|
|
864
|
+
"""This number determines how many moved instructions like fusion kernels are
|
|
865
|
+
required for a region to be captured as a function to be launched as a GPU
|
|
866
|
+
graph.
|
|
867
|
+
"""
|
|
868
|
+
xla_gpu_graph_enable_concurrent_region: builtins.bool
|
|
869
|
+
"""Identify concurrent regions in GPU graphs and execute them concurrently."""
|
|
870
|
+
xla_gpu_redzone_scratch_max_megabytes: builtins.int
|
|
871
|
+
"""Size threshold (in megabytes) for the GPU redzone scratch allocator."""
|
|
872
|
+
xla_gpu_redzone_padding_bytes: builtins.int
|
|
873
|
+
"""Amount of padding the redzone allocator will put on one side of each buffer
|
|
874
|
+
it allocates. (So the buffer's total size will be increased by 2x this
|
|
875
|
+
value.)
|
|
876
|
+
|
|
877
|
+
Higher values make it more likely that we'll catch an out-of-bounds read or
|
|
878
|
+
write. Smaller values consume less memory during autotuning. Note that a
|
|
879
|
+
fused cudnn conv has up to 6 total buffers (4 inputs, 1 output, and 1
|
|
880
|
+
scratch), so this can be multiplied by quite a lot.
|
|
881
|
+
"""
|
|
882
|
+
xla_cpu_use_acl: builtins.bool
|
|
883
|
+
"""Generate calls to Arm Compute Library in the CPU backend."""
|
|
884
|
+
xla_cpu_strict_dot_conv_math: builtins.bool
|
|
885
|
+
"""By default, XLA:CPU will run fp16 dot/conv as fp32, as this is generally
|
|
886
|
+
(much) faster on our hardware. Set this flag to disable this behavior.
|
|
887
|
+
"""
|
|
888
|
+
xla_gpu_use_runtime_fusion: builtins.bool
|
|
889
|
+
"""An option to enable using cuDNN runtime compiled fusion kernels which is
|
|
890
|
+
available and recommended for Ampere+ GPUs.
|
|
891
|
+
"""
|
|
892
|
+
xla_dump_latency_hiding_schedule: builtins.bool
|
|
893
|
+
xla_cpu_enable_mlir_tiling_and_fusion: builtins.bool
|
|
894
|
+
"""By default, MLIR lowering will use Linalg elementwise fusion. If this flag
|
|
895
|
+
is enabled, the pipeline will use tiling, fusion, peeling, vectorization
|
|
896
|
+
instead.
|
|
897
|
+
"""
|
|
898
|
+
xla_cpu_enable_custom_matmul_tiling: builtins.bool
|
|
899
|
+
"""XLA:CPU-Next tiling parameters for matmul."""
|
|
900
|
+
xla_cpu_matmul_tiling_m_dim: builtins.int
|
|
901
|
+
xla_cpu_matmul_tiling_n_dim: builtins.int
|
|
902
|
+
xla_cpu_matmul_tiling_k_dim: builtins.int
|
|
903
|
+
xla_cpu_enable_mlir_fusion_outlining: builtins.bool
|
|
904
|
+
xla_cpu_enable_experimental_deallocation: builtins.bool
|
|
905
|
+
"""If set, use the experimental deallocation pass from mlir-hlo."""
|
|
906
|
+
xla_gpu_enable_latency_hiding_scheduler: builtins.bool
|
|
907
|
+
xla_gpu_enable_highest_priority_async_stream: builtins.bool
|
|
908
|
+
xla_gpu_enable_analytical_latency_estimator: builtins.bool
|
|
909
|
+
xla_gpu_lhs_enable_gpu_async_tracker: builtins.bool
|
|
910
|
+
xla_gpu_pgle_profile_file_or_directory_path: builtins.str
|
|
911
|
+
xla_gpu_memory_limit_slop_factor: builtins.int
|
|
912
|
+
xla_gpu_enable_pipelined_collectives: builtins.bool
|
|
913
|
+
xla_gpu_enable_pipelined_all_reduce: builtins.bool
|
|
914
|
+
xla_gpu_enable_pipelined_all_gather: builtins.bool
|
|
915
|
+
xla_gpu_enable_pipelined_reduce_scatter: builtins.bool
|
|
916
|
+
xla_gpu_enable_pipelined_p2p: builtins.bool
|
|
917
|
+
xla_gpu_run_post_layout_collective_pipeliner: builtins.bool
|
|
918
|
+
xla_gpu_collective_permute_decomposer_threshold: builtins.int
|
|
919
|
+
"""The minimum data size in bytes to trigger collective-permute-decomposer
|
|
920
|
+
transformation.
|
|
921
|
+
"""
|
|
922
|
+
xla_partitioning_algorithm: global___DebugOptions.PartitioningAlgorithm.ValueType
|
|
923
|
+
"""The partitioning algorithm to be used in the PartitionAssignment pass."""
|
|
924
|
+
xla_gpu_enable_triton_gemm: builtins.bool
|
|
925
|
+
xla_gpu_enable_cudnn_int8x32_convolution_reordering: builtins.bool
|
|
926
|
+
xla_gpu_triton_gemm_any: builtins.bool
|
|
927
|
+
"""Creates triton fusion for all supported gemms.
|
|
928
|
+
To make sure only triton gemm is chosen by the autotuner run with
|
|
929
|
+
`xla_gpu_cublas_fallback` set to false.
|
|
930
|
+
"""
|
|
931
|
+
xla_gpu_exhaustive_tiling_search: builtins.bool
|
|
932
|
+
xla_gpu_enable_priority_fusion: builtins.bool
|
|
933
|
+
xla_gpu_dump_autotune_results_to: builtins.str
|
|
934
|
+
"""File to write autotune results to. It will be a binary file unless the name
|
|
935
|
+
ends with .txt or .textproto. Warning: The results are written at every
|
|
936
|
+
compilation, possibly multiple times per process. This only works on CUDA.
|
|
937
|
+
"""
|
|
938
|
+
xla_gpu_load_autotune_results_from: builtins.str
|
|
939
|
+
"""File to load autotune results from. It will be considered a binary file
|
|
940
|
+
unless the name ends with .txt or .textproto. At most one loading will
|
|
941
|
+
happen during the lifetime of one process, even if the first one is
|
|
942
|
+
unsuccessful or different file paths are passed here. This only works on
|
|
943
|
+
CUDA.
|
|
944
|
+
"""
|
|
945
|
+
xla_gpu_target_config_filename: builtins.str
|
|
946
|
+
"""Description of the target platform in GpuTargetConfigProto format; if
|
|
947
|
+
provided, deviceless compilation is assumed, and the current device is
|
|
948
|
+
ignored.
|
|
949
|
+
"""
|
|
950
|
+
xla_gpu_auto_spmd_partitioning_memory_budget_gb: builtins.int
|
|
951
|
+
"""Memory budget in GB per device for AutoSharding."""
|
|
952
|
+
xla_gpu_auto_spmd_partitioning_memory_budget_ratio: builtins.float
|
|
953
|
+
"""See the definition of the
|
|
954
|
+
xla_gpu_auto_spmd_partitioning_memory_budget_ratio flag for the meaning of
|
|
955
|
+
this field.
|
|
956
|
+
"""
|
|
957
|
+
xla_gpu_triton_gemm_disable_reduced_precision_reduction: builtins.bool
|
|
958
|
+
xla_gpu_triton_fusion_level: builtins.int
|
|
959
|
+
xla_gpu_dump_autotuned_gemm_fusions: builtins.bool
|
|
960
|
+
xla_gpu_override_gemm_autotuner: builtins.str
|
|
961
|
+
xla_gpu_copy_insertion_use_region_analysis: builtins.bool
|
|
962
|
+
xla_gpu_collect_cost_model_stats: builtins.bool
|
|
963
|
+
"""If true, each fusion instruction will have a cost model runtime estimate in
|
|
964
|
+
backend config after compilation.
|
|
965
|
+
"""
|
|
966
|
+
xla_gpu_enable_split_k_autotuning: builtins.bool
|
|
967
|
+
xla_gpu_enable_reduction_epilogue_fusion: builtins.bool
|
|
968
|
+
"""Whether reduction epilogue fusion is enabled in fusion passes."""
|
|
969
|
+
xla_gpu_enable_nccl_clique_optimization: builtins.bool
|
|
970
|
+
"""Allow early return when acquiring NCCL cliques."""
|
|
971
|
+
xla_gpu_mock_custom_calls: builtins.bool
|
|
972
|
+
"""Replace custom calls with noop operations."""
|
|
973
|
+
xla_gpu_cublas_fallback: builtins.bool
|
|
974
|
+
"""Allow Triton GEMM autotuning to fall back to cuBLAS when that is
|
|
975
|
+
faster.
|
|
976
|
+
"""
|
|
977
|
+
xla_gpu_enable_while_loop_double_buffering: builtins.bool
|
|
978
|
+
"""Enable double buffering for loops."""
|
|
979
|
+
xla_gpu_enable_while_loop_unrolling: global___DebugOptions.WhileLoopUnrolling.ValueType
|
|
980
|
+
"""Determine the while loop unrolling scheme."""
|
|
981
|
+
xla_gpu_ensure_minor_dot_contraction_dims: builtins.bool
|
|
982
|
+
"""Change the layout of the second triton dot operand to be column major.
|
|
983
|
+
Only works for (bf16 x bf16) -> bf16.
|
|
984
|
+
"""
|
|
985
|
+
xla_gpu_filter_kernels_spilling_registers_on_autotuning: builtins.bool
|
|
986
|
+
"""Filter out kernels that spill registers during autotuning."""
|
|
987
|
+
xla_debug_buffer_assignment_show_max: builtins.int
|
|
988
|
+
"""Maximum number of buffers to print when debugging buffer assignment."""
|
|
989
|
+
xla_gpu_llvm_verification_level: builtins.int
|
|
990
|
+
xla_gpu_enable_cub_radix_sort: builtins.bool
|
|
991
|
+
"""Enable radix sort using CUB."""
|
|
992
|
+
xla_gpu_threshold_for_windowed_einsum_mib: builtins.int
|
|
993
|
+
"""Threshold to enable windowed einsum (collective matmul) in MB."""
|
|
994
|
+
xla_gpu_enable_triton_hopper: builtins.bool
|
|
995
|
+
"""Enables currently disabled features within Triton for Hopper."""
|
|
996
|
+
xla_gpu_enable_nccl_user_buffers: builtins.bool
|
|
997
|
+
"""Enable NCCL user buffers."""
|
|
998
|
+
xla_gpu_enable_nccl_comm_splitting: builtins.bool
|
|
999
|
+
"""Enable NCCL communicator splitting."""
|
|
1000
|
+
xla_gpu_enable_nccl_per_stream_comms: builtins.bool
|
|
1001
|
+
"""Enable NCCL per stream communicators."""
|
|
1002
|
+
xla_gpu_enable_libnvptxcompiler: builtins.bool
|
|
1003
|
+
"""If enabled, uses the libnvptxcompiler library to compile PTX to cuBIN."""
|
|
1004
|
+
xla_gpu_enable_dot_strength_reduction: builtins.bool
|
|
1005
|
+
xla_gpu_multi_streamed_windowed_einsum: builtins.bool
|
|
1006
|
+
"""Whether to use multiple compute streams to run windowed einsum."""
|
|
1007
|
+
xla_gpu_enable_bf16_6way_gemm: builtins.bool
|
|
1008
|
+
"""If enabled, uses bf16_6way gemm to compute F32 gemm."""
|
|
1009
|
+
xla_gpu_enable_bf16_3way_gemm: builtins.bool
|
|
1010
|
+
"""If enabled, uses bf16_3way gemm to compute F32 gemm."""
|
|
1011
|
+
xla_gpu_nccl_collective_max_nchannels: builtins.int
|
|
1012
|
+
"""Specify the maximum number of channels(SMs) NCCL
|
|
1013
|
+
will use for collective operations.
|
|
1014
|
+
"""
|
|
1015
|
+
xla_gpu_nccl_p2p_max_nchannels: builtins.int
|
|
1016
|
+
"""Specify the maximum number of channels(SMs) NCCL
|
|
1017
|
+
will use for p2p operations.
|
|
1018
|
+
"""
|
|
1019
|
+
xla_gpu_mlir_emitter_level: builtins.int
|
|
1020
|
+
"""Choose the level of mlir emitters that are enabled.
|
|
1021
|
+
Current levels:
|
|
1022
|
+
0: Disabled.
|
|
1023
|
+
1: Loop emitter
|
|
1024
|
+
2: + Loop-like emitters
|
|
1025
|
+
3: + Transpose
|
|
1026
|
+
4: + Reduce
|
|
1027
|
+
"""
|
|
1028
|
+
xla_gpu_gemm_rewrite_size_threshold: builtins.int
|
|
1029
|
+
"""Threshold to rewrite matmul to cuBLAS or Triton (minimum combined number of
|
|
1030
|
+
elements of both matrices in non-batch dimensions to be considered for a
|
|
1031
|
+
rewrite).
|
|
1032
|
+
"""
|
|
1033
|
+
xla_gpu_require_complete_aot_autotune_results: builtins.bool
|
|
1034
|
+
"""If true, will require complete AOT autotuning results; in the case of
|
|
1035
|
+
missing AOT result, the model will not be compiled or executed, a
|
|
1036
|
+
`NotFound` error will be returned.
|
|
1037
|
+
"""
|
|
1038
|
+
xla_gpu_cudnn_gemm_fusion_level: builtins.int
|
|
1039
|
+
"""Let GEMM fusion autotuning probe cuDNN as a backend.
|
|
1040
|
+
Current levels:
|
|
1041
|
+
0: Disabled.
|
|
1042
|
+
1: Fusions of GEMM, elementwise, transpose/reshape operations.
|
|
1043
|
+
2: + Broadcasts, slicing.
|
|
1044
|
+
3: + Nontrivial noncontracting dimension reshapes/transposes.
|
|
1045
|
+
"""
|
|
1046
|
+
xla_gpu_use_memcpy_local_p2p: builtins.bool
|
|
1047
|
+
"""This instructs the runtime whether to use
|
|
1048
|
+
memcpy for p2p communication when source and
|
|
1049
|
+
target are located within a node(nvlink).
|
|
1050
|
+
"""
|
|
1051
|
+
xla_gpu_autotune_max_solutions: builtins.int
|
|
1052
|
+
"""If non-zero, limits the number of solutions to be used by GEMM autotuner.
|
|
1053
|
+
This might be useful if underlying math library returns too many GEMM
|
|
1054
|
+
solutions.
|
|
1055
|
+
"""
|
|
1056
|
+
xla_dump_large_constants: builtins.bool
|
|
1057
|
+
"""If true, large constants will be printed out when dumping HLOs."""
|
|
1058
|
+
xla_gpu_verify_triton_fusion_numerics: builtins.bool
|
|
1059
|
+
"""If true, will verify that the numerical results of Triton fusions match
|
|
1060
|
+
the results of regular emitters.
|
|
1061
|
+
"""
|
|
1062
|
+
xla_gpu_dump_autotune_logs_to: builtins.str
|
|
1063
|
+
"""File to write autotune logs to. It will stored in txt format."""
|
|
1064
|
+
xla_reduce_window_rewrite_base_length: builtins.int
|
|
1065
|
+
"""Base length to rewrite the reduce window to, no rewrite if set to 0."""
|
|
1066
|
+
xla_gpu_enable_host_memory_offloading: builtins.bool
|
|
1067
|
+
"""If true, will enable host memory offloading on a device."""
|
|
1068
|
+
xla_gpu_exclude_nondeterministic_ops: builtins.bool
|
|
1069
|
+
"""Excludes non-deterministic ops from compiled executables.
|
|
1070
|
+
Unlike --xla_gpu_deterministic_ops does not disable autotuning - the
|
|
1071
|
+
compilation itself can be non-deterministic.
|
|
1072
|
+
At present, the HLO op SelectAndScatter does not have a
|
|
1073
|
+
deterministic XLA:GPU implementation.
|
|
1074
|
+
Compilation errors out if SelectAndScatter is encountered.
|
|
1075
|
+
Scatter ops can non-deterministic by default; these get converted to
|
|
1076
|
+
a deterministic implementation.
|
|
1077
|
+
"""
|
|
1078
|
+
xla_gpu_nccl_terminate_on_error: builtins.bool
|
|
1079
|
+
"""If true, Nccl errors will terminate the process."""
|
|
1080
|
+
xla_gpu_shard_autotuning: builtins.bool
|
|
1081
|
+
xla_gpu_enable_approx_costly_collectives: builtins.bool
|
|
1082
|
+
xla_gpu_kernel_cache_file: builtins.str
|
|
1083
|
+
xla_gpu_unsafe_pipelined_loop_annotator: builtins.bool
|
|
1084
|
+
"""Recognises rotate-right patterns (slice, slice, concat) within a while
|
|
1085
|
+
loop and labels the while loop as a pipelined while loop. This is an
|
|
1086
|
+
unsafe flag.
|
|
1087
|
+
"""
|
|
1088
|
+
xla_gpu_per_fusion_autotune_cache_dir: builtins.str
|
|
1089
|
+
xla_cmd_buffer_trace_cache_size: builtins.int
|
|
1090
|
+
"""The command buffer trace cache size, increasing the cache size may
|
|
1091
|
+
sometimes reduces the chances of doing command buffer tracing for
|
|
1092
|
+
updating command buffer instance.
|
|
1093
|
+
"""
|
|
1094
|
+
xla_gpu_temp_buffer_use_separate_color: builtins.bool
|
|
1095
|
+
"""Enable this flag will use a separate memory space color for
|
|
1096
|
+
temp buffer, and then will use separate memory allocator to allocate it,
|
|
1097
|
+
as there is no other memory allocation interference,
|
|
1098
|
+
it will allocate temp buffer to some fix address on every iteration,
|
|
1099
|
+
which is good for cuda-graph perf.
|
|
1100
|
+
"""
|
|
1101
|
+
xla_syntax_sugar_async_ops: builtins.bool
|
|
1102
|
+
"""This flag is used for controlling HLO dumping and NVTX marker. If turned
|
|
1103
|
+
on, both HLO dumping and NVTX marker will use syntactic sugar wrappers
|
|
1104
|
+
as op names, while the actual op names will be shown if turned off.
|
|
1105
|
+
|
|
1106
|
+
Here is an example HLO excerpt with the flag off:
|
|
1107
|
+
|
|
1108
|
+
async_computation {
|
|
1109
|
+
param_0 = f32[1,4,8]{1,0,2} parameter(0)
|
|
1110
|
+
ROOT all-to-all.3.1 = f32[1,4,8]{1,0,2} all-to-all(param_0),
|
|
1111
|
+
replica_groups={{0,1,2,3,4,5,6,7}}, dimensions={2}
|
|
1112
|
+
}
|
|
1113
|
+
...
|
|
1114
|
+
|
|
1115
|
+
all-to-all-start =
|
|
1116
|
+
((f32[1,4,8]{1,0,2}), f32[1,4,8]{1,0,2}) async-start(bitcast.24.0),
|
|
1117
|
+
calls=async_computation, backend_config={...}
|
|
1118
|
+
all-to-all-done = f32[1,4,8]{1,0,2} async-done(all-to-all-start)
|
|
1119
|
+
|
|
1120
|
+
and with the flag on:
|
|
1121
|
+
|
|
1122
|
+
all-to-all-start = ((f32[1,4,8]{1,0,2}), f32[1,4,8]{1,0,2})
|
|
1123
|
+
all-to-all-start(bitcast.24.0),
|
|
1124
|
+
replica_groups={{0,1,2,3,4,5,6,7}}, dimensions={2},
|
|
1125
|
+
backend_config={...}
|
|
1126
|
+
all-to-all-done = f32[1,4,8]{1,0,2} all-to-all-done(all-to-all-start)
|
|
1127
|
+
"""
|
|
1128
|
+
xla_gpu_autotune_gemm_rtol: builtins.float
|
|
1129
|
+
"""Relative precision for comparing different GEMM solutions"""
|
|
1130
|
+
xla_enable_command_buffers_during_profiling: builtins.bool
|
|
1131
|
+
"""Allow launching command buffers while profiling active.
|
|
1132
|
+
When disabled, execute in op-by-op mode.
|
|
1133
|
+
TODO(b/355487968): Remove this option when validation complete.
|
|
1134
|
+
"""
|
|
1135
|
+
xla_gpu_cudnn_gemm_max_plans: builtins.int
|
|
1136
|
+
"""Limit for the number of kernel configurations (plans) to use during
|
|
1137
|
+
autotuning of cuDNN GEMM fusions. The more - the slower the autotuning
|
|
1138
|
+
but potentially higher the performance.
|
|
1139
|
+
"""
|
|
1140
|
+
xla_gpu_enable_libnvjitlink: builtins.bool
|
|
1141
|
+
"""If enabled, uses the libnvjitlink library for PTX compilation and linking"""
|
|
1142
|
+
xla_gpu_enable_triton_gemm_int4: builtins.bool
|
|
1143
|
+
"""If enabled, generates triton gemm kernels for int4 inputs."""
|
|
1144
|
+
xla_gpu_async_dot: builtins.bool
|
|
1145
|
+
"""If true, XLA will wrap `dot` operations into async computations in an
|
|
1146
|
+
effort to parallelize matrix operations.
|
|
1147
|
+
"""
|
|
1148
|
+
xla_gpu_enable_pgle_accuracy_checker: builtins.bool
|
|
1149
|
+
"""Enables strict PGLE checking. If an FDO profile is specified and latency
|
|
1150
|
+
hiding scheduler encounters missing instructions in the profile
|
|
1151
|
+
compilation will halt.
|
|
1152
|
+
"""
|
|
1153
|
+
xla_gpu_executable_warn_stuck_timeout_seconds: builtins.int
|
|
1154
|
+
"""Timeouts for RendezvousSingle stuck warning and termination."""
|
|
1155
|
+
xla_gpu_executable_terminate_timeout_seconds: builtins.int
|
|
1156
|
+
xla_experimental_ignore_channel_id: builtins.bool
|
|
1157
|
+
"""Whether to ignore channel ids(including verifier channel id checks)
|
|
1158
|
+
for collectives in the given HLO.
|
|
1159
|
+
"""
|
|
1160
|
+
@property
|
|
1161
|
+
def xla_disable_hlo_passes(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
|
|
1162
|
+
"""List of HLO passes to disable/enable. These names must exactly match the
|
|
1163
|
+
pass names as specified by the HloPassInterface::name() method.
|
|
1164
|
+
|
|
1165
|
+
At least one of xla_disable_hlo_passes and xla_enable_hlo_passes_only must
|
|
1166
|
+
be empty.
|
|
1167
|
+
"""
|
|
1168
|
+
|
|
1169
|
+
@property
|
|
1170
|
+
def xla_enable_hlo_passes_only(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]: ...
|
|
1171
|
+
@property
|
|
1172
|
+
def xla_gpu_ptx_file(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
|
|
1173
|
+
"""Paths to files with ptx code."""
|
|
1174
|
+
|
|
1175
|
+
@property
|
|
1176
|
+
def xla_gpu_llvm_ir_file(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
|
|
1177
|
+
"""Paths to files with LLVM code."""
|
|
1178
|
+
|
|
1179
|
+
@property
|
|
1180
|
+
def xla_gpu_disable_async_collectives(
|
|
1181
|
+
self,
|
|
1182
|
+
) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[global___DebugOptions.CollectiveOpType.ValueType]: ...
|
|
1183
|
+
@property
|
|
1184
|
+
def xla_gpu_enable_command_buffer(
|
|
1185
|
+
self,
|
|
1186
|
+
) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[global___DebugOptions.CommandBufferCmdType.ValueType]:
|
|
1187
|
+
"""Determine the types of commands that are recorded into command buffers."""
|
|
1188
|
+
|
|
1189
|
+
@property
|
|
1190
|
+
def legacy_command_buffer_custom_call_targets(
|
|
1191
|
+
self,
|
|
1192
|
+
) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]:
|
|
1193
|
+
"""Custom call targets with legacy registry API (non FFI API),
|
|
1194
|
+
that support recording to command buffer custom command,
|
|
1195
|
+
i.e., custom call target supports cuda-graph capturing for CUDA devices.
|
|
1196
|
+
This flag is read if CUSTOM_CALL command type is recorded into
|
|
1197
|
+
command buffer.
|
|
1198
|
+
"""
|
|
1199
|
+
|
|
1200
|
+
@property
|
|
1201
|
+
def xla_backend_extra_options(self) -> google.protobuf.internal.containers.ScalarMap[builtins.str, builtins.str]:
|
|
1202
|
+
"""Next id: 331
|
|
1203
|
+
|
|
1204
|
+
Extra options to pass to the compilation backend (e.g. LLVM); specific
|
|
1205
|
+
interpretation of these values is left to the backend.
|
|
1206
|
+
"""
|
|
1207
|
+
|
|
1208
|
+
def __init__(
|
|
1209
|
+
self,
|
|
1210
|
+
*,
|
|
1211
|
+
xla_cpu_enable_concurrency_optimized_scheduler: builtins.bool | None = ...,
|
|
1212
|
+
xla_cpu_enable_fast_math: builtins.bool | None = ...,
|
|
1213
|
+
xla_cpu_enable_fast_min_max: builtins.bool | None = ...,
|
|
1214
|
+
xla_cpu_fast_math_honor_division: builtins.bool | None = ...,
|
|
1215
|
+
xla_cpu_fast_math_honor_functions: builtins.bool | None = ...,
|
|
1216
|
+
xla_cpu_fast_math_honor_infs: builtins.bool | None = ...,
|
|
1217
|
+
xla_cpu_fast_math_honor_nans: builtins.bool | None = ...,
|
|
1218
|
+
xla_cpu_use_thunk_runtime: builtins.bool | None = ...,
|
|
1219
|
+
xla_cpu_parallel_codegen_split_count: builtins.int | None = ...,
|
|
1220
|
+
xla_cpu_prefer_vector_width: builtins.int | None = ...,
|
|
1221
|
+
xla_gpu_experimental_autotune_cache_mode: global___DebugOptions.AutotuneCacheMode.ValueType | None = ...,
|
|
1222
|
+
xla_gpu_experimental_disable_binary_libraries: builtins.bool | None = ...,
|
|
1223
|
+
xla_gpu_experimental_enable_triton_softmax_priority_fusion: builtins.bool | None = ...,
|
|
1224
|
+
xla_gpu_unsupported_enable_triton_gemm: builtins.bool | None = ...,
|
|
1225
|
+
xla_hlo_graph_addresses: builtins.bool | None = ...,
|
|
1226
|
+
xla_hlo_profile: builtins.bool | None = ...,
|
|
1227
|
+
xla_disable_hlo_passes: collections.abc.Iterable[builtins.str] | None = ...,
|
|
1228
|
+
xla_enable_hlo_passes_only: collections.abc.Iterable[builtins.str] | None = ...,
|
|
1229
|
+
xla_disable_all_hlo_passes: builtins.bool | None = ...,
|
|
1230
|
+
xla_backend_optimization_level: builtins.int | None = ...,
|
|
1231
|
+
xla_embed_ir_in_executable: builtins.bool | None = ...,
|
|
1232
|
+
xla_eliminate_hlo_implicit_broadcast: builtins.bool | None = ...,
|
|
1233
|
+
xla_cpu_multi_thread_eigen: builtins.bool | None = ...,
|
|
1234
|
+
xla_gpu_cuda_data_dir: builtins.str | None = ...,
|
|
1235
|
+
xla_gpu_ftz: builtins.bool | None = ...,
|
|
1236
|
+
xla_llvm_enable_alias_scope_metadata: builtins.bool | None = ...,
|
|
1237
|
+
xla_llvm_enable_noalias_metadata: builtins.bool | None = ...,
|
|
1238
|
+
xla_llvm_enable_invariant_load_metadata: builtins.bool | None = ...,
|
|
1239
|
+
xla_llvm_disable_expensive_passes: builtins.bool | None = ...,
|
|
1240
|
+
xla_test_all_output_layouts: builtins.bool | None = ...,
|
|
1241
|
+
xla_test_all_input_layouts: builtins.bool | None = ...,
|
|
1242
|
+
xla_hlo_graph_sharding_color: builtins.bool | None = ...,
|
|
1243
|
+
xla_cpu_use_mkl_dnn: builtins.bool | None = ...,
|
|
1244
|
+
xla_gpu_enable_fast_min_max: builtins.bool | None = ...,
|
|
1245
|
+
xla_allow_excess_precision: builtins.bool | None = ...,
|
|
1246
|
+
xla_gpu_crash_on_verification_failures: builtins.bool | None = ...,
|
|
1247
|
+
xla_gpu_autotune_level: builtins.int | None = ...,
|
|
1248
|
+
xla_force_host_platform_device_count: builtins.int | None = ...,
|
|
1249
|
+
xla_gpu_disable_gpuasm_optimizations: builtins.bool | None = ...,
|
|
1250
|
+
xla_gpu_shape_checks: global___DebugOptions.ShapeChecks.ValueType | None = ...,
|
|
1251
|
+
xla_hlo_evaluator_use_fast_path: builtins.bool | None = ...,
|
|
1252
|
+
xla_allow_scalar_index_dynamic_ops: builtins.bool | None = ...,
|
|
1253
|
+
xla_step_marker_location: global___DebugOptions.StepMarkerLocation.ValueType | None = ...,
|
|
1254
|
+
xla_dump_to: builtins.str | None = ...,
|
|
1255
|
+
xla_dump_hlo_module_re: builtins.str | None = ...,
|
|
1256
|
+
xla_dump_hlo_pass_re: builtins.str | None = ...,
|
|
1257
|
+
xla_dump_hlo_as_text: builtins.bool | None = ...,
|
|
1258
|
+
xla_dump_hlo_as_proto: builtins.bool | None = ...,
|
|
1259
|
+
xla_dump_hlo_as_dot: builtins.bool | None = ...,
|
|
1260
|
+
xla_dump_hlo_as_url: builtins.bool | None = ...,
|
|
1261
|
+
xla_dump_hlo_as_html: builtins.bool | None = ...,
|
|
1262
|
+
xla_dump_fusion_visualization: builtins.bool | None = ...,
|
|
1263
|
+
xla_dump_hlo_snapshots: builtins.bool | None = ...,
|
|
1264
|
+
xla_dump_include_timestamp: builtins.bool | None = ...,
|
|
1265
|
+
xla_dump_max_hlo_modules: builtins.int | None = ...,
|
|
1266
|
+
xla_dump_module_metadata: builtins.bool | None = ...,
|
|
1267
|
+
xla_dump_compress_protos: builtins.bool | None = ...,
|
|
1268
|
+
xla_dump_hlo_as_long_text: builtins.bool | None = ...,
|
|
1269
|
+
xla_gpu_force_conv_nchw: builtins.bool | None = ...,
|
|
1270
|
+
xla_gpu_force_conv_nhwc: builtins.bool | None = ...,
|
|
1271
|
+
xla_gpu_ptx_file: collections.abc.Iterable[builtins.str] | None = ...,
|
|
1272
|
+
xla_gpu_dump_llvmir: builtins.bool | None = ...,
|
|
1273
|
+
xla_dump_enable_mlir_pretty_form: builtins.bool | None = ...,
|
|
1274
|
+
xla_gpu_algorithm_denylist_path: builtins.str | None = ...,
|
|
1275
|
+
xla_tpu_detect_nan: builtins.bool | None = ...,
|
|
1276
|
+
xla_tpu_detect_inf: builtins.bool | None = ...,
|
|
1277
|
+
xla_cpu_enable_xprof_traceme: builtins.bool | None = ...,
|
|
1278
|
+
xla_gpu_unsafe_fallback_to_driver_on_ptxas_not_found: builtins.bool | None = ...,
|
|
1279
|
+
xla_gpu_asm_extra_flags: builtins.str | None = ...,
|
|
1280
|
+
xla_multiheap_size_constraint_per_heap: builtins.int | None = ...,
|
|
1281
|
+
xla_detailed_logging: builtins.bool | None = ...,
|
|
1282
|
+
xla_enable_dumping: builtins.bool | None = ...,
|
|
1283
|
+
xla_gpu_force_compilation_parallelism: builtins.int | None = ...,
|
|
1284
|
+
xla_gpu_enable_llvm_module_compilation_parallelism: builtins.bool | None = ...,
|
|
1285
|
+
xla_gpu_deterministic_ops: builtins.bool | None = ...,
|
|
1286
|
+
xla_gpu_llvm_ir_file: collections.abc.Iterable[builtins.str] | None = ...,
|
|
1287
|
+
xla_gpu_disable_async_collectives: (
|
|
1288
|
+
collections.abc.Iterable[global___DebugOptions.CollectiveOpType.ValueType] | None
|
|
1289
|
+
) = ...,
|
|
1290
|
+
xla_gpu_all_reduce_combine_threshold_bytes: builtins.int | None = ...,
|
|
1291
|
+
xla_gpu_all_gather_combine_threshold_bytes: builtins.int | None = ...,
|
|
1292
|
+
xla_gpu_reduce_scatter_combine_threshold_bytes: builtins.int | None = ...,
|
|
1293
|
+
xla_gpu_enable_all_gather_combine_by_dim: builtins.bool | None = ...,
|
|
1294
|
+
xla_gpu_enable_reduce_scatter_combine_by_dim: builtins.bool | None = ...,
|
|
1295
|
+
xla_gpu_enable_reassociation_for_converted_ar: builtins.bool | None = ...,
|
|
1296
|
+
xla_gpu_all_reduce_blueconnect_num_devices_per_host: builtins.int | None = ...,
|
|
1297
|
+
xla_gpu_enable_while_loop_reduce_scatter_code_motion: builtins.bool | None = ...,
|
|
1298
|
+
xla_gpu_collective_inflation_factor: builtins.int | None = ...,
|
|
1299
|
+
xla_llvm_force_inline_before_split: builtins.bool | None = ...,
|
|
1300
|
+
xla_gpu_enable_cudnn_frontend: builtins.bool | None = ...,
|
|
1301
|
+
xla_gpu_enable_cudnn_fmha: builtins.bool | None = ...,
|
|
1302
|
+
xla_gpu_fused_attention_use_cudnn_rng: builtins.bool | None = ...,
|
|
1303
|
+
xla_gpu_enable_cudnn_layer_norm: builtins.bool | None = ...,
|
|
1304
|
+
xla_dump_disable_metadata: builtins.bool | None = ...,
|
|
1305
|
+
xla_dump_hlo_pipeline_re: builtins.str | None = ...,
|
|
1306
|
+
xla_gpu_strict_conv_algorithm_picker: builtins.bool | None = ...,
|
|
1307
|
+
xla_gpu_enable_custom_fusions: builtins.bool | None = ...,
|
|
1308
|
+
xla_gpu_enable_custom_fusions_re: builtins.str | None = ...,
|
|
1309
|
+
xla_gpu_enable_dynamic_slice_fusion: builtins.bool | None = ...,
|
|
1310
|
+
xla_gpu_nccl_termination_timeout_seconds: builtins.int | None = ...,
|
|
1311
|
+
xla_gpu_enable_shared_constants: builtins.bool | None = ...,
|
|
1312
|
+
xla_gpu_enable_cublaslt: builtins.bool | None = ...,
|
|
1313
|
+
xla_gpu_enable_command_buffer: (
|
|
1314
|
+
collections.abc.Iterable[global___DebugOptions.CommandBufferCmdType.ValueType] | None
|
|
1315
|
+
) = ...,
|
|
1316
|
+
xla_gpu_graph_min_graph_size: builtins.int | None = ...,
|
|
1317
|
+
xla_gpu_graph_enable_concurrent_region: builtins.bool | None = ...,
|
|
1318
|
+
xla_gpu_redzone_scratch_max_megabytes: builtins.int | None = ...,
|
|
1319
|
+
xla_gpu_redzone_padding_bytes: builtins.int | None = ...,
|
|
1320
|
+
xla_cpu_use_acl: builtins.bool | None = ...,
|
|
1321
|
+
xla_cpu_strict_dot_conv_math: builtins.bool | None = ...,
|
|
1322
|
+
xla_gpu_use_runtime_fusion: builtins.bool | None = ...,
|
|
1323
|
+
xla_dump_latency_hiding_schedule: builtins.bool | None = ...,
|
|
1324
|
+
xla_cpu_enable_mlir_tiling_and_fusion: builtins.bool | None = ...,
|
|
1325
|
+
xla_cpu_enable_custom_matmul_tiling: builtins.bool | None = ...,
|
|
1326
|
+
xla_cpu_matmul_tiling_m_dim: builtins.int | None = ...,
|
|
1327
|
+
xla_cpu_matmul_tiling_n_dim: builtins.int | None = ...,
|
|
1328
|
+
xla_cpu_matmul_tiling_k_dim: builtins.int | None = ...,
|
|
1329
|
+
xla_cpu_enable_mlir_fusion_outlining: builtins.bool | None = ...,
|
|
1330
|
+
xla_cpu_enable_experimental_deallocation: builtins.bool | None = ...,
|
|
1331
|
+
xla_gpu_enable_latency_hiding_scheduler: builtins.bool | None = ...,
|
|
1332
|
+
xla_gpu_enable_highest_priority_async_stream: builtins.bool | None = ...,
|
|
1333
|
+
xla_gpu_enable_analytical_latency_estimator: builtins.bool | None = ...,
|
|
1334
|
+
xla_gpu_lhs_enable_gpu_async_tracker: builtins.bool | None = ...,
|
|
1335
|
+
xla_gpu_pgle_profile_file_or_directory_path: builtins.str | None = ...,
|
|
1336
|
+
xla_gpu_memory_limit_slop_factor: builtins.int | None = ...,
|
|
1337
|
+
xla_gpu_enable_pipelined_collectives: builtins.bool | None = ...,
|
|
1338
|
+
xla_gpu_enable_pipelined_all_reduce: builtins.bool | None = ...,
|
|
1339
|
+
xla_gpu_enable_pipelined_all_gather: builtins.bool | None = ...,
|
|
1340
|
+
xla_gpu_enable_pipelined_reduce_scatter: builtins.bool | None = ...,
|
|
1341
|
+
xla_gpu_enable_pipelined_p2p: builtins.bool | None = ...,
|
|
1342
|
+
xla_gpu_run_post_layout_collective_pipeliner: builtins.bool | None = ...,
|
|
1343
|
+
xla_gpu_collective_permute_decomposer_threshold: builtins.int | None = ...,
|
|
1344
|
+
xla_partitioning_algorithm: global___DebugOptions.PartitioningAlgorithm.ValueType | None = ...,
|
|
1345
|
+
xla_gpu_enable_triton_gemm: builtins.bool | None = ...,
|
|
1346
|
+
xla_gpu_enable_cudnn_int8x32_convolution_reordering: builtins.bool | None = ...,
|
|
1347
|
+
xla_gpu_triton_gemm_any: builtins.bool | None = ...,
|
|
1348
|
+
xla_gpu_exhaustive_tiling_search: builtins.bool | None = ...,
|
|
1349
|
+
xla_gpu_enable_priority_fusion: builtins.bool | None = ...,
|
|
1350
|
+
xla_gpu_dump_autotune_results_to: builtins.str | None = ...,
|
|
1351
|
+
xla_gpu_load_autotune_results_from: builtins.str | None = ...,
|
|
1352
|
+
xla_gpu_target_config_filename: builtins.str | None = ...,
|
|
1353
|
+
xla_gpu_auto_spmd_partitioning_memory_budget_gb: builtins.int | None = ...,
|
|
1354
|
+
xla_gpu_auto_spmd_partitioning_memory_budget_ratio: builtins.float | None = ...,
|
|
1355
|
+
xla_gpu_triton_gemm_disable_reduced_precision_reduction: builtins.bool | None = ...,
|
|
1356
|
+
xla_gpu_triton_fusion_level: builtins.int | None = ...,
|
|
1357
|
+
xla_gpu_dump_autotuned_gemm_fusions: builtins.bool | None = ...,
|
|
1358
|
+
xla_gpu_override_gemm_autotuner: builtins.str | None = ...,
|
|
1359
|
+
xla_gpu_copy_insertion_use_region_analysis: builtins.bool | None = ...,
|
|
1360
|
+
xla_gpu_collect_cost_model_stats: builtins.bool | None = ...,
|
|
1361
|
+
xla_gpu_enable_split_k_autotuning: builtins.bool | None = ...,
|
|
1362
|
+
xla_gpu_enable_reduction_epilogue_fusion: builtins.bool | None = ...,
|
|
1363
|
+
xla_gpu_enable_nccl_clique_optimization: builtins.bool | None = ...,
|
|
1364
|
+
xla_gpu_mock_custom_calls: builtins.bool | None = ...,
|
|
1365
|
+
xla_gpu_cublas_fallback: builtins.bool | None = ...,
|
|
1366
|
+
xla_gpu_enable_while_loop_double_buffering: builtins.bool | None = ...,
|
|
1367
|
+
xla_gpu_enable_while_loop_unrolling: global___DebugOptions.WhileLoopUnrolling.ValueType | None = ...,
|
|
1368
|
+
xla_gpu_ensure_minor_dot_contraction_dims: builtins.bool | None = ...,
|
|
1369
|
+
xla_gpu_filter_kernels_spilling_registers_on_autotuning: builtins.bool | None = ...,
|
|
1370
|
+
xla_debug_buffer_assignment_show_max: builtins.int | None = ...,
|
|
1371
|
+
xla_gpu_llvm_verification_level: builtins.int | None = ...,
|
|
1372
|
+
xla_gpu_enable_cub_radix_sort: builtins.bool | None = ...,
|
|
1373
|
+
xla_gpu_threshold_for_windowed_einsum_mib: builtins.int | None = ...,
|
|
1374
|
+
xla_gpu_enable_triton_hopper: builtins.bool | None = ...,
|
|
1375
|
+
xla_gpu_enable_nccl_user_buffers: builtins.bool | None = ...,
|
|
1376
|
+
xla_gpu_enable_nccl_comm_splitting: builtins.bool | None = ...,
|
|
1377
|
+
xla_gpu_enable_nccl_per_stream_comms: builtins.bool | None = ...,
|
|
1378
|
+
xla_gpu_enable_libnvptxcompiler: builtins.bool | None = ...,
|
|
1379
|
+
xla_gpu_enable_dot_strength_reduction: builtins.bool | None = ...,
|
|
1380
|
+
xla_gpu_multi_streamed_windowed_einsum: builtins.bool | None = ...,
|
|
1381
|
+
xla_gpu_enable_bf16_6way_gemm: builtins.bool | None = ...,
|
|
1382
|
+
xla_gpu_enable_bf16_3way_gemm: builtins.bool | None = ...,
|
|
1383
|
+
xla_gpu_nccl_collective_max_nchannels: builtins.int | None = ...,
|
|
1384
|
+
xla_gpu_nccl_p2p_max_nchannels: builtins.int | None = ...,
|
|
1385
|
+
xla_gpu_mlir_emitter_level: builtins.int | None = ...,
|
|
1386
|
+
xla_gpu_gemm_rewrite_size_threshold: builtins.int | None = ...,
|
|
1387
|
+
xla_gpu_require_complete_aot_autotune_results: builtins.bool | None = ...,
|
|
1388
|
+
xla_gpu_cudnn_gemm_fusion_level: builtins.int | None = ...,
|
|
1389
|
+
xla_gpu_use_memcpy_local_p2p: builtins.bool | None = ...,
|
|
1390
|
+
xla_gpu_autotune_max_solutions: builtins.int | None = ...,
|
|
1391
|
+
xla_dump_large_constants: builtins.bool | None = ...,
|
|
1392
|
+
xla_gpu_verify_triton_fusion_numerics: builtins.bool | None = ...,
|
|
1393
|
+
xla_gpu_dump_autotune_logs_to: builtins.str | None = ...,
|
|
1394
|
+
xla_reduce_window_rewrite_base_length: builtins.int | None = ...,
|
|
1395
|
+
xla_gpu_enable_host_memory_offloading: builtins.bool | None = ...,
|
|
1396
|
+
xla_gpu_exclude_nondeterministic_ops: builtins.bool | None = ...,
|
|
1397
|
+
xla_gpu_nccl_terminate_on_error: builtins.bool | None = ...,
|
|
1398
|
+
xla_gpu_shard_autotuning: builtins.bool | None = ...,
|
|
1399
|
+
xla_gpu_enable_approx_costly_collectives: builtins.bool | None = ...,
|
|
1400
|
+
xla_gpu_kernel_cache_file: builtins.str | None = ...,
|
|
1401
|
+
xla_gpu_unsafe_pipelined_loop_annotator: builtins.bool | None = ...,
|
|
1402
|
+
xla_gpu_per_fusion_autotune_cache_dir: builtins.str | None = ...,
|
|
1403
|
+
xla_cmd_buffer_trace_cache_size: builtins.int | None = ...,
|
|
1404
|
+
xla_gpu_temp_buffer_use_separate_color: builtins.bool | None = ...,
|
|
1405
|
+
legacy_command_buffer_custom_call_targets: collections.abc.Iterable[builtins.str] | None = ...,
|
|
1406
|
+
xla_syntax_sugar_async_ops: builtins.bool | None = ...,
|
|
1407
|
+
xla_gpu_autotune_gemm_rtol: builtins.float | None = ...,
|
|
1408
|
+
xla_enable_command_buffers_during_profiling: builtins.bool | None = ...,
|
|
1409
|
+
xla_gpu_cudnn_gemm_max_plans: builtins.int | None = ...,
|
|
1410
|
+
xla_gpu_enable_libnvjitlink: builtins.bool | None = ...,
|
|
1411
|
+
xla_gpu_enable_triton_gemm_int4: builtins.bool | None = ...,
|
|
1412
|
+
xla_gpu_async_dot: builtins.bool | None = ...,
|
|
1413
|
+
xla_gpu_enable_pgle_accuracy_checker: builtins.bool | None = ...,
|
|
1414
|
+
xla_gpu_executable_warn_stuck_timeout_seconds: builtins.int | None = ...,
|
|
1415
|
+
xla_gpu_executable_terminate_timeout_seconds: builtins.int | None = ...,
|
|
1416
|
+
xla_experimental_ignore_channel_id: builtins.bool | None = ...,
|
|
1417
|
+
xla_backend_extra_options: collections.abc.Mapping[builtins.str, builtins.str] | None = ...,
|
|
1418
|
+
) -> None: ...
|
|
1419
|
+
def ClearField(
|
|
1420
|
+
self,
|
|
1421
|
+
field_name: typing.Literal[
|
|
1422
|
+
"legacy_command_buffer_custom_call_targets",
|
|
1423
|
+
b"legacy_command_buffer_custom_call_targets",
|
|
1424
|
+
"xla_allow_excess_precision",
|
|
1425
|
+
b"xla_allow_excess_precision",
|
|
1426
|
+
"xla_allow_scalar_index_dynamic_ops",
|
|
1427
|
+
b"xla_allow_scalar_index_dynamic_ops",
|
|
1428
|
+
"xla_backend_extra_options",
|
|
1429
|
+
b"xla_backend_extra_options",
|
|
1430
|
+
"xla_backend_optimization_level",
|
|
1431
|
+
b"xla_backend_optimization_level",
|
|
1432
|
+
"xla_cmd_buffer_trace_cache_size",
|
|
1433
|
+
b"xla_cmd_buffer_trace_cache_size",
|
|
1434
|
+
"xla_cpu_enable_concurrency_optimized_scheduler",
|
|
1435
|
+
b"xla_cpu_enable_concurrency_optimized_scheduler",
|
|
1436
|
+
"xla_cpu_enable_custom_matmul_tiling",
|
|
1437
|
+
b"xla_cpu_enable_custom_matmul_tiling",
|
|
1438
|
+
"xla_cpu_enable_experimental_deallocation",
|
|
1439
|
+
b"xla_cpu_enable_experimental_deallocation",
|
|
1440
|
+
"xla_cpu_enable_fast_math",
|
|
1441
|
+
b"xla_cpu_enable_fast_math",
|
|
1442
|
+
"xla_cpu_enable_fast_min_max",
|
|
1443
|
+
b"xla_cpu_enable_fast_min_max",
|
|
1444
|
+
"xla_cpu_enable_mlir_fusion_outlining",
|
|
1445
|
+
b"xla_cpu_enable_mlir_fusion_outlining",
|
|
1446
|
+
"xla_cpu_enable_mlir_tiling_and_fusion",
|
|
1447
|
+
b"xla_cpu_enable_mlir_tiling_and_fusion",
|
|
1448
|
+
"xla_cpu_enable_xprof_traceme",
|
|
1449
|
+
b"xla_cpu_enable_xprof_traceme",
|
|
1450
|
+
"xla_cpu_fast_math_honor_division",
|
|
1451
|
+
b"xla_cpu_fast_math_honor_division",
|
|
1452
|
+
"xla_cpu_fast_math_honor_functions",
|
|
1453
|
+
b"xla_cpu_fast_math_honor_functions",
|
|
1454
|
+
"xla_cpu_fast_math_honor_infs",
|
|
1455
|
+
b"xla_cpu_fast_math_honor_infs",
|
|
1456
|
+
"xla_cpu_fast_math_honor_nans",
|
|
1457
|
+
b"xla_cpu_fast_math_honor_nans",
|
|
1458
|
+
"xla_cpu_matmul_tiling_k_dim",
|
|
1459
|
+
b"xla_cpu_matmul_tiling_k_dim",
|
|
1460
|
+
"xla_cpu_matmul_tiling_m_dim",
|
|
1461
|
+
b"xla_cpu_matmul_tiling_m_dim",
|
|
1462
|
+
"xla_cpu_matmul_tiling_n_dim",
|
|
1463
|
+
b"xla_cpu_matmul_tiling_n_dim",
|
|
1464
|
+
"xla_cpu_multi_thread_eigen",
|
|
1465
|
+
b"xla_cpu_multi_thread_eigen",
|
|
1466
|
+
"xla_cpu_parallel_codegen_split_count",
|
|
1467
|
+
b"xla_cpu_parallel_codegen_split_count",
|
|
1468
|
+
"xla_cpu_prefer_vector_width",
|
|
1469
|
+
b"xla_cpu_prefer_vector_width",
|
|
1470
|
+
"xla_cpu_strict_dot_conv_math",
|
|
1471
|
+
b"xla_cpu_strict_dot_conv_math",
|
|
1472
|
+
"xla_cpu_use_acl",
|
|
1473
|
+
b"xla_cpu_use_acl",
|
|
1474
|
+
"xla_cpu_use_mkl_dnn",
|
|
1475
|
+
b"xla_cpu_use_mkl_dnn",
|
|
1476
|
+
"xla_cpu_use_thunk_runtime",
|
|
1477
|
+
b"xla_cpu_use_thunk_runtime",
|
|
1478
|
+
"xla_debug_buffer_assignment_show_max",
|
|
1479
|
+
b"xla_debug_buffer_assignment_show_max",
|
|
1480
|
+
"xla_detailed_logging",
|
|
1481
|
+
b"xla_detailed_logging",
|
|
1482
|
+
"xla_disable_all_hlo_passes",
|
|
1483
|
+
b"xla_disable_all_hlo_passes",
|
|
1484
|
+
"xla_disable_hlo_passes",
|
|
1485
|
+
b"xla_disable_hlo_passes",
|
|
1486
|
+
"xla_dump_compress_protos",
|
|
1487
|
+
b"xla_dump_compress_protos",
|
|
1488
|
+
"xla_dump_disable_metadata",
|
|
1489
|
+
b"xla_dump_disable_metadata",
|
|
1490
|
+
"xla_dump_enable_mlir_pretty_form",
|
|
1491
|
+
b"xla_dump_enable_mlir_pretty_form",
|
|
1492
|
+
"xla_dump_fusion_visualization",
|
|
1493
|
+
b"xla_dump_fusion_visualization",
|
|
1494
|
+
"xla_dump_hlo_as_dot",
|
|
1495
|
+
b"xla_dump_hlo_as_dot",
|
|
1496
|
+
"xla_dump_hlo_as_html",
|
|
1497
|
+
b"xla_dump_hlo_as_html",
|
|
1498
|
+
"xla_dump_hlo_as_long_text",
|
|
1499
|
+
b"xla_dump_hlo_as_long_text",
|
|
1500
|
+
"xla_dump_hlo_as_proto",
|
|
1501
|
+
b"xla_dump_hlo_as_proto",
|
|
1502
|
+
"xla_dump_hlo_as_text",
|
|
1503
|
+
b"xla_dump_hlo_as_text",
|
|
1504
|
+
"xla_dump_hlo_as_url",
|
|
1505
|
+
b"xla_dump_hlo_as_url",
|
|
1506
|
+
"xla_dump_hlo_module_re",
|
|
1507
|
+
b"xla_dump_hlo_module_re",
|
|
1508
|
+
"xla_dump_hlo_pass_re",
|
|
1509
|
+
b"xla_dump_hlo_pass_re",
|
|
1510
|
+
"xla_dump_hlo_pipeline_re",
|
|
1511
|
+
b"xla_dump_hlo_pipeline_re",
|
|
1512
|
+
"xla_dump_hlo_snapshots",
|
|
1513
|
+
b"xla_dump_hlo_snapshots",
|
|
1514
|
+
"xla_dump_include_timestamp",
|
|
1515
|
+
b"xla_dump_include_timestamp",
|
|
1516
|
+
"xla_dump_large_constants",
|
|
1517
|
+
b"xla_dump_large_constants",
|
|
1518
|
+
"xla_dump_latency_hiding_schedule",
|
|
1519
|
+
b"xla_dump_latency_hiding_schedule",
|
|
1520
|
+
"xla_dump_max_hlo_modules",
|
|
1521
|
+
b"xla_dump_max_hlo_modules",
|
|
1522
|
+
"xla_dump_module_metadata",
|
|
1523
|
+
b"xla_dump_module_metadata",
|
|
1524
|
+
"xla_dump_to",
|
|
1525
|
+
b"xla_dump_to",
|
|
1526
|
+
"xla_eliminate_hlo_implicit_broadcast",
|
|
1527
|
+
b"xla_eliminate_hlo_implicit_broadcast",
|
|
1528
|
+
"xla_embed_ir_in_executable",
|
|
1529
|
+
b"xla_embed_ir_in_executable",
|
|
1530
|
+
"xla_enable_command_buffers_during_profiling",
|
|
1531
|
+
b"xla_enable_command_buffers_during_profiling",
|
|
1532
|
+
"xla_enable_dumping",
|
|
1533
|
+
b"xla_enable_dumping",
|
|
1534
|
+
"xla_enable_hlo_passes_only",
|
|
1535
|
+
b"xla_enable_hlo_passes_only",
|
|
1536
|
+
"xla_experimental_ignore_channel_id",
|
|
1537
|
+
b"xla_experimental_ignore_channel_id",
|
|
1538
|
+
"xla_force_host_platform_device_count",
|
|
1539
|
+
b"xla_force_host_platform_device_count",
|
|
1540
|
+
"xla_gpu_algorithm_denylist_path",
|
|
1541
|
+
b"xla_gpu_algorithm_denylist_path",
|
|
1542
|
+
"xla_gpu_all_gather_combine_threshold_bytes",
|
|
1543
|
+
b"xla_gpu_all_gather_combine_threshold_bytes",
|
|
1544
|
+
"xla_gpu_all_reduce_blueconnect_num_devices_per_host",
|
|
1545
|
+
b"xla_gpu_all_reduce_blueconnect_num_devices_per_host",
|
|
1546
|
+
"xla_gpu_all_reduce_combine_threshold_bytes",
|
|
1547
|
+
b"xla_gpu_all_reduce_combine_threshold_bytes",
|
|
1548
|
+
"xla_gpu_asm_extra_flags",
|
|
1549
|
+
b"xla_gpu_asm_extra_flags",
|
|
1550
|
+
"xla_gpu_async_dot",
|
|
1551
|
+
b"xla_gpu_async_dot",
|
|
1552
|
+
"xla_gpu_auto_spmd_partitioning_memory_budget_gb",
|
|
1553
|
+
b"xla_gpu_auto_spmd_partitioning_memory_budget_gb",
|
|
1554
|
+
"xla_gpu_auto_spmd_partitioning_memory_budget_ratio",
|
|
1555
|
+
b"xla_gpu_auto_spmd_partitioning_memory_budget_ratio",
|
|
1556
|
+
"xla_gpu_autotune_gemm_rtol",
|
|
1557
|
+
b"xla_gpu_autotune_gemm_rtol",
|
|
1558
|
+
"xla_gpu_autotune_level",
|
|
1559
|
+
b"xla_gpu_autotune_level",
|
|
1560
|
+
"xla_gpu_autotune_max_solutions",
|
|
1561
|
+
b"xla_gpu_autotune_max_solutions",
|
|
1562
|
+
"xla_gpu_collect_cost_model_stats",
|
|
1563
|
+
b"xla_gpu_collect_cost_model_stats",
|
|
1564
|
+
"xla_gpu_collective_inflation_factor",
|
|
1565
|
+
b"xla_gpu_collective_inflation_factor",
|
|
1566
|
+
"xla_gpu_collective_permute_decomposer_threshold",
|
|
1567
|
+
b"xla_gpu_collective_permute_decomposer_threshold",
|
|
1568
|
+
"xla_gpu_copy_insertion_use_region_analysis",
|
|
1569
|
+
b"xla_gpu_copy_insertion_use_region_analysis",
|
|
1570
|
+
"xla_gpu_crash_on_verification_failures",
|
|
1571
|
+
b"xla_gpu_crash_on_verification_failures",
|
|
1572
|
+
"xla_gpu_cublas_fallback",
|
|
1573
|
+
b"xla_gpu_cublas_fallback",
|
|
1574
|
+
"xla_gpu_cuda_data_dir",
|
|
1575
|
+
b"xla_gpu_cuda_data_dir",
|
|
1576
|
+
"xla_gpu_cudnn_gemm_fusion_level",
|
|
1577
|
+
b"xla_gpu_cudnn_gemm_fusion_level",
|
|
1578
|
+
"xla_gpu_cudnn_gemm_max_plans",
|
|
1579
|
+
b"xla_gpu_cudnn_gemm_max_plans",
|
|
1580
|
+
"xla_gpu_deterministic_ops",
|
|
1581
|
+
b"xla_gpu_deterministic_ops",
|
|
1582
|
+
"xla_gpu_disable_async_collectives",
|
|
1583
|
+
b"xla_gpu_disable_async_collectives",
|
|
1584
|
+
"xla_gpu_disable_gpuasm_optimizations",
|
|
1585
|
+
b"xla_gpu_disable_gpuasm_optimizations",
|
|
1586
|
+
"xla_gpu_dump_autotune_logs_to",
|
|
1587
|
+
b"xla_gpu_dump_autotune_logs_to",
|
|
1588
|
+
"xla_gpu_dump_autotune_results_to",
|
|
1589
|
+
b"xla_gpu_dump_autotune_results_to",
|
|
1590
|
+
"xla_gpu_dump_autotuned_gemm_fusions",
|
|
1591
|
+
b"xla_gpu_dump_autotuned_gemm_fusions",
|
|
1592
|
+
"xla_gpu_dump_llvmir",
|
|
1593
|
+
b"xla_gpu_dump_llvmir",
|
|
1594
|
+
"xla_gpu_enable_all_gather_combine_by_dim",
|
|
1595
|
+
b"xla_gpu_enable_all_gather_combine_by_dim",
|
|
1596
|
+
"xla_gpu_enable_analytical_latency_estimator",
|
|
1597
|
+
b"xla_gpu_enable_analytical_latency_estimator",
|
|
1598
|
+
"xla_gpu_enable_approx_costly_collectives",
|
|
1599
|
+
b"xla_gpu_enable_approx_costly_collectives",
|
|
1600
|
+
"xla_gpu_enable_bf16_3way_gemm",
|
|
1601
|
+
b"xla_gpu_enable_bf16_3way_gemm",
|
|
1602
|
+
"xla_gpu_enable_bf16_6way_gemm",
|
|
1603
|
+
b"xla_gpu_enable_bf16_6way_gemm",
|
|
1604
|
+
"xla_gpu_enable_command_buffer",
|
|
1605
|
+
b"xla_gpu_enable_command_buffer",
|
|
1606
|
+
"xla_gpu_enable_cub_radix_sort",
|
|
1607
|
+
b"xla_gpu_enable_cub_radix_sort",
|
|
1608
|
+
"xla_gpu_enable_cublaslt",
|
|
1609
|
+
b"xla_gpu_enable_cublaslt",
|
|
1610
|
+
"xla_gpu_enable_cudnn_fmha",
|
|
1611
|
+
b"xla_gpu_enable_cudnn_fmha",
|
|
1612
|
+
"xla_gpu_enable_cudnn_frontend",
|
|
1613
|
+
b"xla_gpu_enable_cudnn_frontend",
|
|
1614
|
+
"xla_gpu_enable_cudnn_int8x32_convolution_reordering",
|
|
1615
|
+
b"xla_gpu_enable_cudnn_int8x32_convolution_reordering",
|
|
1616
|
+
"xla_gpu_enable_cudnn_layer_norm",
|
|
1617
|
+
b"xla_gpu_enable_cudnn_layer_norm",
|
|
1618
|
+
"xla_gpu_enable_custom_fusions",
|
|
1619
|
+
b"xla_gpu_enable_custom_fusions",
|
|
1620
|
+
"xla_gpu_enable_custom_fusions_re",
|
|
1621
|
+
b"xla_gpu_enable_custom_fusions_re",
|
|
1622
|
+
"xla_gpu_enable_dot_strength_reduction",
|
|
1623
|
+
b"xla_gpu_enable_dot_strength_reduction",
|
|
1624
|
+
"xla_gpu_enable_dynamic_slice_fusion",
|
|
1625
|
+
b"xla_gpu_enable_dynamic_slice_fusion",
|
|
1626
|
+
"xla_gpu_enable_fast_min_max",
|
|
1627
|
+
b"xla_gpu_enable_fast_min_max",
|
|
1628
|
+
"xla_gpu_enable_highest_priority_async_stream",
|
|
1629
|
+
b"xla_gpu_enable_highest_priority_async_stream",
|
|
1630
|
+
"xla_gpu_enable_host_memory_offloading",
|
|
1631
|
+
b"xla_gpu_enable_host_memory_offloading",
|
|
1632
|
+
"xla_gpu_enable_latency_hiding_scheduler",
|
|
1633
|
+
b"xla_gpu_enable_latency_hiding_scheduler",
|
|
1634
|
+
"xla_gpu_enable_libnvjitlink",
|
|
1635
|
+
b"xla_gpu_enable_libnvjitlink",
|
|
1636
|
+
"xla_gpu_enable_libnvptxcompiler",
|
|
1637
|
+
b"xla_gpu_enable_libnvptxcompiler",
|
|
1638
|
+
"xla_gpu_enable_llvm_module_compilation_parallelism",
|
|
1639
|
+
b"xla_gpu_enable_llvm_module_compilation_parallelism",
|
|
1640
|
+
"xla_gpu_enable_nccl_clique_optimization",
|
|
1641
|
+
b"xla_gpu_enable_nccl_clique_optimization",
|
|
1642
|
+
"xla_gpu_enable_nccl_comm_splitting",
|
|
1643
|
+
b"xla_gpu_enable_nccl_comm_splitting",
|
|
1644
|
+
"xla_gpu_enable_nccl_per_stream_comms",
|
|
1645
|
+
b"xla_gpu_enable_nccl_per_stream_comms",
|
|
1646
|
+
"xla_gpu_enable_nccl_user_buffers",
|
|
1647
|
+
b"xla_gpu_enable_nccl_user_buffers",
|
|
1648
|
+
"xla_gpu_enable_pgle_accuracy_checker",
|
|
1649
|
+
b"xla_gpu_enable_pgle_accuracy_checker",
|
|
1650
|
+
"xla_gpu_enable_pipelined_all_gather",
|
|
1651
|
+
b"xla_gpu_enable_pipelined_all_gather",
|
|
1652
|
+
"xla_gpu_enable_pipelined_all_reduce",
|
|
1653
|
+
b"xla_gpu_enable_pipelined_all_reduce",
|
|
1654
|
+
"xla_gpu_enable_pipelined_collectives",
|
|
1655
|
+
b"xla_gpu_enable_pipelined_collectives",
|
|
1656
|
+
"xla_gpu_enable_pipelined_p2p",
|
|
1657
|
+
b"xla_gpu_enable_pipelined_p2p",
|
|
1658
|
+
"xla_gpu_enable_pipelined_reduce_scatter",
|
|
1659
|
+
b"xla_gpu_enable_pipelined_reduce_scatter",
|
|
1660
|
+
"xla_gpu_enable_priority_fusion",
|
|
1661
|
+
b"xla_gpu_enable_priority_fusion",
|
|
1662
|
+
"xla_gpu_enable_reassociation_for_converted_ar",
|
|
1663
|
+
b"xla_gpu_enable_reassociation_for_converted_ar",
|
|
1664
|
+
"xla_gpu_enable_reduce_scatter_combine_by_dim",
|
|
1665
|
+
b"xla_gpu_enable_reduce_scatter_combine_by_dim",
|
|
1666
|
+
"xla_gpu_enable_reduction_epilogue_fusion",
|
|
1667
|
+
b"xla_gpu_enable_reduction_epilogue_fusion",
|
|
1668
|
+
"xla_gpu_enable_shared_constants",
|
|
1669
|
+
b"xla_gpu_enable_shared_constants",
|
|
1670
|
+
"xla_gpu_enable_split_k_autotuning",
|
|
1671
|
+
b"xla_gpu_enable_split_k_autotuning",
|
|
1672
|
+
"xla_gpu_enable_triton_gemm",
|
|
1673
|
+
b"xla_gpu_enable_triton_gemm",
|
|
1674
|
+
"xla_gpu_enable_triton_gemm_int4",
|
|
1675
|
+
b"xla_gpu_enable_triton_gemm_int4",
|
|
1676
|
+
"xla_gpu_enable_triton_hopper",
|
|
1677
|
+
b"xla_gpu_enable_triton_hopper",
|
|
1678
|
+
"xla_gpu_enable_while_loop_double_buffering",
|
|
1679
|
+
b"xla_gpu_enable_while_loop_double_buffering",
|
|
1680
|
+
"xla_gpu_enable_while_loop_reduce_scatter_code_motion",
|
|
1681
|
+
b"xla_gpu_enable_while_loop_reduce_scatter_code_motion",
|
|
1682
|
+
"xla_gpu_enable_while_loop_unrolling",
|
|
1683
|
+
b"xla_gpu_enable_while_loop_unrolling",
|
|
1684
|
+
"xla_gpu_ensure_minor_dot_contraction_dims",
|
|
1685
|
+
b"xla_gpu_ensure_minor_dot_contraction_dims",
|
|
1686
|
+
"xla_gpu_exclude_nondeterministic_ops",
|
|
1687
|
+
b"xla_gpu_exclude_nondeterministic_ops",
|
|
1688
|
+
"xla_gpu_executable_terminate_timeout_seconds",
|
|
1689
|
+
b"xla_gpu_executable_terminate_timeout_seconds",
|
|
1690
|
+
"xla_gpu_executable_warn_stuck_timeout_seconds",
|
|
1691
|
+
b"xla_gpu_executable_warn_stuck_timeout_seconds",
|
|
1692
|
+
"xla_gpu_exhaustive_tiling_search",
|
|
1693
|
+
b"xla_gpu_exhaustive_tiling_search",
|
|
1694
|
+
"xla_gpu_experimental_autotune_cache_mode",
|
|
1695
|
+
b"xla_gpu_experimental_autotune_cache_mode",
|
|
1696
|
+
"xla_gpu_experimental_disable_binary_libraries",
|
|
1697
|
+
b"xla_gpu_experimental_disable_binary_libraries",
|
|
1698
|
+
"xla_gpu_experimental_enable_triton_softmax_priority_fusion",
|
|
1699
|
+
b"xla_gpu_experimental_enable_triton_softmax_priority_fusion",
|
|
1700
|
+
"xla_gpu_filter_kernels_spilling_registers_on_autotuning",
|
|
1701
|
+
b"xla_gpu_filter_kernels_spilling_registers_on_autotuning",
|
|
1702
|
+
"xla_gpu_force_compilation_parallelism",
|
|
1703
|
+
b"xla_gpu_force_compilation_parallelism",
|
|
1704
|
+
"xla_gpu_force_conv_nchw",
|
|
1705
|
+
b"xla_gpu_force_conv_nchw",
|
|
1706
|
+
"xla_gpu_force_conv_nhwc",
|
|
1707
|
+
b"xla_gpu_force_conv_nhwc",
|
|
1708
|
+
"xla_gpu_ftz",
|
|
1709
|
+
b"xla_gpu_ftz",
|
|
1710
|
+
"xla_gpu_fused_attention_use_cudnn_rng",
|
|
1711
|
+
b"xla_gpu_fused_attention_use_cudnn_rng",
|
|
1712
|
+
"xla_gpu_gemm_rewrite_size_threshold",
|
|
1713
|
+
b"xla_gpu_gemm_rewrite_size_threshold",
|
|
1714
|
+
"xla_gpu_graph_enable_concurrent_region",
|
|
1715
|
+
b"xla_gpu_graph_enable_concurrent_region",
|
|
1716
|
+
"xla_gpu_graph_min_graph_size",
|
|
1717
|
+
b"xla_gpu_graph_min_graph_size",
|
|
1718
|
+
"xla_gpu_kernel_cache_file",
|
|
1719
|
+
b"xla_gpu_kernel_cache_file",
|
|
1720
|
+
"xla_gpu_lhs_enable_gpu_async_tracker",
|
|
1721
|
+
b"xla_gpu_lhs_enable_gpu_async_tracker",
|
|
1722
|
+
"xla_gpu_llvm_ir_file",
|
|
1723
|
+
b"xla_gpu_llvm_ir_file",
|
|
1724
|
+
"xla_gpu_llvm_verification_level",
|
|
1725
|
+
b"xla_gpu_llvm_verification_level",
|
|
1726
|
+
"xla_gpu_load_autotune_results_from",
|
|
1727
|
+
b"xla_gpu_load_autotune_results_from",
|
|
1728
|
+
"xla_gpu_memory_limit_slop_factor",
|
|
1729
|
+
b"xla_gpu_memory_limit_slop_factor",
|
|
1730
|
+
"xla_gpu_mlir_emitter_level",
|
|
1731
|
+
b"xla_gpu_mlir_emitter_level",
|
|
1732
|
+
"xla_gpu_mock_custom_calls",
|
|
1733
|
+
b"xla_gpu_mock_custom_calls",
|
|
1734
|
+
"xla_gpu_multi_streamed_windowed_einsum",
|
|
1735
|
+
b"xla_gpu_multi_streamed_windowed_einsum",
|
|
1736
|
+
"xla_gpu_nccl_collective_max_nchannels",
|
|
1737
|
+
b"xla_gpu_nccl_collective_max_nchannels",
|
|
1738
|
+
"xla_gpu_nccl_p2p_max_nchannels",
|
|
1739
|
+
b"xla_gpu_nccl_p2p_max_nchannels",
|
|
1740
|
+
"xla_gpu_nccl_terminate_on_error",
|
|
1741
|
+
b"xla_gpu_nccl_terminate_on_error",
|
|
1742
|
+
"xla_gpu_nccl_termination_timeout_seconds",
|
|
1743
|
+
b"xla_gpu_nccl_termination_timeout_seconds",
|
|
1744
|
+
"xla_gpu_override_gemm_autotuner",
|
|
1745
|
+
b"xla_gpu_override_gemm_autotuner",
|
|
1746
|
+
"xla_gpu_per_fusion_autotune_cache_dir",
|
|
1747
|
+
b"xla_gpu_per_fusion_autotune_cache_dir",
|
|
1748
|
+
"xla_gpu_pgle_profile_file_or_directory_path",
|
|
1749
|
+
b"xla_gpu_pgle_profile_file_or_directory_path",
|
|
1750
|
+
"xla_gpu_ptx_file",
|
|
1751
|
+
b"xla_gpu_ptx_file",
|
|
1752
|
+
"xla_gpu_reduce_scatter_combine_threshold_bytes",
|
|
1753
|
+
b"xla_gpu_reduce_scatter_combine_threshold_bytes",
|
|
1754
|
+
"xla_gpu_redzone_padding_bytes",
|
|
1755
|
+
b"xla_gpu_redzone_padding_bytes",
|
|
1756
|
+
"xla_gpu_redzone_scratch_max_megabytes",
|
|
1757
|
+
b"xla_gpu_redzone_scratch_max_megabytes",
|
|
1758
|
+
"xla_gpu_require_complete_aot_autotune_results",
|
|
1759
|
+
b"xla_gpu_require_complete_aot_autotune_results",
|
|
1760
|
+
"xla_gpu_run_post_layout_collective_pipeliner",
|
|
1761
|
+
b"xla_gpu_run_post_layout_collective_pipeliner",
|
|
1762
|
+
"xla_gpu_shape_checks",
|
|
1763
|
+
b"xla_gpu_shape_checks",
|
|
1764
|
+
"xla_gpu_shard_autotuning",
|
|
1765
|
+
b"xla_gpu_shard_autotuning",
|
|
1766
|
+
"xla_gpu_strict_conv_algorithm_picker",
|
|
1767
|
+
b"xla_gpu_strict_conv_algorithm_picker",
|
|
1768
|
+
"xla_gpu_target_config_filename",
|
|
1769
|
+
b"xla_gpu_target_config_filename",
|
|
1770
|
+
"xla_gpu_temp_buffer_use_separate_color",
|
|
1771
|
+
b"xla_gpu_temp_buffer_use_separate_color",
|
|
1772
|
+
"xla_gpu_threshold_for_windowed_einsum_mib",
|
|
1773
|
+
b"xla_gpu_threshold_for_windowed_einsum_mib",
|
|
1774
|
+
"xla_gpu_triton_fusion_level",
|
|
1775
|
+
b"xla_gpu_triton_fusion_level",
|
|
1776
|
+
"xla_gpu_triton_gemm_any",
|
|
1777
|
+
b"xla_gpu_triton_gemm_any",
|
|
1778
|
+
"xla_gpu_triton_gemm_disable_reduced_precision_reduction",
|
|
1779
|
+
b"xla_gpu_triton_gemm_disable_reduced_precision_reduction",
|
|
1780
|
+
"xla_gpu_unsafe_fallback_to_driver_on_ptxas_not_found",
|
|
1781
|
+
b"xla_gpu_unsafe_fallback_to_driver_on_ptxas_not_found",
|
|
1782
|
+
"xla_gpu_unsafe_pipelined_loop_annotator",
|
|
1783
|
+
b"xla_gpu_unsafe_pipelined_loop_annotator",
|
|
1784
|
+
"xla_gpu_unsupported_enable_triton_gemm",
|
|
1785
|
+
b"xla_gpu_unsupported_enable_triton_gemm",
|
|
1786
|
+
"xla_gpu_use_memcpy_local_p2p",
|
|
1787
|
+
b"xla_gpu_use_memcpy_local_p2p",
|
|
1788
|
+
"xla_gpu_use_runtime_fusion",
|
|
1789
|
+
b"xla_gpu_use_runtime_fusion",
|
|
1790
|
+
"xla_gpu_verify_triton_fusion_numerics",
|
|
1791
|
+
b"xla_gpu_verify_triton_fusion_numerics",
|
|
1792
|
+
"xla_hlo_evaluator_use_fast_path",
|
|
1793
|
+
b"xla_hlo_evaluator_use_fast_path",
|
|
1794
|
+
"xla_hlo_graph_addresses",
|
|
1795
|
+
b"xla_hlo_graph_addresses",
|
|
1796
|
+
"xla_hlo_graph_sharding_color",
|
|
1797
|
+
b"xla_hlo_graph_sharding_color",
|
|
1798
|
+
"xla_hlo_profile",
|
|
1799
|
+
b"xla_hlo_profile",
|
|
1800
|
+
"xla_llvm_disable_expensive_passes",
|
|
1801
|
+
b"xla_llvm_disable_expensive_passes",
|
|
1802
|
+
"xla_llvm_enable_alias_scope_metadata",
|
|
1803
|
+
b"xla_llvm_enable_alias_scope_metadata",
|
|
1804
|
+
"xla_llvm_enable_invariant_load_metadata",
|
|
1805
|
+
b"xla_llvm_enable_invariant_load_metadata",
|
|
1806
|
+
"xla_llvm_enable_noalias_metadata",
|
|
1807
|
+
b"xla_llvm_enable_noalias_metadata",
|
|
1808
|
+
"xla_llvm_force_inline_before_split",
|
|
1809
|
+
b"xla_llvm_force_inline_before_split",
|
|
1810
|
+
"xla_multiheap_size_constraint_per_heap",
|
|
1811
|
+
b"xla_multiheap_size_constraint_per_heap",
|
|
1812
|
+
"xla_partitioning_algorithm",
|
|
1813
|
+
b"xla_partitioning_algorithm",
|
|
1814
|
+
"xla_reduce_window_rewrite_base_length",
|
|
1815
|
+
b"xla_reduce_window_rewrite_base_length",
|
|
1816
|
+
"xla_step_marker_location",
|
|
1817
|
+
b"xla_step_marker_location",
|
|
1818
|
+
"xla_syntax_sugar_async_ops",
|
|
1819
|
+
b"xla_syntax_sugar_async_ops",
|
|
1820
|
+
"xla_test_all_input_layouts",
|
|
1821
|
+
b"xla_test_all_input_layouts",
|
|
1822
|
+
"xla_test_all_output_layouts",
|
|
1823
|
+
b"xla_test_all_output_layouts",
|
|
1824
|
+
"xla_tpu_detect_inf",
|
|
1825
|
+
b"xla_tpu_detect_inf",
|
|
1826
|
+
"xla_tpu_detect_nan",
|
|
1827
|
+
b"xla_tpu_detect_nan",
|
|
1828
|
+
],
|
|
1829
|
+
) -> None: ...
|
|
1830
|
+
|
|
1831
|
+
global___DebugOptions = DebugOptions
|
|
1832
|
+
|
|
1833
|
+
@typing.final
|
|
1834
|
+
class GpuCompilationEnvironment(google.protobuf.message.Message):
|
|
1835
|
+
"""Contains flags which affects the GPU compilation result.
|
|
1836
|
+
These flags are part of Debug Options as of now, and will be migrated to
|
|
1837
|
+
this proto.
|
|
1838
|
+
"""
|
|
1839
|
+
|
|
1840
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
1841
|
+
|
|
1842
|
+
DUMMY_FLAG_FIELD_NUMBER: builtins.int
|
|
1843
|
+
dummy_flag: builtins.int
|
|
1844
|
+
"""Temporary dummy flag is added to test the flow.
|
|
1845
|
+
To be removed when we add flags here.
|
|
1846
|
+
"""
|
|
1847
|
+
def __init__(self, *, dummy_flag: builtins.int | None = ...) -> None: ...
|
|
1848
|
+
def ClearField(self, field_name: typing.Literal["dummy_flag", b"dummy_flag"]) -> None: ...
|
|
1849
|
+
|
|
1850
|
+
global___GpuCompilationEnvironment = GpuCompilationEnvironment
|
|
1851
|
+
|
|
1852
|
+
@typing.final
|
|
1853
|
+
class ShardableValueUpdatePairProto(google.protobuf.message.Message):
|
|
1854
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
1855
|
+
|
|
1856
|
+
INPUT_PARAMETER_NUMBER_FIELD_NUMBER: builtins.int
|
|
1857
|
+
PARAMETER_SHAPE_INDEX_FIELD_NUMBER: builtins.int
|
|
1858
|
+
OUTPUT_SHAPE_INDEX_FIELD_NUMBER: builtins.int
|
|
1859
|
+
input_parameter_number: builtins.int
|
|
1860
|
+
@property
|
|
1861
|
+
def parameter_shape_index(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]: ...
|
|
1862
|
+
@property
|
|
1863
|
+
def output_shape_index(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]: ...
|
|
1864
|
+
def __init__(
|
|
1865
|
+
self,
|
|
1866
|
+
*,
|
|
1867
|
+
input_parameter_number: builtins.int | None = ...,
|
|
1868
|
+
parameter_shape_index: collections.abc.Iterable[builtins.int] | None = ...,
|
|
1869
|
+
output_shape_index: collections.abc.Iterable[builtins.int] | None = ...,
|
|
1870
|
+
) -> None: ...
|
|
1871
|
+
def ClearField(
|
|
1872
|
+
self,
|
|
1873
|
+
field_name: typing.Literal[
|
|
1874
|
+
"input_parameter_number",
|
|
1875
|
+
b"input_parameter_number",
|
|
1876
|
+
"output_shape_index",
|
|
1877
|
+
b"output_shape_index",
|
|
1878
|
+
"parameter_shape_index",
|
|
1879
|
+
b"parameter_shape_index",
|
|
1880
|
+
],
|
|
1881
|
+
) -> None: ...
|
|
1882
|
+
|
|
1883
|
+
global___ShardableValueUpdatePairProto = ShardableValueUpdatePairProto
|
|
1884
|
+
|
|
1885
|
+
@typing.final
|
|
1886
|
+
class ExecutionOptions(google.protobuf.message.Message):
|
|
1887
|
+
"""These settings control how XLA compiles and/or runs code. Not all settings
|
|
1888
|
+
will have an effect on every platform.
|
|
1889
|
+
|
|
1890
|
+
When adding new fields, keep in mind that boolean fields default to false.
|
|
1891
|
+
Next id: 25.
|
|
1892
|
+
"""
|
|
1893
|
+
|
|
1894
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
1895
|
+
|
|
1896
|
+
SHAPE_WITH_OUTPUT_LAYOUT_FIELD_NUMBER: builtins.int
|
|
1897
|
+
SEED_FIELD_NUMBER: builtins.int
|
|
1898
|
+
DEBUG_OPTIONS_FIELD_NUMBER: builtins.int
|
|
1899
|
+
DEVICE_HANDLES_FIELD_NUMBER: builtins.int
|
|
1900
|
+
NUM_REPLICAS_FIELD_NUMBER: builtins.int
|
|
1901
|
+
DEVICE_ASSIGNMENT_FIELD_NUMBER: builtins.int
|
|
1902
|
+
ALIAS_PASSTHROUGH_PARAMS_FIELD_NUMBER: builtins.int
|
|
1903
|
+
NUM_PARTITIONS_FIELD_NUMBER: builtins.int
|
|
1904
|
+
LAUNCH_ID_FIELD_NUMBER: builtins.int
|
|
1905
|
+
USE_SPMD_PARTITIONING_FIELD_NUMBER: builtins.int
|
|
1906
|
+
USE_AUTO_SPMD_PARTITIONING_FIELD_NUMBER: builtins.int
|
|
1907
|
+
AUTO_SPMD_PARTITIONING_MESH_SHAPE_FIELD_NUMBER: builtins.int
|
|
1908
|
+
AUTO_SPMD_PARTITIONING_MESH_IDS_FIELD_NUMBER: builtins.int
|
|
1909
|
+
DEDUPLICATE_HLO_FIELD_NUMBER: builtins.int
|
|
1910
|
+
ALLOW_SPMD_SHARDING_PROPAGATION_TO_PARAMETERS_FIELD_NUMBER: builtins.int
|
|
1911
|
+
ALLOW_SPMD_SHARDING_PROPAGATION_TO_OUTPUT_FIELD_NUMBER: builtins.int
|
|
1912
|
+
PARAM_REQUIRES_BROADCAST_VIA_COLLECTIVES_FIELD_NUMBER: builtins.int
|
|
1913
|
+
ALLOW_SEPARATE_SHARDING_PROGRAMS_FIELD_NUMBER: builtins.int
|
|
1914
|
+
SHARDABLE_VALUE_UPDATE_PAIRS_FIELD_NUMBER: builtins.int
|
|
1915
|
+
FDO_PROFILE_FIELD_NUMBER: builtins.int
|
|
1916
|
+
DEVICE_MEMORY_SIZE_FIELD_NUMBER: builtins.int
|
|
1917
|
+
USE_SHARDY_PARTITIONER_FIELD_NUMBER: builtins.int
|
|
1918
|
+
seed: builtins.int
|
|
1919
|
+
"""Used to seed random-number generators used in this computation. If this is
|
|
1920
|
+
0, we generate a seed ourselves.
|
|
1921
|
+
|
|
1922
|
+
TODO(b/32083678): Changing the seed unnecessarily forces a recompilation.
|
|
1923
|
+
"""
|
|
1924
|
+
num_replicas: builtins.int
|
|
1925
|
+
"""Number of replicas of the computation to run. If zero, uses the default
|
|
1926
|
+
number of replicas for the XLA service.
|
|
1927
|
+
"""
|
|
1928
|
+
alias_passthrough_params: builtins.bool
|
|
1929
|
+
"""Alias input and output buffers for parameters that are passed-through XLA
|
|
1930
|
+
modules without being changed.
|
|
1931
|
+
"""
|
|
1932
|
+
num_partitions: builtins.int
|
|
1933
|
+
"""Number of partitions of the computation to run (model parallelism).
|
|
1934
|
+
If zero, uses the default number of partitions for the XLA service.
|
|
1935
|
+
"""
|
|
1936
|
+
launch_id: builtins.int
|
|
1937
|
+
"""Used to identify a set of programs that should be launch together."""
|
|
1938
|
+
use_spmd_partitioning: builtins.bool
|
|
1939
|
+
"""Indicates whether to use SPMD (true) or MPMD (false) partitioning when
|
|
1940
|
+
num_partitions > 1 and XLA is requested to partition the input program.
|
|
1941
|
+
"""
|
|
1942
|
+
use_auto_spmd_partitioning: builtins.bool
|
|
1943
|
+
"""Whether to automatically generate XLA shardings for SPMD partitioner."""
|
|
1944
|
+
deduplicate_hlo: builtins.bool
|
|
1945
|
+
"""If set, deduplicate hlo into function calls to reduce binary size. Only
|
|
1946
|
+
works on TPU.
|
|
1947
|
+
"""
|
|
1948
|
+
allow_separate_sharding_programs: builtins.bool
|
|
1949
|
+
"""If enabled, the compiler may generate sharding and unsharding programs as
|
|
1950
|
+
separate HLO modules, and modify the main program's input and output to
|
|
1951
|
+
be sharded.
|
|
1952
|
+
"""
|
|
1953
|
+
fdo_profile: builtins.bytes
|
|
1954
|
+
"""Profiling data for feedback directed optimizations. Note that this is not
|
|
1955
|
+
the only way to feed FDO data into the compiler and individual backends
|
|
1956
|
+
may choose to get FDO data by other means.
|
|
1957
|
+
"""
|
|
1958
|
+
device_memory_size: builtins.int
|
|
1959
|
+
"""Amount of device memory available for the executable to use."""
|
|
1960
|
+
use_shardy_partitioner: builtins.bool
|
|
1961
|
+
"""Use Shardy, a new partitioner, to replace the existing
|
|
1962
|
+
ShardingPropagation and SpmdPartitioner. See go/xla-sdy-pipeline for
|
|
1963
|
+
details.
|
|
1964
|
+
"""
|
|
1965
|
+
@property
|
|
1966
|
+
def shape_with_output_layout(self) -> tensorflow.compiler.xla.xla_data_pb2.ShapeProto:
|
|
1967
|
+
"""This optional field's layout is used as a hint when storing the output of
|
|
1968
|
+
this computation. Subsequent transfers of this output array to the client
|
|
1969
|
+
may be faster when using this layout.
|
|
1970
|
+
|
|
1971
|
+
We use a Shape here to accommodate computations that return a tuple.
|
|
1972
|
+
"""
|
|
1973
|
+
|
|
1974
|
+
@property
|
|
1975
|
+
def debug_options(self) -> global___DebugOptions: ...
|
|
1976
|
+
@property
|
|
1977
|
+
def device_handles(
|
|
1978
|
+
self,
|
|
1979
|
+
) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[tensorflow.compiler.xla.xla_data_pb2.DeviceHandle]:
|
|
1980
|
+
"""This optional field specifies a particular set of devices to run the
|
|
1981
|
+
computation on. The computation will be partitioned across these devices.
|
|
1982
|
+
If not provided, the default device will be chosen.
|
|
1983
|
+
"""
|
|
1984
|
+
|
|
1985
|
+
@property
|
|
1986
|
+
def device_assignment(self) -> tensorflow.compiler.xla.xla_data_pb2.DeviceAssignmentProto:
|
|
1987
|
+
"""This optional field specifies the device assignment if known at compile
|
|
1988
|
+
time.
|
|
1989
|
+
"""
|
|
1990
|
+
|
|
1991
|
+
@property
|
|
1992
|
+
def auto_spmd_partitioning_mesh_shape(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]:
|
|
1993
|
+
"""Device mesh shape used to create the sharding search space when
|
|
1994
|
+
use_auto_spmd_partitioning=true.
|
|
1995
|
+
"""
|
|
1996
|
+
|
|
1997
|
+
@property
|
|
1998
|
+
def auto_spmd_partitioning_mesh_ids(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]:
|
|
1999
|
+
"""Device mesh ids compatible with the above mesh_shape used when
|
|
2000
|
+
use_auto_spmd_partitioning=true.
|
|
2001
|
+
"""
|
|
2002
|
+
|
|
2003
|
+
@property
|
|
2004
|
+
def allow_spmd_sharding_propagation_to_parameters(
|
|
2005
|
+
self,
|
|
2006
|
+
) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.bool]:
|
|
2007
|
+
"""Allows sharding propagation to propagate to the parameters. This changes
|
|
2008
|
+
the input shape of the computation (which is undesirable), but it can be
|
|
2009
|
+
used to allow to run partial compilation to determine what would be the
|
|
2010
|
+
input sharding of a computation if XLA would be allowed to propagate the
|
|
2011
|
+
sharding which can be used by higher level framework as a way to query
|
|
2012
|
+
intermediate sharding of operations when multiple computation would be
|
|
2013
|
+
chained and merged together.
|
|
2014
|
+
This is a vector of bool, because the user can control which parameters can
|
|
2015
|
+
have the sharding substituted. If only one boolean value is passed in the
|
|
2016
|
+
vector that is interpreted as the value to be applied for every parameter.
|
|
2017
|
+
"""
|
|
2018
|
+
|
|
2019
|
+
@property
|
|
2020
|
+
def allow_spmd_sharding_propagation_to_output(
|
|
2021
|
+
self,
|
|
2022
|
+
) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.bool]:
|
|
2023
|
+
"""Allows sharding propagation to propagate to the outputs. This changes the
|
|
2024
|
+
output shape of the computation (which is undesirable), but it can be used
|
|
2025
|
+
to allow to run partial compilation to determine what would be the output
|
|
2026
|
+
sharding of a computation if XLA would be allowed to propagate the sharding
|
|
2027
|
+
which can be used by higher level framework as a way to query intermediate
|
|
2028
|
+
sharding of operations when multiple computation would be chained and
|
|
2029
|
+
merged together.
|
|
2030
|
+
This is a vector of bool, because the user can control (if the output of
|
|
2031
|
+
the computation is a tuple) which elements of the tuple can have the
|
|
2032
|
+
sharding substituted and which don't. If only one boolean value is passed
|
|
2033
|
+
in the vector that's interpreted as the value to be applied for every
|
|
2034
|
+
single element of the output tuple. One value per element of the tuple
|
|
2035
|
+
means that each value is attached to one of the output elements.
|
|
2036
|
+
"""
|
|
2037
|
+
|
|
2038
|
+
@property
|
|
2039
|
+
def param_requires_broadcast_via_collectives(
|
|
2040
|
+
self,
|
|
2041
|
+
) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.bool]:
|
|
2042
|
+
"""Whether to broadcast args across all replicas. One entry per arg."""
|
|
2043
|
+
|
|
2044
|
+
@property
|
|
2045
|
+
def shardable_value_update_pairs(
|
|
2046
|
+
self,
|
|
2047
|
+
) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ShardableValueUpdatePairProto]:
|
|
2048
|
+
"""The list of input/output pairs in the main program that could be sharded."""
|
|
2049
|
+
|
|
2050
|
+
def __init__(
|
|
2051
|
+
self,
|
|
2052
|
+
*,
|
|
2053
|
+
shape_with_output_layout: tensorflow.compiler.xla.xla_data_pb2.ShapeProto | None = ...,
|
|
2054
|
+
seed: builtins.int | None = ...,
|
|
2055
|
+
debug_options: global___DebugOptions | None = ...,
|
|
2056
|
+
device_handles: collections.abc.Iterable[tensorflow.compiler.xla.xla_data_pb2.DeviceHandle] | None = ...,
|
|
2057
|
+
num_replicas: builtins.int | None = ...,
|
|
2058
|
+
device_assignment: tensorflow.compiler.xla.xla_data_pb2.DeviceAssignmentProto | None = ...,
|
|
2059
|
+
alias_passthrough_params: builtins.bool | None = ...,
|
|
2060
|
+
num_partitions: builtins.int | None = ...,
|
|
2061
|
+
launch_id: builtins.int | None = ...,
|
|
2062
|
+
use_spmd_partitioning: builtins.bool | None = ...,
|
|
2063
|
+
use_auto_spmd_partitioning: builtins.bool | None = ...,
|
|
2064
|
+
auto_spmd_partitioning_mesh_shape: collections.abc.Iterable[builtins.int] | None = ...,
|
|
2065
|
+
auto_spmd_partitioning_mesh_ids: collections.abc.Iterable[builtins.int] | None = ...,
|
|
2066
|
+
deduplicate_hlo: builtins.bool | None = ...,
|
|
2067
|
+
allow_spmd_sharding_propagation_to_parameters: collections.abc.Iterable[builtins.bool] | None = ...,
|
|
2068
|
+
allow_spmd_sharding_propagation_to_output: collections.abc.Iterable[builtins.bool] | None = ...,
|
|
2069
|
+
param_requires_broadcast_via_collectives: collections.abc.Iterable[builtins.bool] | None = ...,
|
|
2070
|
+
allow_separate_sharding_programs: builtins.bool | None = ...,
|
|
2071
|
+
shardable_value_update_pairs: collections.abc.Iterable[global___ShardableValueUpdatePairProto] | None = ...,
|
|
2072
|
+
fdo_profile: builtins.bytes | None = ...,
|
|
2073
|
+
device_memory_size: builtins.int | None = ...,
|
|
2074
|
+
use_shardy_partitioner: builtins.bool | None = ...,
|
|
2075
|
+
) -> None: ...
|
|
2076
|
+
def HasField(
|
|
2077
|
+
self,
|
|
2078
|
+
field_name: typing.Literal[
|
|
2079
|
+
"debug_options",
|
|
2080
|
+
b"debug_options",
|
|
2081
|
+
"device_assignment",
|
|
2082
|
+
b"device_assignment",
|
|
2083
|
+
"shape_with_output_layout",
|
|
2084
|
+
b"shape_with_output_layout",
|
|
2085
|
+
],
|
|
2086
|
+
) -> builtins.bool: ...
|
|
2087
|
+
def ClearField(
|
|
2088
|
+
self,
|
|
2089
|
+
field_name: typing.Literal[
|
|
2090
|
+
"alias_passthrough_params",
|
|
2091
|
+
b"alias_passthrough_params",
|
|
2092
|
+
"allow_separate_sharding_programs",
|
|
2093
|
+
b"allow_separate_sharding_programs",
|
|
2094
|
+
"allow_spmd_sharding_propagation_to_output",
|
|
2095
|
+
b"allow_spmd_sharding_propagation_to_output",
|
|
2096
|
+
"allow_spmd_sharding_propagation_to_parameters",
|
|
2097
|
+
b"allow_spmd_sharding_propagation_to_parameters",
|
|
2098
|
+
"auto_spmd_partitioning_mesh_ids",
|
|
2099
|
+
b"auto_spmd_partitioning_mesh_ids",
|
|
2100
|
+
"auto_spmd_partitioning_mesh_shape",
|
|
2101
|
+
b"auto_spmd_partitioning_mesh_shape",
|
|
2102
|
+
"debug_options",
|
|
2103
|
+
b"debug_options",
|
|
2104
|
+
"deduplicate_hlo",
|
|
2105
|
+
b"deduplicate_hlo",
|
|
2106
|
+
"device_assignment",
|
|
2107
|
+
b"device_assignment",
|
|
2108
|
+
"device_handles",
|
|
2109
|
+
b"device_handles",
|
|
2110
|
+
"device_memory_size",
|
|
2111
|
+
b"device_memory_size",
|
|
2112
|
+
"fdo_profile",
|
|
2113
|
+
b"fdo_profile",
|
|
2114
|
+
"launch_id",
|
|
2115
|
+
b"launch_id",
|
|
2116
|
+
"num_partitions",
|
|
2117
|
+
b"num_partitions",
|
|
2118
|
+
"num_replicas",
|
|
2119
|
+
b"num_replicas",
|
|
2120
|
+
"param_requires_broadcast_via_collectives",
|
|
2121
|
+
b"param_requires_broadcast_via_collectives",
|
|
2122
|
+
"seed",
|
|
2123
|
+
b"seed",
|
|
2124
|
+
"shape_with_output_layout",
|
|
2125
|
+
b"shape_with_output_layout",
|
|
2126
|
+
"shardable_value_update_pairs",
|
|
2127
|
+
b"shardable_value_update_pairs",
|
|
2128
|
+
"use_auto_spmd_partitioning",
|
|
2129
|
+
b"use_auto_spmd_partitioning",
|
|
2130
|
+
"use_shardy_partitioner",
|
|
2131
|
+
b"use_shardy_partitioner",
|
|
2132
|
+
"use_spmd_partitioning",
|
|
2133
|
+
b"use_spmd_partitioning",
|
|
2134
|
+
],
|
|
2135
|
+
) -> None: ...
|
|
2136
|
+
|
|
2137
|
+
global___ExecutionOptions = ExecutionOptions
|
|
2138
|
+
|
|
2139
|
+
@typing.final
|
|
2140
|
+
class HloModuleConfigProto(google.protobuf.message.Message):
|
|
2141
|
+
"""Serialization of HloModuleConfig. See the C++ class definition for
|
|
2142
|
+
descriptions of each field.
|
|
2143
|
+
There are no guarantees of backwards or forwards compatibility.
|
|
2144
|
+
Next id: 36.
|
|
2145
|
+
"""
|
|
2146
|
+
|
|
2147
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
2148
|
+
|
|
2149
|
+
class _FusionConfigCollection:
|
|
2150
|
+
ValueType = typing.NewType("ValueType", builtins.int)
|
|
2151
|
+
V: typing_extensions.TypeAlias = ValueType
|
|
2152
|
+
|
|
2153
|
+
class _FusionConfigCollectionEnumTypeWrapper(
|
|
2154
|
+
google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[HloModuleConfigProto._FusionConfigCollection.ValueType],
|
|
2155
|
+
builtins.type,
|
|
2156
|
+
):
|
|
2157
|
+
DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor
|
|
2158
|
+
OFF: HloModuleConfigProto._FusionConfigCollection.ValueType # 0
|
|
2159
|
+
"""Do not collect configuration."""
|
|
2160
|
+
PER_EDGE: HloModuleConfigProto._FusionConfigCollection.ValueType # 1
|
|
2161
|
+
"""Collect per-edge configuration."""
|
|
2162
|
+
PER_NODE: HloModuleConfigProto._FusionConfigCollection.ValueType # 2
|
|
2163
|
+
"""Collect per-node configuration."""
|
|
2164
|
+
|
|
2165
|
+
class FusionConfigCollection(_FusionConfigCollection, metaclass=_FusionConfigCollectionEnumTypeWrapper): ...
|
|
2166
|
+
OFF: HloModuleConfigProto.FusionConfigCollection.ValueType # 0
|
|
2167
|
+
"""Do not collect configuration."""
|
|
2168
|
+
PER_EDGE: HloModuleConfigProto.FusionConfigCollection.ValueType # 1
|
|
2169
|
+
"""Collect per-edge configuration."""
|
|
2170
|
+
PER_NODE: HloModuleConfigProto.FusionConfigCollection.ValueType # 2
|
|
2171
|
+
"""Collect per-node configuration."""
|
|
2172
|
+
|
|
2173
|
+
@typing.final
|
|
2174
|
+
class BoolList(google.protobuf.message.Message):
|
|
2175
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
2176
|
+
|
|
2177
|
+
VALS_FIELD_NUMBER: builtins.int
|
|
2178
|
+
@property
|
|
2179
|
+
def vals(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.bool]: ...
|
|
2180
|
+
def __init__(self, *, vals: collections.abc.Iterable[builtins.bool] | None = ...) -> None: ...
|
|
2181
|
+
def ClearField(self, field_name: typing.Literal["vals", b"vals"]) -> None: ...
|
|
2182
|
+
|
|
2183
|
+
@typing.final
|
|
2184
|
+
class Int64List(google.protobuf.message.Message):
|
|
2185
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
2186
|
+
|
|
2187
|
+
VALS_FIELD_NUMBER: builtins.int
|
|
2188
|
+
@property
|
|
2189
|
+
def vals(self) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]: ...
|
|
2190
|
+
def __init__(self, *, vals: collections.abc.Iterable[builtins.int] | None = ...) -> None: ...
|
|
2191
|
+
def ClearField(self, field_name: typing.Literal["vals", b"vals"]) -> None: ...
|
|
2192
|
+
|
|
2193
|
+
@typing.final
|
|
2194
|
+
class Int64ListList(google.protobuf.message.Message):
|
|
2195
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
2196
|
+
|
|
2197
|
+
LISTS_FIELD_NUMBER: builtins.int
|
|
2198
|
+
@property
|
|
2199
|
+
def lists(
|
|
2200
|
+
self,
|
|
2201
|
+
) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___HloModuleConfigProto.Int64List]: ...
|
|
2202
|
+
def __init__(self, *, lists: collections.abc.Iterable[global___HloModuleConfigProto.Int64List] | None = ...) -> None: ...
|
|
2203
|
+
def ClearField(self, field_name: typing.Literal["lists", b"lists"]) -> None: ...
|
|
2204
|
+
|
|
2205
|
+
@typing.final
|
|
2206
|
+
class DotConfigEntry(google.protobuf.message.Message):
|
|
2207
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
2208
|
+
|
|
2209
|
+
KEY_FIELD_NUMBER: builtins.int
|
|
2210
|
+
VALUE_FIELD_NUMBER: builtins.int
|
|
2211
|
+
key: builtins.str
|
|
2212
|
+
@property
|
|
2213
|
+
def value(self) -> global___HloModuleConfigProto.Int64List: ...
|
|
2214
|
+
def __init__(
|
|
2215
|
+
self, *, key: builtins.str | None = ..., value: global___HloModuleConfigProto.Int64List | None = ...
|
|
2216
|
+
) -> None: ...
|
|
2217
|
+
def HasField(self, field_name: typing.Literal["value", b"value"]) -> builtins.bool: ...
|
|
2218
|
+
def ClearField(self, field_name: typing.Literal["key", b"key", "value", b"value"]) -> None: ...
|
|
2219
|
+
|
|
2220
|
+
@typing.final
|
|
2221
|
+
class AnalysisAllowanceMapEntry(google.protobuf.message.Message):
|
|
2222
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
2223
|
+
|
|
2224
|
+
KEY_FIELD_NUMBER: builtins.int
|
|
2225
|
+
VALUE_FIELD_NUMBER: builtins.int
|
|
2226
|
+
key: builtins.str
|
|
2227
|
+
value: builtins.int
|
|
2228
|
+
def __init__(self, *, key: builtins.str | None = ..., value: builtins.int | None = ...) -> None: ...
|
|
2229
|
+
def ClearField(self, field_name: typing.Literal["key", b"key", "value", b"value"]) -> None: ...
|
|
2230
|
+
|
|
2231
|
+
ENTRY_COMPUTATION_LAYOUT_FIELD_NUMBER: builtins.int
|
|
2232
|
+
SEED_FIELD_NUMBER: builtins.int
|
|
2233
|
+
LAUNCH_ID_FIELD_NUMBER: builtins.int
|
|
2234
|
+
REPLICA_COUNT_FIELD_NUMBER: builtins.int
|
|
2235
|
+
NUM_PARTITIONS_FIELD_NUMBER: builtins.int
|
|
2236
|
+
PARAM_REQUIRES_BROADCAST_VIA_COLLECTIVES_FIELD_NUMBER: builtins.int
|
|
2237
|
+
USE_SPMD_PARTITIONING_FIELD_NUMBER: builtins.int
|
|
2238
|
+
USE_AUTO_SPMD_PARTITIONING_FIELD_NUMBER: builtins.int
|
|
2239
|
+
AUTO_SPMD_PARTITIONING_MESH_SHAPE_FIELD_NUMBER: builtins.int
|
|
2240
|
+
AUTO_SPMD_PARTITIONING_MESH_IDS_FIELD_NUMBER: builtins.int
|
|
2241
|
+
DEDUPLICATE_HLO_FIELD_NUMBER: builtins.int
|
|
2242
|
+
INTRA_OP_PARALLELISM_THREADS_FIELD_NUMBER: builtins.int
|
|
2243
|
+
DEVICE_TYPE_FIELD_NUMBER: builtins.int
|
|
2244
|
+
DEBUG_OPTIONS_FIELD_NUMBER: builtins.int
|
|
2245
|
+
STATIC_DEVICE_ASSIGNMENT_FIELD_NUMBER: builtins.int
|
|
2246
|
+
PRE_SIMULATION_DEVICE_ASSIGNMENT_FIELD_NUMBER: builtins.int
|
|
2247
|
+
ALLOW_SEPARATE_SHARDING_PROGRAMS_FIELD_NUMBER: builtins.int
|
|
2248
|
+
SHARDABLE_VALUE_UPDATE_PAIRS_FIELD_NUMBER: builtins.int
|
|
2249
|
+
ALIAS_PASSTHROUGH_PARAMS_FIELD_NUMBER: builtins.int
|
|
2250
|
+
CONTENT_AWARE_COMPUTATION_SORTING_FIELD_NUMBER: builtins.int
|
|
2251
|
+
FUSION_CONFIG_COLLECTION_FIELD_NUMBER: builtins.int
|
|
2252
|
+
FUSION_CONFIG_FIELD_NUMBER: builtins.int
|
|
2253
|
+
DOT_CONFIG_FIELD_NUMBER: builtins.int
|
|
2254
|
+
LAYOUT_CONFIG_FIELD_NUMBER: builtins.int
|
|
2255
|
+
MEMORY_SPACE_ASSIGNMENT_CONFIG_FIELD_NUMBER: builtins.int
|
|
2256
|
+
PHASE_ORDERING_CONFIG_FIELD_NUMBER: builtins.int
|
|
2257
|
+
PHASE_INDEX_FIELD_NUMBER: builtins.int
|
|
2258
|
+
ALLOW_SPMD_SHARDING_PROPAGATION_TO_PARAMETERS_FIELD_NUMBER: builtins.int
|
|
2259
|
+
ALLOW_SPMD_SHARDING_PROPAGATION_TO_OUTPUT_FIELD_NUMBER: builtins.int
|
|
2260
|
+
ANALYSIS_ALLOWANCE_MAP_FIELD_NUMBER: builtins.int
|
|
2261
|
+
MATRIX_UNIT_OPERAND_PRECISION_FIELD_NUMBER: builtins.int
|
|
2262
|
+
FDO_PROFILE_FIELD_NUMBER: builtins.int
|
|
2263
|
+
DEVICE_MEMORY_SIZE_FIELD_NUMBER: builtins.int
|
|
2264
|
+
USE_SHARDY_PARTITIONER_FIELD_NUMBER: builtins.int
|
|
2265
|
+
seed: builtins.int
|
|
2266
|
+
launch_id: builtins.int
|
|
2267
|
+
replica_count: builtins.int
|
|
2268
|
+
num_partitions: builtins.int
|
|
2269
|
+
use_spmd_partitioning: builtins.bool
|
|
2270
|
+
use_auto_spmd_partitioning: builtins.bool
|
|
2271
|
+
deduplicate_hlo: builtins.bool
|
|
2272
|
+
intra_op_parallelism_threads: builtins.int
|
|
2273
|
+
device_type: builtins.str
|
|
2274
|
+
allow_separate_sharding_programs: builtins.bool
|
|
2275
|
+
alias_passthrough_params: builtins.bool
|
|
2276
|
+
content_aware_computation_sorting: builtins.bool
|
|
2277
|
+
fusion_config_collection: global___HloModuleConfigProto.FusionConfigCollection.ValueType
|
|
2278
|
+
phase_index: builtins.int
|
|
2279
|
+
matrix_unit_operand_precision: tensorflow.compiler.xla.xla_data_pb2.PrecisionConfig.Precision.ValueType
|
|
2280
|
+
fdo_profile: builtins.bytes
|
|
2281
|
+
device_memory_size: builtins.int
|
|
2282
|
+
use_shardy_partitioner: builtins.bool
|
|
2283
|
+
@property
|
|
2284
|
+
def entry_computation_layout(self) -> tensorflow.compiler.xla.xla_data_pb2.ProgramShapeProto: ...
|
|
2285
|
+
@property
|
|
2286
|
+
def param_requires_broadcast_via_collectives(
|
|
2287
|
+
self,
|
|
2288
|
+
) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.bool]: ...
|
|
2289
|
+
@property
|
|
2290
|
+
def auto_spmd_partitioning_mesh_shape(
|
|
2291
|
+
self,
|
|
2292
|
+
) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]: ...
|
|
2293
|
+
@property
|
|
2294
|
+
def auto_spmd_partitioning_mesh_ids(
|
|
2295
|
+
self,
|
|
2296
|
+
) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]: ...
|
|
2297
|
+
@property
|
|
2298
|
+
def debug_options(self) -> global___DebugOptions: ...
|
|
2299
|
+
@property
|
|
2300
|
+
def static_device_assignment(self) -> tensorflow.compiler.xla.xla_data_pb2.DeviceAssignmentProto: ...
|
|
2301
|
+
@property
|
|
2302
|
+
def pre_simulation_device_assignment(self) -> tensorflow.compiler.xla.xla_data_pb2.DeviceAssignmentProto:
|
|
2303
|
+
"""The original device assignment before being changed by a simulator.
|
|
2304
|
+
Simulators, like HybridSim, may change the device assignment to a smaller
|
|
2305
|
+
topology, to make simulation easier.
|
|
2306
|
+
"""
|
|
2307
|
+
|
|
2308
|
+
@property
|
|
2309
|
+
def shardable_value_update_pairs(
|
|
2310
|
+
self,
|
|
2311
|
+
) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ShardableValueUpdatePairProto]: ...
|
|
2312
|
+
@property
|
|
2313
|
+
def fusion_config(
|
|
2314
|
+
self,
|
|
2315
|
+
) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___HloModuleConfigProto.BoolList]: ...
|
|
2316
|
+
@property
|
|
2317
|
+
def dot_config(
|
|
2318
|
+
self,
|
|
2319
|
+
) -> google.protobuf.internal.containers.MessageMap[builtins.str, global___HloModuleConfigProto.Int64List]: ...
|
|
2320
|
+
@property
|
|
2321
|
+
def layout_config(
|
|
2322
|
+
self,
|
|
2323
|
+
) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___HloModuleConfigProto.Int64ListList]: ...
|
|
2324
|
+
@property
|
|
2325
|
+
def memory_space_assignment_config(
|
|
2326
|
+
self,
|
|
2327
|
+
) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.int]: ...
|
|
2328
|
+
@property
|
|
2329
|
+
def phase_ordering_config(
|
|
2330
|
+
self,
|
|
2331
|
+
) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___HloModuleConfigProto.BoolList]: ...
|
|
2332
|
+
@property
|
|
2333
|
+
def allow_spmd_sharding_propagation_to_parameters(
|
|
2334
|
+
self,
|
|
2335
|
+
) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.bool]: ...
|
|
2336
|
+
@property
|
|
2337
|
+
def allow_spmd_sharding_propagation_to_output(
|
|
2338
|
+
self,
|
|
2339
|
+
) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.bool]: ...
|
|
2340
|
+
@property
|
|
2341
|
+
def analysis_allowance_map(self) -> google.protobuf.internal.containers.ScalarMap[builtins.str, builtins.int]: ...
|
|
2342
|
+
def __init__(
|
|
2343
|
+
self,
|
|
2344
|
+
*,
|
|
2345
|
+
entry_computation_layout: tensorflow.compiler.xla.xla_data_pb2.ProgramShapeProto | None = ...,
|
|
2346
|
+
seed: builtins.int | None = ...,
|
|
2347
|
+
launch_id: builtins.int | None = ...,
|
|
2348
|
+
replica_count: builtins.int | None = ...,
|
|
2349
|
+
num_partitions: builtins.int | None = ...,
|
|
2350
|
+
param_requires_broadcast_via_collectives: collections.abc.Iterable[builtins.bool] | None = ...,
|
|
2351
|
+
use_spmd_partitioning: builtins.bool | None = ...,
|
|
2352
|
+
use_auto_spmd_partitioning: builtins.bool | None = ...,
|
|
2353
|
+
auto_spmd_partitioning_mesh_shape: collections.abc.Iterable[builtins.int] | None = ...,
|
|
2354
|
+
auto_spmd_partitioning_mesh_ids: collections.abc.Iterable[builtins.int] | None = ...,
|
|
2355
|
+
deduplicate_hlo: builtins.bool | None = ...,
|
|
2356
|
+
intra_op_parallelism_threads: builtins.int | None = ...,
|
|
2357
|
+
device_type: builtins.str | None = ...,
|
|
2358
|
+
debug_options: global___DebugOptions | None = ...,
|
|
2359
|
+
static_device_assignment: tensorflow.compiler.xla.xla_data_pb2.DeviceAssignmentProto | None = ...,
|
|
2360
|
+
pre_simulation_device_assignment: tensorflow.compiler.xla.xla_data_pb2.DeviceAssignmentProto | None = ...,
|
|
2361
|
+
allow_separate_sharding_programs: builtins.bool | None = ...,
|
|
2362
|
+
shardable_value_update_pairs: collections.abc.Iterable[global___ShardableValueUpdatePairProto] | None = ...,
|
|
2363
|
+
alias_passthrough_params: builtins.bool | None = ...,
|
|
2364
|
+
content_aware_computation_sorting: builtins.bool | None = ...,
|
|
2365
|
+
fusion_config_collection: global___HloModuleConfigProto.FusionConfigCollection.ValueType | None = ...,
|
|
2366
|
+
fusion_config: collections.abc.Iterable[global___HloModuleConfigProto.BoolList] | None = ...,
|
|
2367
|
+
dot_config: collections.abc.Mapping[builtins.str, global___HloModuleConfigProto.Int64List] | None = ...,
|
|
2368
|
+
layout_config: collections.abc.Iterable[global___HloModuleConfigProto.Int64ListList] | None = ...,
|
|
2369
|
+
memory_space_assignment_config: collections.abc.Iterable[builtins.int] | None = ...,
|
|
2370
|
+
phase_ordering_config: collections.abc.Iterable[global___HloModuleConfigProto.BoolList] | None = ...,
|
|
2371
|
+
phase_index: builtins.int | None = ...,
|
|
2372
|
+
allow_spmd_sharding_propagation_to_parameters: collections.abc.Iterable[builtins.bool] | None = ...,
|
|
2373
|
+
allow_spmd_sharding_propagation_to_output: collections.abc.Iterable[builtins.bool] | None = ...,
|
|
2374
|
+
analysis_allowance_map: collections.abc.Mapping[builtins.str, builtins.int] | None = ...,
|
|
2375
|
+
matrix_unit_operand_precision: tensorflow.compiler.xla.xla_data_pb2.PrecisionConfig.Precision.ValueType | None = ...,
|
|
2376
|
+
fdo_profile: builtins.bytes | None = ...,
|
|
2377
|
+
device_memory_size: builtins.int | None = ...,
|
|
2378
|
+
use_shardy_partitioner: builtins.bool | None = ...,
|
|
2379
|
+
) -> None: ...
|
|
2380
|
+
def HasField(
|
|
2381
|
+
self,
|
|
2382
|
+
field_name: typing.Literal[
|
|
2383
|
+
"debug_options",
|
|
2384
|
+
b"debug_options",
|
|
2385
|
+
"entry_computation_layout",
|
|
2386
|
+
b"entry_computation_layout",
|
|
2387
|
+
"pre_simulation_device_assignment",
|
|
2388
|
+
b"pre_simulation_device_assignment",
|
|
2389
|
+
"static_device_assignment",
|
|
2390
|
+
b"static_device_assignment",
|
|
2391
|
+
],
|
|
2392
|
+
) -> builtins.bool: ...
|
|
2393
|
+
def ClearField(
|
|
2394
|
+
self,
|
|
2395
|
+
field_name: typing.Literal[
|
|
2396
|
+
"alias_passthrough_params",
|
|
2397
|
+
b"alias_passthrough_params",
|
|
2398
|
+
"allow_separate_sharding_programs",
|
|
2399
|
+
b"allow_separate_sharding_programs",
|
|
2400
|
+
"allow_spmd_sharding_propagation_to_output",
|
|
2401
|
+
b"allow_spmd_sharding_propagation_to_output",
|
|
2402
|
+
"allow_spmd_sharding_propagation_to_parameters",
|
|
2403
|
+
b"allow_spmd_sharding_propagation_to_parameters",
|
|
2404
|
+
"analysis_allowance_map",
|
|
2405
|
+
b"analysis_allowance_map",
|
|
2406
|
+
"auto_spmd_partitioning_mesh_ids",
|
|
2407
|
+
b"auto_spmd_partitioning_mesh_ids",
|
|
2408
|
+
"auto_spmd_partitioning_mesh_shape",
|
|
2409
|
+
b"auto_spmd_partitioning_mesh_shape",
|
|
2410
|
+
"content_aware_computation_sorting",
|
|
2411
|
+
b"content_aware_computation_sorting",
|
|
2412
|
+
"debug_options",
|
|
2413
|
+
b"debug_options",
|
|
2414
|
+
"deduplicate_hlo",
|
|
2415
|
+
b"deduplicate_hlo",
|
|
2416
|
+
"device_memory_size",
|
|
2417
|
+
b"device_memory_size",
|
|
2418
|
+
"device_type",
|
|
2419
|
+
b"device_type",
|
|
2420
|
+
"dot_config",
|
|
2421
|
+
b"dot_config",
|
|
2422
|
+
"entry_computation_layout",
|
|
2423
|
+
b"entry_computation_layout",
|
|
2424
|
+
"fdo_profile",
|
|
2425
|
+
b"fdo_profile",
|
|
2426
|
+
"fusion_config",
|
|
2427
|
+
b"fusion_config",
|
|
2428
|
+
"fusion_config_collection",
|
|
2429
|
+
b"fusion_config_collection",
|
|
2430
|
+
"intra_op_parallelism_threads",
|
|
2431
|
+
b"intra_op_parallelism_threads",
|
|
2432
|
+
"launch_id",
|
|
2433
|
+
b"launch_id",
|
|
2434
|
+
"layout_config",
|
|
2435
|
+
b"layout_config",
|
|
2436
|
+
"matrix_unit_operand_precision",
|
|
2437
|
+
b"matrix_unit_operand_precision",
|
|
2438
|
+
"memory_space_assignment_config",
|
|
2439
|
+
b"memory_space_assignment_config",
|
|
2440
|
+
"num_partitions",
|
|
2441
|
+
b"num_partitions",
|
|
2442
|
+
"param_requires_broadcast_via_collectives",
|
|
2443
|
+
b"param_requires_broadcast_via_collectives",
|
|
2444
|
+
"phase_index",
|
|
2445
|
+
b"phase_index",
|
|
2446
|
+
"phase_ordering_config",
|
|
2447
|
+
b"phase_ordering_config",
|
|
2448
|
+
"pre_simulation_device_assignment",
|
|
2449
|
+
b"pre_simulation_device_assignment",
|
|
2450
|
+
"replica_count",
|
|
2451
|
+
b"replica_count",
|
|
2452
|
+
"seed",
|
|
2453
|
+
b"seed",
|
|
2454
|
+
"shardable_value_update_pairs",
|
|
2455
|
+
b"shardable_value_update_pairs",
|
|
2456
|
+
"static_device_assignment",
|
|
2457
|
+
b"static_device_assignment",
|
|
2458
|
+
"use_auto_spmd_partitioning",
|
|
2459
|
+
b"use_auto_spmd_partitioning",
|
|
2460
|
+
"use_shardy_partitioner",
|
|
2461
|
+
b"use_shardy_partitioner",
|
|
2462
|
+
"use_spmd_partitioning",
|
|
2463
|
+
b"use_spmd_partitioning",
|
|
2464
|
+
],
|
|
2465
|
+
) -> None: ...
|
|
2466
|
+
|
|
2467
|
+
global___HloModuleConfigProto = HloModuleConfigProto
|
|
2468
|
+
|
|
2469
|
+
@typing.final
|
|
2470
|
+
class HloModuleProtoWithConfig(google.protobuf.message.Message):
|
|
2471
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
2472
|
+
|
|
2473
|
+
HLO_MODULE_FIELD_NUMBER: builtins.int
|
|
2474
|
+
CONFIG_FIELD_NUMBER: builtins.int
|
|
2475
|
+
@property
|
|
2476
|
+
def hlo_module(self) -> tensorflow.compiler.xla.service.hlo_pb2.HloModuleProto: ...
|
|
2477
|
+
@property
|
|
2478
|
+
def config(self) -> global___HloModuleConfigProto: ...
|
|
2479
|
+
def __init__(
|
|
2480
|
+
self,
|
|
2481
|
+
*,
|
|
2482
|
+
hlo_module: tensorflow.compiler.xla.service.hlo_pb2.HloModuleProto | None = ...,
|
|
2483
|
+
config: global___HloModuleConfigProto | None = ...,
|
|
2484
|
+
) -> None: ...
|
|
2485
|
+
def HasField(self, field_name: typing.Literal["config", b"config", "hlo_module", b"hlo_module"]) -> builtins.bool: ...
|
|
2486
|
+
def ClearField(self, field_name: typing.Literal["config", b"config", "hlo_module", b"hlo_module"]) -> None: ...
|
|
2487
|
+
|
|
2488
|
+
global___HloModuleProtoWithConfig = HloModuleProtoWithConfig
|
|
2489
|
+
|
|
2490
|
+
@typing.final
|
|
2491
|
+
class ScheduleProto(google.protobuf.message.Message):
|
|
2492
|
+
"""A trace estimated by the Latency Hiding Scheduler."""
|
|
2493
|
+
|
|
2494
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
2495
|
+
|
|
2496
|
+
@typing.final
|
|
2497
|
+
class Instruction(google.protobuf.message.Message):
|
|
2498
|
+
DESCRIPTOR: google.protobuf.descriptor.Descriptor
|
|
2499
|
+
|
|
2500
|
+
ID_FIELD_NUMBER: builtins.int
|
|
2501
|
+
START_TIMESTAMP_CYCLES_FIELD_NUMBER: builtins.int
|
|
2502
|
+
END_TIMESTAMP_CYCLES_FIELD_NUMBER: builtins.int
|
|
2503
|
+
id: builtins.int
|
|
2504
|
+
"""Instruction id (matches the id in HloInstructionProto)."""
|
|
2505
|
+
start_timestamp_cycles: builtins.float
|
|
2506
|
+
"""Start and end timestamps in cycles."""
|
|
2507
|
+
end_timestamp_cycles: builtins.float
|
|
2508
|
+
def __init__(
|
|
2509
|
+
self,
|
|
2510
|
+
*,
|
|
2511
|
+
id: builtins.int | None = ...,
|
|
2512
|
+
start_timestamp_cycles: builtins.float | None = ...,
|
|
2513
|
+
end_timestamp_cycles: builtins.float | None = ...,
|
|
2514
|
+
) -> None: ...
|
|
2515
|
+
def ClearField(
|
|
2516
|
+
self,
|
|
2517
|
+
field_name: typing.Literal[
|
|
2518
|
+
"end_timestamp_cycles", b"end_timestamp_cycles", "id", b"id", "start_timestamp_cycles", b"start_timestamp_cycles"
|
|
2519
|
+
],
|
|
2520
|
+
) -> None: ...
|
|
2521
|
+
|
|
2522
|
+
INSTRUCTIONS_FIELD_NUMBER: builtins.int
|
|
2523
|
+
COMPUTATION_ID_FIELD_NUMBER: builtins.int
|
|
2524
|
+
HLO_MODULE_FIELD_NUMBER: builtins.int
|
|
2525
|
+
CYCLES_PER_MICROSECOND_FIELD_NUMBER: builtins.int
|
|
2526
|
+
computation_id: builtins.int
|
|
2527
|
+
"""Computation id (matches the id in HloComputationProto)."""
|
|
2528
|
+
cycles_per_microsecond: builtins.int
|
|
2529
|
+
@property
|
|
2530
|
+
def instructions(
|
|
2531
|
+
self,
|
|
2532
|
+
) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___ScheduleProto.Instruction]: ...
|
|
2533
|
+
@property
|
|
2534
|
+
def hlo_module(self) -> tensorflow.compiler.xla.service.hlo_pb2.HloModuleProto: ...
|
|
2535
|
+
def __init__(
|
|
2536
|
+
self,
|
|
2537
|
+
*,
|
|
2538
|
+
instructions: collections.abc.Iterable[global___ScheduleProto.Instruction] | None = ...,
|
|
2539
|
+
computation_id: builtins.int | None = ...,
|
|
2540
|
+
hlo_module: tensorflow.compiler.xla.service.hlo_pb2.HloModuleProto | None = ...,
|
|
2541
|
+
cycles_per_microsecond: builtins.int | None = ...,
|
|
2542
|
+
) -> None: ...
|
|
2543
|
+
def HasField(self, field_name: typing.Literal["hlo_module", b"hlo_module"]) -> builtins.bool: ...
|
|
2544
|
+
def ClearField(
|
|
2545
|
+
self,
|
|
2546
|
+
field_name: typing.Literal[
|
|
2547
|
+
"computation_id",
|
|
2548
|
+
b"computation_id",
|
|
2549
|
+
"cycles_per_microsecond",
|
|
2550
|
+
b"cycles_per_microsecond",
|
|
2551
|
+
"hlo_module",
|
|
2552
|
+
b"hlo_module",
|
|
2553
|
+
"instructions",
|
|
2554
|
+
b"instructions",
|
|
2555
|
+
],
|
|
2556
|
+
) -> None: ...
|
|
2557
|
+
|
|
2558
|
+
global___ScheduleProto = ScheduleProto
|