tfds-nightly 4.9.9.dev202508170045__py3-none-any.whl → 4.9.9.dev202508190044__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -25,17 +25,18 @@ class LazyImportsTest(testing.TestCase, parameterized.TestCase):
25
25
  # The following deps are not in the test list because the datasets that
26
26
  # require them need to have their tests run in isolation:
27
27
  # * crepe (NSynth)
28
+ # * gcld3 (c4)
29
+ # * langdetect (c4)
28
30
  # * librosa (NSynth)
31
+ # * nltk (c4)
32
+ # * tldextract (c4)
29
33
  @parameterized.parameters(
30
34
  "bs4",
31
35
  "cv2",
32
- "gcld3",
33
36
  "gcsfs_store",
34
- "langdetect",
35
37
  "lxml",
36
38
  "matplotlib",
37
39
  "mwparserfromhell",
38
- "nltk",
39
40
  "os",
40
41
  "pandas",
41
42
  "pretty_midi",
@@ -44,7 +45,6 @@ class LazyImportsTest(testing.TestCase, parameterized.TestCase):
44
45
  "scipy",
45
46
  "skimage",
46
47
  "tifffile",
47
- "tldextract",
48
48
  "zarr",
49
49
  )
50
50
  def test_import(self, module_name):
@@ -29,9 +29,7 @@ class A:
29
29
  ```
30
30
 
31
31
  The functions exposed below are dummy decorators. This allows not having to load
32
- TensorFlow. The functions are monkey patched when needed in
33
- scripts/documentation/build_api_docs.py with actual TensorFlow documentation
34
- decorators.
32
+ TensorFlow.
35
33
  """
36
34
 
37
35
  from typing import Any, TypeVar
@@ -27,7 +27,7 @@ from tensorflow_datasets.core import deprecated
27
27
  from tensorflow_datasets.core import download
28
28
  from tensorflow_datasets.core import features
29
29
  from tensorflow_datasets.core import folder_dataset
30
- from tensorflow_datasets.core import transform # pylint: disable=unused-import
30
+ from tensorflow_datasets.core import transform
31
31
  from tensorflow_datasets.core import visualization
32
32
  from tensorflow_datasets.core.as_dataframe import as_dataframe
33
33
  from tensorflow_datasets.core.dataset_utils import as_numpy
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: tfds-nightly
3
- Version: 4.9.9.dev202508170045
3
+ Version: 4.9.9.dev202508190044
4
4
  Summary: tensorflow/datasets is a library of datasets ready to use with TensorFlow.
5
5
  Home-page: https://github.com/tensorflow/datasets
6
6
  Download-URL: https://github.com/tensorflow/datasets/tags
@@ -63,11 +63,6 @@ Requires-Dist: apache-beam<2.65.0; extra == "tests-all"
63
63
  Requires-Dist: gcsfs; extra == "tests-all"
64
64
  Requires-Dist: zarr<3.0.0; extra == "tests-all"
65
65
  Requires-Dist: apache-beam<2.65.0; extra == "tests-all"
66
- Requires-Dist: gcld3; extra == "tests-all"
67
- Requires-Dist: langdetect; extra == "tests-all"
68
- Requires-Dist: nltk==3.8.1; extra == "tests-all"
69
- Requires-Dist: tldextract; extra == "tests-all"
70
- Requires-Dist: apache-beam<2.65.0; extra == "tests-all"
71
66
  Requires-Dist: matplotlib; extra == "tests-all"
72
67
  Requires-Dist: Pillow; extra == "tests-all"
73
68
  Requires-Dist: pydub; extra == "tests-all"
@@ -132,10 +127,6 @@ Requires-Dist: pyyaml; extra == "huggingface"
132
127
  Requires-Dist: scipy; extra == "huggingface"
133
128
  Requires-Dist: gcsfs; extra == "huggingface"
134
129
  Requires-Dist: zarr<3.0.0; extra == "huggingface"
135
- Requires-Dist: gcld3; extra == "huggingface"
136
- Requires-Dist: langdetect; extra == "huggingface"
137
- Requires-Dist: nltk==3.8.1; extra == "huggingface"
138
- Requires-Dist: tldextract; extra == "huggingface"
139
130
  Requires-Dist: matplotlib; extra == "huggingface"
140
131
  Requires-Dist: Pillow; extra == "huggingface"
141
132
  Requires-Dist: pydub; extra == "huggingface"
@@ -182,6 +173,7 @@ Requires-Dist: gcld3; extra == "c4"
182
173
  Requires-Dist: langdetect; extra == "c4"
183
174
  Requires-Dist: nltk==3.8.1; extra == "c4"
184
175
  Requires-Dist: tldextract; extra == "c4"
176
+ Requires-Dist: tensorflow<2.20.0; extra == "c4"
185
177
  Provides-Extra: c4-wsrs
186
178
  Requires-Dist: apache-beam<2.65.0; extra == "c4-wsrs"
187
179
  Provides-Extra: cats-vs-dogs
@@ -210,9 +202,9 @@ Requires-Dist: pydub; extra == "librispeech"
210
202
  Provides-Extra: lsun
211
203
  Requires-Dist: tensorflow-io[tensorflow]; extra == "lsun"
212
204
  Provides-Extra: nsynth
213
- Requires-Dist: crepe>=0.0.11; extra == "nsynth"
205
+ Requires-Dist: crepe; extra == "nsynth"
214
206
  Requires-Dist: librosa; extra == "nsynth"
215
- Requires-Dist: scikit-learn==0.20.3; extra == "nsynth"
207
+ Requires-Dist: tensorflow<2.20.0; extra == "nsynth"
216
208
  Provides-Extra: ogbg-molpcba
217
209
  Requires-Dist: pandas; extra == "ogbg-molpcba"
218
210
  Requires-Dist: networkx; extra == "ogbg-molpcba"
@@ -4,7 +4,7 @@ tensorflow_datasets/conftest.py,sha256=95ocbFU_SvVgWHI7NVkH1wXDrb-OGZrd_ihiVVkpk
4
4
  tensorflow_datasets/import_public_api_test.py,sha256=NPjteyVJL2ZDPEznMWBpP6hHbCaWiaZPxOcgnyr-HqY,852
5
5
  tensorflow_datasets/import_test.py,sha256=6rtuyyIwF9QNPyu6c-s5t-aA0fKPy2R9ondIHJVOkhY,801
6
6
  tensorflow_datasets/import_without_tf_test.py,sha256=znenZUnnltG7Jh2-PhhMEl_APadgiz1qPXi5P3Z85xo,3459
7
- tensorflow_datasets/public_api.py,sha256=I2YpOEEw0uND1EujpL_Sa3TSgcz_hY_I4F_iTZe9A3c,3896
7
+ tensorflow_datasets/public_api.py,sha256=omCZpyf4uc9jgFTEvZ35ws4S-nDqMY1Yd_IyQwG3aCs,3863
8
8
  tensorflow_datasets/setup_teardown.py,sha256=4qTNXhVE9nvp6U77HEX7_OJ49_eCTFJLmp3suWGiM6g,1450
9
9
  tensorflow_datasets/typing.py,sha256=UThdRvwKbnXQkllNBhMCVHmCaIhBLHZtxQtfl0TE8e8,1397
10
10
  tensorflow_datasets/version.py,sha256=Yid9O3W4SuoFxeA25B4U4gQnr0lhvGeQKMY3tQ_F6No,1114
@@ -85,7 +85,7 @@ tensorflow_datasets/core/hashing_test.py,sha256=duT2nnr6jT7ienXvjWa7-s32Oy6saskJ
85
85
  tensorflow_datasets/core/lazy_builder_import.py,sha256=ASBXRekXOeQuaY3KrL1zvUYJ6g1yxHztc8EsNo0GK1s,2019
86
86
  tensorflow_datasets/core/lazy_builder_import_test.py,sha256=I8Z8UHjaGeKCct3mf_nC-kMq0IXqNxKcnPNLXmzzCrQ,2081
87
87
  tensorflow_datasets/core/lazy_imports_lib.py,sha256=Q-c3qGEZJDqviEQUiro2iBpMw7KA-YbWQ3VFIcJ7rK8,5961
88
- tensorflow_datasets/core/lazy_imports_lib_test.py,sha256=cbdamDUJIY5YORm6coyCMIralgsL_gCUfa2Dzdj6ZPY,1695
88
+ tensorflow_datasets/core/lazy_imports_lib_test.py,sha256=j4d2n3gkJZ-BMPJF9_C2Hptk0k_nTWZSZufKEaoP9KA,1703
89
89
  tensorflow_datasets/core/load.py,sha256=1FQVnKwn8OVS_IgDbs9XN7aIVxQnyfrS0pI2X9dh77M,37765
90
90
  tensorflow_datasets/core/load_test.py,sha256=EEa8GuSIrEbn0RcGrWS3hmmatKBqBA3QOQWpQ1WjVgA,6490
91
91
  tensorflow_datasets/core/naming.py,sha256=B_P77QDA4lkG2FUl4PrzZR0U6qqae_fLxruGBw3ZSVc,25614
@@ -247,7 +247,7 @@ tensorflow_datasets/core/utils/conversion_utils.py,sha256=V8kFmJu38op7-8ufZvEn0f
247
247
  tensorflow_datasets/core/utils/conversion_utils_test.py,sha256=rP_nbzQWzmZc_GXp3Y6TirwIGJqiQbF-JtY3B1tOuN0,5346
248
248
  tensorflow_datasets/core/utils/croissant_utils.py,sha256=9C8sScaEqSRsThqpQQc48GDNR1KFmDkS8hmKIvfZCB0,5181
249
249
  tensorflow_datasets/core/utils/croissant_utils_test.py,sha256=UdkAVYDTPm1L0zmMESScurV_IMA5K3qAKmL_umeMJZI,4497
250
- tensorflow_datasets/core/utils/docs.py,sha256=0AeGyhPXQc_DMsTX7ocbb8IuJv_RTrG7StYdfZiwSzU,1549
250
+ tensorflow_datasets/core/utils/docs.py,sha256=nRE4d8wxYZav8AcT3dkiY0yplAJBx1hygWxkeKj_V7I,1412
251
251
  tensorflow_datasets/core/utils/dtype_utils.py,sha256=LvDe1hbgQem57RiqXjG9U5Roj8-1KkBMmSYTtgctx2U,3246
252
252
  tensorflow_datasets/core/utils/dtype_utils_test.py,sha256=-Qe2fQzDO5sjS36ZL-dY9w0tNrJXokIoSRFEQCv5dQA,3259
253
253
  tensorflow_datasets/core/utils/error_utils.py,sha256=lnquUa_VGRjn7-G_5x-PvWGgnnO6GAWsi9I7xeVuGxQ,3204
@@ -2005,8 +2005,6 @@ tensorflow_datasets/scripts/deployment/copy_dataset_info_files.py,sha256=uLuvwOW
2005
2005
  tensorflow_datasets/scripts/deployment/export_community_datasets.py,sha256=h6IEu7Y19cl6adiT3ve5HUbKb5hUuysKdMqIgRc8CcM,3643
2006
2006
  tensorflow_datasets/scripts/deployment/export_community_datasets_test.py,sha256=eIE0dR9KGkXWX2kcanfN6rKdLYSsFb5QCVXfe7nd7gI,3234
2007
2007
  tensorflow_datasets/scripts/documentation/__init__.py,sha256=Z8UWkv0wbzS4AzaLgSpYVGApYv5j57RWY0vN5Z553BQ,613
2008
- tensorflow_datasets/scripts/documentation/build_api_docs.py,sha256=NXbW1ijbIUogxpWFDaSngw8_84TJuFv7C_Jr7hLxeyk,2893
2009
- tensorflow_datasets/scripts/documentation/build_api_docs_test.py,sha256=m8_XuS_05NYaErMJeCQoqdrvKUFw-JKBi7wdDeqI1jA,2065
2010
2008
  tensorflow_datasets/scripts/documentation/build_catalog.py,sha256=SYJoNW-VxvL8xx85uYlFBwbr1k64HcmRBfxsj9-sdYA,8680
2011
2009
  tensorflow_datasets/scripts/documentation/build_catalog_test.py,sha256=qjnqK6lhBh-uNrjLQkEs3AbKFBo5uz_sxhhdT4ibOyA,2532
2012
2010
  tensorflow_datasets/scripts/documentation/build_community_catalog.py,sha256=gh84xnKbL_ndR4GGbgBNLJ0nxjFwiAPLuhUvzeKPZAo,19902
@@ -2474,10 +2472,10 @@ tensorflow_datasets/vision_language/wit/wit_test.py,sha256=PXS8DMNW-MDrT2p5oy4Ic
2474
2472
  tensorflow_datasets/vision_language/wit_kaggle/__init__.py,sha256=vGwSGeM8WE4Q-l0-eEE1sBojmk6YT0l1OO60AWa4Q40,719
2475
2473
  tensorflow_datasets/vision_language/wit_kaggle/wit_kaggle.py,sha256=q-vX_FBzIwsFxL4sY9vuyQ3UQD2PLM4yhUR4U6l-qao,16903
2476
2474
  tensorflow_datasets/vision_language/wit_kaggle/wit_kaggle_test.py,sha256=ZymHT1NkmD-pUnh3BmM3_g30c5afsWYnmqDD9dVyDSA,1778
2477
- tfds_nightly-4.9.9.dev202508170045.dist-info/licenses/AUTHORS,sha256=nvBG4WwfgjuOu1oZkuQKw9kg7X6rve679ObS-YDDmXg,309
2478
- tfds_nightly-4.9.9.dev202508170045.dist-info/licenses/LICENSE,sha256=z8d0m5b2O9McPEK1xHG_dWgUBT6EfBDz6wA0F7xSPTA,11358
2479
- tfds_nightly-4.9.9.dev202508170045.dist-info/METADATA,sha256=YZ4_31y9BZJ-lAZP6FTq_Y2WCyDHQhs2pn-ieKTSyMA,11694
2480
- tfds_nightly-4.9.9.dev202508170045.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
2481
- tfds_nightly-4.9.9.dev202508170045.dist-info/entry_points.txt,sha256=eHEL7nF5y1uCY2FgkuYIdE062epJXlAQTSdq89px4p4,73
2482
- tfds_nightly-4.9.9.dev202508170045.dist-info/top_level.txt,sha256=bAevmk9209s_oxVZVlN6hSDIVS423qrMQvmcWSvW4do,20
2483
- tfds_nightly-4.9.9.dev202508170045.dist-info/RECORD,,
2475
+ tfds_nightly-4.9.9.dev202508190044.dist-info/licenses/AUTHORS,sha256=nvBG4WwfgjuOu1oZkuQKw9kg7X6rve679ObS-YDDmXg,309
2476
+ tfds_nightly-4.9.9.dev202508190044.dist-info/licenses/LICENSE,sha256=z8d0m5b2O9McPEK1xHG_dWgUBT6EfBDz6wA0F7xSPTA,11358
2477
+ tfds_nightly-4.9.9.dev202508190044.dist-info/METADATA,sha256=V2S5ib8SHZzto6RkHAkgCcIHBTX6PNAQ5SrbHA16P2Q,11291
2478
+ tfds_nightly-4.9.9.dev202508190044.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
2479
+ tfds_nightly-4.9.9.dev202508190044.dist-info/entry_points.txt,sha256=eHEL7nF5y1uCY2FgkuYIdE062epJXlAQTSdq89px4p4,73
2480
+ tfds_nightly-4.9.9.dev202508190044.dist-info/top_level.txt,sha256=bAevmk9209s_oxVZVlN6hSDIVS423qrMQvmcWSvW4do,20
2481
+ tfds_nightly-4.9.9.dev202508190044.dist-info/RECORD,,
@@ -1,91 +0,0 @@
1
- # coding=utf-8
2
- # Copyright 2025 The TensorFlow Datasets Authors.
3
- #
4
- # Licensed under the Apache License, Version 2.0 (the "License");
5
- # you may not use this file except in compliance with the License.
6
- # You may obtain a copy of the License at
7
- #
8
- # http://www.apache.org/licenses/LICENSE-2.0
9
- #
10
- # Unless required by applicable law or agreed to in writing, software
11
- # distributed under the License is distributed on an "AS IS" BASIS,
12
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13
- # See the License for the specific language governing permissions and
14
- # limitations under the License.
15
-
16
- """generates api_docs for tensorflow_datasets."""
17
- import os
18
-
19
- from absl import app
20
- from absl import flags
21
- from absl import logging
22
-
23
- import tensorflow_datasets as tfds
24
- # Testing is lazily imported, so we first force its import.
25
- from tensorflow_datasets.testing import * # pylint: disable=wildcard-import
26
- from tensorflow_docs.api_generator import generate_lib
27
-
28
- from tensorflow.tools.docs import doc_controls # pylint: disable=g-direct-tensorflow-import
29
-
30
- # Force the definition of all documentation decorators declared in
31
- # third_party/py/tensorflow_datasets/core/utils/docs.py to TensorFlow decorators
32
- try:
33
- tfds.core.utils.docs.deprecated = doc_controls.set_deprecated
34
- tfds.core.utils.docs.doc_private = doc_controls.doc_private
35
- tfds.core.utils.docs.do_not_doc = doc_controls.do_not_generate_docs
36
- # Same as `do_not_doc`, but also applied to children
37
- tfds.core.utils.docs.do_not_doc_inheritable = (
38
- doc_controls.do_not_doc_inheritable
39
- )
40
- # Document the parent, but not the children
41
- tfds.core.utils.docs.do_not_doc_in_subclasses = (
42
- doc_controls.do_not_doc_in_subclasses
43
- )
44
- except AttributeError:
45
- logging.info("Could not set TensorFlow documentation decorators.")
46
-
47
- FLAGS = flags.FLAGS
48
-
49
- flags.DEFINE_string(
50
- "output_dir", "/tmp/datasets_api", "Where to output the docs"
51
- )
52
- flags.DEFINE_string(
53
- "code_url_prefix",
54
- "https://github.com/tensorflow/datasets/tree/master/tensorflow_datasets/",
55
- "The url prefix for links to code.",
56
- )
57
-
58
- flags.DEFINE_bool(
59
- "search_hints", True, "Include metadata search hints in the generated files"
60
- )
61
-
62
- flags.DEFINE_string(
63
- "site_path", "datasets/api_docs/python", "Path prefix in the _toc.yaml"
64
- )
65
-
66
-
67
- def execute(output_dir, code_url_prefix, search_hints, site_path):
68
- """Builds API docs for tensorflow_datasets."""
69
- doc_generator = generate_lib.DocGenerator(
70
- root_title="TensorFlow Datasets",
71
- py_modules=[("tfds", tfds)],
72
- base_dir=os.path.dirname(tfds.__file__),
73
- search_hints=search_hints,
74
- code_url_prefix=code_url_prefix,
75
- site_path=site_path,
76
- )
77
-
78
- doc_generator.build(output_dir)
79
-
80
-
81
- def main(unused_argv):
82
- execute(
83
- FLAGS.output_dir,
84
- FLAGS.code_url_prefix,
85
- FLAGS.search_hints,
86
- FLAGS.site_path,
87
- )
88
-
89
-
90
- if __name__ == "__main__":
91
- app.run(main)
@@ -1,65 +0,0 @@
1
- # coding=utf-8
2
- # Copyright 2025 The TensorFlow Datasets Authors.
3
- #
4
- # Licensed under the Apache License, Version 2.0 (the "License");
5
- # you may not use this file except in compliance with the License.
6
- # You may obtain a copy of the License at
7
- #
8
- # http://www.apache.org/licenses/LICENSE-2.0
9
- #
10
- # Unless required by applicable law or agreed to in writing, software
11
- # distributed under the License is distributed on an "AS IS" BASIS,
12
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13
- # See the License for the specific language governing permissions and
14
- # limitations under the License.
15
-
16
- """Smoke Test for docs generation."""
17
-
18
- import os
19
- import shutil
20
- import tempfile
21
-
22
- from absl.testing import absltest
23
-
24
- from tensorflow_datasets.core.utils import docs
25
- from tensorflow_datasets.scripts.documentation import build_api_docs
26
-
27
-
28
- class BuildDocsTest(absltest.TestCase):
29
-
30
- def setUp(self):
31
- super(BuildDocsTest, self).setUp()
32
- self.workdir = tempfile.mkdtemp()
33
- if os.path.exists(self.workdir):
34
- shutil.rmtree(self.workdir)
35
- os.makedirs(self.workdir)
36
-
37
- def test_api_gen(self):
38
- build_api_docs.execute(
39
- output_dir=self.workdir,
40
- code_url_prefix="",
41
- search_hints=True,
42
- site_path="datasets/api_docs/python",
43
- )
44
-
45
- # Check that the "defined in" section is working
46
- with open(os.path.join(self.workdir, "tfds.md")) as f:
47
- content = f.read()
48
- self.assertIn("__init__.py", content)
49
-
50
- # Check that the `testing` folder is generated.
51
- with open(os.path.join(self.workdir, "tfds/testing.md")) as f:
52
- content = f.read()
53
- self.assertIn("__init__.py", content)
54
-
55
- def test_tensorflow_decorators_are_used_and_not_dummy_decorator(self):
56
- self.assertEqual(docs.deprecated.__name__, "set_deprecated")
57
- self.assertEqual(docs.doc_private.__name__, "doc_private")
58
- self.assertEqual(docs.do_not_doc.__name__, "do_not_generate_docs")
59
- self.assertEqual(
60
- docs.do_not_doc_inheritable.__name__, "do_not_doc_inheritable"
61
- )
62
-
63
-
64
- if __name__ == "__main__":
65
- absltest.main()