datachain 0.23.0__py3-none-any.whl → 0.24.0__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of datachain might be problematic. Click here for more details.

@@ -49,6 +49,7 @@ from datachain.error import (
49
49
  DatasetInvalidVersionError,
50
50
  DatasetNotFoundError,
51
51
  DatasetVersionNotFoundError,
52
+ NamespaceNotFoundError,
52
53
  ProjectNotFoundError,
53
54
  QueryScriptCancelError,
54
55
  QueryScriptRunError,
@@ -1107,21 +1108,26 @@ class Catalog:
1107
1108
  namespace_name: str,
1108
1109
  project_name: str,
1109
1110
  version: Optional[str] = None,
1111
+ pull_dataset: bool = False,
1112
+ update: bool = False,
1110
1113
  ) -> DatasetRecord:
1111
- try:
1112
- project = self.metastore.get_project(project_name, namespace_name)
1113
- ds = self.get_dataset(name, project)
1114
- if version and not ds.has_version(version):
1115
- raise DatasetVersionNotFoundError(
1116
- f"Dataset {name} does not have version {version}"
1117
- )
1118
- return ds
1114
+ if self.metastore.is_local_dataset(namespace_name) or not update:
1115
+ try:
1116
+ project = self.metastore.get_project(project_name, namespace_name)
1117
+ ds = self.get_dataset(name, project)
1118
+ if not version or ds.has_version(version):
1119
+ return ds
1120
+ except (NamespaceNotFoundError, ProjectNotFoundError, DatasetNotFoundError):
1121
+ pass
1122
+
1123
+ if self.metastore.is_local_dataset(namespace_name):
1124
+ raise DatasetNotFoundError(
1125
+ f"Dataset {name}"
1126
+ + (f" version {version} " if version else " ")
1127
+ + "not found"
1128
+ )
1119
1129
 
1120
- except (
1121
- ProjectNotFoundError,
1122
- DatasetNotFoundError,
1123
- DatasetVersionNotFoundError,
1124
- ):
1130
+ if pull_dataset:
1125
1131
  print("Dataset not found in local catalog, trying to get from studio")
1126
1132
  remote_ds_uri = create_dataset_uri(
1127
1133
  name, namespace_name, project_name, version
@@ -1136,6 +1142,8 @@ class Catalog:
1136
1142
  name, self.metastore.get_project(project_name, namespace_name)
1137
1143
  )
1138
1144
 
1145
+ return self.get_remote_dataset(namespace_name, project_name, name)
1146
+
1139
1147
  def get_dataset_with_version_uuid(self, uuid: str) -> DatasetRecord:
1140
1148
  """Returns dataset that contains version with specific uuid"""
1141
1149
  for dataset in self.ls_datasets():
@@ -1152,6 +1160,10 @@ class Catalog:
1152
1160
 
1153
1161
  info_response = studio_client.dataset_info(namespace, project, name)
1154
1162
  if not info_response.ok:
1163
+ if info_response.status == 404:
1164
+ raise DatasetNotFoundError(
1165
+ f"Dataset {namespace}.{project}.{name} not found"
1166
+ )
1155
1167
  raise DataChainError(info_response.message)
1156
1168
 
1157
1169
  dataset_info = info_response.data
datachain/dataset.py CHANGED
@@ -12,6 +12,9 @@ from typing import (
12
12
  )
13
13
  from urllib.parse import urlparse
14
14
 
15
+ from packaging.specifiers import SpecifierSet
16
+ from packaging.version import Version
17
+
15
18
  from datachain import semver
16
19
  from datachain.error import DatasetVersionNotFoundError, InvalidDatasetNameError
17
20
  from datachain.namespace import Namespace
@@ -661,13 +664,39 @@ class DatasetRecord:
661
664
  return None
662
665
  return max(versions).version
663
666
 
664
- @property
665
- def prev_version(self) -> Optional[str]:
666
- """Returns previous version of a dataset"""
667
- if len(self.versions) == 1:
667
+ def latest_compatible_version(self, version_spec: str) -> Optional[str]:
668
+ """
669
+ Returns the latest version that matches the given version specifier.
670
+
671
+ Supports Python version specifiers like:
672
+ - ">=1.0.0,<2.0.0" (compatible release range)
673
+ - "~=1.4.2" (compatible release clause)
674
+ - "==1.2.*" (prefix matching)
675
+ - ">1.0.0" (exclusive ordered comparison)
676
+ - ">=1.0.0" (inclusive ordered comparison)
677
+ - "!=1.3.0" (version exclusion)
678
+
679
+ Args:
680
+ version_spec: Version specifier string following PEP 440
681
+
682
+ Returns:
683
+ Latest compatible version string, or None if no compatible version found
684
+ """
685
+ spec_set = SpecifierSet(version_spec)
686
+
687
+ # Convert dataset versions to packaging.Version objects
688
+ # and filter compatible ones
689
+ compatible_versions = []
690
+ for v in self.versions:
691
+ pkg_version = Version(v.version)
692
+ if spec_set.contains(pkg_version):
693
+ compatible_versions.append(v)
694
+
695
+ if not compatible_versions:
668
696
  return None
669
697
 
670
- return sorted(self.versions)[-2].version
698
+ # Return the latest compatible version
699
+ return max(compatible_versions).version
671
700
 
672
701
  @classmethod
673
702
  def from_dict(cls, d: dict[str, Any]) -> "DatasetRecord":
@@ -7,9 +7,6 @@ from datachain.error import (
7
7
  ProjectNotFoundError,
8
8
  )
9
9
  from datachain.lib.dataset_info import DatasetInfo
10
- from datachain.lib.file import (
11
- File,
12
- )
13
10
  from datachain.lib.projects import get as get_project
14
11
  from datachain.lib.settings import Settings
15
12
  from datachain.lib.signal_schema import SignalSchema
@@ -34,7 +31,6 @@ def read_dataset(
34
31
  version: Optional[Union[str, int]] = None,
35
32
  session: Optional[Session] = None,
36
33
  settings: Optional[dict] = None,
37
- fallback_to_studio: bool = True,
38
34
  delta: Optional[bool] = False,
39
35
  delta_on: Optional[Union[str, Sequence[str]]] = (
40
36
  "file.path",
@@ -44,6 +40,7 @@ def read_dataset(
44
40
  delta_result_on: Optional[Union[str, Sequence[str]]] = None,
45
41
  delta_compare: Optional[Union[str, Sequence[str]]] = None,
46
42
  delta_retry: Optional[Union[bool, str]] = None,
43
+ update: bool = False,
47
44
  ) -> "DataChain":
48
45
  """Get data from a saved Dataset. It returns the chain itself.
49
46
  If dataset or version is not found locally, it will try to pull it from Studio.
@@ -55,11 +52,12 @@ def read_dataset(
55
52
  set; otherwise, default values will be applied.
56
53
  namespace : optional name of namespace in which dataset to read is created
57
54
  project : optional name of project in which dataset to read is created
58
- version : dataset version
55
+ version : dataset version. Supports:
56
+ - Exact version strings: "1.2.3"
57
+ - Legacy integer versions: 1, 2, 3 (finds latest major version)
58
+ - Version specifiers (PEP 440): ">=1.0.0,<2.0.0", "~=1.4.2", "==1.2.*", etc.
59
59
  session : Session to use for the chain.
60
60
  settings : Settings to use for the chain.
61
- fallback_to_studio : Try to pull dataset from Studio if not found locally.
62
- Default is True.
63
61
  delta: If True, only process new or changed files instead of reprocessing
64
62
  everything. This saves time by skipping files that were already processed in
65
63
  previous versions. The optimization is working when a new version of the
@@ -79,6 +77,10 @@ def read_dataset(
79
77
  (error mode)
80
78
  - True: Reprocess records missing from the result dataset (missing mode)
81
79
  - None: No retry processing (default)
80
+ update: If True always checks for newer versions available on Studio, even if
81
+ some version of the dataset exists locally already. If False (default), it
82
+ will only fetch the dataset from Studio if it is not found locally.
83
+
82
84
 
83
85
  Example:
84
86
  ```py
@@ -92,11 +94,22 @@ def read_dataset(
92
94
  ```
93
95
 
94
96
  ```py
95
- chain = dc.read_dataset("my_cats", fallback_to_studio=False)
97
+ chain = dc.read_dataset("my_cats", version="1.0.0")
96
98
  ```
97
99
 
98
100
  ```py
99
- chain = dc.read_dataset("my_cats", version="1.0.0")
101
+ # Using version specifiers (PEP 440)
102
+ chain = dc.read_dataset("my_cats", version=">=1.0.0,<2.0.0")
103
+ ```
104
+
105
+ ```py
106
+ # Legacy integer version support (finds latest in major version)
107
+ chain = dc.read_dataset("my_cats", version=1) # Latest 1.x.x version
108
+ ```
109
+
110
+ ```py
111
+ # Always check for newer versions matching a version specifier from Studio
112
+ chain = dc.read_dataset("my_cats", version=">=1.0.0", update=True)
100
113
  ```
101
114
 
102
115
  ```py
@@ -113,7 +126,6 @@ def read_dataset(
113
126
  version="1.0.0",
114
127
  session=session,
115
128
  settings=settings,
116
- fallback_to_studio=True,
117
129
  )
118
130
  ```
119
131
  """
@@ -121,6 +133,8 @@ def read_dataset(
121
133
 
122
134
  from .datachain import DataChain
123
135
 
136
+ telemetry.send_event_once("class", "datachain_init", name=name, version=version)
137
+
124
138
  session = Session.get(session)
125
139
  catalog = session.catalog
126
140
 
@@ -131,31 +145,37 @@ def read_dataset(
131
145
  )
132
146
 
133
147
  if version is not None:
148
+ dataset = session.catalog.get_dataset_with_remote_fallback(
149
+ name, namespace_name, project_name, update=update
150
+ )
151
+
152
+ # Convert legacy integer versions to version specifiers
153
+ # For backward compatibility we still allow users to put version as integer
154
+ # in which case we convert it to a version specifier that finds the latest
155
+ # version where major part is equal to that input version.
156
+ # For example if user sets version=2, we convert it to ">=2.0.0,<3.0.0"
157
+ # which will find something like 2.4.3 (assuming 2.4.3 is the biggest among
158
+ # all 2.* dataset versions)
159
+ if isinstance(version, int):
160
+ version_spec = f">={version}.0.0,<{version + 1}.0.0"
161
+ else:
162
+ version_spec = str(version)
163
+
164
+ from packaging.specifiers import InvalidSpecifier, SpecifierSet
165
+
134
166
  try:
135
- # for backward compatibility we still allow users to put version as integer
136
- # in which case we are trying to find latest version where major part is
137
- # equal to that input version. For example if user sets version=2, we could
138
- # continue with something like 2.4.3 (assuming 2.4.3 is the biggest among
139
- # all 2.* dataset versions). If dataset doesn't have any versions where
140
- # major part is equal to that input, exception is thrown.
141
- major = int(version)
142
- try:
143
- ds_project = get_project(project_name, namespace_name, session=session)
144
- except ProjectNotFoundError:
145
- raise DatasetNotFoundError(
146
- f"Dataset {name} not found in namespace {namespace_name} and",
147
- f" project {project_name}",
148
- ) from None
149
-
150
- dataset = session.catalog.get_dataset(name, ds_project)
151
- latest_major = dataset.latest_major_version(major)
152
- if not latest_major:
167
+ # Try to parse as version specifier
168
+ SpecifierSet(version_spec)
169
+ # If it's a valid specifier set, find the latest compatible version
170
+ latest_compatible = dataset.latest_compatible_version(version_spec)
171
+ if not latest_compatible:
153
172
  raise DatasetVersionNotFoundError(
154
- f"Dataset {name} does not have version {version}"
173
+ f"No dataset {name} version matching specifier {version_spec}"
155
174
  )
156
- version = latest_major
157
- except ValueError:
158
- # version is in new semver string format, continuing as normal
175
+ version = latest_compatible
176
+ except InvalidSpecifier:
177
+ # If not a valid specifier, treat as exact version string
178
+ # This handles cases like "1.2.3" which are exact versions, not specifiers
159
179
  pass
160
180
 
161
181
  if settings:
@@ -169,11 +189,8 @@ def read_dataset(
169
189
  namespace_name=namespace_name,
170
190
  version=version, # type: ignore[arg-type]
171
191
  session=session,
172
- indexing_column_types=File._datachain_column_types,
173
- fallback_to_studio=fallback_to_studio,
174
192
  )
175
193
 
176
- telemetry.send_event_once("class", "datachain_init", name=name, version=version)
177
194
  signals_schema = SignalSchema({"sys": Sys})
178
195
  if query.feature_schema:
179
196
  signals_schema |= SignalSchema.deserialize(query.feature_schema)
@@ -127,12 +127,8 @@ def read_listing_dataset(
127
127
  if version is None:
128
128
  version = dataset.latest_version
129
129
 
130
- query = DatasetQuery(
131
- name=name,
132
- session=session,
133
- indexing_column_types=File._datachain_column_types,
134
- fallback_to_studio=False,
135
- )
130
+ query = DatasetQuery(name=name, session=session)
131
+
136
132
  if settings:
137
133
  cfg = {**settings}
138
134
  if "prefetch" not in cfg:
datachain/lib/projects.py CHANGED
@@ -54,7 +54,7 @@ def get(name: str, namespace: str, session: Optional[Session]) -> Project:
54
54
  ```py
55
55
  import datachain as dc
56
56
  from datachain.lib.projects import get as get_project
57
- project = get_project("my-project", "local")
57
+ project = get_project("my-project", "local")
58
58
  ```
59
59
  """
60
60
  return Session.get(session).catalog.metastore.get_project(name, namespace)
@@ -1099,13 +1099,9 @@ class DatasetQuery:
1099
1099
  namespace_name: Optional[str] = None,
1100
1100
  catalog: Optional["Catalog"] = None,
1101
1101
  session: Optional[Session] = None,
1102
- indexing_column_types: Optional[dict[str, Any]] = None,
1103
1102
  in_memory: bool = False,
1104
- fallback_to_studio: bool = True,
1105
1103
  update: bool = False,
1106
1104
  ) -> None:
1107
- from datachain.remote.studio import is_token_set
1108
-
1109
1105
  self.session = Session.get(session, catalog=catalog, in_memory=in_memory)
1110
1106
  self.catalog = catalog or self.session.catalog
1111
1107
  self.steps: list[Step] = []
@@ -1137,18 +1133,16 @@ class DatasetQuery:
1137
1133
  # not setting query step yet as listing dataset might not exist at
1138
1134
  # this point
1139
1135
  self.list_ds_name = name
1140
- elif fallback_to_studio and is_token_set():
1136
+ else:
1141
1137
  self._set_starting_step(
1142
1138
  self.catalog.get_dataset_with_remote_fallback(
1143
1139
  name,
1144
1140
  namespace_name=namespace_name,
1145
1141
  project_name=project_name,
1146
1142
  version=version,
1143
+ pull_dataset=True,
1147
1144
  )
1148
1145
  )
1149
- else:
1150
- project = self.catalog.metastore.get_project(project_name, namespace_name)
1151
- self._set_starting_step(self.catalog.get_dataset(name, project=project))
1152
1146
 
1153
1147
  def _set_starting_step(self, ds: "DatasetRecord") -> None:
1154
1148
  if not self.version:
@@ -78,10 +78,11 @@ def _parse_dates(obj: dict, date_fields: list[str]):
78
78
 
79
79
 
80
80
  class Response(Generic[T]):
81
- def __init__(self, data: T, ok: bool, message: str) -> None:
81
+ def __init__(self, data: T, ok: bool, message: str, status: int) -> None:
82
82
  self.data = data
83
83
  self.ok = ok
84
84
  self.message = message
85
+ self.status = status
85
86
 
86
87
  def __repr__(self):
87
88
  return (
@@ -186,7 +187,7 @@ class StudioClient:
186
187
  message = "Indexing in progress"
187
188
  else:
188
189
  message = content.get("message", "")
189
- return Response(response_data, ok, message)
190
+ return Response(response_data, ok, message, response.status_code)
190
191
 
191
192
  @retry_with_backoff(retries=3, errors=(HTTPError, Timeout))
192
193
  def _send_request(
@@ -236,7 +237,7 @@ class StudioClient:
236
237
  else:
237
238
  message = ""
238
239
 
239
- return Response(data, ok, message)
240
+ return Response(data, ok, message, response.status_code)
240
241
 
241
242
  @staticmethod
242
243
  def _unpacker_hook(code, data):
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: datachain
3
- Version: 0.23.0
3
+ Version: 0.24.0
4
4
  Summary: Wrangle unstructured AI data at scale
5
5
  Author-email: Dmitry Petrov <support@dvc.org>
6
6
  License-Expression: Apache-2.0
@@ -3,7 +3,7 @@ datachain/__main__.py,sha256=hG3Y4ARGEqe1AWwNMd259rBlqtphx1Wk39YbueQ0yV8,91
3
3
  datachain/asyn.py,sha256=RH_jFwJcTXxhEFomaI9yL6S3Onau6NZ6FSKfKFGtrJE,9689
4
4
  datachain/cache.py,sha256=ESVRaCJXEThMIfGEFVHx6wJPOZA7FYk9V6WxjyuqUBY,3626
5
5
  datachain/config.py,sha256=g8qbNV0vW2VEKpX-dGZ9pAn0DAz6G2ZFcr7SAV3PoSM,4272
6
- datachain/dataset.py,sha256=Dn8zx_4zuiW28dF-cNtAXFLS7Her0015YYdLlFGwMas,24195
6
+ datachain/dataset.py,sha256=wDrukmkDnYP0X8bAGY-7O1NDE3DWCFqrH8VVDpXM9Ok,25263
7
7
  datachain/delta.py,sha256=4RqLLc9dJLF8x9GG9IDgi86DwuPerZQ4HAUnNBeACw8,8446
8
8
  datachain/error.py,sha256=OWwWMkzZYJrkcoEDGhJHMf7SfKvxcsOLRF94mjPf29I,1609
9
9
  datachain/job.py,sha256=x5PB6d5sqx00hePNNkirESlOVAvnmkEM5ygUgQmAhsk,1262
@@ -21,7 +21,7 @@ datachain/studio.py,sha256=bLok-eJNFRHQScEyAyA_Fas52dmijd5r-73KudWxV4k,13337
21
21
  datachain/telemetry.py,sha256=0A4IOPPp9VlP5pyW9eBfaTK3YhHGzHl7dQudQjUAx9A,994
22
22
  datachain/utils.py,sha256=DNqOi-Ydb7InyWvD9m7_yailxz6-YGpZzh00biQaHNo,15305
23
23
  datachain/catalog/__init__.py,sha256=cMZzSz3VoUi-6qXSVaHYN-agxQuAcz2XSqnEPZ55crE,353
24
- datachain/catalog/catalog.py,sha256=j8aH0zAHJRCJ_-uOYapK0spp9Go_Ehwxz_1FI3p-Q5I,64733
24
+ datachain/catalog/catalog.py,sha256=z4GbRMHeW0YA20Sjh7QuPy1Rj4RkX547WN9Pp5wAD6o,65277
25
25
  datachain/catalog/datasource.py,sha256=IkGMh0Ttg6Q-9DWfU_H05WUnZepbGa28HYleECi6K7I,1353
26
26
  datachain/catalog/loader.py,sha256=UXjYD6BNRoupPvkiz3-b04jepXhtLHCA4gzKFnXxOtQ,5987
27
27
  datachain/cli/__init__.py,sha256=WvBqnwjG8Wp9xGCn-4eqfoZ3n7Sj1HJemCi4MayJh_c,8221
@@ -82,7 +82,7 @@ datachain/lib/listing_info.py,sha256=9ua40Hw0aiQByUw3oAEeNzMavJYfW0Uhe8YdCTK-m_g
82
82
  datachain/lib/meta_formats.py,sha256=zdyg6XLk3QIsSk3I7s0Ez5kaCJSlE3uq7JiGxf7UwtU,6348
83
83
  datachain/lib/model_store.py,sha256=DNIv8Y6Jtk1_idNLzIpsThOsdW2BMAudyUCbPUcgcxk,2515
84
84
  datachain/lib/namespaces.py,sha256=it52UbbwB8dzhesO2pMs_nThXiPQ1Ph9sD9I3GQkg5s,2099
85
- datachain/lib/projects.py,sha256=C-HTzTLUbIB735_iBSV6MjWnntV6gaKCEIkMSR1YEQw,2596
85
+ datachain/lib/projects.py,sha256=8lN0qV8czX1LGtWURCUvRlSJk-RpO9w9Rra_pOZus6g,2595
86
86
  datachain/lib/pytorch.py,sha256=oBBd6cxYrcwaFz7IQajKqhGqDdNnwUZWs0wJPRizrjk,7712
87
87
  datachain/lib/settings.py,sha256=9wi0FoHxRxNiyn99pR28IYsMkoo47jQxeXuObQr2Ar0,2929
88
88
  datachain/lib/signal_schema.py,sha256=dVEqqrQQ_BS3yzU_49-Gari7IjVyMl1UT8h1WIsZabs,36489
@@ -104,10 +104,10 @@ datachain/lib/dc/__init__.py,sha256=HD0NYrdy44u6kkpvgGjJcvGz-UGTHui2azghcT8ZUg0,
104
104
  datachain/lib/dc/csv.py,sha256=q6a9BpapGwP6nwy6c5cklxQumep2fUp9l2LAjtTJr6s,4411
105
105
  datachain/lib/dc/database.py,sha256=g5M6NjYR1T0vKte-abV-3Ejnm-HqxTIMir5cRi_SziE,6051
106
106
  datachain/lib/dc/datachain.py,sha256=dFI7JX5-41HLgA-TUR99dtR1lvk2vokaMC3mbIW1XT4,85814
107
- datachain/lib/dc/datasets.py,sha256=H7s_4B68MMpRvLrK874QS8xSq1FeT1f6j7qPhnP2i64,12490
107
+ datachain/lib/dc/datasets.py,sha256=U4xqAfs6FdW8HIJjeayQaIg1dunaIsVXYGqfq_sDSv0,13274
108
108
  datachain/lib/dc/hf.py,sha256=PJl2wiLjdRsMz0SYbLT-6H8b-D5i2WjeH7li8HHOk_0,2145
109
109
  datachain/lib/dc/json.py,sha256=dNijfJ-H92vU3soyR7X1IiDrWhm6yZIGG3bSnZkPdAE,2733
110
- datachain/lib/dc/listings.py,sha256=eVBUP25W81dv46DLqkv8K0X7N3nxhoZm77gFrByeT_E,4660
110
+ datachain/lib/dc/listings.py,sha256=V379Cb-7ZyquM0w7sWArQZkzInZy4GB7QQ1ZfowKzQY,4544
111
111
  datachain/lib/dc/pandas.py,sha256=ObueUXDUFKJGu380GmazdG02ARpKAHPhSaymfmOH13E,1489
112
112
  datachain/lib/dc/parquet.py,sha256=zYcSgrWwyEDW9UxGUSVdIVsCu15IGEf0xL8KfWQqK94,1782
113
113
  datachain/lib/dc/records.py,sha256=FpPbApWopUri1gIaSMsfXN4fevja4mjmfb6Q5eiaGxI,3116
@@ -125,7 +125,7 @@ datachain/model/ultralytics/pose.py,sha256=pBlmt63Qe68FKmexHimUGlNbNOoOlMHXG4fzX
125
125
  datachain/model/ultralytics/segment.py,sha256=63bDCj43E6iZ0hFI5J6uQfksdCmjEp6sEm1XzVaE8pw,2986
126
126
  datachain/query/__init__.py,sha256=7DhEIjAA8uZJfejruAVMZVcGFmvUpffuZJwgRqNwe-c,263
127
127
  datachain/query/batch.py,sha256=-goxLpE0EUvaDHu66rstj53UnfHpYfBUGux8GSpJ93k,4306
128
- datachain/query/dataset.py,sha256=SjFUh77rBTpgBZG4cfMJiJ2DhiCubGVk2cG1RYX4oyA,61571
128
+ datachain/query/dataset.py,sha256=C60VM0pScsrWcMqLNdX-tU0HE1SnEE9lRN3TU8CfTu4,61223
129
129
  datachain/query/dispatch.py,sha256=A0nPxn6mEN5d9dDo6S8m16Ji_9IvJLXrgF2kqXdi4fs,15546
130
130
  datachain/query/metrics.py,sha256=DOK5HdNVaRugYPjl8qnBONvTkwjMloLqAr7Mi3TjCO0,858
131
131
  datachain/query/params.py,sha256=O_j89mjYRLOwWNhYZl-z7mi-rkdP7WyFmaDufsdTryE,863
@@ -135,7 +135,7 @@ datachain/query/session.py,sha256=gKblltJAVQAVSTswAgWGDgGbpmFlFzFVkIQojDCjgXM,68
135
135
  datachain/query/udf.py,sha256=e753bDJzTNjGFQn1WGTvOAWSwjDbrFI1-_DDWkWN2ls,1343
136
136
  datachain/query/utils.py,sha256=HaSDNH_XGvp_NIcXjcB7j4vJRPi4_tbztDWclYelHY4,1208
137
137
  datachain/remote/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
138
- datachain/remote/studio.py,sha256=aSAh7CCHrTp7U-642jHFkwY0Fer4YRAHVVpWomab3zY,15110
138
+ datachain/remote/studio.py,sha256=oJp2KD9eO8zQDnPfNpAALZYsOlBfqVKKRTeCkEpcsYk,15196
139
139
  datachain/sql/__init__.py,sha256=6SQRdbljO3d2hx3EAVXEZrHQKv5jth0Jh98PogT59No,262
140
140
  datachain/sql/selectable.py,sha256=cTc60qVoAwqqss0Vop8Lt5Z-ROnM1XrQmL_GLjRxhXs,1765
141
141
  datachain/sql/types.py,sha256=ASSPkmM5EzdRindqj2O7WHLXq8VHAgFYedG8lYfGvVI,14045
@@ -157,9 +157,9 @@ datachain/sql/sqlite/vector.py,sha256=ncW4eu2FlJhrP_CIpsvtkUabZlQdl2D5Lgwy_cbfqR
157
157
  datachain/toolkit/__init__.py,sha256=eQ58Q5Yf_Fgv1ZG0IO5dpB4jmP90rk8YxUWmPc1M2Bo,68
158
158
  datachain/toolkit/split.py,sha256=ktGWzY4kyzjWyR86dhvzw-Zhl0lVk_LOX3NciTac6qo,2914
159
159
  datachain/torch/__init__.py,sha256=gIS74PoEPy4TB3X6vx9nLO0Y3sLJzsA8ckn8pRWihJM,579
160
- datachain-0.23.0.dist-info/licenses/LICENSE,sha256=8DnqK5yoPI_E50bEg_zsHKZHY2HqPy4rYN338BHQaRA,11344
161
- datachain-0.23.0.dist-info/METADATA,sha256=cS34xnVaaqHRes2oZ1uJnJVtDZ8L_jIBq6bXyTno8aI,13281
162
- datachain-0.23.0.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
163
- datachain-0.23.0.dist-info/entry_points.txt,sha256=0GMJS6B_KWq0m3VT98vQI2YZodAMkn4uReZ_okga9R4,49
164
- datachain-0.23.0.dist-info/top_level.txt,sha256=lZPpdU_2jJABLNIg2kvEOBi8PtsYikbN1OdMLHk8bTg,10
165
- datachain-0.23.0.dist-info/RECORD,,
160
+ datachain-0.24.0.dist-info/licenses/LICENSE,sha256=8DnqK5yoPI_E50bEg_zsHKZHY2HqPy4rYN338BHQaRA,11344
161
+ datachain-0.24.0.dist-info/METADATA,sha256=QWSVON3r5d5d18gRMs9G5DNV4z-kBBY47dMYUEFR0b0,13281
162
+ datachain-0.24.0.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
163
+ datachain-0.24.0.dist-info/entry_points.txt,sha256=0GMJS6B_KWq0m3VT98vQI2YZodAMkn4uReZ_okga9R4,49
164
+ datachain-0.24.0.dist-info/top_level.txt,sha256=lZPpdU_2jJABLNIg2kvEOBi8PtsYikbN1OdMLHk8bTg,10
165
+ datachain-0.24.0.dist-info/RECORD,,