datachain 0.23.0__py3-none-any.whl → 0.24.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of datachain might be problematic. Click here for more details.
- datachain/catalog/catalog.py +25 -13
- datachain/dataset.py +34 -5
- datachain/lib/dc/datasets.py +52 -35
- datachain/lib/dc/listings.py +2 -6
- datachain/lib/projects.py +1 -1
- datachain/query/dataset.py +2 -8
- datachain/remote/studio.py +4 -3
- {datachain-0.23.0.dist-info → datachain-0.24.0.dist-info}/METADATA +1 -1
- {datachain-0.23.0.dist-info → datachain-0.24.0.dist-info}/RECORD +13 -13
- {datachain-0.23.0.dist-info → datachain-0.24.0.dist-info}/WHEEL +0 -0
- {datachain-0.23.0.dist-info → datachain-0.24.0.dist-info}/entry_points.txt +0 -0
- {datachain-0.23.0.dist-info → datachain-0.24.0.dist-info}/licenses/LICENSE +0 -0
- {datachain-0.23.0.dist-info → datachain-0.24.0.dist-info}/top_level.txt +0 -0
datachain/catalog/catalog.py
CHANGED
|
@@ -49,6 +49,7 @@ from datachain.error import (
|
|
|
49
49
|
DatasetInvalidVersionError,
|
|
50
50
|
DatasetNotFoundError,
|
|
51
51
|
DatasetVersionNotFoundError,
|
|
52
|
+
NamespaceNotFoundError,
|
|
52
53
|
ProjectNotFoundError,
|
|
53
54
|
QueryScriptCancelError,
|
|
54
55
|
QueryScriptRunError,
|
|
@@ -1107,21 +1108,26 @@ class Catalog:
|
|
|
1107
1108
|
namespace_name: str,
|
|
1108
1109
|
project_name: str,
|
|
1109
1110
|
version: Optional[str] = None,
|
|
1111
|
+
pull_dataset: bool = False,
|
|
1112
|
+
update: bool = False,
|
|
1110
1113
|
) -> DatasetRecord:
|
|
1111
|
-
|
|
1112
|
-
|
|
1113
|
-
|
|
1114
|
-
|
|
1115
|
-
|
|
1116
|
-
|
|
1117
|
-
|
|
1118
|
-
|
|
1114
|
+
if self.metastore.is_local_dataset(namespace_name) or not update:
|
|
1115
|
+
try:
|
|
1116
|
+
project = self.metastore.get_project(project_name, namespace_name)
|
|
1117
|
+
ds = self.get_dataset(name, project)
|
|
1118
|
+
if not version or ds.has_version(version):
|
|
1119
|
+
return ds
|
|
1120
|
+
except (NamespaceNotFoundError, ProjectNotFoundError, DatasetNotFoundError):
|
|
1121
|
+
pass
|
|
1122
|
+
|
|
1123
|
+
if self.metastore.is_local_dataset(namespace_name):
|
|
1124
|
+
raise DatasetNotFoundError(
|
|
1125
|
+
f"Dataset {name}"
|
|
1126
|
+
+ (f" version {version} " if version else " ")
|
|
1127
|
+
+ "not found"
|
|
1128
|
+
)
|
|
1119
1129
|
|
|
1120
|
-
|
|
1121
|
-
ProjectNotFoundError,
|
|
1122
|
-
DatasetNotFoundError,
|
|
1123
|
-
DatasetVersionNotFoundError,
|
|
1124
|
-
):
|
|
1130
|
+
if pull_dataset:
|
|
1125
1131
|
print("Dataset not found in local catalog, trying to get from studio")
|
|
1126
1132
|
remote_ds_uri = create_dataset_uri(
|
|
1127
1133
|
name, namespace_name, project_name, version
|
|
@@ -1136,6 +1142,8 @@ class Catalog:
|
|
|
1136
1142
|
name, self.metastore.get_project(project_name, namespace_name)
|
|
1137
1143
|
)
|
|
1138
1144
|
|
|
1145
|
+
return self.get_remote_dataset(namespace_name, project_name, name)
|
|
1146
|
+
|
|
1139
1147
|
def get_dataset_with_version_uuid(self, uuid: str) -> DatasetRecord:
|
|
1140
1148
|
"""Returns dataset that contains version with specific uuid"""
|
|
1141
1149
|
for dataset in self.ls_datasets():
|
|
@@ -1152,6 +1160,10 @@ class Catalog:
|
|
|
1152
1160
|
|
|
1153
1161
|
info_response = studio_client.dataset_info(namespace, project, name)
|
|
1154
1162
|
if not info_response.ok:
|
|
1163
|
+
if info_response.status == 404:
|
|
1164
|
+
raise DatasetNotFoundError(
|
|
1165
|
+
f"Dataset {namespace}.{project}.{name} not found"
|
|
1166
|
+
)
|
|
1155
1167
|
raise DataChainError(info_response.message)
|
|
1156
1168
|
|
|
1157
1169
|
dataset_info = info_response.data
|
datachain/dataset.py
CHANGED
|
@@ -12,6 +12,9 @@ from typing import (
|
|
|
12
12
|
)
|
|
13
13
|
from urllib.parse import urlparse
|
|
14
14
|
|
|
15
|
+
from packaging.specifiers import SpecifierSet
|
|
16
|
+
from packaging.version import Version
|
|
17
|
+
|
|
15
18
|
from datachain import semver
|
|
16
19
|
from datachain.error import DatasetVersionNotFoundError, InvalidDatasetNameError
|
|
17
20
|
from datachain.namespace import Namespace
|
|
@@ -661,13 +664,39 @@ class DatasetRecord:
|
|
|
661
664
|
return None
|
|
662
665
|
return max(versions).version
|
|
663
666
|
|
|
664
|
-
|
|
665
|
-
|
|
666
|
-
|
|
667
|
-
|
|
667
|
+
def latest_compatible_version(self, version_spec: str) -> Optional[str]:
|
|
668
|
+
"""
|
|
669
|
+
Returns the latest version that matches the given version specifier.
|
|
670
|
+
|
|
671
|
+
Supports Python version specifiers like:
|
|
672
|
+
- ">=1.0.0,<2.0.0" (compatible release range)
|
|
673
|
+
- "~=1.4.2" (compatible release clause)
|
|
674
|
+
- "==1.2.*" (prefix matching)
|
|
675
|
+
- ">1.0.0" (exclusive ordered comparison)
|
|
676
|
+
- ">=1.0.0" (inclusive ordered comparison)
|
|
677
|
+
- "!=1.3.0" (version exclusion)
|
|
678
|
+
|
|
679
|
+
Args:
|
|
680
|
+
version_spec: Version specifier string following PEP 440
|
|
681
|
+
|
|
682
|
+
Returns:
|
|
683
|
+
Latest compatible version string, or None if no compatible version found
|
|
684
|
+
"""
|
|
685
|
+
spec_set = SpecifierSet(version_spec)
|
|
686
|
+
|
|
687
|
+
# Convert dataset versions to packaging.Version objects
|
|
688
|
+
# and filter compatible ones
|
|
689
|
+
compatible_versions = []
|
|
690
|
+
for v in self.versions:
|
|
691
|
+
pkg_version = Version(v.version)
|
|
692
|
+
if spec_set.contains(pkg_version):
|
|
693
|
+
compatible_versions.append(v)
|
|
694
|
+
|
|
695
|
+
if not compatible_versions:
|
|
668
696
|
return None
|
|
669
697
|
|
|
670
|
-
|
|
698
|
+
# Return the latest compatible version
|
|
699
|
+
return max(compatible_versions).version
|
|
671
700
|
|
|
672
701
|
@classmethod
|
|
673
702
|
def from_dict(cls, d: dict[str, Any]) -> "DatasetRecord":
|
datachain/lib/dc/datasets.py
CHANGED
|
@@ -7,9 +7,6 @@ from datachain.error import (
|
|
|
7
7
|
ProjectNotFoundError,
|
|
8
8
|
)
|
|
9
9
|
from datachain.lib.dataset_info import DatasetInfo
|
|
10
|
-
from datachain.lib.file import (
|
|
11
|
-
File,
|
|
12
|
-
)
|
|
13
10
|
from datachain.lib.projects import get as get_project
|
|
14
11
|
from datachain.lib.settings import Settings
|
|
15
12
|
from datachain.lib.signal_schema import SignalSchema
|
|
@@ -34,7 +31,6 @@ def read_dataset(
|
|
|
34
31
|
version: Optional[Union[str, int]] = None,
|
|
35
32
|
session: Optional[Session] = None,
|
|
36
33
|
settings: Optional[dict] = None,
|
|
37
|
-
fallback_to_studio: bool = True,
|
|
38
34
|
delta: Optional[bool] = False,
|
|
39
35
|
delta_on: Optional[Union[str, Sequence[str]]] = (
|
|
40
36
|
"file.path",
|
|
@@ -44,6 +40,7 @@ def read_dataset(
|
|
|
44
40
|
delta_result_on: Optional[Union[str, Sequence[str]]] = None,
|
|
45
41
|
delta_compare: Optional[Union[str, Sequence[str]]] = None,
|
|
46
42
|
delta_retry: Optional[Union[bool, str]] = None,
|
|
43
|
+
update: bool = False,
|
|
47
44
|
) -> "DataChain":
|
|
48
45
|
"""Get data from a saved Dataset. It returns the chain itself.
|
|
49
46
|
If dataset or version is not found locally, it will try to pull it from Studio.
|
|
@@ -55,11 +52,12 @@ def read_dataset(
|
|
|
55
52
|
set; otherwise, default values will be applied.
|
|
56
53
|
namespace : optional name of namespace in which dataset to read is created
|
|
57
54
|
project : optional name of project in which dataset to read is created
|
|
58
|
-
version : dataset version
|
|
55
|
+
version : dataset version. Supports:
|
|
56
|
+
- Exact version strings: "1.2.3"
|
|
57
|
+
- Legacy integer versions: 1, 2, 3 (finds latest major version)
|
|
58
|
+
- Version specifiers (PEP 440): ">=1.0.0,<2.0.0", "~=1.4.2", "==1.2.*", etc.
|
|
59
59
|
session : Session to use for the chain.
|
|
60
60
|
settings : Settings to use for the chain.
|
|
61
|
-
fallback_to_studio : Try to pull dataset from Studio if not found locally.
|
|
62
|
-
Default is True.
|
|
63
61
|
delta: If True, only process new or changed files instead of reprocessing
|
|
64
62
|
everything. This saves time by skipping files that were already processed in
|
|
65
63
|
previous versions. The optimization is working when a new version of the
|
|
@@ -79,6 +77,10 @@ def read_dataset(
|
|
|
79
77
|
(error mode)
|
|
80
78
|
- True: Reprocess records missing from the result dataset (missing mode)
|
|
81
79
|
- None: No retry processing (default)
|
|
80
|
+
update: If True always checks for newer versions available on Studio, even if
|
|
81
|
+
some version of the dataset exists locally already. If False (default), it
|
|
82
|
+
will only fetch the dataset from Studio if it is not found locally.
|
|
83
|
+
|
|
82
84
|
|
|
83
85
|
Example:
|
|
84
86
|
```py
|
|
@@ -92,11 +94,22 @@ def read_dataset(
|
|
|
92
94
|
```
|
|
93
95
|
|
|
94
96
|
```py
|
|
95
|
-
chain = dc.read_dataset("my_cats",
|
|
97
|
+
chain = dc.read_dataset("my_cats", version="1.0.0")
|
|
96
98
|
```
|
|
97
99
|
|
|
98
100
|
```py
|
|
99
|
-
|
|
101
|
+
# Using version specifiers (PEP 440)
|
|
102
|
+
chain = dc.read_dataset("my_cats", version=">=1.0.0,<2.0.0")
|
|
103
|
+
```
|
|
104
|
+
|
|
105
|
+
```py
|
|
106
|
+
# Legacy integer version support (finds latest in major version)
|
|
107
|
+
chain = dc.read_dataset("my_cats", version=1) # Latest 1.x.x version
|
|
108
|
+
```
|
|
109
|
+
|
|
110
|
+
```py
|
|
111
|
+
# Always check for newer versions matching a version specifier from Studio
|
|
112
|
+
chain = dc.read_dataset("my_cats", version=">=1.0.0", update=True)
|
|
100
113
|
```
|
|
101
114
|
|
|
102
115
|
```py
|
|
@@ -113,7 +126,6 @@ def read_dataset(
|
|
|
113
126
|
version="1.0.0",
|
|
114
127
|
session=session,
|
|
115
128
|
settings=settings,
|
|
116
|
-
fallback_to_studio=True,
|
|
117
129
|
)
|
|
118
130
|
```
|
|
119
131
|
"""
|
|
@@ -121,6 +133,8 @@ def read_dataset(
|
|
|
121
133
|
|
|
122
134
|
from .datachain import DataChain
|
|
123
135
|
|
|
136
|
+
telemetry.send_event_once("class", "datachain_init", name=name, version=version)
|
|
137
|
+
|
|
124
138
|
session = Session.get(session)
|
|
125
139
|
catalog = session.catalog
|
|
126
140
|
|
|
@@ -131,31 +145,37 @@ def read_dataset(
|
|
|
131
145
|
)
|
|
132
146
|
|
|
133
147
|
if version is not None:
|
|
148
|
+
dataset = session.catalog.get_dataset_with_remote_fallback(
|
|
149
|
+
name, namespace_name, project_name, update=update
|
|
150
|
+
)
|
|
151
|
+
|
|
152
|
+
# Convert legacy integer versions to version specifiers
|
|
153
|
+
# For backward compatibility we still allow users to put version as integer
|
|
154
|
+
# in which case we convert it to a version specifier that finds the latest
|
|
155
|
+
# version where major part is equal to that input version.
|
|
156
|
+
# For example if user sets version=2, we convert it to ">=2.0.0,<3.0.0"
|
|
157
|
+
# which will find something like 2.4.3 (assuming 2.4.3 is the biggest among
|
|
158
|
+
# all 2.* dataset versions)
|
|
159
|
+
if isinstance(version, int):
|
|
160
|
+
version_spec = f">={version}.0.0,<{version + 1}.0.0"
|
|
161
|
+
else:
|
|
162
|
+
version_spec = str(version)
|
|
163
|
+
|
|
164
|
+
from packaging.specifiers import InvalidSpecifier, SpecifierSet
|
|
165
|
+
|
|
134
166
|
try:
|
|
135
|
-
#
|
|
136
|
-
|
|
137
|
-
#
|
|
138
|
-
|
|
139
|
-
|
|
140
|
-
# major part is equal to that input, exception is thrown.
|
|
141
|
-
major = int(version)
|
|
142
|
-
try:
|
|
143
|
-
ds_project = get_project(project_name, namespace_name, session=session)
|
|
144
|
-
except ProjectNotFoundError:
|
|
145
|
-
raise DatasetNotFoundError(
|
|
146
|
-
f"Dataset {name} not found in namespace {namespace_name} and",
|
|
147
|
-
f" project {project_name}",
|
|
148
|
-
) from None
|
|
149
|
-
|
|
150
|
-
dataset = session.catalog.get_dataset(name, ds_project)
|
|
151
|
-
latest_major = dataset.latest_major_version(major)
|
|
152
|
-
if not latest_major:
|
|
167
|
+
# Try to parse as version specifier
|
|
168
|
+
SpecifierSet(version_spec)
|
|
169
|
+
# If it's a valid specifier set, find the latest compatible version
|
|
170
|
+
latest_compatible = dataset.latest_compatible_version(version_spec)
|
|
171
|
+
if not latest_compatible:
|
|
153
172
|
raise DatasetVersionNotFoundError(
|
|
154
|
-
f"
|
|
173
|
+
f"No dataset {name} version matching specifier {version_spec}"
|
|
155
174
|
)
|
|
156
|
-
version =
|
|
157
|
-
except
|
|
158
|
-
#
|
|
175
|
+
version = latest_compatible
|
|
176
|
+
except InvalidSpecifier:
|
|
177
|
+
# If not a valid specifier, treat as exact version string
|
|
178
|
+
# This handles cases like "1.2.3" which are exact versions, not specifiers
|
|
159
179
|
pass
|
|
160
180
|
|
|
161
181
|
if settings:
|
|
@@ -169,11 +189,8 @@ def read_dataset(
|
|
|
169
189
|
namespace_name=namespace_name,
|
|
170
190
|
version=version, # type: ignore[arg-type]
|
|
171
191
|
session=session,
|
|
172
|
-
indexing_column_types=File._datachain_column_types,
|
|
173
|
-
fallback_to_studio=fallback_to_studio,
|
|
174
192
|
)
|
|
175
193
|
|
|
176
|
-
telemetry.send_event_once("class", "datachain_init", name=name, version=version)
|
|
177
194
|
signals_schema = SignalSchema({"sys": Sys})
|
|
178
195
|
if query.feature_schema:
|
|
179
196
|
signals_schema |= SignalSchema.deserialize(query.feature_schema)
|
datachain/lib/dc/listings.py
CHANGED
|
@@ -127,12 +127,8 @@ def read_listing_dataset(
|
|
|
127
127
|
if version is None:
|
|
128
128
|
version = dataset.latest_version
|
|
129
129
|
|
|
130
|
-
query = DatasetQuery(
|
|
131
|
-
|
|
132
|
-
session=session,
|
|
133
|
-
indexing_column_types=File._datachain_column_types,
|
|
134
|
-
fallback_to_studio=False,
|
|
135
|
-
)
|
|
130
|
+
query = DatasetQuery(name=name, session=session)
|
|
131
|
+
|
|
136
132
|
if settings:
|
|
137
133
|
cfg = {**settings}
|
|
138
134
|
if "prefetch" not in cfg:
|
datachain/lib/projects.py
CHANGED
|
@@ -54,7 +54,7 @@ def get(name: str, namespace: str, session: Optional[Session]) -> Project:
|
|
|
54
54
|
```py
|
|
55
55
|
import datachain as dc
|
|
56
56
|
from datachain.lib.projects import get as get_project
|
|
57
|
-
project
|
|
57
|
+
project = get_project("my-project", "local")
|
|
58
58
|
```
|
|
59
59
|
"""
|
|
60
60
|
return Session.get(session).catalog.metastore.get_project(name, namespace)
|
datachain/query/dataset.py
CHANGED
|
@@ -1099,13 +1099,9 @@ class DatasetQuery:
|
|
|
1099
1099
|
namespace_name: Optional[str] = None,
|
|
1100
1100
|
catalog: Optional["Catalog"] = None,
|
|
1101
1101
|
session: Optional[Session] = None,
|
|
1102
|
-
indexing_column_types: Optional[dict[str, Any]] = None,
|
|
1103
1102
|
in_memory: bool = False,
|
|
1104
|
-
fallback_to_studio: bool = True,
|
|
1105
1103
|
update: bool = False,
|
|
1106
1104
|
) -> None:
|
|
1107
|
-
from datachain.remote.studio import is_token_set
|
|
1108
|
-
|
|
1109
1105
|
self.session = Session.get(session, catalog=catalog, in_memory=in_memory)
|
|
1110
1106
|
self.catalog = catalog or self.session.catalog
|
|
1111
1107
|
self.steps: list[Step] = []
|
|
@@ -1137,18 +1133,16 @@ class DatasetQuery:
|
|
|
1137
1133
|
# not setting query step yet as listing dataset might not exist at
|
|
1138
1134
|
# this point
|
|
1139
1135
|
self.list_ds_name = name
|
|
1140
|
-
|
|
1136
|
+
else:
|
|
1141
1137
|
self._set_starting_step(
|
|
1142
1138
|
self.catalog.get_dataset_with_remote_fallback(
|
|
1143
1139
|
name,
|
|
1144
1140
|
namespace_name=namespace_name,
|
|
1145
1141
|
project_name=project_name,
|
|
1146
1142
|
version=version,
|
|
1143
|
+
pull_dataset=True,
|
|
1147
1144
|
)
|
|
1148
1145
|
)
|
|
1149
|
-
else:
|
|
1150
|
-
project = self.catalog.metastore.get_project(project_name, namespace_name)
|
|
1151
|
-
self._set_starting_step(self.catalog.get_dataset(name, project=project))
|
|
1152
1146
|
|
|
1153
1147
|
def _set_starting_step(self, ds: "DatasetRecord") -> None:
|
|
1154
1148
|
if not self.version:
|
datachain/remote/studio.py
CHANGED
|
@@ -78,10 +78,11 @@ def _parse_dates(obj: dict, date_fields: list[str]):
|
|
|
78
78
|
|
|
79
79
|
|
|
80
80
|
class Response(Generic[T]):
|
|
81
|
-
def __init__(self, data: T, ok: bool, message: str) -> None:
|
|
81
|
+
def __init__(self, data: T, ok: bool, message: str, status: int) -> None:
|
|
82
82
|
self.data = data
|
|
83
83
|
self.ok = ok
|
|
84
84
|
self.message = message
|
|
85
|
+
self.status = status
|
|
85
86
|
|
|
86
87
|
def __repr__(self):
|
|
87
88
|
return (
|
|
@@ -186,7 +187,7 @@ class StudioClient:
|
|
|
186
187
|
message = "Indexing in progress"
|
|
187
188
|
else:
|
|
188
189
|
message = content.get("message", "")
|
|
189
|
-
return Response(response_data, ok, message)
|
|
190
|
+
return Response(response_data, ok, message, response.status_code)
|
|
190
191
|
|
|
191
192
|
@retry_with_backoff(retries=3, errors=(HTTPError, Timeout))
|
|
192
193
|
def _send_request(
|
|
@@ -236,7 +237,7 @@ class StudioClient:
|
|
|
236
237
|
else:
|
|
237
238
|
message = ""
|
|
238
239
|
|
|
239
|
-
return Response(data, ok, message)
|
|
240
|
+
return Response(data, ok, message, response.status_code)
|
|
240
241
|
|
|
241
242
|
@staticmethod
|
|
242
243
|
def _unpacker_hook(code, data):
|
|
@@ -3,7 +3,7 @@ datachain/__main__.py,sha256=hG3Y4ARGEqe1AWwNMd259rBlqtphx1Wk39YbueQ0yV8,91
|
|
|
3
3
|
datachain/asyn.py,sha256=RH_jFwJcTXxhEFomaI9yL6S3Onau6NZ6FSKfKFGtrJE,9689
|
|
4
4
|
datachain/cache.py,sha256=ESVRaCJXEThMIfGEFVHx6wJPOZA7FYk9V6WxjyuqUBY,3626
|
|
5
5
|
datachain/config.py,sha256=g8qbNV0vW2VEKpX-dGZ9pAn0DAz6G2ZFcr7SAV3PoSM,4272
|
|
6
|
-
datachain/dataset.py,sha256=
|
|
6
|
+
datachain/dataset.py,sha256=wDrukmkDnYP0X8bAGY-7O1NDE3DWCFqrH8VVDpXM9Ok,25263
|
|
7
7
|
datachain/delta.py,sha256=4RqLLc9dJLF8x9GG9IDgi86DwuPerZQ4HAUnNBeACw8,8446
|
|
8
8
|
datachain/error.py,sha256=OWwWMkzZYJrkcoEDGhJHMf7SfKvxcsOLRF94mjPf29I,1609
|
|
9
9
|
datachain/job.py,sha256=x5PB6d5sqx00hePNNkirESlOVAvnmkEM5ygUgQmAhsk,1262
|
|
@@ -21,7 +21,7 @@ datachain/studio.py,sha256=bLok-eJNFRHQScEyAyA_Fas52dmijd5r-73KudWxV4k,13337
|
|
|
21
21
|
datachain/telemetry.py,sha256=0A4IOPPp9VlP5pyW9eBfaTK3YhHGzHl7dQudQjUAx9A,994
|
|
22
22
|
datachain/utils.py,sha256=DNqOi-Ydb7InyWvD9m7_yailxz6-YGpZzh00biQaHNo,15305
|
|
23
23
|
datachain/catalog/__init__.py,sha256=cMZzSz3VoUi-6qXSVaHYN-agxQuAcz2XSqnEPZ55crE,353
|
|
24
|
-
datachain/catalog/catalog.py,sha256=
|
|
24
|
+
datachain/catalog/catalog.py,sha256=z4GbRMHeW0YA20Sjh7QuPy1Rj4RkX547WN9Pp5wAD6o,65277
|
|
25
25
|
datachain/catalog/datasource.py,sha256=IkGMh0Ttg6Q-9DWfU_H05WUnZepbGa28HYleECi6K7I,1353
|
|
26
26
|
datachain/catalog/loader.py,sha256=UXjYD6BNRoupPvkiz3-b04jepXhtLHCA4gzKFnXxOtQ,5987
|
|
27
27
|
datachain/cli/__init__.py,sha256=WvBqnwjG8Wp9xGCn-4eqfoZ3n7Sj1HJemCi4MayJh_c,8221
|
|
@@ -82,7 +82,7 @@ datachain/lib/listing_info.py,sha256=9ua40Hw0aiQByUw3oAEeNzMavJYfW0Uhe8YdCTK-m_g
|
|
|
82
82
|
datachain/lib/meta_formats.py,sha256=zdyg6XLk3QIsSk3I7s0Ez5kaCJSlE3uq7JiGxf7UwtU,6348
|
|
83
83
|
datachain/lib/model_store.py,sha256=DNIv8Y6Jtk1_idNLzIpsThOsdW2BMAudyUCbPUcgcxk,2515
|
|
84
84
|
datachain/lib/namespaces.py,sha256=it52UbbwB8dzhesO2pMs_nThXiPQ1Ph9sD9I3GQkg5s,2099
|
|
85
|
-
datachain/lib/projects.py,sha256=
|
|
85
|
+
datachain/lib/projects.py,sha256=8lN0qV8czX1LGtWURCUvRlSJk-RpO9w9Rra_pOZus6g,2595
|
|
86
86
|
datachain/lib/pytorch.py,sha256=oBBd6cxYrcwaFz7IQajKqhGqDdNnwUZWs0wJPRizrjk,7712
|
|
87
87
|
datachain/lib/settings.py,sha256=9wi0FoHxRxNiyn99pR28IYsMkoo47jQxeXuObQr2Ar0,2929
|
|
88
88
|
datachain/lib/signal_schema.py,sha256=dVEqqrQQ_BS3yzU_49-Gari7IjVyMl1UT8h1WIsZabs,36489
|
|
@@ -104,10 +104,10 @@ datachain/lib/dc/__init__.py,sha256=HD0NYrdy44u6kkpvgGjJcvGz-UGTHui2azghcT8ZUg0,
|
|
|
104
104
|
datachain/lib/dc/csv.py,sha256=q6a9BpapGwP6nwy6c5cklxQumep2fUp9l2LAjtTJr6s,4411
|
|
105
105
|
datachain/lib/dc/database.py,sha256=g5M6NjYR1T0vKte-abV-3Ejnm-HqxTIMir5cRi_SziE,6051
|
|
106
106
|
datachain/lib/dc/datachain.py,sha256=dFI7JX5-41HLgA-TUR99dtR1lvk2vokaMC3mbIW1XT4,85814
|
|
107
|
-
datachain/lib/dc/datasets.py,sha256=
|
|
107
|
+
datachain/lib/dc/datasets.py,sha256=U4xqAfs6FdW8HIJjeayQaIg1dunaIsVXYGqfq_sDSv0,13274
|
|
108
108
|
datachain/lib/dc/hf.py,sha256=PJl2wiLjdRsMz0SYbLT-6H8b-D5i2WjeH7li8HHOk_0,2145
|
|
109
109
|
datachain/lib/dc/json.py,sha256=dNijfJ-H92vU3soyR7X1IiDrWhm6yZIGG3bSnZkPdAE,2733
|
|
110
|
-
datachain/lib/dc/listings.py,sha256=
|
|
110
|
+
datachain/lib/dc/listings.py,sha256=V379Cb-7ZyquM0w7sWArQZkzInZy4GB7QQ1ZfowKzQY,4544
|
|
111
111
|
datachain/lib/dc/pandas.py,sha256=ObueUXDUFKJGu380GmazdG02ARpKAHPhSaymfmOH13E,1489
|
|
112
112
|
datachain/lib/dc/parquet.py,sha256=zYcSgrWwyEDW9UxGUSVdIVsCu15IGEf0xL8KfWQqK94,1782
|
|
113
113
|
datachain/lib/dc/records.py,sha256=FpPbApWopUri1gIaSMsfXN4fevja4mjmfb6Q5eiaGxI,3116
|
|
@@ -125,7 +125,7 @@ datachain/model/ultralytics/pose.py,sha256=pBlmt63Qe68FKmexHimUGlNbNOoOlMHXG4fzX
|
|
|
125
125
|
datachain/model/ultralytics/segment.py,sha256=63bDCj43E6iZ0hFI5J6uQfksdCmjEp6sEm1XzVaE8pw,2986
|
|
126
126
|
datachain/query/__init__.py,sha256=7DhEIjAA8uZJfejruAVMZVcGFmvUpffuZJwgRqNwe-c,263
|
|
127
127
|
datachain/query/batch.py,sha256=-goxLpE0EUvaDHu66rstj53UnfHpYfBUGux8GSpJ93k,4306
|
|
128
|
-
datachain/query/dataset.py,sha256=
|
|
128
|
+
datachain/query/dataset.py,sha256=C60VM0pScsrWcMqLNdX-tU0HE1SnEE9lRN3TU8CfTu4,61223
|
|
129
129
|
datachain/query/dispatch.py,sha256=A0nPxn6mEN5d9dDo6S8m16Ji_9IvJLXrgF2kqXdi4fs,15546
|
|
130
130
|
datachain/query/metrics.py,sha256=DOK5HdNVaRugYPjl8qnBONvTkwjMloLqAr7Mi3TjCO0,858
|
|
131
131
|
datachain/query/params.py,sha256=O_j89mjYRLOwWNhYZl-z7mi-rkdP7WyFmaDufsdTryE,863
|
|
@@ -135,7 +135,7 @@ datachain/query/session.py,sha256=gKblltJAVQAVSTswAgWGDgGbpmFlFzFVkIQojDCjgXM,68
|
|
|
135
135
|
datachain/query/udf.py,sha256=e753bDJzTNjGFQn1WGTvOAWSwjDbrFI1-_DDWkWN2ls,1343
|
|
136
136
|
datachain/query/utils.py,sha256=HaSDNH_XGvp_NIcXjcB7j4vJRPi4_tbztDWclYelHY4,1208
|
|
137
137
|
datachain/remote/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
138
|
-
datachain/remote/studio.py,sha256=
|
|
138
|
+
datachain/remote/studio.py,sha256=oJp2KD9eO8zQDnPfNpAALZYsOlBfqVKKRTeCkEpcsYk,15196
|
|
139
139
|
datachain/sql/__init__.py,sha256=6SQRdbljO3d2hx3EAVXEZrHQKv5jth0Jh98PogT59No,262
|
|
140
140
|
datachain/sql/selectable.py,sha256=cTc60qVoAwqqss0Vop8Lt5Z-ROnM1XrQmL_GLjRxhXs,1765
|
|
141
141
|
datachain/sql/types.py,sha256=ASSPkmM5EzdRindqj2O7WHLXq8VHAgFYedG8lYfGvVI,14045
|
|
@@ -157,9 +157,9 @@ datachain/sql/sqlite/vector.py,sha256=ncW4eu2FlJhrP_CIpsvtkUabZlQdl2D5Lgwy_cbfqR
|
|
|
157
157
|
datachain/toolkit/__init__.py,sha256=eQ58Q5Yf_Fgv1ZG0IO5dpB4jmP90rk8YxUWmPc1M2Bo,68
|
|
158
158
|
datachain/toolkit/split.py,sha256=ktGWzY4kyzjWyR86dhvzw-Zhl0lVk_LOX3NciTac6qo,2914
|
|
159
159
|
datachain/torch/__init__.py,sha256=gIS74PoEPy4TB3X6vx9nLO0Y3sLJzsA8ckn8pRWihJM,579
|
|
160
|
-
datachain-0.
|
|
161
|
-
datachain-0.
|
|
162
|
-
datachain-0.
|
|
163
|
-
datachain-0.
|
|
164
|
-
datachain-0.
|
|
165
|
-
datachain-0.
|
|
160
|
+
datachain-0.24.0.dist-info/licenses/LICENSE,sha256=8DnqK5yoPI_E50bEg_zsHKZHY2HqPy4rYN338BHQaRA,11344
|
|
161
|
+
datachain-0.24.0.dist-info/METADATA,sha256=QWSVON3r5d5d18gRMs9G5DNV4z-kBBY47dMYUEFR0b0,13281
|
|
162
|
+
datachain-0.24.0.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
|
|
163
|
+
datachain-0.24.0.dist-info/entry_points.txt,sha256=0GMJS6B_KWq0m3VT98vQI2YZodAMkn4uReZ_okga9R4,49
|
|
164
|
+
datachain-0.24.0.dist-info/top_level.txt,sha256=lZPpdU_2jJABLNIg2kvEOBi8PtsYikbN1OdMLHk8bTg,10
|
|
165
|
+
datachain-0.24.0.dist-info/RECORD,,
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|