datachain 0.6.11__py3-none-any.whl → 0.7.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of datachain might be problematic. Click here for more details.
- datachain/asyn.py +36 -4
- datachain/data_storage/warehouse.py +4 -1
- datachain/lib/dc.py +6 -1
- datachain/lib/file.py +5 -0
- datachain/lib/settings.py +11 -1
- datachain/lib/udf.py +45 -18
- datachain/node.py +1 -1
- datachain/query/dataset.py +25 -27
- {datachain-0.6.11.dist-info → datachain-0.7.0.dist-info}/METADATA +1 -1
- {datachain-0.6.11.dist-info → datachain-0.7.0.dist-info}/RECORD +14 -14
- {datachain-0.6.11.dist-info → datachain-0.7.0.dist-info}/LICENSE +0 -0
- {datachain-0.6.11.dist-info → datachain-0.7.0.dist-info}/WHEEL +0 -0
- {datachain-0.6.11.dist-info → datachain-0.7.0.dist-info}/entry_points.txt +0 -0
- {datachain-0.6.11.dist-info → datachain-0.7.0.dist-info}/top_level.txt +0 -0
datachain/asyn.py
CHANGED
|
@@ -1,5 +1,13 @@
|
|
|
1
1
|
import asyncio
|
|
2
|
-
|
|
2
|
+
import threading
|
|
3
|
+
from collections.abc import (
|
|
4
|
+
AsyncIterable,
|
|
5
|
+
Awaitable,
|
|
6
|
+
Coroutine,
|
|
7
|
+
Generator,
|
|
8
|
+
Iterable,
|
|
9
|
+
Iterator,
|
|
10
|
+
)
|
|
3
11
|
from concurrent.futures import ThreadPoolExecutor
|
|
4
12
|
from heapq import heappop, heappush
|
|
5
13
|
from typing import Any, Callable, Generic, Optional, TypeVar
|
|
@@ -47,6 +55,7 @@ class AsyncMapper(Generic[InputT, ResultT]):
|
|
|
47
55
|
self.loop = get_loop() if loop is None else loop
|
|
48
56
|
self.pool = ThreadPoolExecutor(workers)
|
|
49
57
|
self._tasks: set[asyncio.Task] = set()
|
|
58
|
+
self._shutdown_producer = threading.Event()
|
|
50
59
|
|
|
51
60
|
def start_task(self, coro: Coroutine) -> asyncio.Task:
|
|
52
61
|
task = self.loop.create_task(coro)
|
|
@@ -54,9 +63,31 @@ class AsyncMapper(Generic[InputT, ResultT]):
|
|
|
54
63
|
task.add_done_callback(self._tasks.discard)
|
|
55
64
|
return task
|
|
56
65
|
|
|
57
|
-
|
|
66
|
+
def _produce(self) -> None:
|
|
58
67
|
for item in self.iterable:
|
|
59
|
-
|
|
68
|
+
if self._shutdown_producer.is_set():
|
|
69
|
+
return
|
|
70
|
+
fut = asyncio.run_coroutine_threadsafe(self.work_queue.put(item), self.loop)
|
|
71
|
+
fut.result() # wait until the item is in the queue
|
|
72
|
+
|
|
73
|
+
async def produce(self) -> None:
|
|
74
|
+
await self.to_thread(self._produce)
|
|
75
|
+
|
|
76
|
+
def shutdown_producer(self) -> None:
|
|
77
|
+
"""
|
|
78
|
+
Signal the producer to stop and drain any remaining items from the work_queue.
|
|
79
|
+
|
|
80
|
+
This method sets an internal event, `_shutdown_producer`, which tells the
|
|
81
|
+
producer that it should stop adding items to the queue. To ensure that the
|
|
82
|
+
producer notices this signal promptly, we also attempt to drain any items
|
|
83
|
+
currently in the queue, clearing it so that the event can be checked without
|
|
84
|
+
delay.
|
|
85
|
+
"""
|
|
86
|
+
self._shutdown_producer.set()
|
|
87
|
+
q = self.work_queue
|
|
88
|
+
while not q.empty():
|
|
89
|
+
q.get_nowait()
|
|
90
|
+
q.task_done()
|
|
60
91
|
|
|
61
92
|
async def worker(self) -> None:
|
|
62
93
|
while (item := await self.work_queue.get()) is not None:
|
|
@@ -132,7 +163,7 @@ class AsyncMapper(Generic[InputT, ResultT]):
|
|
|
132
163
|
self.result_queue.get_nowait()
|
|
133
164
|
await self.result_queue.put(None)
|
|
134
165
|
|
|
135
|
-
def iterate(self, timeout=None) ->
|
|
166
|
+
def iterate(self, timeout=None) -> Generator[ResultT, None, None]:
|
|
136
167
|
init = asyncio.run_coroutine_threadsafe(self.init(), self.loop)
|
|
137
168
|
init.result(timeout=1)
|
|
138
169
|
async_run = asyncio.run_coroutine_threadsafe(self.run(), self.loop)
|
|
@@ -145,6 +176,7 @@ class AsyncMapper(Generic[InputT, ResultT]):
|
|
|
145
176
|
if exc := async_run.exception():
|
|
146
177
|
raise exc
|
|
147
178
|
finally:
|
|
179
|
+
self.shutdown_producer()
|
|
148
180
|
if not async_run.done():
|
|
149
181
|
async_run.cancel()
|
|
150
182
|
|
|
@@ -232,7 +232,10 @@ class AbstractWarehouse(ABC, Serializable):
|
|
|
232
232
|
if limit < page_size:
|
|
233
233
|
paginated_query = paginated_query.limit(None).limit(limit)
|
|
234
234
|
|
|
235
|
-
|
|
235
|
+
# Ensure we're using a thread-local connection
|
|
236
|
+
with self.clone() as wh:
|
|
237
|
+
# Cursor results are not thread-safe, so we convert them to a list
|
|
238
|
+
results = list(wh.dataset_rows_select(paginated_query.offset(offset)))
|
|
236
239
|
|
|
237
240
|
processed = False
|
|
238
241
|
for row in results:
|
datachain/lib/dc.py
CHANGED
|
@@ -334,6 +334,7 @@ class DataChain:
|
|
|
334
334
|
parallel=None,
|
|
335
335
|
workers=None,
|
|
336
336
|
min_task_size=None,
|
|
337
|
+
prefetch: Optional[int] = None,
|
|
337
338
|
sys: Optional[bool] = None,
|
|
338
339
|
) -> "Self":
|
|
339
340
|
"""Change settings for chain.
|
|
@@ -360,7 +361,7 @@ class DataChain:
|
|
|
360
361
|
if sys is None:
|
|
361
362
|
sys = self._sys
|
|
362
363
|
settings = copy.copy(self._settings)
|
|
363
|
-
settings.add(Settings(cache, parallel, workers, min_task_size))
|
|
364
|
+
settings.add(Settings(cache, parallel, workers, min_task_size, prefetch))
|
|
364
365
|
return self._evolve(settings=settings, _sys=sys)
|
|
365
366
|
|
|
366
367
|
def reset_settings(self, settings: Optional[Settings] = None) -> "Self":
|
|
@@ -882,6 +883,8 @@ class DataChain:
|
|
|
882
883
|
```
|
|
883
884
|
"""
|
|
884
885
|
udf_obj = self._udf_to_obj(Mapper, func, params, output, signal_map)
|
|
886
|
+
if (prefetch := self._settings.prefetch) is not None:
|
|
887
|
+
udf_obj.prefetch = prefetch
|
|
885
888
|
|
|
886
889
|
return self._evolve(
|
|
887
890
|
query=self._query.add_signals(
|
|
@@ -919,6 +922,8 @@ class DataChain:
|
|
|
919
922
|
```
|
|
920
923
|
"""
|
|
921
924
|
udf_obj = self._udf_to_obj(Generator, func, params, output, signal_map)
|
|
925
|
+
if (prefetch := self._settings.prefetch) is not None:
|
|
926
|
+
udf_obj.prefetch = prefetch
|
|
922
927
|
return self._evolve(
|
|
923
928
|
query=self._query.generate(
|
|
924
929
|
udf_obj.to_udf_wrapper(),
|
datachain/lib/file.py
CHANGED
|
@@ -268,6 +268,11 @@ class File(DataModel):
|
|
|
268
268
|
client = self._catalog.get_client(self.source)
|
|
269
269
|
client.download(self, callback=self._download_cb)
|
|
270
270
|
|
|
271
|
+
async def _prefetch(self) -> None:
|
|
272
|
+
if self._caching_enabled:
|
|
273
|
+
client = self._catalog.get_client(self.source)
|
|
274
|
+
await client._download(self, callback=self._download_cb)
|
|
275
|
+
|
|
271
276
|
def get_local_path(self) -> Optional[str]:
|
|
272
277
|
"""Return path to a file in a local cache.
|
|
273
278
|
|
datachain/lib/settings.py
CHANGED
|
@@ -7,11 +7,19 @@ class SettingsError(DataChainParamsError):
|
|
|
7
7
|
|
|
8
8
|
|
|
9
9
|
class Settings:
|
|
10
|
-
def __init__(
|
|
10
|
+
def __init__(
|
|
11
|
+
self,
|
|
12
|
+
cache=None,
|
|
13
|
+
parallel=None,
|
|
14
|
+
workers=None,
|
|
15
|
+
min_task_size=None,
|
|
16
|
+
prefetch=None,
|
|
17
|
+
):
|
|
11
18
|
self._cache = cache
|
|
12
19
|
self.parallel = parallel
|
|
13
20
|
self._workers = workers
|
|
14
21
|
self.min_task_size = min_task_size
|
|
22
|
+
self.prefetch = prefetch
|
|
15
23
|
|
|
16
24
|
if not isinstance(cache, bool) and cache is not None:
|
|
17
25
|
raise SettingsError(
|
|
@@ -66,3 +74,5 @@ class Settings:
|
|
|
66
74
|
self.parallel = settings.parallel or self.parallel
|
|
67
75
|
self._workers = settings._workers or self._workers
|
|
68
76
|
self.min_task_size = settings.min_task_size or self.min_task_size
|
|
77
|
+
if settings.prefetch is not None:
|
|
78
|
+
self.prefetch = settings.prefetch
|
datachain/lib/udf.py
CHANGED
|
@@ -1,3 +1,4 @@
|
|
|
1
|
+
import contextlib
|
|
1
2
|
import sys
|
|
2
3
|
import traceback
|
|
3
4
|
from collections.abc import Iterable, Iterator, Mapping, Sequence
|
|
@@ -7,6 +8,7 @@ import attrs
|
|
|
7
8
|
from fsspec.callbacks import DEFAULT_CALLBACK, Callback
|
|
8
9
|
from pydantic import BaseModel
|
|
9
10
|
|
|
11
|
+
from datachain.asyn import AsyncMapper
|
|
10
12
|
from datachain.dataset import RowDict
|
|
11
13
|
from datachain.lib.convert.flatten import flatten
|
|
12
14
|
from datachain.lib.data_model import DataValue
|
|
@@ -21,6 +23,8 @@ from datachain.query.batch import (
|
|
|
21
23
|
)
|
|
22
24
|
|
|
23
25
|
if TYPE_CHECKING:
|
|
26
|
+
from collections import abc
|
|
27
|
+
|
|
24
28
|
from typing_extensions import Self
|
|
25
29
|
|
|
26
30
|
from datachain.catalog import Catalog
|
|
@@ -276,9 +280,18 @@ class UDFBase(AbstractUDF):
|
|
|
276
280
|
return result_objs
|
|
277
281
|
|
|
278
282
|
|
|
283
|
+
async def _prefetch_input(row):
|
|
284
|
+
for obj in row:
|
|
285
|
+
if isinstance(obj, File):
|
|
286
|
+
await obj._prefetch()
|
|
287
|
+
return row
|
|
288
|
+
|
|
289
|
+
|
|
279
290
|
class Mapper(UDFBase):
|
|
280
291
|
"""Inherit from this class to pass to `DataChain.map()`."""
|
|
281
292
|
|
|
293
|
+
prefetch: int = 2
|
|
294
|
+
|
|
282
295
|
def run(
|
|
283
296
|
self,
|
|
284
297
|
udf_fields: "Sequence[str]",
|
|
@@ -290,16 +303,22 @@ class Mapper(UDFBase):
|
|
|
290
303
|
) -> Iterator[Iterable[UDFResult]]:
|
|
291
304
|
self.catalog = catalog
|
|
292
305
|
self.setup()
|
|
293
|
-
|
|
294
|
-
|
|
295
|
-
|
|
296
|
-
|
|
297
|
-
|
|
298
|
-
|
|
299
|
-
|
|
300
|
-
|
|
301
|
-
|
|
302
|
-
|
|
306
|
+
prepared_inputs: abc.Generator[Sequence[Any], None, None] = (
|
|
307
|
+
self._prepare_row_and_id(row, udf_fields, cache, download_cb)
|
|
308
|
+
for row in udf_inputs
|
|
309
|
+
)
|
|
310
|
+
if self.prefetch > 0:
|
|
311
|
+
prepared_inputs = AsyncMapper(
|
|
312
|
+
_prefetch_input, prepared_inputs, workers=self.prefetch
|
|
313
|
+
).iterate()
|
|
314
|
+
|
|
315
|
+
with contextlib.closing(prepared_inputs):
|
|
316
|
+
for id_, *udf_args in prepared_inputs:
|
|
317
|
+
result_objs = self.process_safe(udf_args)
|
|
318
|
+
udf_output = self._flatten_row(result_objs)
|
|
319
|
+
output = [{"sys__id": id_} | dict(zip(self.signal_names, udf_output))]
|
|
320
|
+
processed_cb.relative_update(1)
|
|
321
|
+
yield output
|
|
303
322
|
|
|
304
323
|
self.teardown()
|
|
305
324
|
|
|
@@ -349,6 +368,7 @@ class Generator(UDFBase):
|
|
|
349
368
|
"""Inherit from this class to pass to `DataChain.gen()`."""
|
|
350
369
|
|
|
351
370
|
is_output_batched = True
|
|
371
|
+
prefetch: int = 2
|
|
352
372
|
|
|
353
373
|
def run(
|
|
354
374
|
self,
|
|
@@ -361,14 +381,21 @@ class Generator(UDFBase):
|
|
|
361
381
|
) -> Iterator[Iterable[UDFResult]]:
|
|
362
382
|
self.catalog = catalog
|
|
363
383
|
self.setup()
|
|
364
|
-
|
|
365
|
-
|
|
366
|
-
|
|
367
|
-
|
|
368
|
-
|
|
369
|
-
|
|
370
|
-
|
|
371
|
-
|
|
384
|
+
prepared_inputs: abc.Generator[Sequence[Any], None, None] = (
|
|
385
|
+
self._prepare_row(row, udf_fields, cache, download_cb) for row in udf_inputs
|
|
386
|
+
)
|
|
387
|
+
if self.prefetch > 0:
|
|
388
|
+
prepared_inputs = AsyncMapper(
|
|
389
|
+
_prefetch_input, prepared_inputs, workers=self.prefetch
|
|
390
|
+
).iterate()
|
|
391
|
+
|
|
392
|
+
with contextlib.closing(prepared_inputs):
|
|
393
|
+
for row in prepared_inputs:
|
|
394
|
+
result_objs = self.process_safe(row)
|
|
395
|
+
udf_outputs = (self._flatten_row(row) for row in result_objs)
|
|
396
|
+
output = (dict(zip(self.signal_names, row)) for row in udf_outputs)
|
|
397
|
+
processed_cb.relative_update(1)
|
|
398
|
+
yield output
|
|
372
399
|
|
|
373
400
|
self.teardown()
|
|
374
401
|
|
datachain/node.py
CHANGED
datachain/query/dataset.py
CHANGED
|
@@ -473,33 +473,31 @@ class UDFStep(Step, ABC):
|
|
|
473
473
|
# Otherwise process single-threaded (faster for smaller UDFs)
|
|
474
474
|
warehouse = self.catalog.warehouse
|
|
475
475
|
|
|
476
|
-
|
|
477
|
-
|
|
478
|
-
|
|
479
|
-
|
|
480
|
-
|
|
481
|
-
|
|
482
|
-
|
|
483
|
-
|
|
484
|
-
|
|
485
|
-
|
|
486
|
-
|
|
487
|
-
|
|
488
|
-
|
|
489
|
-
|
|
490
|
-
|
|
491
|
-
|
|
492
|
-
|
|
493
|
-
|
|
494
|
-
|
|
495
|
-
|
|
496
|
-
|
|
497
|
-
|
|
498
|
-
|
|
499
|
-
|
|
500
|
-
|
|
501
|
-
processed_cb.close()
|
|
502
|
-
generated_cb.close()
|
|
476
|
+
udf_inputs = batching(warehouse.dataset_select_paginated, query)
|
|
477
|
+
download_cb = get_download_callback()
|
|
478
|
+
processed_cb = get_processed_callback()
|
|
479
|
+
generated_cb = get_generated_callback(self.is_generator)
|
|
480
|
+
try:
|
|
481
|
+
udf_results = self.udf.run(
|
|
482
|
+
udf_fields,
|
|
483
|
+
udf_inputs,
|
|
484
|
+
self.catalog,
|
|
485
|
+
self.is_generator,
|
|
486
|
+
self.cache,
|
|
487
|
+
download_cb,
|
|
488
|
+
processed_cb,
|
|
489
|
+
)
|
|
490
|
+
process_udf_outputs(
|
|
491
|
+
warehouse,
|
|
492
|
+
udf_table,
|
|
493
|
+
udf_results,
|
|
494
|
+
self.udf,
|
|
495
|
+
cb=generated_cb,
|
|
496
|
+
)
|
|
497
|
+
finally:
|
|
498
|
+
download_cb.close()
|
|
499
|
+
processed_cb.close()
|
|
500
|
+
generated_cb.close()
|
|
503
501
|
|
|
504
502
|
warehouse.insert_rows_done(udf_table)
|
|
505
503
|
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
datachain/__init__.py,sha256=nnTyB5MpCfBZ6D85JPz-5hUT7i-68Is-47Bxgew8lRw,930
|
|
2
2
|
datachain/__main__.py,sha256=hG3Y4ARGEqe1AWwNMd259rBlqtphx1Wk39YbueQ0yV8,91
|
|
3
|
-
datachain/asyn.py,sha256=
|
|
3
|
+
datachain/asyn.py,sha256=5aKrjnUxk0mtnZeFKNJd1DCE0MsnSoyJBZkr0y9H_a0,9313
|
|
4
4
|
datachain/cache.py,sha256=s0YHN7qurmQv-eC265TjeureK84TebWWAnL07cxchZQ,2997
|
|
5
5
|
datachain/cli.py,sha256=hdVt_HJumQVgtaBAtBVJm-uPyYVogMXNVLmRcZyWHgk,36677
|
|
6
6
|
datachain/cli_utils.py,sha256=jrn9ejGXjybeO1ur3fjdSiAyCHZrX0qsLLbJzN9ErPM,2418
|
|
@@ -9,7 +9,7 @@ datachain/dataset.py,sha256=0IN-5y723y-bnFlieKtOFZLCjwX_yplFo3q0DV7LRPw,14821
|
|
|
9
9
|
datachain/error.py,sha256=bxAAL32lSeMgzsQDEHbGTGORj-mPzzpCRvWDPueJNN4,1092
|
|
10
10
|
datachain/job.py,sha256=Jt4sNutMHJReaGsj3r3scueN5aESLGfhimAa8pUP7Is,1271
|
|
11
11
|
datachain/listing.py,sha256=TgKg25ZWAP5enzKgw2_2GUPJVdnQUh6uySHB5SJrUY4,7773
|
|
12
|
-
datachain/node.py,sha256=
|
|
12
|
+
datachain/node.py,sha256=o8Sqy92QkzzcLK6XmIFLyDSE6Rw6kUTmGRhEmfLFdhg,5211
|
|
13
13
|
datachain/nodes_fetcher.py,sha256=F-73-h19HHNGtHFBGKk7p3mc0ALm4a9zGnzhtuUjnp4,1107
|
|
14
14
|
datachain/nodes_thread_pool.py,sha256=uPo-xl8zG5m9YgODjPFBpbcqqHjI-dcxH87yAbj_qco,3192
|
|
15
15
|
datachain/progress.py,sha256=5KotcvvzAUL_RF0GEj4JY0IB1lyImnmHxe89YkT1XO4,4330
|
|
@@ -37,14 +37,14 @@ datachain/data_storage/metastore.py,sha256=5b7o_CSHC2djottebYn-Hq5q0yaSLOKPIRCna
|
|
|
37
37
|
datachain/data_storage/schema.py,sha256=scANMQqozita3HjEtq7eupMgh6yYkrZHoXtfuL2RoQg,9879
|
|
38
38
|
datachain/data_storage/serializer.py,sha256=6G2YtOFqqDzJf1KbvZraKGXl2XHZyVml2krunWUum5o,927
|
|
39
39
|
datachain/data_storage/sqlite.py,sha256=CspRUlYsIcubgzvcQxTACnmcuKESSLZcqCl0dcrtRiA,27471
|
|
40
|
-
datachain/data_storage/warehouse.py,sha256=
|
|
40
|
+
datachain/data_storage/warehouse.py,sha256=yXNU0U3exzR1E6dqbYYmL4RhXWsbYWVdZ3jONGcVniY,30914
|
|
41
41
|
datachain/lib/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
42
42
|
datachain/lib/arrow.py,sha256=-hu9tic79a01SY2UBqkA3U6wUr6tnE3T3q5q_BnO93A,9156
|
|
43
43
|
datachain/lib/clip.py,sha256=lm5CzVi4Cj1jVLEKvERKArb-egb9j1Ls-fwTItT6vlI,6150
|
|
44
44
|
datachain/lib/data_model.py,sha256=dau4AlZBhOFvF7pEKMeqCeRkcFFg5KFvTBWW_2CdH5g,2371
|
|
45
45
|
datachain/lib/dataset_info.py,sha256=q0EW9tj5jXGSD9Lzct9zbH4P1lfIGd_cIWqhnMxv7Q0,2464
|
|
46
|
-
datachain/lib/dc.py,sha256=
|
|
47
|
-
datachain/lib/file.py,sha256
|
|
46
|
+
datachain/lib/dc.py,sha256=u0RQJPG0zwxsoYS-4wrbDBPuLYZajwIi1YX37khKfkI,87942
|
|
47
|
+
datachain/lib/file.py,sha256=-XMkL6ED1sE7TMhWoMRTEuOXswZJw8X6AEmJDONFP74,15019
|
|
48
48
|
datachain/lib/hf.py,sha256=BW2NPpqxkpPwkSaGlppT8Rbs8zPpyYC-tR6htY08c-0,5817
|
|
49
49
|
datachain/lib/image.py,sha256=AMXYwQsmarZjRbPCZY3M1jDsM2WAB_b3cTY4uOIuXNU,2675
|
|
50
50
|
datachain/lib/listing.py,sha256=cVkCp7TRVpcZKSx-Bbk9t51bQI9Mw0o86W6ZPhAsuzM,3667
|
|
@@ -52,11 +52,11 @@ datachain/lib/listing_info.py,sha256=9ua40Hw0aiQByUw3oAEeNzMavJYfW0Uhe8YdCTK-m_g
|
|
|
52
52
|
datachain/lib/meta_formats.py,sha256=anK2bDVbaeCCh0yvKUBaW2MVos3zRgdaSV8uSduzPcU,6680
|
|
53
53
|
datachain/lib/model_store.py,sha256=DNIv8Y6Jtk1_idNLzIpsThOsdW2BMAudyUCbPUcgcxk,2515
|
|
54
54
|
datachain/lib/pytorch.py,sha256=W-ARi2xH1f1DUkVfRuerW-YWYgSaJASmNCxtz2lrJGI,6072
|
|
55
|
-
datachain/lib/settings.py,sha256=
|
|
55
|
+
datachain/lib/settings.py,sha256=ZELRCTLbi5vzRPiDX6cQ9LLg9TefJ_A05gIGni0lll8,2535
|
|
56
56
|
datachain/lib/signal_schema.py,sha256=xwkE5bxJxUhZTjrA6jqN87XbSXPikCbL6eOPL9WyrKM,24556
|
|
57
57
|
datachain/lib/tar.py,sha256=3WIzao6yD5fbLqXLTt9GhPGNonbFIs_fDRu-9vgLgsA,1038
|
|
58
58
|
datachain/lib/text.py,sha256=UNHm8fhidk7wdrWqacEWaA6I9ykfYqarQ2URby7jc7M,1261
|
|
59
|
-
datachain/lib/udf.py,sha256
|
|
59
|
+
datachain/lib/udf.py,sha256=-j0krjNAELTqRI0dB1N65AmawtcIY5vN---AuUcW8Us,13637
|
|
60
60
|
datachain/lib/udf_signature.py,sha256=GXw24A-Olna6DWCdgy2bC-gZh_gLGPQ-KvjuI6pUjC0,7281
|
|
61
61
|
datachain/lib/utils.py,sha256=6NwgWLl5JrgtD4rsSFEe-yR2ntEwJMJEtAZ3FIxK3fg,1529
|
|
62
62
|
datachain/lib/vfile.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
@@ -77,7 +77,7 @@ datachain/lib/models/pose.py,sha256=peuJPNSiGuTXfCfGIABwv8PGYistvTTBmtf-8X8E_eA,
|
|
|
77
77
|
datachain/lib/models/yolo.py,sha256=eftoJDUa8iOpFTF1EkKVAd5Q-3HRd6X4eCIZ9h5p4nI,972
|
|
78
78
|
datachain/query/__init__.py,sha256=7DhEIjAA8uZJfejruAVMZVcGFmvUpffuZJwgRqNwe-c,263
|
|
79
79
|
datachain/query/batch.py,sha256=5fEhORFe7li12SdYddaSK3LyqksMfCHhwN1_A6TfsA4,3485
|
|
80
|
-
datachain/query/dataset.py,sha256=
|
|
80
|
+
datachain/query/dataset.py,sha256=sQny-ZemB2HueC4mPg-7qSaqUD85MMO-DQyVVP8K1CA,53765
|
|
81
81
|
datachain/query/dispatch.py,sha256=wjjTWw6sFQbB9SKRh78VbfvwSMgJXCfqJklS3-9KnCU,12025
|
|
82
82
|
datachain/query/metrics.py,sha256=r5b0ygYhokbXp8Mg3kCH8iFSRw0jxzyeBe-C-J_bKFc,938
|
|
83
83
|
datachain/query/params.py,sha256=O_j89mjYRLOwWNhYZl-z7mi-rkdP7WyFmaDufsdTryE,863
|
|
@@ -106,9 +106,9 @@ datachain/sql/sqlite/vector.py,sha256=ncW4eu2FlJhrP_CIpsvtkUabZlQdl2D5Lgwy_cbfqR
|
|
|
106
106
|
datachain/toolkit/__init__.py,sha256=eQ58Q5Yf_Fgv1ZG0IO5dpB4jmP90rk8YxUWmPc1M2Bo,68
|
|
107
107
|
datachain/toolkit/split.py,sha256=6FcEJgUsJsUcCqKW5aXuJy4DvbcQ7_dFbsfNPhn8EVg,2377
|
|
108
108
|
datachain/torch/__init__.py,sha256=gIS74PoEPy4TB3X6vx9nLO0Y3sLJzsA8ckn8pRWihJM,579
|
|
109
|
-
datachain-0.
|
|
110
|
-
datachain-0.
|
|
111
|
-
datachain-0.
|
|
112
|
-
datachain-0.
|
|
113
|
-
datachain-0.
|
|
114
|
-
datachain-0.
|
|
109
|
+
datachain-0.7.0.dist-info/LICENSE,sha256=8DnqK5yoPI_E50bEg_zsHKZHY2HqPy4rYN338BHQaRA,11344
|
|
110
|
+
datachain-0.7.0.dist-info/METADATA,sha256=Cm0v22C-aT14JBG4NjNQmM7kvhYZFAf6rh0yDAMRWFU,18037
|
|
111
|
+
datachain-0.7.0.dist-info/WHEEL,sha256=R06PA3UVYHThwHvxuRWMqaGcr-PuniXahwjmQRFMEkY,91
|
|
112
|
+
datachain-0.7.0.dist-info/entry_points.txt,sha256=0GMJS6B_KWq0m3VT98vQI2YZodAMkn4uReZ_okga9R4,49
|
|
113
|
+
datachain-0.7.0.dist-info/top_level.txt,sha256=lZPpdU_2jJABLNIg2kvEOBi8PtsYikbN1OdMLHk8bTg,10
|
|
114
|
+
datachain-0.7.0.dist-info/RECORD,,
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|