nedo-vision-worker-core 0.3.4__py3-none-any.whl → 0.3.5__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of nedo-vision-worker-core might be problematic. Click here for more details.

@@ -7,7 +7,7 @@ A library for running AI vision processing and detection in the Nedo Vision plat
7
7
  from .core_service import CoreService
8
8
  from .callbacks import DetectionType, CallbackTrigger, DetectionData, IntervalMetadata
9
9
 
10
- __version__ = "0.3.4"
10
+ __version__ = "0.3.5"
11
11
  __all__ = [
12
12
  "CoreService",
13
13
  "DetectionType",
@@ -39,6 +39,11 @@ class PipelineManager:
39
39
 
40
40
  logging.info(f"🚀 Starting Pipeline processing for pipeline: {pipeline_id} | Source: {worker_source_id} ({pipeline.name})")
41
41
 
42
+ # Acquire the video stream (starts it if not already running)
43
+ if not self.video_manager.acquire_stream(worker_source_id, pipeline_id):
44
+ logging.error(f"❌ Failed to acquire stream {worker_source_id} for pipeline {pipeline_id}")
45
+ return
46
+
42
47
  processor = PipelineProcessor(pipeline, detector, False)
43
48
  processor.frame_drawer.location_name = pipeline.location_name
44
49
  self.processors[pipeline_id] = processor # Store processor instance
@@ -79,6 +84,10 @@ class PipelineManager:
79
84
  self._stopping_pipelines.add(pipeline_id)
80
85
 
81
86
  try:
87
+ # Get worker_source_id before removing metadata
88
+ pipeline = self.pipeline_metadata.get(pipeline_id)
89
+ worker_source_id = pipeline.worker_source_id if pipeline else None
90
+
82
91
  # Stop AI processing
83
92
  processor = self.processors.pop(pipeline_id, None)
84
93
  if processor:
@@ -92,6 +101,10 @@ class PipelineManager:
92
101
  # Remove metadata
93
102
  self.pipeline_metadata.pop(pipeline_id, None)
94
103
 
104
+ # Release the video stream (stops it if no more pipelines use it)
105
+ if worker_source_id:
106
+ self.video_manager.release_stream(worker_source_id, pipeline_id)
107
+
95
108
  logging.info(f"✅ Pipeline {pipeline_id} stopped successfully.")
96
109
 
97
110
  except Exception as e:
@@ -459,9 +459,10 @@ class PipelineProcessor:
459
459
  logging.error(f" Cannot get stream URL for {worker_source_id}")
460
460
  return False
461
461
 
462
- video_manager.remove_stream(worker_source_id)
462
+ # Use internal methods to restart the stream without affecting reference counting
463
+ video_manager._stop_stream(worker_source_id)
463
464
  time.sleep(1.0)
464
- video_manager.add_stream(worker_source_id, stream_url)
465
+ video_manager._start_stream(worker_source_id, stream_url)
465
466
  time.sleep(2.0)
466
467
 
467
468
  if not video_manager.has_stream(worker_source_id):
@@ -103,8 +103,22 @@ class PipelineSyncThread(threading.Thread):
103
103
 
104
104
  for pid in pipeline_ids:
105
105
  db_pipeline = db_pipelines[pid]
106
+
107
+ # Check if pipeline should be stopped (status changed to stop/restart in DB)
108
+ if db_pipeline.pipeline_status_code in ['stop', 'restart']:
109
+ if self.pipeline_manager.is_running(pid):
110
+ logging.info(f"⏹️ Stopping pipeline due to status change: {pid}")
111
+ self.pipeline_manager.stop_pipeline(pid)
112
+ continue
113
+
106
114
  processor = self.pipeline_manager.processors.get(pid)
107
115
  if not processor:
116
+ # Pipeline exists in both sets but processor doesn't exist - shouldn't happen
117
+ # but if it does, try to start it if status is 'run'
118
+ if db_pipeline.pipeline_status_code == 'run':
119
+ logging.warning(f"⚠️ Pipeline {pid} exists locally but has no processor. Restarting...")
120
+ detector = self.model_manager.get_detector(db_pipeline.ai_model_id)
121
+ self.pipeline_manager.start_pipeline(db_pipeline, detector)
108
122
  continue
109
123
 
110
124
  local_detector = processor.detector
@@ -114,29 +128,12 @@ class PipelineSyncThread(threading.Thread):
114
128
  processor.enable_debug()
115
129
 
116
130
  def update_pipeline(self, pid: str, db_pipeline: object, local_detector: object) -> None:
117
- """Updates a single pipeline if necessary."""
131
+ """Updates a single pipeline if necessary (only called for running pipelines)."""
118
132
  processor = self.pipeline_manager.processors.get(pid)
119
133
  if not processor:
120
134
  return
121
135
 
122
- # Stop/start based on status change
123
- if db_pipeline.pipeline_status_code != processor._pipeline.pipeline_status_code:
124
- if db_pipeline.pipeline_status_code == 'run':
125
- logging.info(f"▶️ Resuming pipeline: {pid}")
126
- self.pipeline_manager.start_pipeline(db_pipeline, self.model_manager.get_detector(db_pipeline.ai_model_id))
127
- elif db_pipeline.pipeline_status_code in ['stop', 'restart']:
128
- logging.info(f"⏹️ Stopping pipeline: {pid}")
129
- self.pipeline_manager.stop_pipeline(pid)
130
- if db_pipeline.pipeline_status_code == 'restart':
131
- # This will be picked up by the 'add_new_pipelines' logic in the next cycle
132
- return
133
- else:
134
- processor.update_config(db_pipeline) # Update config for non-running pipelines
135
- return
136
- elif db_pipeline.pipeline_status_code != 'run':
137
- processor.update_config(db_pipeline)
138
- return
139
-
136
+ # At this point, we know db_pipeline.pipeline_status_code == 'run' (checked in caller)
140
137
  # Check for significant changes that require a restart
141
138
  db_detector = self.model_manager.get_detector(db_pipeline.ai_model_id)
142
139
 
@@ -146,7 +143,7 @@ class PipelineSyncThread(threading.Thread):
146
143
  local_detector != db_detector
147
144
  ])
148
145
 
149
- if requires_restart and db_pipeline.pipeline_status_code == 'run':
146
+ if requires_restart:
150
147
  logging.info(f"🔄 Restarting pipeline due to significant changes: {pid}")
151
148
  self.pipeline_manager.stop_pipeline(pid)
152
149
  self.pipeline_manager.start_pipeline(db_pipeline, db_detector)
@@ -29,40 +29,67 @@ class StreamSyncThread(threading.Thread):
29
29
  try:
30
30
  sources = self.worker_source_repo.get_worker_sources()
31
31
  db_sources = {
32
- source.id: (
32
+ source.id:
33
33
  source.url if source.type_code == "live"
34
34
  else source.url if source.type_code == "direct"
35
- else self._get_source_file_path(source.file_path),
36
- source.status_code
37
- ) for source in sources
35
+ else self._get_source_file_path(source.file_path)
36
+ for source in sources
38
37
  } # Store latest sources
38
+
39
+ # Get both active streams and pending streams
39
40
  active_stream_ids = set(self.manager.get_active_stream_ids())
41
+ with self.manager._lock:
42
+ pending_stream_ids = set(self.manager.pending_streams.keys())
43
+ registered_stream_ids = active_stream_ids | pending_stream_ids
40
44
 
41
- # **1️⃣ Add new streams**
42
- for source_id, (url, status_code) in db_sources.items():
43
- if source_id not in active_stream_ids and status_code == "connected":
44
- logging.info(f"🟢 Adding new stream: {source_id} ({url})")
45
- self.manager.add_stream(source_id, url)
45
+ # **1️⃣ Register new streams (lazy loading - don't start them yet)**
46
+ for source_id, url in db_sources.items():
47
+ if source_id not in registered_stream_ids:
48
+ logging.info(f"🟢 Registering new stream: {source_id} ({url})")
49
+ self.manager.register_stream(source_id, url)
46
50
 
47
- # **2️⃣ Remove deleted streams**
48
- for stream_id in active_stream_ids:
49
- if stream_id not in db_sources or db_sources[stream_id][1] != "connected":
50
- logging.info(f"🔴 Removing deleted stream: {stream_id}")
51
- self.manager.remove_stream(stream_id)
51
+ # **2️⃣ Unregister deleted or disconnected streams**
52
+ for stream_id in registered_stream_ids:
53
+ if stream_id not in db_sources:
54
+ logging.info(f"🔴 Unregistering stream: {stream_id}")
55
+ self.manager.unregister_stream(stream_id)
52
56
 
53
- active_stream_ids = set(self.manager.get_active_stream_ids())
57
+ # Refresh registered streams
58
+ with self.manager._lock:
59
+ pending_stream_ids = set(self.manager.pending_streams.keys())
60
+ registered_stream_ids = active_stream_ids | pending_stream_ids
54
61
 
55
62
  # **3️⃣ Update streams if URL has changed**
56
- for source_id, (url, status_code) in db_sources.items():
57
- if source_id in active_stream_ids:
58
- existing_url = self.manager.get_stream_url(source_id)
63
+ for source_id, url in db_sources.items():
64
+ if source_id in registered_stream_ids:
65
+ # Check if it's an active stream or pending stream
66
+ with self.manager._lock:
67
+ is_pending = source_id in self.manager.pending_streams
68
+ if is_pending:
69
+ existing_url = self.manager.pending_streams.get(source_id)
70
+ else:
71
+ existing_url = None
72
+
73
+ if existing_url is None:
74
+ # It's an active stream, get URL from stream manager
75
+ existing_url = self.manager.get_stream_url(source_id)
76
+
77
+ # Only update if URL actually changed
59
78
  if existing_url != url:
60
- logging.info(f"🟡 Updating stream {source_id}: New URL {url}")
61
- self.manager.remove_stream(source_id)
62
- # Add a small delay for device cleanup
63
- if self._is_direct_device(url) or self._is_direct_device(existing_url):
64
- time.sleep(0.5) # Allow device to be properly released
65
- self.manager.add_stream(source_id, url)
79
+ if is_pending:
80
+ # It's pending, just update the URL
81
+ with self.manager._lock:
82
+ self.manager.pending_streams[source_id] = url
83
+ logging.info(f"🟡 Updated pending stream {source_id} URL")
84
+ else:
85
+ # It's active, need to restart it
86
+ logging.info(f"🟡 Updating active stream {source_id}: New URL {url}")
87
+ # Unregister and re-register with new URL
88
+ self.manager.unregister_stream(source_id)
89
+ # Add a small delay for device cleanup
90
+ if self._is_direct_device(url) or self._is_direct_device(existing_url):
91
+ time.sleep(0.5) # Allow device to be properly released
92
+ self.manager.register_stream(source_id, url)
66
93
 
67
94
  except Exception as e:
68
95
  logging.error(f"⚠️ Error syncing streams from database: {e}")
@@ -1,7 +1,6 @@
1
1
  import logging
2
2
  import time
3
3
  import threading
4
- import cv2
5
4
  from typing import Any, Dict, Optional
6
5
 
7
6
  from .VideoStream import VideoStream
@@ -18,6 +17,11 @@ class VideoStreamManager:
18
17
  self.direct_device_streams: Dict[Any, Dict[str, Any]] = {}
19
18
  # Per-direct-device locks: {worker_source_id: threading.Lock}
20
19
  self.direct_device_locks: Dict[Any, threading.Lock] = {}
20
+
21
+ # Reference counting for lazy loading: {worker_source_id: set of pipeline_ids}
22
+ self.stream_references: Dict[Any, set] = {}
23
+ # Store URLs for streams that aren't started yet: {worker_source_id: url}
24
+ self.pending_streams: Dict[Any, str] = {}
21
25
 
22
26
  self.shared_device_manager = SharedVideoDeviceManager()
23
27
 
@@ -38,16 +42,71 @@ class VideoStreamManager:
38
42
  # -----------------------
39
43
  # Public API
40
44
  # -----------------------
41
- def add_stream(self, worker_source_id, url):
42
- """Adds and starts a stream (regular file/RTSP or a shared direct device) if not already present."""
45
+ def register_stream(self, worker_source_id, url):
46
+ """Register a stream URL without starting it (lazy loading)."""
43
47
  with self._lock:
48
+ if worker_source_id not in self.pending_streams:
49
+ self.pending_streams[worker_source_id] = url
50
+ logging.debug(f"📝 Registered stream {worker_source_id} for lazy loading")
51
+
52
+ def unregister_stream(self, worker_source_id):
53
+ """Unregister a stream that's no longer available in the database."""
54
+ with self._lock:
55
+ # If it's pending, remove it
56
+ if worker_source_id in self.pending_streams:
57
+ del self.pending_streams[worker_source_id]
58
+ logging.debug(f"🗑️ Unregistered pending stream {worker_source_id}")
59
+
60
+ # If it's active and has no references, remove it
61
+ if worker_source_id in self.stream_references:
62
+ if len(self.stream_references[worker_source_id]) == 0:
63
+ self._stop_stream(worker_source_id)
64
+
65
+ def acquire_stream(self, worker_source_id, pipeline_id):
66
+ """Request access to a stream for a pipeline. Starts the stream if not already running."""
67
+ with self._lock:
68
+ # Initialize reference set if needed
69
+ if worker_source_id not in self.stream_references:
70
+ self.stream_references[worker_source_id] = set()
71
+
72
+ # Add pipeline reference
73
+ self.stream_references[worker_source_id].add(pipeline_id)
74
+ logging.info(f"🔗 Pipeline {pipeline_id} acquired stream {worker_source_id} (refs: {len(self.stream_references[worker_source_id])})")
75
+
76
+ # If stream is already running, we're done
44
77
  if worker_source_id in self.streams or worker_source_id in self.direct_device_streams:
45
- logging.warning("⚠️ Stream %s is already active.", worker_source_id)
78
+ return True
79
+
80
+ # Get URL from pending streams
81
+ url = self.pending_streams.get(worker_source_id)
82
+ if not url:
83
+ logging.error(f"❌ Cannot acquire stream {worker_source_id}: URL not registered")
84
+ return False
85
+
86
+ # Start the stream (outside lock to avoid blocking)
87
+ return self._start_stream(worker_source_id, url)
88
+
89
+ def release_stream(self, worker_source_id, pipeline_id):
90
+ """Release a stream reference from a pipeline. Stops the stream if no more references."""
91
+ with self._lock:
92
+ if worker_source_id not in self.stream_references:
46
93
  return
47
-
94
+
95
+ # Remove pipeline reference
96
+ self.stream_references[worker_source_id].discard(pipeline_id)
97
+ ref_count = len(self.stream_references[worker_source_id])
98
+ logging.info(f"🔓 Pipeline {pipeline_id} released stream {worker_source_id} (refs: {ref_count})")
99
+
100
+ # If no more references, stop the stream
101
+ if ref_count == 0:
102
+ logging.info(f"💤 Stream {worker_source_id} has no more references, stopping...")
103
+ self._stop_stream(worker_source_id)
104
+
105
+ def _start_stream(self, worker_source_id, url):
106
+ """Internal method to actually start a stream."""
48
107
  if self._is_direct_device(url):
49
108
  self._add_direct_device_stream(worker_source_id, url)
50
- return
109
+ return True
51
110
 
52
111
  # Regular stream
53
112
  stream = VideoStream(url)
@@ -55,18 +114,19 @@ class VideoStreamManager:
55
114
  stream.start() # start thread
56
115
  with self._lock:
57
116
  self.streams[worker_source_id] = stream
58
- logging.info("✅ Added and started video stream: %s", worker_source_id)
117
+ logging.info("✅ Started video stream: %s", worker_source_id)
118
+ return True
59
119
  except Exception as e:
60
120
  logging.error("❌ Failed to start regular stream %s: %s", worker_source_id, e)
61
-
62
- def remove_stream(self, worker_source_id):
63
- """Stops and removes a stream (regular or direct device)."""
64
- if not worker_source_id:
65
- return
66
-
121
+ return False
122
+
123
+ def _stop_stream(self, worker_source_id):
124
+ """Internal method to stop a stream."""
67
125
  # Direct device?
68
126
  with self._lock:
69
127
  is_direct = worker_source_id in self.direct_device_streams
128
+ # Clean up references
129
+ self.stream_references.pop(worker_source_id, None)
70
130
 
71
131
  if is_direct:
72
132
  self._remove_direct_device_stream(worker_source_id)
@@ -77,20 +137,16 @@ class VideoStreamManager:
77
137
  stream = self.streams.pop(worker_source_id, None)
78
138
 
79
139
  if stream is None:
80
- logging.warning("⚠️ Stream %s not found in manager.", worker_source_id)
81
140
  return
82
141
 
83
- logging.info("🛑 Removing video stream: %s", worker_source_id)
142
+ logging.info("🛑 Stopping video stream: %s", worker_source_id)
84
143
  try:
85
- # Expectation: VideoStream.stop() should signal and join internally.
86
144
  stream.stop()
87
145
  except Exception as e:
88
146
  logging.error("❌ Error stopping stream %s: %s", worker_source_id, e)
89
147
  finally:
90
148
  stream = None
91
149
 
92
- logging.info("✅ Stream %s removed successfully.", worker_source_id)
93
-
94
150
  def start_all(self):
95
151
  """Starts all regular streams that are not alive. (Direct devices are publisher-driven.)"""
96
152
  logging.info("🔄 Starting all video streams...")
@@ -114,13 +170,13 @@ class VideoStreamManager:
114
170
 
115
171
  for wid in regular_ids:
116
172
  try:
117
- self.remove_stream(wid)
173
+ self._stop_stream(wid)
118
174
  except Exception as e:
119
175
  logging.error("Error stopping regular stream %s: %s", wid, e)
120
176
 
121
177
  for wid in direct_ids:
122
178
  try:
123
- self.remove_stream(wid)
179
+ self._stop_stream(wid)
124
180
  except Exception as e:
125
181
  logging.error("Error stopping direct device stream %s: %s", wid, e)
126
182
 
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: nedo-vision-worker-core
3
- Version: 0.3.4
3
+ Version: 0.3.5
4
4
  Summary: Nedo Vision Worker Core Library for AI Vision Processing
5
5
  Author-email: Willy Achmat Fauzi <willy.achmat@gmail.com>
6
6
  Maintainer-email: Willy Achmat Fauzi <willy.achmat@gmail.com>
@@ -1,4 +1,4 @@
1
- nedo_vision_worker_core/__init__.py,sha256=0iPoN0eHddv9i7aZ8YJKl2Rn_P_BTsxTJMPggoVd_kc,1924
1
+ nedo_vision_worker_core/__init__.py,sha256=GmU149mopLuJKQ3rEY_-OetVB1aD4BIryIFVa7thDAw,1924
2
2
  nedo_vision_worker_core/cli.py,sha256=8YuKWsIgICUYXE_QtwyU3WzGhVjTWiAo5uzpFOmjNc8,5766
3
3
  nedo_vision_worker_core/core_service.py,sha256=dnHNjbslOeyeWqHDFnk_yKdfTICYzLyRIcuZNwF0Zf4,11323
4
4
  nedo_vision_worker_core/doctor.py,sha256=K_-hVV2-mdEefZ4Cfu5hMCiOxBiI1aXY8VtkkpK80Lc,10651
@@ -59,10 +59,10 @@ nedo_vision_worker_core/models/worker_source_pipeline_debug.py,sha256=6S7TkN37Fr
59
59
  nedo_vision_worker_core/models/worker_source_pipeline_detection.py,sha256=p6CJsiVCKprTYrNxJsiTB8njXdHkjZKVEyBceRVE6fY,560
60
60
  nedo_vision_worker_core/pipeline/ModelManager.py,sha256=K7lmVOo-KL7bnWtyafilZs23bzd6loCgfUz7xuAmlVw,6195
61
61
  nedo_vision_worker_core/pipeline/PipelineConfigManager.py,sha256=X55i9GyXcW9ylO6cj2UMAZFSxxPViacL4H4DZl60CAY,1157
62
- nedo_vision_worker_core/pipeline/PipelineManager.py,sha256=dAYLK5AXXUkfO7cyxLEGGa9SCQTP-9iAO_otffRgVsI,5482
62
+ nedo_vision_worker_core/pipeline/PipelineManager.py,sha256=GGW3fBmDQwNcPJn5yU_BNz7hsR_H7rKoGMImCtC4T9s,6154
63
63
  nedo_vision_worker_core/pipeline/PipelinePrepocessor.py,sha256=cCiVSHHqsKCtKYURdYoEjHJX2GnT6zd8kQ6ZukjQ3V0,1271
64
- nedo_vision_worker_core/pipeline/PipelineProcessor.py,sha256=92ef6fRpibY3VXNeydpD1kprRCRH3fye_cpztAEVrLQ,26350
65
- nedo_vision_worker_core/pipeline/PipelineSyncThread.py,sha256=LX_LXU9MTdD5rP-ElHcYR431_qTXNaKOWcCsE4E9Gd0,8613
64
+ nedo_vision_worker_core/pipeline/PipelineProcessor.py,sha256=4pAcyVpiPM8KClm6hjENQgjDbPH9_na93-ehBDGw_5U,26446
65
+ nedo_vision_worker_core/pipeline/PipelineSyncThread.py,sha256=CAn-4F95J_VAv_7_ClrT-z9226U8ATrOGg5jOYqNuyM,8601
66
66
  nedo_vision_worker_core/pipeline/__init__.py,sha256=Nqnn8clbgv-5l0PgxcTOldg8mkMKrFn4TvPL-rYUUGg,1
67
67
  nedo_vision_worker_core/preprocessing/ImageResizer.py,sha256=RvOazxe6dJQuiy0ZH4lIGbdFfiu0FLUVCHoMvxkDNT4,1324
68
68
  nedo_vision_worker_core/preprocessing/ImageRoi.py,sha256=iO7oQ-SdUSA_kTIVBuq_mdycXsiJNfiFD3J7-VTxiQ4,2141
@@ -81,9 +81,9 @@ nedo_vision_worker_core/services/VideoSharingDaemon.py,sha256=iY6afEKTOsphfHvmZT
81
81
  nedo_vision_worker_core/services/VideoSharingDaemonManager.py,sha256=sc8VZo5iwoOdR8uTiel5BKz6-eZ7wwLy3IwV_3tsAu0,10340
82
82
  nedo_vision_worker_core/streams/RTMPStreamer.py,sha256=Dblfutc1UVHj159KUHFYZ8xFEVhHVknZn_nAqKR6uCs,8695
83
83
  nedo_vision_worker_core/streams/SharedVideoDeviceManager.py,sha256=vSslwxbhKH6FPndR1HcSFIVWtF-iiOQMlSa4VvFa6M4,16265
84
- nedo_vision_worker_core/streams/StreamSyncThread.py,sha256=WmYAY9wFiFhLlxGdnvKGIjAqLwCBayNKdmAWzkbU0jM,3763
84
+ nedo_vision_worker_core/streams/StreamSyncThread.py,sha256=ETT0N_P90ksn6Q5pb7NvMadqCuoicz_g52lcDkHIp88,5382
85
85
  nedo_vision_worker_core/streams/VideoStream.py,sha256=nGtJ4FAZ1Ek-8hVRopEt0bLWLpa10OtyUwdDEuXLObQ,13343
86
- nedo_vision_worker_core/streams/VideoStreamManager.py,sha256=i7uemC3tUdsrAhL8lvoxCYa0fwUh5oAVD6DDaxflXGE,12057
86
+ nedo_vision_worker_core/streams/VideoStreamManager.py,sha256=g5cz-YXPewSubBXxCg4mfzsuGKoOHXu-SrMxaGjYPHw,14956
87
87
  nedo_vision_worker_core/streams/__init__.py,sha256=Nqnn8clbgv-5l0PgxcTOldg8mkMKrFn4TvPL-rYUUGg,1
88
88
  nedo_vision_worker_core/tracker/SFSORT.py,sha256=0kggw0l4yPZ55AKHdqVX6mu9ehHmJed7jcJ3JQoC4sk,14061
89
89
  nedo_vision_worker_core/tracker/TrackerManager.py,sha256=xtDMI657W2s7HM2lMGtwU0x5Hq74BZpLHd-5xk-278I,6152
@@ -96,8 +96,8 @@ nedo_vision_worker_core/util/PersonRestrictedAreaMatcher.py,sha256=iuzCU32BQKaZ3
96
96
  nedo_vision_worker_core/util/PlatformDetector.py,sha256=GGL8UfeMQITR22EMYIRWnuOEnSqo7Dr5mb0PaFrl8AM,3006
97
97
  nedo_vision_worker_core/util/TablePrinter.py,sha256=wzLGgb1GFMeIbAP6HmKcZD33j4D-IlyqlyeR7C5yD7w,1137
98
98
  nedo_vision_worker_core/util/__init__.py,sha256=Nqnn8clbgv-5l0PgxcTOldg8mkMKrFn4TvPL-rYUUGg,1
99
- nedo_vision_worker_core-0.3.4.dist-info/METADATA,sha256=3_UyMqyQc4m_h1qyECq1k7w3vWnN2Hr-ixnLK_BPEp0,14370
100
- nedo_vision_worker_core-0.3.4.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
101
- nedo_vision_worker_core-0.3.4.dist-info/entry_points.txt,sha256=pIPafsvPnBw-fpBKBmc1NQCQ6PQY3ad8mZ6mn8_p5FI,70
102
- nedo_vision_worker_core-0.3.4.dist-info/top_level.txt,sha256=y8kusXjVYqtG8MSHYWTrk8bRrvjOrphKXYyzu943TTQ,24
103
- nedo_vision_worker_core-0.3.4.dist-info/RECORD,,
99
+ nedo_vision_worker_core-0.3.5.dist-info/METADATA,sha256=IkDINswFijRo2j9RTaziXJdaivHiTXDe5M6X5fBAh3o,14370
100
+ nedo_vision_worker_core-0.3.5.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
101
+ nedo_vision_worker_core-0.3.5.dist-info/entry_points.txt,sha256=pIPafsvPnBw-fpBKBmc1NQCQ6PQY3ad8mZ6mn8_p5FI,70
102
+ nedo_vision_worker_core-0.3.5.dist-info/top_level.txt,sha256=y8kusXjVYqtG8MSHYWTrk8bRrvjOrphKXYyzu943TTQ,24
103
+ nedo_vision_worker_core-0.3.5.dist-info/RECORD,,