abstract-block-dumper 0.0.9__py3-none-any.whl → 0.1.1__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,8 +1,10 @@
1
+ from collections.abc import Iterator
1
2
  from datetime import timedelta
2
3
  from typing import Any
3
4
 
4
5
  from django.conf import settings
5
6
  from django.db import transaction
7
+ from django.db.models import Max
6
8
  from django.db.models.query import QuerySet
7
9
  from django.utils import timezone
8
10
 
@@ -10,19 +12,21 @@ import abstract_block_dumper._internal.services.utils as abd_utils
10
12
  import abstract_block_dumper.models as abd_models
11
13
 
12
14
 
13
- def get_ready_to_retry_attempts() -> QuerySet[abd_models.TaskAttempt]:
14
- return abd_models.TaskAttempt.objects.filter(
15
- next_retry_at__isnull=False,
16
- next_retry_at__lte=timezone.now(),
17
- attempt_count__lt=abd_utils.get_max_attempt_limit(),
18
- ).exclude(
19
- status=abd_models.TaskAttempt.Status.SUCCESS,
15
+ def get_ready_to_retry_attempts() -> Iterator[abd_models.TaskAttempt]:
16
+ return (
17
+ abd_models.TaskAttempt.objects.filter(
18
+ next_retry_at__isnull=False,
19
+ next_retry_at__lte=timezone.now(),
20
+ attempt_count__lt=abd_utils.get_max_attempt_limit(),
21
+ )
22
+ .exclude(
23
+ status=abd_models.TaskAttempt.Status.SUCCESS,
24
+ )
25
+ .iterator()
20
26
  )
21
27
 
22
28
 
23
29
  def executed_block_numbers(executable_path: str, args_json: str, from_block: int, to_block: int) -> set[int]:
24
- # Use iterator() to avoid Django's QuerySet caching which causes memory leaks
25
- # during long-running backfill operations
26
30
  block_numbers = (
27
31
  abd_models.TaskAttempt.objects.filter(
28
32
  executable_path=executable_path,
@@ -151,7 +155,5 @@ def task_create_or_get_pending(
151
155
 
152
156
 
153
157
  def get_the_latest_executed_block_number() -> int | None:
154
- qs = abd_models.TaskAttempt.objects.order_by("-block_number").first()
155
- if qs:
156
- return qs.block_number
157
- return None
158
+ result = abd_models.TaskAttempt.objects.aggregate(max_block=Max("block_number"))
159
+ return result["max_block"]
File without changes
@@ -0,0 +1,78 @@
1
+ import bittensor as bt
2
+ import structlog
3
+
4
+ import abstract_block_dumper._internal.services.utils as abd_utils
5
+
6
+ logger = structlog.get_logger(__name__)
7
+
8
+
9
+ # Blocks older than this threshold from current head require archive network
10
+ ARCHIVE_BLOCK_THRESHOLD = 300
11
+
12
+
13
+ class BittensorConnectionClient:
14
+ """
15
+ Manages connections to regular and archive Bittensor subtensor networks.
16
+ """
17
+
18
+ def __init__(self, network: str) -> None:
19
+ self.network = network
20
+ self._subtensor: bt.Subtensor | None = None
21
+ self._archive_subtensor: bt.Subtensor | None = None
22
+ self._current_block_cache: int | None = None
23
+
24
+ def get_for_block(self, block_number: int) -> bt.Subtensor:
25
+ """Get the appropriate subtensor client for the given block number."""
26
+ raise NotImplementedError
27
+
28
+ @property
29
+ def subtensor(self) -> bt.Subtensor:
30
+ """Get the regular subtensor connection, creating it if needed."""
31
+ if self._subtensor is None:
32
+ self._subtensor = abd_utils.get_bittensor_client(self.network)
33
+ return self._subtensor
34
+
35
+ @subtensor.setter
36
+ def subtensor(self, value: bt.Subtensor | None) -> None:
37
+ """Set or reset the subtensor connection."""
38
+ self._subtensor = value
39
+
40
+ @property
41
+ def archive_subtensor(self) -> bt.Subtensor:
42
+ """Get the archive subtensor connection, creating it if needed."""
43
+ if self._archive_subtensor is None:
44
+ self._archive_subtensor = abd_utils.get_bittensor_client("archive")
45
+ return self._archive_subtensor
46
+
47
+ @archive_subtensor.setter
48
+ def archive_subtensor(self, value: bt.Subtensor | None) -> None:
49
+ """Set or reset the archive subtensor connection."""
50
+ self._archive_subtensor = value
51
+
52
+ def get_subtensor_for_block(self, block_number: int) -> bt.Subtensor:
53
+ """
54
+ Get the appropriate subtensor for the given block number.
55
+
56
+ Uses archive network for blocks older than ARCHIVE_BLOCK_THRESHOLD
57
+ from the current head.
58
+ """
59
+ if self._current_block_cache is None:
60
+ self._current_block_cache = self.subtensor.get_current_block()
61
+
62
+ blocks_behind = self._current_block_cache - block_number
63
+
64
+ if blocks_behind > ARCHIVE_BLOCK_THRESHOLD:
65
+ logger.debug(
66
+ "Using archive network for old block",
67
+ block_number=block_number,
68
+ blocks_behind=blocks_behind,
69
+ )
70
+ return self.archive_subtensor
71
+ return self.subtensor
72
+
73
+ def refresh_connections(self) -> None:
74
+ """Reset all subtensor connections to force re-establishment."""
75
+ self._subtensor = None
76
+ self._archive_subtensor = None
77
+ self._current_block_cache = None
78
+ logger.info("Subtensor connections reset")
@@ -15,7 +15,7 @@ import structlog
15
15
 
16
16
  import abstract_block_dumper._internal.dal.django_dal as abd_dal
17
17
  import abstract_block_dumper._internal.services.utils as abd_utils
18
- from abstract_block_dumper._internal.services.block_processor import BlockProcessor, block_processor_factory
18
+ from abstract_block_dumper._internal.services.block_processor import BaseBlockProcessor, block_processor_factory
19
19
  from abstract_block_dumper._internal.services.metrics import (
20
20
  BlockProcessingTimer,
21
21
  increment_archive_network_usage,
@@ -59,7 +59,7 @@ class BackfillScheduler:
59
59
 
60
60
  def __init__(
61
61
  self,
62
- block_processor: BlockProcessor,
62
+ block_processor: BaseBlockProcessor,
63
63
  network: str,
64
64
  from_block: int,
65
65
  to_block: int,
@@ -1,4 +1,6 @@
1
+ import itertools
1
2
  import time
3
+ from typing import Protocol
2
4
 
3
5
  import structlog
4
6
  from django.db import transaction
@@ -12,6 +14,25 @@ from abstract_block_dumper.models import TaskAttempt
12
14
  logger = structlog.get_logger(__name__)
13
15
 
14
16
 
17
+ class BaseBlockProcessor(Protocol):
18
+ """Protocol defining the interface for block processors."""
19
+
20
+ executor: CeleryExecutor
21
+ registry: BaseRegistry
22
+
23
+ def process_block(self, block_number: int) -> None:
24
+ """Process a single block - executes registered tasks for this block only."""
25
+ ...
26
+
27
+ def process_registry_item(self, registry_item: RegistryItem, block_number: int) -> None:
28
+ """Process a single registry item for a given block."""
29
+ ...
30
+
31
+ def recover_failed_retries(self, poll_interval: int, batch_size: int | None = None) -> None:
32
+ """Recover failed tasks that are ready to be retried."""
33
+ ...
34
+
35
+
15
36
  class BlockProcessor:
16
37
  def __init__(self, executor: CeleryExecutor, registry: BaseRegistry) -> None:
17
38
  self.executor = executor
@@ -59,9 +80,9 @@ class BlockProcessor:
59
80
  retry_count = 0
60
81
  retry_attempts = abd_dal.get_ready_to_retry_attempts()
61
82
 
62
- # Apply batch size limit if specified
83
+ # Apply batch size limit if specified (use islice for iterator compatibility)
63
84
  if batch_size is not None:
64
- retry_attempts = retry_attempts[:batch_size]
85
+ retry_attempts = itertools.islice(retry_attempts, batch_size)
65
86
 
66
87
  for retry_attempt in retry_attempts:
67
88
  time.sleep(poll_interval)
@@ -147,7 +168,7 @@ class BlockProcessor:
147
168
  def block_processor_factory(
148
169
  executor: CeleryExecutor | None = None,
149
170
  registry: BaseRegistry | None = None,
150
- ) -> BlockProcessor:
171
+ ) -> BaseBlockProcessor:
151
172
  return BlockProcessor(
152
173
  executor=executor or CeleryExecutor(),
153
174
  registry=registry or task_registry,
@@ -1,12 +1,13 @@
1
1
  import time
2
+ from typing import Protocol
2
3
 
3
- import bittensor as bt
4
4
  import structlog
5
+ from django import db
5
6
  from django.conf import settings
6
7
 
7
8
  import abstract_block_dumper._internal.dal.django_dal as abd_dal
8
- import abstract_block_dumper._internal.services.utils as abd_utils
9
- from abstract_block_dumper._internal.services.block_processor import BlockProcessor, block_processor_factory
9
+ from abstract_block_dumper._internal.providers.bittensor_client import BittensorConnectionClient
10
+ from abstract_block_dumper._internal.services.block_processor import BaseBlockProcessor, block_processor_factory
10
11
  from abstract_block_dumper._internal.services.metrics import (
11
12
  BlockProcessingTimer,
12
13
  increment_blocks_processed,
@@ -15,87 +16,54 @@ from abstract_block_dumper._internal.services.metrics import (
15
16
  set_registered_tasks,
16
17
  )
17
18
 
19
+ # Refresh bittensor connections every N blocks to prevent memory leaks from internal caches
20
+ CONNECTION_REFRESH_INTERVAL = 1000
21
+
18
22
  logger = structlog.get_logger(__name__)
19
23
 
20
- # Blocks older than this threshold from current head require archive network
21
- ARCHIVE_BLOCK_THRESHOLD = 300
24
+
25
+ class BlockStateResolver(Protocol):
26
+ """Protocol defining the interface for block state resolvers."""
27
+
28
+ def get_starting_block(self) -> int:
29
+ """Determine which block to start processing from."""
30
+ ...
31
+
32
+
33
+ class DefaultBlockStateResolver:
34
+ """Default implementation that reads from settings and database."""
35
+
36
+ def __init__(self, bittensor_client: BittensorConnectionClient) -> None:
37
+ self.bittensor_client = bittensor_client
38
+
39
+ def get_starting_block(self) -> int:
40
+ start_setting = getattr(settings, "BLOCK_DUMPER_START_FROM_BLOCK", None)
41
+ if start_setting == "current":
42
+ return self.bittensor_client.subtensor.get_current_block()
43
+ if isinstance(start_setting, int):
44
+ return start_setting
45
+ # Default: resume from DB or current
46
+ return abd_dal.get_the_latest_executed_block_number() or self.bittensor_client.subtensor.get_current_block()
22
47
 
23
48
 
24
49
  class TaskScheduler:
25
50
  def __init__(
26
51
  self,
27
- block_processor: BlockProcessor,
28
- network: str,
52
+ block_processor: BaseBlockProcessor,
53
+ bittensor_client: BittensorConnectionClient,
54
+ state_resolver: BlockStateResolver,
29
55
  poll_interval: int,
30
- realtime_head_only: bool = False,
31
56
  ) -> None:
32
57
  self.block_processor = block_processor
33
- self.network = network
34
58
  self.poll_interval = poll_interval
35
- self.realtime_head_only = realtime_head_only
36
- self.last_processed_block = -1
59
+ self.bittensor_client = bittensor_client
60
+ self.last_processed_block = state_resolver.get_starting_block()
37
61
  self.is_running = False
38
- self._subtensor: bt.Subtensor | None = None
39
- self._archive_subtensor: bt.Subtensor | None = None
40
- self._current_block_cache: int | None = None
41
-
42
- @property
43
- def subtensor(self) -> bt.Subtensor:
44
- """Get the regular subtensor connection, creating it if needed."""
45
- if self._subtensor is None:
46
- self._subtensor = abd_utils.get_bittensor_client(self.network)
47
- return self._subtensor
48
-
49
- @subtensor.setter
50
- def subtensor(self, value: bt.Subtensor | None) -> None:
51
- """Set or reset the subtensor connection."""
52
- self._subtensor = value
53
-
54
- @property
55
- def archive_subtensor(self) -> bt.Subtensor:
56
- """Get the archive subtensor connection, creating it if needed."""
57
- if self._archive_subtensor is None:
58
- self._archive_subtensor = abd_utils.get_bittensor_client("archive")
59
- return self._archive_subtensor
60
-
61
- @archive_subtensor.setter
62
- def archive_subtensor(self, value: bt.Subtensor | None) -> None:
63
- """Set or reset the archive subtensor connection."""
64
- self._archive_subtensor = value
65
-
66
- def get_subtensor_for_block(self, block_number: int) -> bt.Subtensor:
67
- """
68
- Get the appropriate subtensor for the given block number.
69
-
70
- Uses archive network for blocks older than ARCHIVE_BLOCK_THRESHOLD
71
- from the current head.
72
- """
73
- if self._current_block_cache is None:
74
- self._current_block_cache = self.subtensor.get_current_block()
75
-
76
- blocks_behind = self._current_block_cache - block_number
77
-
78
- if blocks_behind > ARCHIVE_BLOCK_THRESHOLD:
79
- logger.debug(
80
- "Using archive network for old block",
81
- block_number=block_number,
82
- blocks_behind=blocks_behind,
83
- )
84
- return self.archive_subtensor
85
- return self.subtensor
86
-
87
- def refresh_connections(self) -> None:
88
- """Reset all subtensor connections to force re-establishment."""
89
- self._subtensor = None
90
- self._archive_subtensor = None
91
- self._current_block_cache = None
92
- logger.info("Subtensor connections reset")
62
+ self._blocks_since_refresh = 0
93
63
 
94
64
  def start(self) -> None:
95
65
  self.is_running = True
96
66
 
97
- self.initialize_last_block()
98
-
99
67
  registered_tasks_count = len(self.block_processor.registry.get_functions())
100
68
  set_registered_tasks(registered_tasks_count)
101
69
 
@@ -107,76 +75,47 @@ class TaskScheduler:
107
75
 
108
76
  while self.is_running:
109
77
  try:
110
- if self._current_block_cache is not None:
111
- self.subtensor = self.get_subtensor_for_block(self._current_block_cache)
112
-
113
- # Update current block cache for archive network decision
114
- self._current_block_cache = self.subtensor.get_current_block()
115
- current_block = self._current_block_cache
116
-
117
- if self.realtime_head_only:
118
- # Only process the current head block, skip if already processed
119
- if current_block != self.last_processed_block:
120
- with BlockProcessingTimer(mode="realtime"):
121
- self.block_processor.process_block(current_block)
122
-
123
- set_current_block("realtime", current_block)
124
- increment_blocks_processed("realtime")
125
- set_block_lag("realtime", 0) # Head-only mode has no lag
126
- self.last_processed_block = current_block
127
-
128
- time.sleep(self.poll_interval)
129
- else:
130
- # Original behavior: process all blocks from last_processed to current
131
- for block_number in range(self.last_processed_block + 1, current_block + 1):
132
- with BlockProcessingTimer(mode="realtime"):
133
- self.block_processor.process_block(block_number)
134
-
135
- # Update metrics
136
- set_current_block("realtime", block_number)
137
- increment_blocks_processed("realtime")
138
- set_block_lag("realtime", current_block - block_number)
139
-
140
- time.sleep(self.poll_interval)
141
- self.last_processed_block = block_number
78
+ current_block = self.bittensor_client.subtensor.get_current_block()
79
+
80
+ # Only process the current head block, skip if already processed
81
+ if current_block != self.last_processed_block:
82
+ with BlockProcessingTimer(mode="realtime"):
83
+ self.block_processor.process_block(current_block)
84
+
85
+ set_current_block("realtime", current_block)
86
+ increment_blocks_processed("realtime")
87
+ set_block_lag("realtime", 0) # Head-only mode has no lag
88
+ self.last_processed_block = current_block
89
+ self._blocks_since_refresh += 1
90
+
91
+ # Periodic memory cleanup
92
+ if self._blocks_since_refresh >= CONNECTION_REFRESH_INTERVAL:
93
+ self._perform_cleanup()
94
+
95
+ time.sleep(self.poll_interval)
142
96
 
143
97
  except KeyboardInterrupt:
144
98
  logger.info("TaskScheduler stopping due to KeyboardInterrupt.")
145
99
  self.stop()
146
100
  break
147
101
  except Exception:
148
- logger.error("Fatal scheduler error", exc_info=True)
149
- # resume the loop even if task failed
102
+ logger.exception("Error in TaskScheduler loop")
150
103
  time.sleep(self.poll_interval)
151
104
 
152
105
  def stop(self) -> None:
153
106
  self.is_running = False
154
107
  logger.info("TaskScheduler stopped.")
155
108
 
156
- def initialize_last_block(self) -> None:
157
- # Safe getattr in case setting is not defined
158
- start_from_block_setting = getattr(settings, "BLOCK_DUMPER_START_FROM_BLOCK", None)
159
-
160
- if start_from_block_setting is not None:
161
- if start_from_block_setting == "current":
162
- self.last_processed_block = self.subtensor.get_current_block()
163
- logger.info("Starting from current blockchain block", block_number=self.last_processed_block)
109
+ def _perform_cleanup(self) -> None:
110
+ """Perform periodic memory cleanup to prevent leaks in long-running processes."""
111
+ # Reset bittensor connections to clear internal caches
112
+ self.bittensor_client.refresh_connections()
164
113
 
165
- elif isinstance(start_from_block_setting, int):
166
- self.last_processed_block = start_from_block_setting
167
- logger.info("Starting from configured block", block_number=self.last_processed_block)
168
- else:
169
- error_msg = f"Invalid BLOCK_DUMPER_START_FROM_BLOCK value: {start_from_block_setting}"
170
- raise ValueError(error_msg)
171
- else:
172
- # Default behavior - resume from database
173
- last_block_number = abd_dal.get_the_latest_executed_block_number()
114
+ # Clear Django's query log (only accumulates if DEBUG=True)
115
+ db.reset_queries()
174
116
 
175
- self.last_processed_block = last_block_number or self.subtensor.get_current_block()
176
- logger.info(
177
- "Resume from the last database block or start from the current block",
178
- last_processed_block=self.last_processed_block,
179
- )
117
+ self._blocks_since_refresh = 0
118
+ logger.debug("Memory cleanup performed", blocks_processed=CONNECTION_REFRESH_INTERVAL)
180
119
 
181
120
 
182
121
  def task_scheduler_factory(network: str = "finney") -> TaskScheduler:
@@ -187,9 +126,11 @@ def task_scheduler_factory(network: str = "finney") -> TaskScheduler:
187
126
  network (str): Bittensor network name. Defaults to "finney"
188
127
 
189
128
  """
129
+ bittensor_client = BittensorConnectionClient(network=network)
130
+ state_resolver = DefaultBlockStateResolver(bittensor_client=bittensor_client)
190
131
  return TaskScheduler(
191
132
  block_processor=block_processor_factory(),
192
- network=network,
193
- poll_interval=getattr(settings, "BLOCK_DUMPER_POLL_INTERVAL", 1),
194
- realtime_head_only=getattr(settings, "BLOCK_DUMPER_REALTIME_HEAD_ONLY", True),
133
+ poll_interval=getattr(settings, "BLOCK_DUMPER_POLL_INTERVAL", 5),
134
+ bittensor_client=bittensor_client,
135
+ state_resolver=state_resolver,
195
136
  )
@@ -17,7 +17,7 @@ def get_bittensor_client(network: str = "finney") -> bt.Subtensor:
17
17
  doesn't change during runtime.
18
18
  """
19
19
  logger.info("Creating new bittensor client for network", network=network)
20
- return bt.subtensor(network=network)
20
+ return bt.Subtensor(network=network)
21
21
 
22
22
 
23
23
  def get_current_celery_task_id() -> str:
@@ -28,7 +28,7 @@ version_tuple: VERSION_TUPLE
28
28
  commit_id: COMMIT_ID
29
29
  __commit_id__: COMMIT_ID
30
30
 
31
- __version__ = version = '0.0.9'
32
- __version_tuple__ = version_tuple = (0, 0, 9)
31
+ __version__ = version = '0.1.1'
32
+ __version_tuple__ = version_tuple = (0, 1, 1)
33
33
 
34
34
  __commit_id__ = commit_id = None
@@ -74,7 +74,9 @@ def schedule_retry(task_attempt: TaskAttempt) -> None:
74
74
 
75
75
 
76
76
  def _celery_task_wrapper(
77
- func: Callable[..., Any], block_number: int, **kwargs: dict[str, Any]
77
+ func: Callable[..., Any],
78
+ block_number: int,
79
+ **kwargs: dict[str, Any],
78
80
  ) -> dict[str, Any] | None:
79
81
  executable_path = abd_utils.get_executable_path(func)
80
82
 
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: abstract-block-dumper
3
- Version: 0.0.9
3
+ Version: 0.1.1
4
4
  Project-URL: Source, https://github.com/bactensor/abstract-block-dumper
5
5
  Project-URL: Issue Tracker, https://github.com/bactensor/abstract-block-dumper/issues
6
6
  Author-email: Reef Technologies <opensource@reef.pl>
@@ -1,5 +1,5 @@
1
1
  abstract_block_dumper/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
2
- abstract_block_dumper/_version.py,sha256=X2FLFwBoUmgJPsOV-1l-SxIXNSTX1TQ036Kf2j9Mc68,704
2
+ abstract_block_dumper/_version.py,sha256=m8HxkqoKGw_wAJtc4ZokpJKNLXqp4zwnNhbnfDtro7w,704
3
3
  abstract_block_dumper/admin.py,sha256=3J3I_QOKFgfMNpTXW-rTQGO_q5Ls6uNuL0FkPVdIsYg,1654
4
4
  abstract_block_dumper/apps.py,sha256=DXATdrjsL3T2IletTbKeD6unr8ScLaxg7wz0nAHTAns,215
5
5
  abstract_block_dumper/models.py,sha256=MO9824dmHB6xF3PrFE_RERh7whVjQtS4tt6QA0wSbg0,2022
@@ -8,15 +8,17 @@ abstract_block_dumper/_internal/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm
8
8
  abstract_block_dumper/_internal/discovery.py,sha256=sISOL8vq6rC0pOndrCfWKDZjyYwzzZIChG-BH9mteq0,745
9
9
  abstract_block_dumper/_internal/exceptions.py,sha256=jVXQ8b3gneno2XYvO0XisJPMlkAWb6H5u10egIpPJ4k,335
10
10
  abstract_block_dumper/_internal/dal/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
11
- abstract_block_dumper/_internal/dal/django_dal.py,sha256=i9jocanfptjXw5lfE2xBYvx5mo1g98IoMjlS-WjGP88,5623
11
+ abstract_block_dumper/_internal/dal/django_dal.py,sha256=QbDsikUthIAhVC_FwSynUUdQL3OWlCo3_Cg65M91Cb4,5618
12
12
  abstract_block_dumper/_internal/dal/memory_registry.py,sha256=m9Yms-cuemi9_5q_Kn_zsJnxDPEiuAUkESIAltD60QY,2943
13
+ abstract_block_dumper/_internal/providers/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
14
+ abstract_block_dumper/_internal/providers/bittensor_client.py,sha256=hkUdkhz8wOZInRloyySo3DDN4JSMEiPwLv0YV5IDZ7o,2798
13
15
  abstract_block_dumper/_internal/services/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
14
- abstract_block_dumper/_internal/services/backfill_scheduler.py,sha256=XgsVYXaz6pR4PBA9giesjhR74x1qLX2281-eQgM5qhs,16311
15
- abstract_block_dumper/_internal/services/block_processor.py,sha256=NC7p1oD38FpaZb6EbykBolP32uY069abumOvXrjOBV0,6644
16
+ abstract_block_dumper/_internal/services/backfill_scheduler.py,sha256=uZqM4bSoMgNRKwBy-W3a7ydh2Z37Sum8cDRGo4T4KhA,16319
17
+ abstract_block_dumper/_internal/services/block_processor.py,sha256=P8_LZR4ZSyNKbtnqFbAUkpT1XmEh9yX5Pgj5J__IwdA,7409
16
18
  abstract_block_dumper/_internal/services/executor.py,sha256=WhpHhOAi4cI-qdEE8-DSt9xZwooOpSc9_uDMQBBoHUM,2317
17
19
  abstract_block_dumper/_internal/services/metrics.py,sha256=Gg-PQYZ98caaS52wm1EqhtPURXlfrVjk2t3-8nccqfo,7821
18
- abstract_block_dumper/_internal/services/scheduler.py,sha256=B62ZoOxSv8i5mverBWv67BzoyBd14f0wWnNQ329pJds,7770
19
- abstract_block_dumper/_internal/services/utils.py,sha256=QSs2hBHWOPgNgKPf_ZmADXuqEiqK5mWZp7JblvQgxZQ,1140
20
+ abstract_block_dumper/_internal/services/scheduler.py,sha256=pkcOlh0XsR-3sStEtTbiSpHFJhMG9tGIgSUFeH-FITk,5020
21
+ abstract_block_dumper/_internal/services/utils.py,sha256=QZxdQyWIcUnezyVmegS4g3x3BoB3-oijYJ_i9nLQWHo,1140
20
22
  abstract_block_dumper/management/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
21
23
  abstract_block_dumper/management/commands/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
22
24
  abstract_block_dumper/management/commands/backfill_blocks_v1.py,sha256=EmNUozAZn8uThjCvusZe7poNrw9RYy-MafMg2wu3XeQ,6392
@@ -25,8 +27,8 @@ abstract_block_dumper/migrations/0001_initial.py,sha256=ImPHC3G6kPkq4Xn_4YVAm4La
25
27
  abstract_block_dumper/migrations/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
26
28
  abstract_block_dumper/v1/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
27
29
  abstract_block_dumper/v1/celery.py,sha256=X4IqVs5i6ZpyY7fy1SqMZgsZy4SXP-jK2qG-FYnjU38,1722
28
- abstract_block_dumper/v1/decorators.py,sha256=4FXsBScT_5Wadl8FadRcZZtyLytZUzSTX4V5DI0IuRs,9820
30
+ abstract_block_dumper/v1/decorators.py,sha256=yQglsy1dU1u7ShwaTqZLahDcybHmetibTIOi53o_ZOM,9829
29
31
  abstract_block_dumper/v1/tasks.py,sha256=u9iMYdDUqzYT3yPrNwZecHnlweZ3yFipV9BcIWHCbus,2647
30
- abstract_block_dumper-0.0.9.dist-info/METADATA,sha256=xWskTf2HCMNuFGA5M2KSyyejjBLKgQ6OBv2Q5VuL5_I,12993
31
- abstract_block_dumper-0.0.9.dist-info/WHEEL,sha256=WLgqFyCfm_KASv4WHyYy0P3pM_m7J5L9k2skdKLirC8,87
32
- abstract_block_dumper-0.0.9.dist-info/RECORD,,
32
+ abstract_block_dumper-0.1.1.dist-info/METADATA,sha256=CVrs2LOv57Y4JOGxJgKsr1Xnnhl0i0wa534T-0hTU58,12993
33
+ abstract_block_dumper-0.1.1.dist-info/WHEEL,sha256=WLgqFyCfm_KASv4WHyYy0P3pM_m7J5L9k2skdKLirC8,87
34
+ abstract_block_dumper-0.1.1.dist-info/RECORD,,