cobweb-launcher 0.0.1__tar.gz

Sign up to get free protection for your applications and to get access to all the features.
Files changed (42) hide show
  1. cobweb-launcher-0.0.1/LICENSE +21 -0
  2. cobweb-launcher-0.0.1/PKG-INFO +41 -0
  3. cobweb-launcher-0.0.1/README.md +27 -0
  4. cobweb-launcher-0.0.1/cobweb/__init__.py +2 -0
  5. cobweb-launcher-0.0.1/cobweb/base/__init__.py +0 -0
  6. cobweb-launcher-0.0.1/cobweb/base/bbb.py +187 -0
  7. cobweb-launcher-0.0.1/cobweb/base/config.py +164 -0
  8. cobweb-launcher-0.0.1/cobweb/base/decorators.py +95 -0
  9. cobweb-launcher-0.0.1/cobweb/base/hash_table.py +60 -0
  10. cobweb-launcher-0.0.1/cobweb/base/interface.py +44 -0
  11. cobweb-launcher-0.0.1/cobweb/base/log.py +96 -0
  12. cobweb-launcher-0.0.1/cobweb/base/queue_tmp.py +60 -0
  13. cobweb-launcher-0.0.1/cobweb/base/request.py +62 -0
  14. cobweb-launcher-0.0.1/cobweb/base/task.py +38 -0
  15. cobweb-launcher-0.0.1/cobweb/base/utils.py +15 -0
  16. cobweb-launcher-0.0.1/cobweb/db/__init__.py +0 -0
  17. cobweb-launcher-0.0.1/cobweb/db/base/__init__.py +0 -0
  18. cobweb-launcher-0.0.1/cobweb/db/base/client_db.py +1 -0
  19. cobweb-launcher-0.0.1/cobweb/db/base/oss_db.py +116 -0
  20. cobweb-launcher-0.0.1/cobweb/db/base/redis_db.py +214 -0
  21. cobweb-launcher-0.0.1/cobweb/db/base/redis_dbv3.py +231 -0
  22. cobweb-launcher-0.0.1/cobweb/db/scheduler/__init__.py +0 -0
  23. cobweb-launcher-0.0.1/cobweb/db/scheduler/default.py +8 -0
  24. cobweb-launcher-0.0.1/cobweb/db/scheduler/textfile.py +29 -0
  25. cobweb-launcher-0.0.1/cobweb/db/storer/__init__.py +0 -0
  26. cobweb-launcher-0.0.1/cobweb/db/storer/console.py +10 -0
  27. cobweb-launcher-0.0.1/cobweb/db/storer/loghub.py +55 -0
  28. cobweb-launcher-0.0.1/cobweb/db/storer/redis.py +16 -0
  29. cobweb-launcher-0.0.1/cobweb/db/storer/textfile.py +16 -0
  30. cobweb-launcher-0.0.1/cobweb/distributed/__init__.py +0 -0
  31. cobweb-launcher-0.0.1/cobweb/distributed/launcher.py +194 -0
  32. cobweb-launcher-0.0.1/cobweb/distributed/models.py +140 -0
  33. cobweb-launcher-0.0.1/cobweb/single/__init__.py +0 -0
  34. cobweb-launcher-0.0.1/cobweb/single/models.py +104 -0
  35. cobweb-launcher-0.0.1/cobweb/single/nest.py +153 -0
  36. cobweb-launcher-0.0.1/cobweb_launcher.egg-info/PKG-INFO +41 -0
  37. cobweb-launcher-0.0.1/cobweb_launcher.egg-info/SOURCES.txt +40 -0
  38. cobweb-launcher-0.0.1/cobweb_launcher.egg-info/dependency_links.txt +1 -0
  39. cobweb-launcher-0.0.1/cobweb_launcher.egg-info/requires.txt +4 -0
  40. cobweb-launcher-0.0.1/cobweb_launcher.egg-info/top_level.txt +1 -0
  41. cobweb-launcher-0.0.1/setup.cfg +4 -0
  42. cobweb-launcher-0.0.1/setup.py +26 -0
@@ -0,0 +1,21 @@
1
+ MIT License
2
+
3
+ Copyright (c) 2024 Juannie
4
+
5
+ Permission is hereby granted, free of charge, to any person obtaining a copy
6
+ of this software and associated documentation files (the "Software"), to deal
7
+ in the Software without restriction, including without limitation the rights
8
+ to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
9
+ copies of the Software, and to permit persons to whom the Software is
10
+ furnished to do so, subject to the following conditions:
11
+
12
+ The above copyright notice and this permission notice shall be included in all
13
+ copies or substantial portions of the Software.
14
+
15
+ THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16
+ IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17
+ FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
18
+ AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19
+ LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
20
+ OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
21
+ SOFTWARE.
@@ -0,0 +1,41 @@
1
+ Metadata-Version: 2.1
2
+ Name: cobweb-launcher
3
+ Version: 0.0.1
4
+ Summary: spider_hole
5
+ Home-page: https://github.com/Juannie-PP/cobweb
6
+ Author: Juannie-PP
7
+ Author-email: 2604868278@qq.com
8
+ License: MIT
9
+ Keywords: cobweb
10
+ Classifier: Programming Language :: Python :: 3
11
+ Requires-Python: >=3.7
12
+ Description-Content-Type: text/markdown
13
+ License-File: LICENSE
14
+
15
+ # cobweb
16
+
17
+ > 通用爬虫框架: 1.单机模式采集框架;2.分布式采集框架
18
+ >
19
+ > 5部分
20
+ >
21
+ > 1. starter -- 启动器
22
+ >
23
+ > 2. scheduler -- 调度器
24
+ >
25
+ > 3. distributor -- 分发器
26
+ >
27
+ > 4. storer -- 存储器
28
+ >
29
+ > 5. utils -- 工具函数
30
+ >
31
+
32
+ need deal
33
+ - 队列优化完善,使用queue的机制wait()同步各模块执行?
34
+ - 日志功能完善,单机模式调度和保存数据写入文件,结构化输出各任务日志
35
+ - 去重过滤(布隆过滤器等)
36
+ - 防丢失(单机模式可以通过日志文件进行检查种子)
37
+ - 自定义数据库的功能
38
+ - excel、mysql、redis数据完善
39
+
40
+
41
+ ![img.png](https://image-luyuan.oss-cn-hangzhou.aliyuncs.com/image/D2388CDC-B9E5-4CE4-9F2C-7D173763B6A8.png)
@@ -0,0 +1,27 @@
1
+ # cobweb
2
+
3
+ > 通用爬虫框架: 1.单机模式采集框架;2.分布式采集框架
4
+ >
5
+ > 5部分
6
+ >
7
+ > 1. starter -- 启动器
8
+ >
9
+ > 2. scheduler -- 调度器
10
+ >
11
+ > 3. distributor -- 分发器
12
+ >
13
+ > 4. storer -- 存储器
14
+ >
15
+ > 5. utils -- 工具函数
16
+ >
17
+
18
+ need deal
19
+ - 队列优化完善,使用queue的机制wait()同步各模块执行?
20
+ - 日志功能完善,单机模式调度和保存数据写入文件,结构化输出各任务日志
21
+ - 去重过滤(布隆过滤器等)
22
+ - 防丢失(单机模式可以通过日志文件进行检查种子)
23
+ - 自定义数据库的功能
24
+ - excel、mysql、redis数据完善
25
+
26
+
27
+ ![img.png](https://image-luyuan.oss-cn-hangzhou.aliyuncs.com/image/D2388CDC-B9E5-4CE4-9F2C-7D173763B6A8.png)
@@ -0,0 +1,2 @@
1
+ from .distributed import launcher
2
+ from .base import task
File without changes
@@ -0,0 +1,187 @@
1
+ # from typing import Iterable
2
+ import json
3
+ import time
4
+ import hashlib
5
+ from log import log
6
+ from utils import struct_queue_name
7
+ from collections import deque, namedtuple
8
+
9
+
10
+ class Queue:
11
+
12
+ def __init__(self):
13
+ self._queue = deque()
14
+
15
+ @property
16
+ def length(self) -> int:
17
+ return len(self._queue)
18
+ #
19
+ # @property
20
+ # def queue_names(self):
21
+ # return tuple(self.__dict__.keys())
22
+ #
23
+ # @property
24
+ # def used_memory(self):
25
+ # return asizeof.asizeof(self)
26
+
27
+ # def create_queue(self, queue_name: str):
28
+ # self.__setattr__(queue_name, deque())
29
+
30
+ # def push_seed(self, seed):
31
+ # self.push("_seed_queue", seed)
32
+
33
+ # def pop_seed(self):
34
+ # return self.pop("_seed_queue")
35
+
36
+ def push(self, data, left: bool = False, direct_insertion: bool = False):
37
+ try:
38
+ if not data:
39
+ return None
40
+ if direct_insertion or isinstance(data, Seed):
41
+ self._queue.appendleft(data) if left else self._queue.append(data)
42
+ elif any(isinstance(data, t) for t in (list, tuple)):
43
+ self._queue.extendleft(data) if left else self._queue.extend(data)
44
+ except AttributeError as e:
45
+ log.exception(e)
46
+
47
+ def pop(self, left: bool = True):
48
+ try:
49
+ return self._queue.popleft() if left else self._queue.pop()
50
+ except IndexError:
51
+ return None
52
+ except AttributeError as e:
53
+ log.exception(e)
54
+ return None
55
+
56
+
57
+ class Seed:
58
+
59
+ def __init__(
60
+ self,
61
+ seed_info=None,
62
+ priority=300,
63
+ version=None,
64
+ retry=0,
65
+ **kwargs
66
+ ):
67
+ if seed_info:
68
+ if any(isinstance(seed_info, t) for t in (str, bytes)):
69
+ try:
70
+ item = json.loads(seed_info)
71
+ for k, v in item.items():
72
+ self.__setattr__(k, v)
73
+ except json.JSONDecodeError:
74
+ self.__setattr__("url", seed_info)
75
+ elif isinstance(seed_info, dict):
76
+ for k, v in seed_info.items():
77
+ self.__setattr__(k, v)
78
+ else:
79
+ raise TypeError(Exception(
80
+ f"seed type error, "
81
+ f"must be str or dict! "
82
+ f"seed_info: {seed_info}"
83
+ ))
84
+ for k, v in kwargs.items():
85
+ self.__setattr__(k, v)
86
+ if not getattr(self, "_priority"):
87
+ self._priority = min(max(1, int(priority)), 999)
88
+ if not getattr(self, "_version"):
89
+ self._version = int(version) or int(time.time())
90
+ if not getattr(self, "_retry"):
91
+ self._retry = retry
92
+ if not getattr(self, "sid"):
93
+ self.init_id()
94
+
95
+ def init_id(self):
96
+ item_string = self.format_seed
97
+ seed_id = hashlib.md5(item_string.encode()).hexdigest()
98
+ self.__setattr__("sid", seed_id)
99
+
100
+ def __setitem__(self, key, value):
101
+ setattr(self, key, value)
102
+
103
+ def __getitem__(self, item):
104
+ return getattr(self, item)
105
+
106
+ def __getattr__(self, name):
107
+ return None
108
+
109
+ def __str__(self):
110
+ return json.dumps(self.__dict__, ensure_ascii=False)
111
+
112
+ def __repr__(self):
113
+ chars = [f"{k}={v}" for k, v in self.__dict__.items()]
114
+ return f'{self.__class__.__name__}({", ".join(chars)})'
115
+
116
+ @property
117
+ def format_seed(self):
118
+ seed = self.__dict__.copy()
119
+ del seed["_priority"]
120
+ del seed["_version"]
121
+ del seed["_retry"]
122
+ return json.dumps(seed, ensure_ascii=False)
123
+
124
+
125
+ class DBItem:
126
+
127
+ def __init__(self, **kwargs):
128
+ self.__setattr__("_index", 0, True)
129
+ for table in self.__class__.__table__:
130
+ if set(kwargs.keys()) == set(table._fields):
131
+ break
132
+ self._index += 1
133
+
134
+ if self._index > len(self.__class__.__table__):
135
+ raise Exception()
136
+
137
+ table = self.__class__.__table__[self._index]
138
+ self.__setattr__("struct_data", table(**kwargs), True)
139
+ self.__setattr__("db_name", self.__class__.__name__, True)
140
+ self.__setattr__("table_name", self.struct_data.__class__.__name__, True)
141
+
142
+ @classmethod
143
+ def init_item(cls, table_name, fields):
144
+ queue_name = struct_queue_name(cls.__name__, table_name)
145
+ if getattr(cls, queue_name, None) is None:
146
+ setattr(cls, queue_name, Queue())
147
+
148
+ if getattr(cls, "__table__", None) is None:
149
+ cls.__table__ = []
150
+
151
+ table = namedtuple(table_name, fields)
152
+
153
+ if table in getattr(cls, "__table__"):
154
+ raise Exception()
155
+ getattr(cls, "__table__").append(table)
156
+
157
+ def queue(self):
158
+ queue_name = struct_queue_name(self.db_name, self.table_name)
159
+ return getattr(self.__class__, queue_name)
160
+
161
+ def __setitem__(self, key, value):
162
+ self.__setattr__(key, value)
163
+
164
+ def __getitem__(self, item):
165
+ return self.struct_data[item]
166
+
167
+ def __getattr__(self, name):
168
+ return None
169
+
170
+ def __setattr__(self, key, value, init=None):
171
+ if init:
172
+ super().__setattr__(key, value)
173
+ elif not getattr(self, "struct_data"):
174
+ raise Exception(f"no struct_data")
175
+ else:
176
+ self.__setattr__(
177
+ "struct_data",
178
+ self.struct_data._replace(**{key: value}),
179
+ init=True
180
+ )
181
+
182
+ def __str__(self):
183
+ return json.dumps(self.struct_data._asdict(), ensure_ascii=False)
184
+
185
+ def __repr__(self):
186
+ return f'{self.__class__.__name__}:{self.struct_data}'
187
+
@@ -0,0 +1,164 @@
1
+ import json
2
+ from collections import namedtuple
3
+ from utils import struct_table_name
4
+
5
+ StorerInfo = namedtuple(
6
+ "StorerInfo",
7
+ "DB, table, fields, length, config"
8
+ )
9
+ SchedulerInfo = namedtuple(
10
+ "SchedulerInfo",
11
+ "DB, table, sql, length, size, config",
12
+ )
13
+ RedisInfo = namedtuple(
14
+ "RedisInfo",
15
+ "host, port, username, password, db",
16
+ defaults=("localhost", 6379, None, None, 0)
17
+ )
18
+
19
+ # redis_info = dict(
20
+ # host="localhost",
21
+ # port=6379,
22
+ # username=None,
23
+ # password=None,
24
+ # db=0
25
+ # )
26
+
27
+
28
+ class SchedulerDB:
29
+
30
+ @staticmethod
31
+ def default():
32
+ from db.scheduler.default import Default
33
+ return SchedulerInfo(DB=Default, table="", sql="", length=100, size=500000, config=None)
34
+
35
+ @staticmethod
36
+ def textfile(table, sql=None, length=100, size=500000):
37
+ from db.scheduler.textfile import Textfile
38
+ return SchedulerInfo(DB=Textfile, table=table, sql=sql, length=length, size=size, config=None)
39
+
40
+ @staticmethod
41
+ def diy(DB, table, sql=None, length=100, size=500000, config=None):
42
+ from base.interface import SchedulerInterface
43
+ if not isinstance(DB, SchedulerInterface):
44
+ raise Exception("DB must be inherit from SchedulerInterface")
45
+ return SchedulerInfo(DB=DB, table=table, sql=sql, length=length, size=size, config=config)
46
+
47
+ # @staticmethod
48
+ # def info(scheduler_info):
49
+ # if not scheduler_info:
50
+ # return SchedulerDB.default()
51
+ #
52
+ # if isinstance(scheduler_info, SchedulerInfo):
53
+ # return scheduler_info
54
+ #
55
+ # if isinstance(scheduler_info, str):
56
+ # scheduler = json.loads(scheduler_info)
57
+ # if isinstance(scheduler, dict):
58
+ # db_name = scheduler["DB"]
59
+ # if db_name in dir(SchedulerDB):
60
+ # del scheduler["DB"]
61
+ # else:
62
+ # db_name = "diy"
63
+ # func = getattr(SchedulerDB, db_name)
64
+ # return func(**scheduler)
65
+
66
+
67
+ class StorerDB:
68
+
69
+ @staticmethod
70
+ def console(table, fields, length=200):
71
+ from db.storer.console import Console
72
+ table = struct_table_name(table)
73
+ return StorerInfo(DB=Console, table=table, fields=fields, length=length, config=None)
74
+
75
+ @staticmethod
76
+ def textfile(table, fields, length=200):
77
+ from db.storer.textfile import Textfile
78
+ table = struct_table_name(table)
79
+ return StorerInfo(DB=Textfile, table=table, fields=fields, length=length, config=None)
80
+
81
+ @staticmethod
82
+ def loghub(table, fields, length=200, config=None):
83
+ from db.storer.loghub import Loghub
84
+ table = struct_table_name(table)
85
+ return StorerInfo(DB=Loghub, table=table, fields=fields, length=length, config=config)
86
+
87
+ @staticmethod
88
+ def diy(DB, table, fields, length=200, config=None):
89
+ from base.interface import StorerInterface
90
+ if not isinstance(DB, StorerInterface):
91
+ raise Exception("DB must be inherit from StorerInterface")
92
+ table = struct_table_name(table)
93
+ return StorerInfo(DB=DB, table=table, fields=fields, length=length, config=config)
94
+
95
+ # @staticmethod
96
+ # def info(storer_info):
97
+ # if not storer_info:
98
+ # return None
99
+ #
100
+ # if isinstance(storer_info, str):
101
+ # storer_info = json.loads(storer_info)
102
+ #
103
+ # if any(isinstance(storer_info, t) for t in (dict, StorerInfo)):
104
+ # storer_info = [storer_info]
105
+ #
106
+ # if not isinstance(storer_info, list):
107
+ # raise Exception("StorerDB.info storer_info")
108
+ #
109
+ # storer_info_list = []
110
+ # for storer in storer_info:
111
+ # if isinstance(storer, StorerInfo):
112
+ # storer_info_list.append(storer)
113
+ # else:
114
+ # db_name = storer["DB"]
115
+ # if db_name in dir(StorerDB):
116
+ # del storer["DB"]
117
+ # else:
118
+ # db_name = "diy"
119
+ # func = getattr(StorerDB, db_name)
120
+ # storer_info_list.append(func(**storer))
121
+ # return storer_info_list
122
+
123
+
124
+
125
+ def deal(config, tag):
126
+ if isinstance(config, dict):
127
+ if tag == 0:
128
+ return RedisInfo(**config)
129
+ elif tag == 1:
130
+ db_name = config["DB"]
131
+ if db_name in dir(SchedulerDB):
132
+ del config["DB"]
133
+ else:
134
+ db_name = "diy"
135
+ func = getattr(SchedulerDB, db_name)
136
+ return func(**config)
137
+ elif tag == 2:
138
+ db_name = config["DB"]
139
+ if db_name in dir(StorerDB):
140
+ del config["DB"]
141
+ else:
142
+ db_name = "diy"
143
+ func = getattr(StorerDB, db_name)
144
+ return func(**config)
145
+ raise ValueError("tag must be in [0, 1, 2]")
146
+ elif any(isinstance(config, t) for t in (StorerInfo, SchedulerInfo, RedisInfo)):
147
+ return config
148
+ raise TypeError("config must be in [StorerInfo, SchedulerInfo, RedisInfo]")
149
+
150
+
151
+ def info(configs, tag = 0):
152
+ if configs is None:
153
+ return SchedulerDB.default() if tag == 1 else None
154
+
155
+ if isinstance(configs, str):
156
+ configs = json.loads(configs)
157
+
158
+ if tag == 0:
159
+ return deal(configs, tag)
160
+
161
+ if not isinstance(configs, list):
162
+ configs = [configs]
163
+
164
+ return [deal(config, tag) for config in configs]
@@ -0,0 +1,95 @@
1
+ import time
2
+ from functools import wraps
3
+
4
+ # from config import DBType
5
+ from log import log
6
+
7
+
8
+ # def find_func_name(func_name, name_list):
9
+ # for name in name_list:
10
+ # if func_name.find(name) == 0:
11
+ # return True
12
+ # return False
13
+
14
+
15
+ # def starter_decorator(execute):
16
+ # @wraps(execute)
17
+ # def wrapper(starter, *args, **kwargs):
18
+ # spider_dynamic_funcs = []
19
+ # scheduler_dynamic_funcs = []
20
+ # starter_functions = inspect.getmembers(starter, lambda a: inspect.isfunction(a))
21
+ # for starter_function in starter_functions:
22
+ # if find_func_name(starter_function[0], starter.scheduler_funcs):
23
+ # scheduler_dynamic_funcs.append(starter_function)
24
+ # elif find_func_name(starter_function[0], starter.spider_funcs):
25
+ # spider_dynamic_funcs.append(starter_function)
26
+ # return execute(starter, scheduler_dynamic_funcs, spider_dynamic_funcs, *args, **kwargs)
27
+ #
28
+ # return wrapper
29
+
30
+
31
+ # def scheduler_decorator(execute):
32
+ # @wraps(execute)
33
+ # def wrapper(scheduler, distribute_task):
34
+ # if not issubclass(scheduler, SchedulerInterface):
35
+ # scheduler.stop = True
36
+ # elif getattr(scheduler, "scheduler", None):
37
+ # execute(scheduler, distribute_task)
38
+ # else:
39
+ # log.error(f"scheduler type: {scheduler.db_type} not have add function!")
40
+ # scheduler.stop = True
41
+ # return wrapper
42
+
43
+
44
+ def storer_decorator(execute):
45
+ @wraps(execute)
46
+ def wrapper(storer, stop_event, last_event, table_name, callback):
47
+ if getattr(storer, "save", None):
48
+ execute(storer, stop_event, last_event, table_name, callback)
49
+ else:
50
+ log.error(f"storer base_type: {storer.data_type} not have add function!")
51
+ storer.stop = True
52
+ return wrapper
53
+
54
+
55
+ def distribute_scheduler_decorators(func):
56
+ @wraps(func)
57
+ def wrapper(distributor, callback):
58
+ try:
59
+ func(distributor, callback)
60
+ except TypeError:
61
+ pass
62
+ distributor.event.set()
63
+ return wrapper
64
+
65
+
66
+ def distribute_spider_decorators(func):
67
+ @wraps(func)
68
+ def wrapper(distributor, stop_event, db, callback):
69
+ while not stop_event.is_set():
70
+ try:
71
+ seed = distributor.queue_client.pop("_seed_queue")
72
+ if not seed:
73
+ time.sleep(3)
74
+ continue
75
+ distributor.spider_in_progress.append(1)
76
+ func(distributor, db, seed, callback)
77
+ except Exception as e:
78
+ print(e)
79
+ finally:
80
+ distributor.spider_in_progress.pop()
81
+
82
+ return wrapper
83
+
84
+
85
+ def distribute_storer_decorators(func):
86
+ @wraps(func)
87
+ def wrapper(distributor, callback, data_type, table_name, last):
88
+ data_list = []
89
+ try:
90
+ func(distributor, callback, data_list, data_type, table_name, last)
91
+ except Exception as e:
92
+ log.info("storage exception! " + str(e))
93
+ # distributor._task_queue.extendleft(data_list)
94
+
95
+ return wrapper
@@ -0,0 +1,60 @@
1
+ # -*- coding: utf-8 -*-
2
+
3
+ class DynamicHashTable:
4
+ def __init__(self):
5
+ self.capacity = 1000 # 初始容量
6
+ self.size = 0 # 元素个数
7
+ self.table = [None] * self.capacity
8
+
9
+ def hash_function(self, key):
10
+ return hash(key) % self.capacity
11
+
12
+ def probe(self, index):
13
+ # 线性探测法
14
+ return (index + 1) % self.capacity
15
+
16
+ def insert(self, key, value):
17
+ index = self.hash_function(key)
18
+ while self.table[index] is not None:
19
+ if self.table[index][0] == key:
20
+ self.table[index][1] = value
21
+ return
22
+ index = self.probe(index)
23
+ self.table[index] = [key, value]
24
+ self.size += 1
25
+
26
+ # 动态扩容
27
+ if self.size / self.capacity >= 0.7:
28
+ self.resize()
29
+
30
+ def get(self, key):
31
+ index = self.hash_function(key)
32
+ while self.table[index] is not None:
33
+ if self.table[index][0] == key:
34
+ return self.table[index][1]
35
+ index = self.probe(index)
36
+ raise KeyError("Key not found")
37
+
38
+ def remove(self, key):
39
+ index = self.hash_function(key)
40
+ while self.table[index] is not None:
41
+ if self.table[index][0] == key:
42
+ self.table[index] = None
43
+ self.size -= 1
44
+ return
45
+ index = self.probe(index)
46
+ raise KeyError("Key not found")
47
+
48
+ def resize(self):
49
+ # 扩容为原容量的两倍
50
+ self.capacity *= 2
51
+ new_table = [None] * self.capacity
52
+ for item in self.table:
53
+ if item is not None:
54
+ key, value = item
55
+ new_index = self.hash_function(key)
56
+ while new_table[new_index] is not None:
57
+ new_index = self.probe(new_index)
58
+ new_table[new_index] = [key, value]
59
+ self.table = new_table
60
+
@@ -0,0 +1,44 @@
1
+ import json
2
+ from abc import ABC, abstractmethod
3
+
4
+
5
+ def parse_config(config):
6
+ if not config:
7
+ return None
8
+ if isinstance(config, str):
9
+ return json.loads(config)
10
+ if isinstance(config, dict):
11
+ return config
12
+ raise TypeError("config type is not in [string, dict]!")
13
+
14
+
15
+ class SchedulerInterface(ABC):
16
+
17
+ def __init__(self, table, sql, length, size, queue, config=None):
18
+ self.sql = sql
19
+ self.table = table
20
+ self.length = length
21
+ self.size = size
22
+ self.queue = queue
23
+ self.config = parse_config(config)
24
+ self.stop = False
25
+
26
+ @abstractmethod
27
+ def schedule(self, *args, **kwargs):
28
+ pass
29
+
30
+
31
+ class StorerInterface(ABC):
32
+
33
+ def __init__(self, table, fields, length, queue, config=None):
34
+ self.table = table
35
+ self.fields = fields
36
+ self.length = length
37
+ self.queue = queue
38
+ self.config = parse_config(config)
39
+ # self.redis_db = redis_db
40
+
41
+ @abstractmethod
42
+ def store(self, *args, **kwargs):
43
+ pass
44
+