crawlo 1.1.2__py3-none-any.whl → 1.1.3__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of crawlo might be problematic. Click here for more details.

Files changed (113) hide show
  1. crawlo/__init__.py +34 -34
  2. crawlo/__version__.py +1 -1
  3. crawlo/cli.py +40 -40
  4. crawlo/commands/__init__.py +13 -13
  5. crawlo/commands/check.py +594 -594
  6. crawlo/commands/genspider.py +151 -151
  7. crawlo/commands/list.py +155 -155
  8. crawlo/commands/run.py +285 -285
  9. crawlo/commands/startproject.py +196 -196
  10. crawlo/commands/stats.py +188 -188
  11. crawlo/commands/utils.py +186 -186
  12. crawlo/config.py +279 -279
  13. crawlo/core/__init__.py +2 -2
  14. crawlo/core/engine.py +171 -171
  15. crawlo/core/enhanced_engine.py +189 -189
  16. crawlo/core/processor.py +40 -40
  17. crawlo/core/scheduler.py +166 -162
  18. crawlo/crawler.py +1027 -1027
  19. crawlo/downloader/__init__.py +242 -242
  20. crawlo/downloader/aiohttp_downloader.py +212 -212
  21. crawlo/downloader/cffi_downloader.py +251 -251
  22. crawlo/downloader/httpx_downloader.py +259 -257
  23. crawlo/event.py +11 -11
  24. crawlo/exceptions.py +82 -78
  25. crawlo/extension/__init__.py +31 -31
  26. crawlo/extension/log_interval.py +49 -49
  27. crawlo/extension/log_stats.py +44 -44
  28. crawlo/extension/logging_extension.py +34 -34
  29. crawlo/filters/__init__.py +154 -154
  30. crawlo/filters/aioredis_filter.py +242 -242
  31. crawlo/filters/memory_filter.py +269 -269
  32. crawlo/items/__init__.py +23 -23
  33. crawlo/items/base.py +21 -21
  34. crawlo/items/fields.py +53 -53
  35. crawlo/items/items.py +104 -104
  36. crawlo/middleware/__init__.py +21 -21
  37. crawlo/middleware/default_header.py +32 -32
  38. crawlo/middleware/download_delay.py +28 -28
  39. crawlo/middleware/middleware_manager.py +135 -135
  40. crawlo/middleware/proxy.py +248 -248
  41. crawlo/middleware/request_ignore.py +30 -30
  42. crawlo/middleware/response_code.py +18 -18
  43. crawlo/middleware/response_filter.py +26 -26
  44. crawlo/middleware/retry.py +125 -125
  45. crawlo/mode_manager.py +200 -200
  46. crawlo/network/__init__.py +21 -21
  47. crawlo/network/request.py +311 -311
  48. crawlo/network/response.py +271 -269
  49. crawlo/pipelines/__init__.py +22 -13
  50. crawlo/pipelines/bloom_dedup_pipeline.py +157 -0
  51. crawlo/pipelines/console_pipeline.py +39 -39
  52. crawlo/pipelines/csv_pipeline.py +316 -316
  53. crawlo/pipelines/database_dedup_pipeline.py +225 -0
  54. crawlo/pipelines/json_pipeline.py +218 -218
  55. crawlo/pipelines/memory_dedup_pipeline.py +116 -0
  56. crawlo/pipelines/mongo_pipeline.py +116 -116
  57. crawlo/pipelines/mysql_pipeline.py +195 -195
  58. crawlo/pipelines/pipeline_manager.py +56 -56
  59. crawlo/pipelines/redis_dedup_pipeline.py +163 -0
  60. crawlo/project.py +153 -153
  61. crawlo/queue/pqueue.py +37 -37
  62. crawlo/queue/queue_manager.py +307 -303
  63. crawlo/queue/redis_priority_queue.py +208 -191
  64. crawlo/settings/__init__.py +7 -7
  65. crawlo/settings/default_settings.py +245 -226
  66. crawlo/settings/setting_manager.py +99 -99
  67. crawlo/spider/__init__.py +639 -639
  68. crawlo/stats_collector.py +59 -59
  69. crawlo/subscriber.py +106 -106
  70. crawlo/task_manager.py +30 -30
  71. crawlo/templates/crawlo.cfg.tmpl +10 -10
  72. crawlo/templates/project/__init__.py.tmpl +3 -3
  73. crawlo/templates/project/items.py.tmpl +17 -17
  74. crawlo/templates/project/middlewares.py.tmpl +86 -86
  75. crawlo/templates/project/pipelines.py.tmpl +341 -335
  76. crawlo/templates/project/run.py.tmpl +251 -238
  77. crawlo/templates/project/settings.py.tmpl +250 -247
  78. crawlo/templates/project/spiders/__init__.py.tmpl +5 -5
  79. crawlo/templates/spider/spider.py.tmpl +177 -177
  80. crawlo/utils/__init__.py +7 -7
  81. crawlo/utils/controlled_spider_mixin.py +439 -335
  82. crawlo/utils/date_tools.py +233 -233
  83. crawlo/utils/db_helper.py +343 -343
  84. crawlo/utils/func_tools.py +82 -82
  85. crawlo/utils/large_scale_config.py +286 -286
  86. crawlo/utils/large_scale_helper.py +343 -343
  87. crawlo/utils/log.py +128 -128
  88. crawlo/utils/queue_helper.py +175 -175
  89. crawlo/utils/request.py +267 -267
  90. crawlo/utils/request_serializer.py +219 -219
  91. crawlo/utils/spider_loader.py +62 -62
  92. crawlo/utils/system.py +11 -11
  93. crawlo/utils/tools.py +4 -4
  94. crawlo/utils/url.py +39 -39
  95. {crawlo-1.1.2.dist-info → crawlo-1.1.3.dist-info}/METADATA +635 -567
  96. crawlo-1.1.3.dist-info/RECORD +113 -0
  97. examples/__init__.py +7 -7
  98. examples/controlled_spider_example.py +205 -0
  99. tests/__init__.py +7 -7
  100. tests/test_final_validation.py +153 -153
  101. tests/test_proxy_health_check.py +32 -32
  102. tests/test_proxy_middleware_integration.py +136 -136
  103. tests/test_proxy_providers.py +56 -56
  104. tests/test_proxy_stats.py +19 -19
  105. tests/test_proxy_strategies.py +59 -59
  106. tests/test_redis_config.py +28 -28
  107. tests/test_redis_queue.py +224 -224
  108. tests/test_request_serialization.py +70 -70
  109. tests/test_scheduler.py +241 -241
  110. crawlo-1.1.2.dist-info/RECORD +0 -108
  111. {crawlo-1.1.2.dist-info → crawlo-1.1.3.dist-info}/WHEEL +0 -0
  112. {crawlo-1.1.2.dist-info → crawlo-1.1.3.dist-info}/entry_points.txt +0 -0
  113. {crawlo-1.1.2.dist-info → crawlo-1.1.3.dist-info}/top_level.txt +0 -0
crawlo/exceptions.py CHANGED
@@ -1,78 +1,82 @@
1
- #!/usr/bin/python
2
- # -*- coding:UTF-8 -*-
3
- class TransformTypeError(TypeError):
4
- pass
5
-
6
-
7
- class OutputError(Exception):
8
- pass
9
-
10
-
11
- class SpiderTypeError(TypeError):
12
- pass
13
-
14
-
15
- class ItemInitError(Exception):
16
- pass
17
-
18
-
19
- class ItemAttributeError(Exception):
20
- pass
21
-
22
-
23
- class DecodeError(Exception):
24
- pass
25
-
26
-
27
- class MiddlewareInitError(Exception):
28
- pass
29
-
30
-
31
- class PipelineInitError(Exception):
32
- pass
33
-
34
-
35
- class InvalidOutputError(Exception):
36
- pass
37
-
38
-
39
- class RequestMethodError(Exception):
40
- pass
41
-
42
-
43
- class IgnoreRequestError(Exception):
44
- def __init__(self, msg):
45
- self.msg = msg
46
- super(IgnoreRequestError, self).__init__(msg)
47
-
48
-
49
- class ItemDiscard(Exception):
50
- def __init__(self, msg):
51
- self.msg = msg
52
- super(ItemDiscard, self).__init__(msg)
53
-
54
-
55
- class NotConfigured(Exception):
56
- pass
57
-
58
-
59
- class NotConfiguredError(Exception):
60
- pass
61
-
62
-
63
- class ExtensionInitError(Exception):
64
- pass
65
-
66
-
67
- class ReceiverTypeError(Exception):
68
- pass
69
-
70
-
71
- class SpiderCreationError(Exception):
72
- """爬虫实例化失败异常"""
73
- pass
74
-
75
-
76
- class ItemValidationError(Exception):
77
- """Item 字段验证错误"""
78
- pass
1
+ #!/usr/bin/python
2
+ # -*- coding:UTF-8 -*-
3
+ class TransformTypeError(TypeError):
4
+ pass
5
+
6
+
7
+ class OutputError(Exception):
8
+ pass
9
+
10
+
11
+ class SpiderTypeError(TypeError):
12
+ pass
13
+
14
+
15
+ class ItemInitError(Exception):
16
+ pass
17
+
18
+
19
+ class ItemAttributeError(Exception):
20
+ pass
21
+
22
+
23
+ class DecodeError(Exception):
24
+ pass
25
+
26
+
27
+ class MiddlewareInitError(Exception):
28
+ pass
29
+
30
+
31
+ class PipelineInitError(Exception):
32
+ pass
33
+
34
+
35
+ class InvalidOutputError(Exception):
36
+ pass
37
+
38
+
39
+ class RequestMethodError(Exception):
40
+ pass
41
+
42
+
43
+ class IgnoreRequestError(Exception):
44
+ def __init__(self, msg):
45
+ self.msg = msg
46
+ super(IgnoreRequestError, self).__init__(msg)
47
+
48
+
49
+ class ItemDiscard(Exception):
50
+ def __init__(self, msg):
51
+ self.msg = msg
52
+ super(ItemDiscard, self).__init__(msg)
53
+
54
+
55
+ class NotConfigured(Exception):
56
+ pass
57
+
58
+
59
+ class NotConfiguredError(Exception):
60
+ pass
61
+
62
+
63
+ class ExtensionInitError(Exception):
64
+ pass
65
+
66
+
67
+ class ReceiverTypeError(Exception):
68
+ pass
69
+
70
+
71
+ class SpiderCreationError(Exception):
72
+ """爬虫实例化失败异常"""
73
+ pass
74
+
75
+
76
+ class ItemValidationError(Exception):
77
+ """Item 字段验证错误"""
78
+ pass
79
+
80
+
81
+ class DropItem(Exception):
82
+ pass
@@ -1,31 +1,31 @@
1
- #!/usr/bin/python
2
- # -*- coding:UTF-8 -*-
3
- from typing import List
4
- from pprint import pformat
5
-
6
- from crawlo.utils.log import get_logger
7
- from crawlo.project import load_class
8
- from crawlo.exceptions import ExtensionInitError
9
-
10
-
11
- class ExtensionManager(object):
12
-
13
- def __init__(self, crawler):
14
- self.crawler = crawler
15
- self.extensions: List = []
16
- extensions = self.crawler.settings.get_list('EXTENSIONS')
17
- self.logger = get_logger(self.__class__.__name__, crawler.settings.get('LOG_LEVEL'))
18
- self._add_extensions(extensions)
19
-
20
- @classmethod
21
- def create_instance(cls, *args, **kwargs):
22
- return cls(*args, **kwargs)
23
-
24
- def _add_extensions(self, extensions):
25
- for extension in extensions:
26
- extension_cls = load_class(extension)
27
- if not hasattr(extension_cls, 'create_instance'):
28
- raise ExtensionInitError(f"extension init failed, Must have method 'create_instance()")
29
- self.extensions.append(extension_cls.create_instance(self.crawler))
30
- if extensions:
31
- self.logger.info(f"enabled extensions: \n {pformat(extensions)}")
1
+ #!/usr/bin/python
2
+ # -*- coding:UTF-8 -*-
3
+ from typing import List
4
+ from pprint import pformat
5
+
6
+ from crawlo.utils.log import get_logger
7
+ from crawlo.project import load_class
8
+ from crawlo.exceptions import ExtensionInitError
9
+
10
+
11
+ class ExtensionManager(object):
12
+
13
+ def __init__(self, crawler):
14
+ self.crawler = crawler
15
+ self.extensions: List = []
16
+ extensions = self.crawler.settings.get_list('EXTENSIONS')
17
+ self.logger = get_logger(self.__class__.__name__, crawler.settings.get('LOG_LEVEL'))
18
+ self._add_extensions(extensions)
19
+
20
+ @classmethod
21
+ def create_instance(cls, *args, **kwargs):
22
+ return cls(*args, **kwargs)
23
+
24
+ def _add_extensions(self, extensions):
25
+ for extension in extensions:
26
+ extension_cls = load_class(extension)
27
+ if not hasattr(extension_cls, 'create_instance'):
28
+ raise ExtensionInitError(f"extension init failed, Must have method 'create_instance()")
29
+ self.extensions.append(extension_cls.create_instance(self.crawler))
30
+ if extensions:
31
+ self.logger.info(f"enabled extensions: \n {pformat(extensions)}")
@@ -1,49 +1,49 @@
1
- #!/usr/bin/python
2
- # -*- coding:UTF-8 -*-
3
- import asyncio
4
-
5
- from crawlo.utils.log import get_logger
6
- from crawlo.event import spider_opened, spider_closed
7
-
8
-
9
- class LogIntervalExtension(object):
10
-
11
- def __init__(self, crawler):
12
- self.task = None
13
- self.stats = crawler.stats
14
- self.item_count = 0
15
- self.response_count = 0
16
- self.seconds = crawler.settings.get('INTERVAL')
17
- self.interval = int(self.seconds / 60) if self.seconds % 60 == 0 else self.seconds
18
- self.interval = "" if self.interval == 1 else self.interval
19
- self.unit = 'min' if self.seconds % 60 == 0 else 's'
20
-
21
- self.logger = get_logger(self.__class__.__name__, crawler.settings.get('LOG_LEVEL'))
22
-
23
- @classmethod
24
- def create_instance(cls, crawler):
25
- o = cls(crawler)
26
- crawler.subscriber.subscribe(o.spider_opened, event=spider_opened)
27
- crawler.subscriber.subscribe(o.spider_closed, event=spider_closed)
28
- return o
29
-
30
- async def spider_opened(self):
31
- self.task = asyncio.create_task(self.interval_log())
32
- await self.task
33
-
34
- async def spider_closed(self):
35
- if self.task:
36
- self.task.cancel()
37
-
38
- async def interval_log(self):
39
- while True:
40
- last_item_count = self.stats.get_value('item_successful_count', default=0)
41
- last_response_count = self.stats.get_value('response_received_count', default=0)
42
- item_rate = last_item_count - self.item_count
43
- response_rate = last_response_count - self.response_count
44
- self.item_count, self.response_count = last_item_count, last_response_count
45
- self.logger.info(
46
- f'Crawled {last_response_count} pages (at {response_rate} pages/{self.interval}{self.unit}),'
47
- f' Got {last_item_count} items (at {item_rate} items/{self.interval}{self.unit}).'
48
- )
49
- await asyncio.sleep(self.seconds)
1
+ #!/usr/bin/python
2
+ # -*- coding:UTF-8 -*-
3
+ import asyncio
4
+
5
+ from crawlo.utils.log import get_logger
6
+ from crawlo.event import spider_opened, spider_closed
7
+
8
+
9
+ class LogIntervalExtension(object):
10
+
11
+ def __init__(self, crawler):
12
+ self.task = None
13
+ self.stats = crawler.stats
14
+ self.item_count = 0
15
+ self.response_count = 0
16
+ self.seconds = crawler.settings.get('INTERVAL')
17
+ self.interval = int(self.seconds / 60) if self.seconds % 60 == 0 else self.seconds
18
+ self.interval = "" if self.interval == 1 else self.interval
19
+ self.unit = 'min' if self.seconds % 60 == 0 else 's'
20
+
21
+ self.logger = get_logger(self.__class__.__name__, crawler.settings.get('LOG_LEVEL'))
22
+
23
+ @classmethod
24
+ def create_instance(cls, crawler):
25
+ o = cls(crawler)
26
+ crawler.subscriber.subscribe(o.spider_opened, event=spider_opened)
27
+ crawler.subscriber.subscribe(o.spider_closed, event=spider_closed)
28
+ return o
29
+
30
+ async def spider_opened(self):
31
+ self.task = asyncio.create_task(self.interval_log())
32
+ await self.task
33
+
34
+ async def spider_closed(self):
35
+ if self.task:
36
+ self.task.cancel()
37
+
38
+ async def interval_log(self):
39
+ while True:
40
+ last_item_count = self.stats.get_value('item_successful_count', default=0)
41
+ last_response_count = self.stats.get_value('response_received_count', default=0)
42
+ item_rate = last_item_count - self.item_count
43
+ response_rate = last_response_count - self.response_count
44
+ self.item_count, self.response_count = last_item_count, last_response_count
45
+ self.logger.info(
46
+ f'Crawled {last_response_count} pages (at {response_rate} pages/{self.interval}{self.unit}),'
47
+ f' Got {last_item_count} items (at {item_rate} items/{self.interval}{self.unit}).'
48
+ )
49
+ await asyncio.sleep(self.seconds)
@@ -1,44 +1,44 @@
1
- #!/usr/bin/python
2
- # -*- coding:UTF-8 -*-
3
- from crawlo import event
4
- from crawlo.utils.date_tools import now, time_diff
5
-
6
-
7
- class LogStats(object):
8
-
9
- def __init__(self, stats):
10
- self._stats = stats
11
-
12
- @classmethod
13
- def create_instance(cls, crawler):
14
- o = cls(crawler.stats)
15
- crawler.subscriber.subscribe(o.spider_opened, event=event.spider_opened)
16
- crawler.subscriber.subscribe(o.spider_closed, event=event.spider_closed)
17
- crawler.subscriber.subscribe(o.item_successful, event=event.item_successful)
18
- crawler.subscriber.subscribe(o.item_discard, event=event.item_discard)
19
- crawler.subscriber.subscribe(o.response_received, event=event.response_received)
20
- crawler.subscriber.subscribe(o.request_scheduled, event=event.request_scheduled)
21
-
22
- return o
23
-
24
- async def spider_opened(self):
25
- self._stats['start_time'] = now(fmt='%Y-%m-%d %H:%M:%S')
26
-
27
- async def spider_closed(self):
28
- self._stats['end_time'] = now(fmt='%Y-%m-%d %H:%M:%S')
29
- self._stats['cost_time(s)'] = time_diff(start=self._stats['start_time'], end=self._stats['end_time'])
30
-
31
- async def item_successful(self, _item, _spider):
32
- self._stats.inc_value('item_successful_count')
33
-
34
- async def item_discard(self, _item, exc, _spider):
35
- self._stats.inc_value('item_discard_count')
36
- reason = exc.msg
37
- if reason:
38
- self._stats.inc_value(f"item_discard/{reason}")
39
-
40
- async def response_received(self, _response, _spider):
41
- self._stats.inc_value('response_received_count')
42
-
43
- async def request_scheduled(self, _request, _spider):
44
- self._stats.inc_value('request_scheduler_count')
1
+ #!/usr/bin/python
2
+ # -*- coding:UTF-8 -*-
3
+ from crawlo import event
4
+ from crawlo.utils.date_tools import now, time_diff
5
+
6
+
7
+ class LogStats(object):
8
+
9
+ def __init__(self, stats):
10
+ self._stats = stats
11
+
12
+ @classmethod
13
+ def create_instance(cls, crawler):
14
+ o = cls(crawler.stats)
15
+ crawler.subscriber.subscribe(o.spider_opened, event=event.spider_opened)
16
+ crawler.subscriber.subscribe(o.spider_closed, event=event.spider_closed)
17
+ crawler.subscriber.subscribe(o.item_successful, event=event.item_successful)
18
+ crawler.subscriber.subscribe(o.item_discard, event=event.item_discard)
19
+ crawler.subscriber.subscribe(o.response_received, event=event.response_received)
20
+ crawler.subscriber.subscribe(o.request_scheduled, event=event.request_scheduled)
21
+
22
+ return o
23
+
24
+ async def spider_opened(self):
25
+ self._stats['start_time'] = now(fmt='%Y-%m-%d %H:%M:%S')
26
+
27
+ async def spider_closed(self):
28
+ self._stats['end_time'] = now(fmt='%Y-%m-%d %H:%M:%S')
29
+ self._stats['cost_time(s)'] = time_diff(start=self._stats['start_time'], end=self._stats['end_time'])
30
+
31
+ async def item_successful(self, _item, _spider):
32
+ self._stats.inc_value('item_successful_count')
33
+
34
+ async def item_discard(self, _item, exc, _spider):
35
+ self._stats.inc_value('item_discard_count')
36
+ reason = exc.msg
37
+ if reason:
38
+ self._stats.inc_value(f"item_discard/{reason}")
39
+
40
+ async def response_received(self, _response, _spider):
41
+ self._stats.inc_value('response_received_count')
42
+
43
+ async def request_scheduled(self, _request, _spider):
44
+ self._stats.inc_value('request_scheduler_count')
@@ -1,35 +1,35 @@
1
- from crawlo.exceptions import NotConfigured
2
- from crawlo.utils.log import get_logger
3
- from crawlo.utils.log import LoggerManager
4
-
5
-
6
- class CustomLoggerExtension:
7
- """
8
- 日志系统初始化扩展
9
- 遵循与 ExtensionManager 一致的接口规范:使用 create_instance
10
- """
11
-
12
- def __init__(self, settings):
13
- self.settings = settings
14
- # 初始化全局日志配置
15
- LoggerManager.configure(settings)
16
-
17
- @classmethod
18
- def create_instance(cls, crawler, *args, **kwargs):
19
- """
20
- 工厂方法:兼容 ExtensionManager 的创建方式
21
- 被 ExtensionManager 调用
22
- """
23
- # 可以通过 settings 控制是否启用
24
- if not crawler.settings.get('LOG_FILE') and not crawler.settings.get('LOG_ENABLE_CUSTOM'):
25
- raise NotConfigured("CustomLoggerExtension: LOG_FILE not set and LOG_ENABLE_CUSTOM=False")
26
-
27
- return cls(crawler.settings)
28
-
29
- def spider_opened(self, spider):
30
- logger = get_logger(__name__)
31
- logger.info(
32
- f"CustomLoggerExtension: Logging initialized. "
33
- f"LOG_FILE={self.settings.get('LOG_FILE')}, "
34
- f"LOG_LEVEL={self.settings.get('LOG_LEVEL')}"
1
+ from crawlo.exceptions import NotConfigured
2
+ from crawlo.utils.log import get_logger
3
+ from crawlo.utils.log import LoggerManager
4
+
5
+
6
+ class CustomLoggerExtension:
7
+ """
8
+ 日志系统初始化扩展
9
+ 遵循与 ExtensionManager 一致的接口规范:使用 create_instance
10
+ """
11
+
12
+ def __init__(self, settings):
13
+ self.settings = settings
14
+ # 初始化全局日志配置
15
+ LoggerManager.configure(settings)
16
+
17
+ @classmethod
18
+ def create_instance(cls, crawler, *args, **kwargs):
19
+ """
20
+ 工厂方法:兼容 ExtensionManager 的创建方式
21
+ 被 ExtensionManager 调用
22
+ """
23
+ # 可以通过 settings 控制是否启用
24
+ if not crawler.settings.get('LOG_FILE') and not crawler.settings.get('LOG_ENABLE_CUSTOM'):
25
+ raise NotConfigured("CustomLoggerExtension: LOG_FILE not set and LOG_ENABLE_CUSTOM=False")
26
+
27
+ return cls(crawler.settings)
28
+
29
+ def spider_opened(self, spider):
30
+ logger = get_logger(__name__)
31
+ logger.info(
32
+ f"CustomLoggerExtension: Logging initialized. "
33
+ f"LOG_FILE={self.settings.get('LOG_FILE')}, "
34
+ f"LOG_LEVEL={self.settings.get('LOG_LEVEL')}"
35
35
  )