crawlo 1.0.3__py3-none-any.whl → 1.0.5__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of crawlo might be problematic. Click here for more details.

Files changed (96) hide show
  1. crawlo/__init__.py +25 -9
  2. crawlo/__version__.py +1 -1
  3. crawlo/core/__init__.py +2 -2
  4. crawlo/core/engine.py +158 -158
  5. crawlo/core/processor.py +40 -40
  6. crawlo/core/scheduler.py +57 -57
  7. crawlo/crawler.py +424 -242
  8. crawlo/downloader/__init__.py +78 -78
  9. crawlo/downloader/aiohttp_downloader.py +200 -259
  10. crawlo/downloader/cffi_downloader.py +277 -0
  11. crawlo/downloader/httpx_downloader.py +246 -187
  12. crawlo/event.py +11 -11
  13. crawlo/exceptions.py +73 -64
  14. crawlo/extension/__init__.py +31 -31
  15. crawlo/extension/log_interval.py +49 -49
  16. crawlo/extension/log_stats.py +44 -44
  17. crawlo/extension/logging_extension.py +35 -0
  18. crawlo/filters/__init__.py +37 -37
  19. crawlo/filters/aioredis_filter.py +150 -158
  20. crawlo/filters/memory_filter.py +202 -202
  21. crawlo/items/__init__.py +62 -62
  22. crawlo/items/items.py +115 -119
  23. crawlo/middleware/__init__.py +21 -21
  24. crawlo/middleware/default_header.py +32 -32
  25. crawlo/middleware/download_delay.py +28 -28
  26. crawlo/middleware/middleware_manager.py +135 -140
  27. crawlo/middleware/proxy.py +246 -0
  28. crawlo/middleware/request_ignore.py +30 -30
  29. crawlo/middleware/response_code.py +18 -18
  30. crawlo/middleware/response_filter.py +26 -26
  31. crawlo/middleware/retry.py +90 -90
  32. crawlo/network/__init__.py +7 -7
  33. crawlo/network/request.py +203 -204
  34. crawlo/network/response.py +166 -166
  35. crawlo/pipelines/__init__.py +13 -13
  36. crawlo/pipelines/console_pipeline.py +39 -39
  37. crawlo/pipelines/mongo_pipeline.py +116 -116
  38. crawlo/pipelines/mysql_batch_pipline.py +273 -134
  39. crawlo/pipelines/mysql_pipeline.py +195 -195
  40. crawlo/pipelines/pipeline_manager.py +56 -56
  41. crawlo/settings/__init__.py +7 -7
  42. crawlo/settings/default_settings.py +169 -93
  43. crawlo/settings/setting_manager.py +99 -99
  44. crawlo/spider/__init__.py +41 -36
  45. crawlo/stats_collector.py +59 -59
  46. crawlo/subscriber.py +106 -106
  47. crawlo/task_manager.py +27 -27
  48. crawlo/templates/item_template.tmpl +21 -21
  49. crawlo/templates/project_template/main.py +32 -32
  50. crawlo/templates/project_template/setting.py +189 -189
  51. crawlo/templates/spider_template.tmpl +30 -30
  52. crawlo/utils/__init__.py +7 -7
  53. crawlo/utils/concurrency_manager.py +124 -124
  54. crawlo/utils/date_tools.py +233 -177
  55. crawlo/utils/db_helper.py +344 -0
  56. crawlo/utils/func_tools.py +82 -82
  57. crawlo/utils/log.py +129 -39
  58. crawlo/utils/pqueue.py +173 -173
  59. crawlo/utils/project.py +59 -59
  60. crawlo/utils/request.py +267 -122
  61. crawlo/utils/system.py +11 -11
  62. crawlo/utils/tools.py +5 -303
  63. crawlo/utils/url.py +39 -39
  64. {crawlo-1.0.3.dist-info → crawlo-1.0.5.dist-info}/METADATA +49 -48
  65. crawlo-1.0.5.dist-info/RECORD +84 -0
  66. {crawlo-1.0.3.dist-info → crawlo-1.0.5.dist-info}/top_level.txt +1 -0
  67. examples/__init__.py +0 -0
  68. examples/gxb/__init__.py +0 -0
  69. examples/gxb/items.py +36 -0
  70. examples/gxb/run.py +15 -0
  71. examples/gxb/settings.py +71 -0
  72. examples/gxb/spider/__init__.py +0 -0
  73. examples/gxb/spider/miit_spider.py +180 -0
  74. examples/gxb/spider/telecom_device_licenses.py +129 -0
  75. tests/__init__.py +7 -7
  76. tests/test_proxy_health_check.py +33 -0
  77. tests/test_proxy_middleware_integration.py +137 -0
  78. tests/test_proxy_providers.py +57 -0
  79. tests/test_proxy_stats.py +20 -0
  80. tests/test_proxy_strategies.py +60 -0
  81. crawlo/downloader/playwright_downloader.py +0 -161
  82. crawlo/filters/redis_filter.py +0 -120
  83. crawlo-1.0.3.dist-info/RECORD +0 -80
  84. tests/baidu_spider/__init__.py +0 -7
  85. tests/baidu_spider/demo.py +0 -94
  86. tests/baidu_spider/items.py +0 -25
  87. tests/baidu_spider/middleware.py +0 -49
  88. tests/baidu_spider/pipeline.py +0 -55
  89. tests/baidu_spider/request_fingerprints.txt +0 -9
  90. tests/baidu_spider/run.py +0 -27
  91. tests/baidu_spider/settings.py +0 -78
  92. tests/baidu_spider/spiders/__init__.py +0 -7
  93. tests/baidu_spider/spiders/bai_du.py +0 -61
  94. tests/baidu_spider/spiders/sina.py +0 -79
  95. {crawlo-1.0.3.dist-info → crawlo-1.0.5.dist-info}/WHEEL +0 -0
  96. {crawlo-1.0.3.dist-info → crawlo-1.0.5.dist-info}/entry_points.txt +0 -0
@@ -1,31 +1,31 @@
1
- #!/usr/bin/python
2
- # -*- coding:UTF-8 -*-
3
- from typing import List
4
- from pprint import pformat
5
-
6
- from crawlo.utils.log import get_logger
7
- from crawlo.utils.project import load_class
8
- from crawlo.exceptions import ExtensionInitError
9
-
10
-
11
- class ExtensionManager(object):
12
-
13
- def __init__(self, crawler):
14
- self.crawler = crawler
15
- self.extensions: List = []
16
- extensions = self.crawler.settings.get_list('EXTENSIONS')
17
- self.logger = get_logger(self.__class__.__name__, crawler.settings.get('LOG_LEVEL'))
18
- self._add_extensions(extensions)
19
-
20
- @classmethod
21
- def create_instance(cls, *args, **kwargs):
22
- return cls(*args, **kwargs)
23
-
24
- def _add_extensions(self, extensions):
25
- for extension in extensions:
26
- extension_cls = load_class(extension)
27
- if not hasattr(extension_cls, 'create_instance'):
28
- raise ExtensionInitError(f"extension init failed, Must have method 'create_instance()")
29
- self.extensions.append(extension_cls.create_instance(self.crawler))
30
- if extensions:
31
- self.logger.info(f"enabled extensions: \n {pformat(extensions)}")
1
+ #!/usr/bin/python
2
+ # -*- coding:UTF-8 -*-
3
+ from typing import List
4
+ from pprint import pformat
5
+
6
+ from crawlo.utils.log import get_logger
7
+ from crawlo.utils.project import load_class
8
+ from crawlo.exceptions import ExtensionInitError
9
+
10
+
11
+ class ExtensionManager(object):
12
+
13
+ def __init__(self, crawler):
14
+ self.crawler = crawler
15
+ self.extensions: List = []
16
+ extensions = self.crawler.settings.get_list('EXTENSIONS')
17
+ self.logger = get_logger(self.__class__.__name__, crawler.settings.get('LOG_LEVEL'))
18
+ self._add_extensions(extensions)
19
+
20
+ @classmethod
21
+ def create_instance(cls, *args, **kwargs):
22
+ return cls(*args, **kwargs)
23
+
24
+ def _add_extensions(self, extensions):
25
+ for extension in extensions:
26
+ extension_cls = load_class(extension)
27
+ if not hasattr(extension_cls, 'create_instance'):
28
+ raise ExtensionInitError(f"extension init failed, Must have method 'create_instance()")
29
+ self.extensions.append(extension_cls.create_instance(self.crawler))
30
+ if extensions:
31
+ self.logger.info(f"enabled extensions: \n {pformat(extensions)}")
@@ -1,49 +1,49 @@
1
- #!/usr/bin/python
2
- # -*- coding:UTF-8 -*-
3
- import asyncio
4
-
5
- from crawlo.utils.log import get_logger
6
- from crawlo.event import spider_opened, spider_closed
7
-
8
-
9
- class LogIntervalExtension(object):
10
-
11
- def __init__(self, crawler):
12
- self.task = None
13
- self.stats = crawler.stats
14
- self.item_count = 0
15
- self.response_count = 0
16
- self.seconds = crawler.settings.get('INTERVAL')
17
- self.interval = int(self.seconds / 60) if self.seconds % 60 == 0 else self.seconds
18
- self.interval = "" if self.interval == 1 else self.interval
19
- self.unit = 'min' if self.seconds % 60 == 0 else 's'
20
-
21
- self.logger = get_logger(self.__class__.__name__, crawler.settings.get('LOG_LEVEL'))
22
-
23
- @classmethod
24
- def create_instance(cls, crawler):
25
- o = cls(crawler)
26
- crawler.subscriber.subscribe(o.spider_opened, event=spider_opened)
27
- crawler.subscriber.subscribe(o.spider_closed, event=spider_closed)
28
- return o
29
-
30
- async def spider_opened(self):
31
- self.task = asyncio.create_task(self.interval_log())
32
- await self.task
33
-
34
- async def spider_closed(self):
35
- if self.task:
36
- self.task.cancel()
37
-
38
- async def interval_log(self):
39
- while True:
40
- last_item_count = self.stats.get_value('item_successful_count', default=0)
41
- last_response_count = self.stats.get_value('response_received_count', default=0)
42
- item_rate = last_item_count - self.item_count
43
- response_rate = last_response_count - self.response_count
44
- self.item_count, self.response_count = last_item_count, last_response_count
45
- self.logger.info(
46
- f'Crawled {last_response_count} pages (at {response_rate} pages/{self.interval}{self.unit}),'
47
- f' Got {last_item_count} items (at {item_rate} items/{self.interval}{self.unit}).'
48
- )
49
- await asyncio.sleep(self.seconds)
1
+ #!/usr/bin/python
2
+ # -*- coding:UTF-8 -*-
3
+ import asyncio
4
+
5
+ from crawlo.utils.log import get_logger
6
+ from crawlo.event import spider_opened, spider_closed
7
+
8
+
9
+ class LogIntervalExtension(object):
10
+
11
+ def __init__(self, crawler):
12
+ self.task = None
13
+ self.stats = crawler.stats
14
+ self.item_count = 0
15
+ self.response_count = 0
16
+ self.seconds = crawler.settings.get('INTERVAL')
17
+ self.interval = int(self.seconds / 60) if self.seconds % 60 == 0 else self.seconds
18
+ self.interval = "" if self.interval == 1 else self.interval
19
+ self.unit = 'min' if self.seconds % 60 == 0 else 's'
20
+
21
+ self.logger = get_logger(self.__class__.__name__, crawler.settings.get('LOG_LEVEL'))
22
+
23
+ @classmethod
24
+ def create_instance(cls, crawler):
25
+ o = cls(crawler)
26
+ crawler.subscriber.subscribe(o.spider_opened, event=spider_opened)
27
+ crawler.subscriber.subscribe(o.spider_closed, event=spider_closed)
28
+ return o
29
+
30
+ async def spider_opened(self):
31
+ self.task = asyncio.create_task(self.interval_log())
32
+ await self.task
33
+
34
+ async def spider_closed(self):
35
+ if self.task:
36
+ self.task.cancel()
37
+
38
+ async def interval_log(self):
39
+ while True:
40
+ last_item_count = self.stats.get_value('item_successful_count', default=0)
41
+ last_response_count = self.stats.get_value('response_received_count', default=0)
42
+ item_rate = last_item_count - self.item_count
43
+ response_rate = last_response_count - self.response_count
44
+ self.item_count, self.response_count = last_item_count, last_response_count
45
+ self.logger.info(
46
+ f'Crawled {last_response_count} pages (at {response_rate} pages/{self.interval}{self.unit}),'
47
+ f' Got {last_item_count} items (at {item_rate} items/{self.interval}{self.unit}).'
48
+ )
49
+ await asyncio.sleep(self.seconds)
@@ -1,44 +1,44 @@
1
- #!/usr/bin/python
2
- # -*- coding:UTF-8 -*-
3
- from crawlo import event
4
- from crawlo.utils.date_tools import get_current_time, time_diff_seconds
5
-
6
-
7
- class LogStats(object):
8
-
9
- def __init__(self, stats):
10
- self._stats = stats
11
-
12
- @classmethod
13
- def create_instance(cls, crawler):
14
- o = cls(crawler.stats)
15
- crawler.subscriber.subscribe(o.spider_opened, event=event.spider_opened)
16
- crawler.subscriber.subscribe(o.spider_closed, event=event.spider_closed)
17
- crawler.subscriber.subscribe(o.item_successful, event=event.item_successful)
18
- crawler.subscriber.subscribe(o.item_discard, event=event.item_discard)
19
- crawler.subscriber.subscribe(o.response_received, event=event.response_received)
20
- crawler.subscriber.subscribe(o.request_scheduled, event=event.request_scheduled)
21
-
22
- return o
23
-
24
- async def spider_opened(self):
25
- self._stats['start_time'] = get_current_time(fmt='%Y-%m-%d %H:%M:%S')
26
-
27
- async def spider_closed(self):
28
- self._stats['end_time'] = get_current_time(fmt='%Y-%m-%d %H:%M:%S')
29
- self._stats['cost_time(s)'] = time_diff_seconds(start_time=self._stats['start_time'], end_time=self._stats['end_time'])
30
-
31
- async def item_successful(self, _item, _spider):
32
- self._stats.inc_value('item_successful_count')
33
-
34
- async def item_discard(self, _item, exc, _spider):
35
- self._stats.inc_value('item_discard_count')
36
- reason = exc.msg
37
- if reason:
38
- self._stats.inc_value(f"item_discard/{reason}")
39
-
40
- async def response_received(self, _response, _spider):
41
- self._stats.inc_value('response_received_count')
42
-
43
- async def request_scheduled(self, _request, _spider):
44
- self._stats.inc_value('request_scheduler_count')
1
+ #!/usr/bin/python
2
+ # -*- coding:UTF-8 -*-
3
+ from crawlo import event
4
+ from crawlo.utils.date_tools import now, time_diff
5
+
6
+
7
+ class LogStats(object):
8
+
9
+ def __init__(self, stats):
10
+ self._stats = stats
11
+
12
+ @classmethod
13
+ def create_instance(cls, crawler):
14
+ o = cls(crawler.stats)
15
+ crawler.subscriber.subscribe(o.spider_opened, event=event.spider_opened)
16
+ crawler.subscriber.subscribe(o.spider_closed, event=event.spider_closed)
17
+ crawler.subscriber.subscribe(o.item_successful, event=event.item_successful)
18
+ crawler.subscriber.subscribe(o.item_discard, event=event.item_discard)
19
+ crawler.subscriber.subscribe(o.response_received, event=event.response_received)
20
+ crawler.subscriber.subscribe(o.request_scheduled, event=event.request_scheduled)
21
+
22
+ return o
23
+
24
+ async def spider_opened(self):
25
+ self._stats['start_time'] = now(fmt='%Y-%m-%d %H:%M:%S')
26
+
27
+ async def spider_closed(self):
28
+ self._stats['end_time'] = now(fmt='%Y-%m-%d %H:%M:%S')
29
+ self._stats['cost_time(s)'] = time_diff(start=self._stats['start_time'], end=self._stats['end_time'])
30
+
31
+ async def item_successful(self, _item, _spider):
32
+ self._stats.inc_value('item_successful_count')
33
+
34
+ async def item_discard(self, _item, exc, _spider):
35
+ self._stats.inc_value('item_discard_count')
36
+ reason = exc.msg
37
+ if reason:
38
+ self._stats.inc_value(f"item_discard/{reason}")
39
+
40
+ async def response_received(self, _response, _spider):
41
+ self._stats.inc_value('response_received_count')
42
+
43
+ async def request_scheduled(self, _request, _spider):
44
+ self._stats.inc_value('request_scheduler_count')
@@ -0,0 +1,35 @@
1
+ from crawlo.exceptions import NotConfigured
2
+ from crawlo.utils.log import get_logger
3
+ from crawlo.utils.log import LoggerManager
4
+
5
+
6
+ class CustomLoggerExtension:
7
+ """
8
+ 日志系统初始化扩展
9
+ 遵循与 ExtensionManager 一致的接口规范:使用 create_instance
10
+ """
11
+
12
+ def __init__(self, settings):
13
+ self.settings = settings
14
+ # 初始化全局日志配置
15
+ LoggerManager.configure(settings)
16
+
17
+ @classmethod
18
+ def create_instance(cls, crawler, *args, **kwargs):
19
+ """
20
+ 工厂方法:兼容 ExtensionManager 的创建方式
21
+ 被 ExtensionManager 调用
22
+ """
23
+ # 可以通过 settings 控制是否启用
24
+ if not crawler.settings.get('LOG_FILE') and not crawler.settings.get('LOG_ENABLE_CUSTOM'):
25
+ raise NotConfigured("CustomLoggerExtension: LOG_FILE not set and LOG_ENABLE_CUSTOM=False")
26
+
27
+ return cls(crawler.settings)
28
+
29
+ def spider_opened(self, spider):
30
+ logger = get_logger(__name__)
31
+ logger.info(
32
+ f"CustomLoggerExtension: Logging initialized. "
33
+ f"LOG_FILE={self.settings.get('LOG_FILE')}, "
34
+ f"LOG_LEVEL={self.settings.get('LOG_LEVEL')}"
35
+ )
@@ -1,37 +1,37 @@
1
- #!/usr/bin/python
2
- # -*- coding:UTF-8 -*-
3
- from abc import ABC, abstractmethod
4
-
5
- from crawlo import Request
6
- from crawlo.utils.request import request_fingerprint
7
-
8
-
9
- class BaseFilter(ABC):
10
-
11
- def __init__(self, logger, stats, debug: bool):
12
- self.logger = logger
13
- self.stats = stats
14
- self.debug = debug
15
-
16
- @classmethod
17
- def create_instance(cls, *args, **kwargs) -> 'BaseFilter':
18
- return cls(*args, **kwargs)
19
-
20
- def requested(self, request: Request):
21
- fp = request_fingerprint(request)
22
- if fp in self:
23
- return True
24
- self.add_fingerprint(fp)
25
- return False
26
-
27
- @abstractmethod
28
- def add_fingerprint(self, fp) -> None:
29
- pass
30
-
31
- def log_stats(self, request: Request) -> None:
32
- if self.debug:
33
- self.logger.debug(f'Filtered duplicate request: {request}')
34
- self.stats.inc_value(f'{self}/filtered_count')
35
-
36
- def __str__(self) -> str:
37
- return f'{self.__class__.__name__}'
1
+ #!/usr/bin/python
2
+ # -*- coding:UTF-8 -*-
3
+ from abc import ABC, abstractmethod
4
+
5
+ from crawlo import Request
6
+ from crawlo.utils.request import request_fingerprint
7
+
8
+
9
+ class BaseFilter(ABC):
10
+
11
+ def __init__(self, logger, stats, debug: bool):
12
+ self.logger = logger
13
+ self.stats = stats
14
+ self.debug = debug
15
+
16
+ @classmethod
17
+ def create_instance(cls, *args, **kwargs) -> 'BaseFilter':
18
+ return cls(*args, **kwargs)
19
+
20
+ def requested(self, request: Request):
21
+ fp = request_fingerprint(request)
22
+ if fp in self:
23
+ return True
24
+ self.add_fingerprint(fp)
25
+ return False
26
+
27
+ @abstractmethod
28
+ def add_fingerprint(self, fp) -> None:
29
+ pass
30
+
31
+ def log_stats(self, request: Request) -> None:
32
+ if self.debug:
33
+ self.logger.debug(f'Filtered duplicate request: {request}')
34
+ self.stats.inc_value(f'{self}/filtered_count')
35
+
36
+ def __str__(self) -> str:
37
+ return f'{self.__class__.__name__}'