crawlo 1.4.6__py3-none-any.whl → 1.4.8__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of crawlo might be problematic. Click here for more details.

Files changed (162) hide show
  1. crawlo/__init__.py +2 -1
  2. crawlo/__version__.py +1 -1
  3. crawlo/cli.py +2 -2
  4. crawlo/commands/check.py +1 -1
  5. crawlo/commands/help.py +5 -3
  6. crawlo/commands/list.py +1 -1
  7. crawlo/commands/run.py +49 -11
  8. crawlo/commands/stats.py +1 -1
  9. crawlo/config.py +12 -4
  10. crawlo/config_validator.py +1 -1
  11. crawlo/core/engine.py +20 -7
  12. crawlo/core/processor.py +1 -1
  13. crawlo/core/scheduler.py +4 -5
  14. crawlo/crawler.py +51 -10
  15. crawlo/downloader/__init__.py +7 -3
  16. crawlo/downloader/aiohttp_downloader.py +18 -18
  17. crawlo/downloader/cffi_downloader.py +5 -2
  18. crawlo/downloader/httpx_downloader.py +9 -3
  19. crawlo/downloader/hybrid_downloader.py +2 -2
  20. crawlo/downloader/playwright_downloader.py +38 -15
  21. crawlo/downloader/selenium_downloader.py +16 -2
  22. crawlo/event.py +42 -8
  23. crawlo/exceptions.py +157 -24
  24. crawlo/extension/__init__.py +10 -9
  25. crawlo/extension/health_check.py +7 -7
  26. crawlo/extension/log_interval.py +6 -6
  27. crawlo/extension/log_stats.py +2 -2
  28. crawlo/extension/logging_extension.py +4 -12
  29. crawlo/extension/memory_monitor.py +5 -5
  30. crawlo/extension/performance_profiler.py +5 -5
  31. crawlo/extension/request_recorder.py +6 -6
  32. crawlo/factories/base.py +1 -1
  33. crawlo/factories/crawler.py +61 -60
  34. crawlo/factories/utils.py +135 -0
  35. crawlo/filters/__init__.py +19 -2
  36. crawlo/filters/aioredis_filter.py +133 -49
  37. crawlo/filters/memory_filter.py +6 -21
  38. crawlo/framework.py +22 -8
  39. crawlo/initialization/built_in.py +24 -67
  40. crawlo/initialization/core.py +65 -19
  41. crawlo/initialization/phases.py +83 -2
  42. crawlo/initialization/registry.py +5 -7
  43. crawlo/initialization/utils.py +49 -0
  44. crawlo/logging/__init__.py +6 -10
  45. crawlo/logging/config.py +106 -22
  46. crawlo/logging/factory.py +12 -8
  47. crawlo/logging/manager.py +19 -27
  48. crawlo/middleware/__init__.py +72 -9
  49. crawlo/middleware/default_header.py +2 -2
  50. crawlo/middleware/download_delay.py +2 -2
  51. crawlo/middleware/middleware_manager.py +6 -6
  52. crawlo/middleware/offsite.py +2 -2
  53. crawlo/middleware/proxy.py +2 -2
  54. crawlo/middleware/request_ignore.py +4 -4
  55. crawlo/middleware/response_code.py +2 -2
  56. crawlo/middleware/response_filter.py +2 -2
  57. crawlo/middleware/retry.py +1 -1
  58. crawlo/mode_manager.py +38 -4
  59. crawlo/network/request.py +54 -26
  60. crawlo/network/response.py +69 -135
  61. crawlo/pipelines/__init__.py +40 -9
  62. crawlo/pipelines/base_pipeline.py +452 -0
  63. crawlo/pipelines/bloom_dedup_pipeline.py +4 -5
  64. crawlo/pipelines/console_pipeline.py +2 -2
  65. crawlo/pipelines/csv_pipeline.py +4 -4
  66. crawlo/pipelines/database_dedup_pipeline.py +4 -5
  67. crawlo/pipelines/json_pipeline.py +4 -4
  68. crawlo/pipelines/memory_dedup_pipeline.py +4 -5
  69. crawlo/pipelines/mongo_pipeline.py +23 -14
  70. crawlo/pipelines/mysql_pipeline.py +31 -39
  71. crawlo/pipelines/pipeline_manager.py +8 -8
  72. crawlo/pipelines/redis_dedup_pipeline.py +13 -14
  73. crawlo/project.py +1 -1
  74. crawlo/queue/__init__.py +10 -0
  75. crawlo/queue/queue_manager.py +79 -13
  76. crawlo/queue/redis_priority_queue.py +196 -47
  77. crawlo/settings/default_settings.py +16 -6
  78. crawlo/spider/__init__.py +6 -5
  79. crawlo/stats_collector.py +2 -2
  80. crawlo/task_manager.py +1 -1
  81. crawlo/templates/crawlo.cfg.tmpl +3 -3
  82. crawlo/templates/project/__init__.py.tmpl +1 -3
  83. crawlo/templates/project/items.py.tmpl +2 -6
  84. crawlo/templates/project/middlewares.py.tmpl +1 -1
  85. crawlo/templates/project/pipelines.py.tmpl +1 -2
  86. crawlo/templates/project/settings.py.tmpl +12 -10
  87. crawlo/templates/project/settings_distributed.py.tmpl +14 -13
  88. crawlo/templates/project/settings_gentle.py.tmpl +21 -23
  89. crawlo/templates/project/settings_high_performance.py.tmpl +21 -23
  90. crawlo/templates/project/settings_minimal.py.tmpl +10 -8
  91. crawlo/templates/project/settings_simple.py.tmpl +21 -23
  92. crawlo/templates/run.py.tmpl +1 -1
  93. crawlo/templates/spider/spider.py.tmpl +4 -12
  94. crawlo/templates/spiders_init.py.tmpl +3 -8
  95. crawlo/tools/__init__.py +0 -103
  96. crawlo/tools/scenario_adapter.py +1 -1
  97. crawlo/utils/__init__.py +25 -1
  98. crawlo/utils/batch_processor.py +23 -6
  99. crawlo/utils/config_manager.py +442 -0
  100. crawlo/utils/controlled_spider_mixin.py +1 -1
  101. crawlo/utils/db_helper.py +1 -1
  102. crawlo/utils/encoding_helper.py +190 -0
  103. crawlo/utils/error_handler.py +2 -2
  104. crawlo/utils/large_scale_helper.py +1 -1
  105. crawlo/utils/leak_detector.py +335 -0
  106. crawlo/utils/mongo_connection_pool.py +157 -0
  107. crawlo/utils/mysql_connection_pool.py +197 -0
  108. crawlo/utils/performance_monitor.py +1 -1
  109. crawlo/utils/redis_checker.py +91 -0
  110. crawlo/utils/redis_connection_pool.py +260 -70
  111. crawlo/utils/redis_key_validator.py +1 -1
  112. crawlo/utils/request.py +24 -2
  113. crawlo/utils/request_serializer.py +1 -1
  114. crawlo/utils/resource_manager.py +337 -0
  115. crawlo/utils/response_helper.py +113 -0
  116. crawlo/utils/selector_helper.py +3 -2
  117. crawlo/utils/singleton.py +70 -0
  118. crawlo/utils/spider_loader.py +1 -1
  119. crawlo/utils/text_helper.py +1 -1
  120. crawlo-1.4.8.dist-info/METADATA +831 -0
  121. {crawlo-1.4.6.dist-info → crawlo-1.4.8.dist-info}/RECORD +131 -145
  122. tests/advanced_tools_example.py +10 -68
  123. tests/distributed_dedup_test.py +467 -0
  124. tests/monitor_redis_dedup.sh +72 -0
  125. tests/ofweek_scrapy/ofweek_scrapy/spiders/__init__.py +4 -4
  126. tests/simple_cli_test.py +55 -0
  127. tests/test_cli_arguments.py +119 -0
  128. tests/test_dedup_fix.py +10 -10
  129. crawlo/logging/async_handler.py +0 -181
  130. crawlo/logging/monitor.py +0 -153
  131. crawlo/logging/sampler.py +0 -167
  132. crawlo/tools/authenticated_proxy.py +0 -241
  133. crawlo/tools/data_formatter.py +0 -226
  134. crawlo/tools/data_validator.py +0 -181
  135. crawlo/tools/encoding_converter.py +0 -127
  136. crawlo/tools/network_diagnostic.py +0 -365
  137. crawlo/tools/request_tools.py +0 -83
  138. crawlo/tools/retry_mechanism.py +0 -224
  139. crawlo/utils/env_config.py +0 -143
  140. crawlo/utils/large_scale_config.py +0 -287
  141. crawlo/utils/log.py +0 -80
  142. crawlo/utils/system.py +0 -11
  143. crawlo/utils/tools.py +0 -5
  144. crawlo/utils/url.py +0 -40
  145. crawlo-1.4.6.dist-info/METADATA +0 -329
  146. tests/env_config_example.py +0 -134
  147. tests/ofweek_scrapy/ofweek_scrapy/spiders/ofweek_spider.py +0 -162
  148. tests/test_authenticated_proxy.py +0 -142
  149. tests/test_comprehensive.py +0 -147
  150. tests/test_dynamic_downloaders_proxy.py +0 -125
  151. tests/test_dynamic_proxy.py +0 -93
  152. tests/test_dynamic_proxy_config.py +0 -147
  153. tests/test_dynamic_proxy_real.py +0 -110
  154. tests/test_env_config.py +0 -122
  155. tests/test_framework_env_usage.py +0 -104
  156. tests/test_large_scale_config.py +0 -113
  157. tests/test_proxy_api.py +0 -265
  158. tests/test_real_scenario_proxy.py +0 -196
  159. tests/tools_example.py +0 -261
  160. {crawlo-1.4.6.dist-info → crawlo-1.4.8.dist-info}/WHEEL +0 -0
  161. {crawlo-1.4.6.dist-info → crawlo-1.4.8.dist-info}/entry_points.txt +0 -0
  162. {crawlo-1.4.6.dist-info → crawlo-1.4.8.dist-info}/top_level.txt +0 -0
@@ -1,147 +0,0 @@
1
- #!/usr/bin/python
2
- # -*- coding: UTF-8 -*-
3
- """
4
- 测试动态下载器(Selenium和Playwright)的代理配置逻辑
5
- """
6
-
7
- from crawlo.tools import AuthenticatedProxy
8
-
9
-
10
- def test_selenium_proxy_logic():
11
- """测试Selenium下载器的代理配置逻辑"""
12
- print("=== 测试Selenium下载器的代理配置逻辑 ===")
13
-
14
- # 代理配置
15
- proxy_config = {
16
- "http": "http://dwe20241014:Dwe0101014@182.201.243.186:58111",
17
- "https": "http://dwe20241014:Dwe0101014@182.201.243.186:58111"
18
- }
19
-
20
- # 创建代理对象
21
- proxy_url = proxy_config["http"]
22
- proxy = AuthenticatedProxy(proxy_url)
23
-
24
- print(f"原始代理URL: {proxy_url}")
25
- print(f"清洁URL: {proxy.clean_url}")
26
- print(f"认证信息: {proxy.get_auth_credentials()}")
27
-
28
- # 模拟Selenium的代理配置逻辑
29
- print(f"\nSelenium代理配置逻辑:")
30
- print(f" 1. 在爬虫设置中配置:")
31
- print(f" settings = {{")
32
- print(f" 'SELENIUM_PROXY': '{proxy.clean_url}',")
33
- print(f" }}")
34
-
35
- # 对于带认证的代理,需要特殊处理
36
- if proxy.username and proxy.password:
37
- print(f"\n 2. 带认证代理的处理逻辑:")
38
- print(f" - 用户名: {proxy.username}")
39
- print(f" - 密码: {proxy.password}")
40
- print(f" - 认证头: {proxy.get_auth_header()}")
41
- print(f" - 处理方式: 通过浏览器扩展或手动输入认证信息")
42
-
43
- print("\nSelenium配置逻辑测试完成!")
44
-
45
-
46
- def test_playwright_proxy_logic():
47
- """测试Playwright下载器的代理配置逻辑"""
48
- print("\n=== 测试Playwright下载器的代理配置逻辑 ===")
49
-
50
- # 代理配置
51
- proxy_config = {
52
- "http": "http://dwe20241014:Dwe0101014@182.201.243.186:58111",
53
- "https": "http://dwe20241014:Dwe0101014@182.201.243.186:58111"
54
- }
55
-
56
- # 创建代理对象
57
- proxy_url = proxy_config["http"]
58
- proxy = AuthenticatedProxy(proxy_url)
59
-
60
- print(f"原始代理URL: {proxy_url}")
61
- print(f"清洁URL: {proxy.clean_url}")
62
- print(f"认证信息: {proxy.get_auth_credentials()}")
63
-
64
- # 模拟Playwright的代理配置逻辑
65
- print(f"\nPlaywright代理配置逻辑:")
66
- print(f" 1. 简单代理配置:")
67
- print(f" settings = {{")
68
- print(f" 'PLAYWRIGHT_PROXY': '{proxy.clean_url}',")
69
- print(f" }}")
70
-
71
- # 对于带认证的代理,Playwright可以直接在代理配置中包含认证信息
72
- if proxy.username and proxy.password:
73
- print(f"\n 2. 带认证的代理配置逻辑:")
74
- print(f" settings = {{")
75
- print(f" 'PLAYWRIGHT_PROXY': {{")
76
- print(f" 'server': '{proxy.clean_url}',")
77
- print(f" 'username': '{proxy.username}',")
78
- print(f" 'password': '{proxy.password}'")
79
- print(f" }}")
80
- print(f" }}")
81
- print(f" 实现方式: 在启动浏览器时传递proxy参数")
82
-
83
- print("\nPlaywright配置逻辑测试完成!")
84
-
85
-
86
- def show_proxy_configuration_examples():
87
- """显示代理配置示例"""
88
- print("\n=== 代理配置示例 ===")
89
-
90
- # 不同类型的代理配置示例
91
- examples = [
92
- {
93
- "name": "带认证HTTP代理",
94
- "url": "http://username:password@proxy.example.com:8080",
95
- "selenium_config": "SELENIUM_PROXY = 'http://proxy.example.com:8080'",
96
- "playwright_config": "PLAYWRIGHT_PROXY = {'server': 'http://proxy.example.com:8080', 'username': 'username', 'password': 'password'}"
97
- },
98
- {
99
- "name": "带认证HTTPS代理",
100
- "url": "https://user:pass@secure-proxy.example.com:443",
101
- "selenium_config": "SELENIUM_PROXY = 'https://secure-proxy.example.com:443'",
102
- "playwright_config": "PLAYWRIGHT_PROXY = {'server': 'https://secure-proxy.example.com:443', 'username': 'user', 'password': 'pass'}"
103
- },
104
- {
105
- "name": "SOCKS5代理",
106
- "url": "socks5://username:password@socks-proxy.example.com:1080",
107
- "selenium_config": "SELENIUM_PROXY = 'socks5://socks-proxy.example.com:1080'",
108
- "playwright_config": "PLAYWRIGHT_PROXY = {'server': 'socks5://socks-proxy.example.com:1080', 'username': 'username', 'password': 'password'}"
109
- },
110
- {
111
- "name": "不带认证代理",
112
- "url": "http://proxy.example.com:8080",
113
- "selenium_config": "SELENIUM_PROXY = 'http://proxy.example.com:8080'",
114
- "playwright_config": "PLAYWRIGHT_PROXY = 'http://proxy.example.com:8080'"
115
- }
116
- ]
117
-
118
- for i, example in enumerate(examples, 1):
119
- print(f"\n示例 {i}: {example['name']}")
120
- print(f" 代理URL: {example['url']}")
121
- proxy = AuthenticatedProxy(example['url'])
122
- print(f" 清洁URL: {proxy.clean_url}")
123
- print(f" 用户名: {proxy.username or '无'}")
124
- print(f" 密码: {proxy.password or '无'}")
125
- print(f" Selenium配置: {example['selenium_config']}")
126
- print(f" Playwright配置: {example['playwright_config']}")
127
-
128
-
129
- def main():
130
- """主函数"""
131
- print("开始测试动态下载器的代理配置逻辑...\n")
132
-
133
- # 测试各个下载器的配置逻辑
134
- test_selenium_proxy_logic()
135
- test_playwright_proxy_logic()
136
- show_proxy_configuration_examples()
137
-
138
- print("\n所有配置逻辑测试完成!")
139
- print("\n总结:")
140
- print("1. Selenium和Playwright都支持代理配置")
141
- print("2. 带认证的代理需要特殊处理")
142
- print("3. Playwright对带认证代理的支持更直接")
143
- print("4. Selenium需要通过扩展或其他方式处理认证")
144
-
145
-
146
- if __name__ == "__main__":
147
- main()
@@ -1,110 +0,0 @@
1
- #!/usr/bin/python
2
- # -*- coding: UTF-8 -*-
3
- """
4
- 实际测试动态下载器(Playwright)通过代理访问网站
5
- """
6
-
7
- import asyncio
8
- from crawlo.spider import Spider
9
- from crawlo.network.request import Request
10
- from crawlo.tools import AuthenticatedProxy
11
-
12
-
13
- class ProxyTestSpider(Spider):
14
- """代理测试爬虫"""
15
- name = "proxy_test_spider" # 添加name属性
16
-
17
- # 自定义配置
18
- custom_settings = {
19
- "DOWNLOADER_TYPE": "playwright",
20
- "PLAYWRIGHT_BROWSER_TYPE": "chromium",
21
- "PLAYWRIGHT_HEADLESS": True,
22
- # 配置带认证的代理
23
- "PLAYWRIGHT_PROXY": {
24
- "server": "http://182.201.243.186:58111",
25
- "username": "dwe20241014",
26
- "password": "Dwe0101014"
27
- }
28
- }
29
-
30
- def __init__(self, **kwargs):
31
- super().__init__(**kwargs)
32
- print("代理测试爬虫初始化完成")
33
- print(f"代理配置: {self.custom_settings.get('PLAYWRIGHT_PROXY')}")
34
-
35
- def start_requests(self):
36
- """开始请求"""
37
- urls = [
38
- "https://httpbin.org/ip", # 查看IP地址
39
- "https://httpbin.org/headers", # 查看请求头
40
- ]
41
-
42
- for url in urls:
43
- request = Request(url, callback=self.parse)
44
- yield request
45
-
46
- def parse(self, response):
47
- """解析响应"""
48
- print(f"\n=== 响应信息 ===")
49
- print(f"URL: {response.url}")
50
- print(f"状态码: {response.status_code}")
51
- print(f"响应内容: {response.text[:500]}")
52
-
53
- # 保存响应内容
54
- filename = response.url.split("/")[-1].replace("?", "_").replace("&", "_")
55
- with open(f"proxy_test_{filename}.html", "w", encoding="utf-8") as f:
56
- f.write(response.text)
57
- print(f"响应已保存到 proxy_test_{filename}.html")
58
-
59
- # 返回一个简单的item
60
- yield {"url": response.url, "status": response.status_code}
61
-
62
-
63
- # 异步运行函数
64
- async def run_spider():
65
- """运行爬虫"""
66
- print("开始测试动态下载器通过代理访问网站...")
67
-
68
- # 创建爬虫实例
69
- spider = ProxyTestSpider()
70
-
71
- # 创建一个简单的crawler模拟器
72
- class MockCrawler:
73
- def __init__(self):
74
- from crawlo.settings.setting_manager import SettingManager
75
- self.settings = SettingManager()
76
- # 应用爬虫的自定义设置
77
- if hasattr(spider, 'custom_settings'):
78
- for key, value in spider.custom_settings.items():
79
- self.settings.set(key, value)
80
-
81
- crawler = MockCrawler()
82
-
83
- # 创建爬虫实例并绑定crawler
84
- spider_instance = spider.create_instance(crawler)
85
-
86
- # 执行初始请求
87
- requests = list(spider_instance.start_requests())
88
- print(f"生成了 {len(requests)} 个请求")
89
-
90
- # 使用Playwright下载器处理请求
91
- try:
92
- from crawlo.downloader import PlaywrightDownloader
93
- downloader = PlaywrightDownloader(crawler)
94
- await downloader.download(requests[0]) # 测试第一个请求
95
- print("Playwright下载器测试成功!")
96
- except Exception as e:
97
- print(f"Playwright下载器测试失败: {e}")
98
- import traceback
99
- traceback.print_exc()
100
-
101
- print("\n测试完成!")
102
-
103
-
104
- async def main():
105
- """主函数"""
106
- await run_spider()
107
-
108
-
109
- if __name__ == "__main__":
110
- asyncio.run(main())
tests/test_env_config.py DELETED
@@ -1,122 +0,0 @@
1
- #!/usr/bin/env python3
2
- # -*- coding: utf-8 -*-
3
- """
4
- 环境变量配置工具测试
5
- """
6
- import os
7
- import sys
8
- import unittest
9
- from unittest.mock import patch
10
-
11
- # 添加项目根目录到Python路径
12
- sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..'))
13
-
14
- from crawlo.utils.env_config import EnvConfigManager, get_env_var, get_redis_config, get_runtime_config
15
-
16
-
17
- class TestEnvConfigManager(unittest.TestCase):
18
- """环境变量配置管理器测试"""
19
-
20
- def test_get_env_var_str(self):
21
- """测试字符串环境变量获取"""
22
- with patch.dict(os.environ, {'TEST_VAR': 'test_value'}):
23
- result = EnvConfigManager.get_env_var('TEST_VAR', 'default', str)
24
- self.assertEqual(result, 'test_value')
25
-
26
- # 测试默认值
27
- result = EnvConfigManager.get_env_var('NON_EXISTENT_VAR', 'default', str)
28
- self.assertEqual(result, 'default')
29
-
30
- def test_get_env_var_int(self):
31
- """测试整数环境变量获取"""
32
- with patch.dict(os.environ, {'TEST_INT': '42'}):
33
- result = EnvConfigManager.get_env_var('TEST_INT', 0, int)
34
- self.assertEqual(result, 42)
35
-
36
- # 测试默认值
37
- result = EnvConfigManager.get_env_var('NON_EXISTENT_VAR', 10, int)
38
- self.assertEqual(result, 10)
39
-
40
- # 测试无效值
41
- with patch.dict(os.environ, {'INVALID_INT': 'not_a_number'}):
42
- result = EnvConfigManager.get_env_var('INVALID_INT', 5, int)
43
- self.assertEqual(result, 5)
44
-
45
- def test_get_env_var_float(self):
46
- """测试浮点数环境变量获取"""
47
- with patch.dict(os.environ, {'TEST_FLOAT': '3.14'}):
48
- result = EnvConfigManager.get_env_var('TEST_FLOAT', 0.0, float)
49
- self.assertEqual(result, 3.14)
50
-
51
- # 测试默认值
52
- result = EnvConfigManager.get_env_var('NON_EXISTENT_VAR', 2.5, float)
53
- self.assertEqual(result, 2.5)
54
-
55
- def test_get_env_var_bool(self):
56
- """测试布尔环境变量获取"""
57
- # 测试 True 值
58
- for true_val in ['1', 'true', 'True', 'TRUE', 'yes', 'on']:
59
- with patch.dict(os.environ, {'TEST_BOOL': true_val}):
60
- result = EnvConfigManager.get_env_var('TEST_BOOL', False, bool)
61
- self.assertTrue(result)
62
-
63
- # 测试 False 值
64
- for false_val in ['0', 'false', 'False', 'FALSE', 'no', 'off', '']:
65
- with patch.dict(os.environ, {'TEST_BOOL': false_val}):
66
- result = EnvConfigManager.get_env_var('TEST_BOOL', True, bool)
67
- self.assertFalse(result)
68
-
69
- # 测试默认值
70
- result = EnvConfigManager.get_env_var('NON_EXISTENT_VAR', True, bool)
71
- self.assertTrue(result)
72
-
73
- def test_get_redis_config(self):
74
- """测试 Redis 配置获取"""
75
- with patch.dict(os.environ, {
76
- 'REDIS_HOST': 'localhost',
77
- 'REDIS_PORT': '6380',
78
- 'REDIS_PASSWORD': 'secret',
79
- 'REDIS_DB': '1'
80
- }):
81
- config = get_redis_config()
82
- self.assertEqual(config['REDIS_HOST'], 'localhost')
83
- self.assertEqual(config['REDIS_PORT'], 6380)
84
- self.assertEqual(config['REDIS_PASSWORD'], 'secret')
85
- self.assertEqual(config['REDIS_DB'], 1)
86
-
87
- # 测试默认值
88
- with patch.dict(os.environ, {}):
89
- config = get_redis_config()
90
- self.assertEqual(config['REDIS_HOST'], '127.0.0.1')
91
- self.assertEqual(config['REDIS_PORT'], 6379)
92
- self.assertEqual(config['REDIS_PASSWORD'], '')
93
- self.assertEqual(config['REDIS_DB'], 0)
94
-
95
- def test_get_runtime_config(self):
96
- """测试运行时配置获取"""
97
- with patch.dict(os.environ, {
98
- 'CRAWLO_MODE': 'distributed',
99
- 'PROJECT_NAME': 'test_project',
100
- 'CONCURRENCY': '20'
101
- }):
102
- config = get_runtime_config()
103
- self.assertEqual(config['CRAWLO_MODE'], 'distributed')
104
- self.assertEqual(config['PROJECT_NAME'], 'test_project')
105
- self.assertEqual(config['CONCURRENCY'], 20)
106
-
107
- # 测试默认值
108
- with patch.dict(os.environ, {}):
109
- config = get_runtime_config()
110
- self.assertEqual(config['CRAWLO_MODE'], 'standalone')
111
- self.assertEqual(config['PROJECT_NAME'], 'crawlo')
112
- self.assertEqual(config['CONCURRENCY'], 8)
113
-
114
- def test_convenience_functions(self):
115
- """测试便捷函数"""
116
- with patch.dict(os.environ, {'TEST_CONVENIENCE': 'convenience_value'}):
117
- result = get_env_var('TEST_CONVENIENCE', 'default')
118
- self.assertEqual(result, 'convenience_value')
119
-
120
-
121
- if __name__ == '__main__':
122
- unittest.main()
@@ -1,104 +0,0 @@
1
- #!/usr/bin/env python3
2
- # -*- coding: utf-8 -*-
3
- """
4
- 框架环境变量使用测试
5
- 验证整个框架中环境变量的正确使用
6
- """
7
- import sys
8
- import os
9
- import unittest
10
- from unittest.mock import patch, MagicMock
11
-
12
- # 添加项目根目录到Python路径
13
- sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..'))
14
-
15
- from crawlo.utils.env_config import get_env_var, get_redis_config, get_runtime_config
16
- from crawlo.settings.setting_manager import SettingManager
17
- from crawlo.settings import default_settings
18
- from crawlo.mode_manager import from_env
19
-
20
-
21
- class TestFrameworkEnvUsage(unittest.TestCase):
22
- """框架环境变量使用测试"""
23
-
24
- def test_default_settings_env_usage(self):
25
- """测试 default_settings.py 中的环境变量使用"""
26
- # 验证 default_settings.py 不直接使用 os.getenv
27
- import inspect
28
- import crawlo.settings.default_settings as default_settings_module
29
-
30
- source_code = inspect.getsource(default_settings_module)
31
- # 检查是否还有直接使用 os.getenv 的地方
32
- self.assertNotIn('os.getenv', source_code,
33
- "default_settings.py 不应该直接使用 os.getenv")
34
-
35
- # 但应该使用 env_config 工具
36
- self.assertIn('get_redis_config', source_code,
37
- "default_settings.py 应该使用 get_redis_config")
38
- self.assertIn('get_runtime_config', source_code,
39
- "default_settings.py 应该使用 get_runtime_config")
40
-
41
- def test_env_config_tool(self):
42
- """测试环境变量配置工具"""
43
- # 测试获取Redis配置
44
- with patch.dict(os.environ, {
45
- 'REDIS_HOST': 'test.redis.com',
46
- 'REDIS_PORT': '6380',
47
- 'REDIS_PASSWORD': 'test_pass',
48
- 'REDIS_DB': '2'
49
- }):
50
- redis_config = get_redis_config()
51
- self.assertEqual(redis_config['REDIS_HOST'], 'test.redis.com')
52
- self.assertEqual(redis_config['REDIS_PORT'], 6380)
53
- self.assertEqual(redis_config['REDIS_PASSWORD'], 'test_pass')
54
- self.assertEqual(redis_config['REDIS_DB'], 2)
55
-
56
- # 测试获取运行时配置
57
- with patch.dict(os.environ, {
58
- 'PROJECT_NAME': 'test_project',
59
- 'CRAWLO_MODE': 'distributed',
60
- 'CONCURRENCY': '16'
61
- }):
62
- runtime_config = get_runtime_config()
63
- self.assertEqual(runtime_config['PROJECT_NAME'], 'test_project')
64
- self.assertEqual(runtime_config['CRAWLO_MODE'], 'distributed')
65
- self.assertEqual(runtime_config['CONCURRENCY'], 16)
66
-
67
- def test_settings_manager_with_env(self):
68
- """测试设置管理器与环境变量的集成"""
69
- # 设置环境变量
70
- env_vars = {
71
- 'PROJECT_NAME': 'env_test_project',
72
- 'CONCURRENCY': '12',
73
- 'REDIS_HOST': 'env.redis.test',
74
- 'REDIS_PORT': '6381'
75
- }
76
-
77
- with patch.dict(os.environ, env_vars):
78
- # 重新导入 default_settings 模块以获取最新的环境变量
79
- import importlib
80
- import crawlo.settings.default_settings
81
- importlib.reload(crawlo.settings.default_settings)
82
-
83
- # 创建设置管理器
84
- settings = SettingManager()
85
- settings.set_settings(crawlo.settings.default_settings)
86
-
87
- # 验证环境变量被正确使用
88
- redis_config = get_redis_config()
89
- self.assertEqual(settings.get('REDIS_HOST'), redis_config['REDIS_HOST'])
90
-
91
- runtime_config = get_runtime_config()
92
- self.assertEqual(settings.get('PROJECT_NAME'), runtime_config['PROJECT_NAME'])
93
-
94
- def test_mode_manager_env_usage(self):
95
- """测试 mode_manager.py 中的环境变量使用"""
96
- # 验证 from_env 函数现在会抛出异常
97
- with self.assertRaises(RuntimeError) as context:
98
- from_env()
99
-
100
- self.assertIn("环境变量配置已移除", str(context.exception))
101
-
102
-
103
- if __name__ == '__main__':
104
- unittest.main()
@@ -1,113 +0,0 @@
1
- #!/usr/bin/env python3
2
- # -*- coding: utf-8 -*-
3
- """
4
- 大规模配置工具测试
5
- 测试 LargeScaleConfig, apply_large_scale_config
6
- """
7
- import sys
8
- import os
9
- import unittest
10
- from unittest.mock import Mock, patch, MagicMock
11
-
12
- # 添加项目根目录到 Python 路径
13
- sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..'))
14
-
15
- from crawlo.utils.large_scale_config import LargeScaleConfig, apply_large_scale_config
16
-
17
-
18
- class TestLargeScaleConfig(unittest.TestCase):
19
- """大规模配置工具测试"""
20
-
21
- def test_conservative_config(self):
22
- """测试保守配置"""
23
- config = LargeScaleConfig.conservative_config(concurrency=8)
24
-
25
- # 验证基本配置
26
- self.assertEqual(config['CONCURRENCY'], 8)
27
- self.assertEqual(config['SCHEDULER_MAX_QUEUE_SIZE'], 80) # 8 * 10
28
- self.assertEqual(config['DOWNLOAD_DELAY'], 0.2)
29
- self.assertEqual(config['MAX_RUNNING_SPIDERS'], 1)
30
-
31
- # 验证连接池配置
32
- self.assertEqual(config['CONNECTION_POOL_LIMIT'], 16) # 8 * 2
33
-
34
- # 验证重试配置
35
- self.assertEqual(config['MAX_RETRY_TIMES'], 2)
36
-
37
- def test_balanced_config(self):
38
- """测试平衡配置"""
39
- config = LargeScaleConfig.balanced_config(concurrency=16)
40
-
41
- # 验证基本配置
42
- self.assertEqual(config['CONCURRENCY'], 16)
43
- self.assertEqual(config['SCHEDULER_MAX_QUEUE_SIZE'], 240) # 16 * 15
44
- self.assertEqual(config['DOWNLOAD_DELAY'], 0.1)
45
- self.assertEqual(config['MAX_RUNNING_SPIDERS'], 2)
46
-
47
- # 验证连接池配置
48
- self.assertEqual(config['CONNECTION_POOL_LIMIT'], 48) # 16 * 3
49
-
50
- # 验证重试配置
51
- self.assertEqual(config['MAX_RETRY_TIMES'], 3)
52
-
53
- def test_aggressive_config(self):
54
- """测试激进配置"""
55
- config = LargeScaleConfig.aggressive_config(concurrency=32)
56
-
57
- # 验证基本配置
58
- self.assertEqual(config['CONCURRENCY'], 32)
59
- self.assertEqual(config['SCHEDULER_MAX_QUEUE_SIZE'], 640) # 32 * 20
60
- self.assertEqual(config['DOWNLOAD_DELAY'], 0.05)
61
- self.assertEqual(config['MAX_RUNNING_SPIDERS'], 3)
62
-
63
- # 验证连接池配置
64
- self.assertEqual(config['CONNECTION_POOL_LIMIT'], 128) # 32 * 4
65
-
66
- # 验证重试配置
67
- self.assertEqual(config['MAX_RETRY_TIMES'], 5)
68
-
69
- def test_memory_optimized_config(self):
70
- """测试内存优化配置"""
71
- config = LargeScaleConfig.memory_optimized_config(concurrency=12)
72
-
73
- # 验证基本配置
74
- self.assertEqual(config['CONCURRENCY'], 12)
75
- self.assertEqual(config['SCHEDULER_MAX_QUEUE_SIZE'], 60) # 12 * 5
76
- self.assertEqual(config['DOWNLOAD_DELAY'], 0.1)
77
- self.assertEqual(config['MAX_RUNNING_SPIDERS'], 1)
78
-
79
- # 验证连接池配置
80
- self.assertEqual(config['CONNECTION_POOL_LIMIT'], 12) # 12 * 1
81
-
82
- # 验证内存限制配置
83
- self.assertEqual(config['DOWNLOAD_MAXSIZE'], 2 * 1024 * 1024) # 2MB
84
- self.assertEqual(config['DOWNLOAD_WARN_SIZE'], 512 * 1024) # 512KB
85
-
86
- # 验证重试配置
87
- self.assertEqual(config['MAX_RETRY_TIMES'], 2)
88
-
89
- def test_apply_large_scale_config(self):
90
- """测试应用大规模配置"""
91
- settings_dict = {}
92
-
93
- # 应用平衡配置
94
- apply_large_scale_config(settings_dict, "balanced", 16)
95
-
96
- # 验证配置已应用
97
- self.assertEqual(settings_dict['CONCURRENCY'], 16)
98
- self.assertEqual(settings_dict['SCHEDULER_MAX_QUEUE_SIZE'], 240)
99
- self.assertEqual(settings_dict['DOWNLOAD_DELAY'], 0.1)
100
-
101
- def test_apply_large_scale_config_invalid_type(self):
102
- """测试应用无效的大规模配置类型"""
103
- settings_dict = {}
104
-
105
- # 应用无效配置类型应该抛出异常
106
- with self.assertRaises(ValueError) as context:
107
- apply_large_scale_config(settings_dict, "invalid_type", 16)
108
-
109
- self.assertIn("不支持的配置类型", str(context.exception))
110
-
111
-
112
- if __name__ == '__main__':
113
- unittest.main()