crawlo 1.4.5__py3-none-any.whl → 1.4.7__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of crawlo might be problematic. Click here for more details.
- crawlo/__init__.py +90 -89
- crawlo/__version__.py +1 -1
- crawlo/cli.py +75 -75
- crawlo/commands/__init__.py +14 -14
- crawlo/commands/check.py +594 -594
- crawlo/commands/genspider.py +186 -186
- crawlo/commands/help.py +140 -138
- crawlo/commands/list.py +155 -155
- crawlo/commands/run.py +379 -341
- crawlo/commands/startproject.py +460 -460
- crawlo/commands/stats.py +187 -187
- crawlo/commands/utils.py +196 -196
- crawlo/config.py +320 -312
- crawlo/config_validator.py +277 -277
- crawlo/core/__init__.py +52 -52
- crawlo/core/engine.py +451 -438
- crawlo/core/processor.py +47 -47
- crawlo/core/scheduler.py +290 -291
- crawlo/crawler.py +698 -657
- crawlo/data/__init__.py +5 -5
- crawlo/data/user_agents.py +194 -194
- crawlo/downloader/__init__.py +280 -276
- crawlo/downloader/aiohttp_downloader.py +233 -233
- crawlo/downloader/cffi_downloader.py +250 -245
- crawlo/downloader/httpx_downloader.py +265 -259
- crawlo/downloader/hybrid_downloader.py +212 -212
- crawlo/downloader/playwright_downloader.py +425 -402
- crawlo/downloader/selenium_downloader.py +486 -472
- crawlo/event.py +45 -11
- crawlo/exceptions.py +215 -82
- crawlo/extension/__init__.py +65 -64
- crawlo/extension/health_check.py +141 -141
- crawlo/extension/log_interval.py +94 -94
- crawlo/extension/log_stats.py +70 -70
- crawlo/extension/logging_extension.py +53 -61
- crawlo/extension/memory_monitor.py +104 -104
- crawlo/extension/performance_profiler.py +133 -133
- crawlo/extension/request_recorder.py +107 -107
- crawlo/factories/__init__.py +27 -27
- crawlo/factories/base.py +68 -68
- crawlo/factories/crawler.py +104 -103
- crawlo/factories/registry.py +84 -84
- crawlo/factories/utils.py +135 -0
- crawlo/filters/__init__.py +170 -153
- crawlo/filters/aioredis_filter.py +348 -264
- crawlo/filters/memory_filter.py +261 -276
- crawlo/framework.py +306 -292
- crawlo/initialization/__init__.py +44 -44
- crawlo/initialization/built_in.py +391 -434
- crawlo/initialization/context.py +141 -141
- crawlo/initialization/core.py +240 -194
- crawlo/initialization/phases.py +230 -149
- crawlo/initialization/registry.py +143 -145
- crawlo/initialization/utils.py +49 -0
- crawlo/interfaces.py +23 -23
- crawlo/items/__init__.py +23 -23
- crawlo/items/base.py +23 -23
- crawlo/items/fields.py +52 -52
- crawlo/items/items.py +104 -104
- crawlo/logging/__init__.py +42 -46
- crawlo/logging/config.py +277 -197
- crawlo/logging/factory.py +175 -171
- crawlo/logging/manager.py +104 -112
- crawlo/middleware/__init__.py +87 -24
- crawlo/middleware/default_header.py +132 -132
- crawlo/middleware/download_delay.py +104 -104
- crawlo/middleware/middleware_manager.py +142 -142
- crawlo/middleware/offsite.py +123 -123
- crawlo/middleware/proxy.py +209 -386
- crawlo/middleware/request_ignore.py +86 -86
- crawlo/middleware/response_code.py +150 -150
- crawlo/middleware/response_filter.py +136 -136
- crawlo/middleware/retry.py +124 -124
- crawlo/mode_manager.py +287 -253
- crawlo/network/__init__.py +21 -21
- crawlo/network/request.py +375 -379
- crawlo/network/response.py +569 -664
- crawlo/pipelines/__init__.py +53 -22
- crawlo/pipelines/base_pipeline.py +452 -0
- crawlo/pipelines/bloom_dedup_pipeline.py +146 -146
- crawlo/pipelines/console_pipeline.py +39 -39
- crawlo/pipelines/csv_pipeline.py +316 -316
- crawlo/pipelines/database_dedup_pipeline.py +197 -197
- crawlo/pipelines/json_pipeline.py +218 -218
- crawlo/pipelines/memory_dedup_pipeline.py +105 -105
- crawlo/pipelines/mongo_pipeline.py +140 -132
- crawlo/pipelines/mysql_pipeline.py +470 -326
- crawlo/pipelines/pipeline_manager.py +100 -100
- crawlo/pipelines/redis_dedup_pipeline.py +155 -156
- crawlo/project.py +347 -347
- crawlo/queue/__init__.py +10 -0
- crawlo/queue/pqueue.py +38 -38
- crawlo/queue/queue_manager.py +591 -525
- crawlo/queue/redis_priority_queue.py +519 -370
- crawlo/settings/__init__.py +7 -7
- crawlo/settings/default_settings.py +285 -270
- crawlo/settings/setting_manager.py +219 -219
- crawlo/spider/__init__.py +657 -657
- crawlo/stats_collector.py +82 -73
- crawlo/subscriber.py +129 -129
- crawlo/task_manager.py +138 -138
- crawlo/templates/crawlo.cfg.tmpl +10 -10
- crawlo/templates/project/__init__.py.tmpl +2 -4
- crawlo/templates/project/items.py.tmpl +13 -17
- crawlo/templates/project/middlewares.py.tmpl +38 -38
- crawlo/templates/project/pipelines.py.tmpl +35 -36
- crawlo/templates/project/settings.py.tmpl +110 -157
- crawlo/templates/project/settings_distributed.py.tmpl +156 -161
- crawlo/templates/project/settings_gentle.py.tmpl +170 -171
- crawlo/templates/project/settings_high_performance.py.tmpl +171 -172
- crawlo/templates/project/settings_minimal.py.tmpl +99 -77
- crawlo/templates/project/settings_simple.py.tmpl +168 -169
- crawlo/templates/project/spiders/__init__.py.tmpl +9 -9
- crawlo/templates/run.py.tmpl +23 -30
- crawlo/templates/spider/spider.py.tmpl +33 -144
- crawlo/templates/spiders_init.py.tmpl +5 -10
- crawlo/tools/__init__.py +86 -189
- crawlo/tools/date_tools.py +289 -289
- crawlo/tools/distributed_coordinator.py +384 -384
- crawlo/tools/scenario_adapter.py +262 -262
- crawlo/tools/text_cleaner.py +232 -232
- crawlo/utils/__init__.py +50 -50
- crawlo/utils/batch_processor.py +276 -259
- crawlo/utils/config_manager.py +442 -0
- crawlo/utils/controlled_spider_mixin.py +439 -439
- crawlo/utils/db_helper.py +250 -244
- crawlo/utils/error_handler.py +410 -410
- crawlo/utils/fingerprint.py +121 -121
- crawlo/utils/func_tools.py +82 -82
- crawlo/utils/large_scale_helper.py +344 -344
- crawlo/utils/leak_detector.py +335 -0
- crawlo/utils/log.py +79 -79
- crawlo/utils/misc.py +81 -81
- crawlo/utils/mongo_connection_pool.py +157 -0
- crawlo/utils/mysql_connection_pool.py +197 -0
- crawlo/utils/performance_monitor.py +285 -285
- crawlo/utils/queue_helper.py +175 -175
- crawlo/utils/redis_checker.py +91 -0
- crawlo/utils/redis_connection_pool.py +578 -388
- crawlo/utils/redis_key_validator.py +198 -198
- crawlo/utils/request.py +278 -256
- crawlo/utils/request_serializer.py +225 -225
- crawlo/utils/resource_manager.py +337 -0
- crawlo/utils/selector_helper.py +137 -137
- crawlo/utils/singleton.py +70 -0
- crawlo/utils/spider_loader.py +201 -201
- crawlo/utils/text_helper.py +94 -94
- crawlo/utils/{url.py → url_utils.py} +39 -39
- crawlo-1.4.7.dist-info/METADATA +689 -0
- crawlo-1.4.7.dist-info/RECORD +347 -0
- examples/__init__.py +7 -7
- tests/__init__.py +7 -7
- tests/advanced_tools_example.py +217 -275
- tests/authenticated_proxy_example.py +110 -106
- tests/baidu_performance_test.py +108 -108
- tests/baidu_test.py +59 -59
- tests/bug_check_test.py +250 -250
- tests/cleaners_example.py +160 -160
- tests/comprehensive_framework_test.py +212 -212
- tests/comprehensive_test.py +81 -81
- tests/comprehensive_testing_summary.md +186 -186
- tests/config_validation_demo.py +142 -142
- tests/controlled_spider_example.py +205 -205
- tests/date_tools_example.py +180 -180
- tests/debug_configure.py +69 -69
- tests/debug_framework_logger.py +84 -84
- tests/debug_log_config.py +126 -126
- tests/debug_log_levels.py +63 -63
- tests/debug_pipelines.py +66 -66
- tests/detailed_log_test.py +233 -233
- tests/direct_selector_helper_test.py +96 -96
- tests/distributed_dedup_test.py +467 -0
- tests/distributed_test.py +66 -66
- tests/distributed_test_debug.py +76 -76
- tests/dynamic_loading_example.py +523 -523
- tests/dynamic_loading_test.py +104 -104
- tests/error_handling_example.py +171 -171
- tests/explain_mysql_update_behavior.py +77 -0
- tests/final_comprehensive_test.py +151 -151
- tests/final_log_test.py +260 -260
- tests/final_validation_test.py +182 -182
- tests/fix_log_test.py +142 -142
- tests/framework_performance_test.py +202 -202
- tests/log_buffering_test.py +111 -111
- tests/log_generation_timing_test.py +153 -153
- tests/monitor_redis_dedup.sh +72 -0
- tests/ofweek_scrapy/ofweek_scrapy/items.py +12 -12
- tests/ofweek_scrapy/ofweek_scrapy/middlewares.py +100 -100
- tests/ofweek_scrapy/ofweek_scrapy/pipelines.py +13 -13
- tests/ofweek_scrapy/ofweek_scrapy/settings.py +84 -84
- tests/ofweek_scrapy/scrapy.cfg +11 -11
- tests/optimized_performance_test.py +211 -211
- tests/performance_comparison.py +244 -244
- tests/queue_blocking_test.py +113 -113
- tests/queue_test.py +89 -89
- tests/redis_key_validation_demo.py +130 -130
- tests/request_params_example.py +150 -150
- tests/response_improvements_example.py +144 -144
- tests/scrapy_comparison/ofweek_scrapy.py +138 -138
- tests/scrapy_comparison/scrapy_test.py +133 -133
- tests/simple_cli_test.py +55 -0
- tests/simple_command_test.py +119 -119
- tests/simple_crawlo_test.py +126 -126
- tests/simple_follow_test.py +38 -38
- tests/simple_log_test2.py +137 -137
- tests/simple_optimization_test.py +128 -128
- tests/simple_queue_type_test.py +41 -41
- tests/simple_response_selector_test.py +94 -94
- tests/simple_selector_helper_test.py +154 -154
- tests/simple_selector_test.py +207 -207
- tests/simple_spider_test.py +49 -49
- tests/simple_url_test.py +73 -73
- tests/simulate_mysql_update_test.py +140 -0
- tests/spider_log_timing_test.py +177 -177
- tests/test_advanced_tools.py +148 -148
- tests/test_all_commands.py +230 -230
- tests/test_all_pipeline_fingerprints.py +133 -133
- tests/test_all_redis_key_configs.py +145 -145
- tests/test_asyncmy_usage.py +57 -0
- tests/test_batch_processor.py +178 -178
- tests/test_cleaners.py +54 -54
- tests/test_cli_arguments.py +119 -0
- tests/test_component_factory.py +174 -174
- tests/test_config_consistency.py +80 -80
- tests/test_config_merge.py +152 -152
- tests/test_config_validator.py +182 -182
- tests/test_controlled_spider_mixin.py +79 -79
- tests/test_crawler_process_import.py +38 -38
- tests/test_crawler_process_spider_modules.py +47 -47
- tests/test_crawlo_proxy_integration.py +114 -108
- tests/test_date_tools.py +123 -123
- tests/test_dedup_fix.py +220 -220
- tests/test_dedup_pipeline_consistency.py +124 -124
- tests/test_default_header_middleware.py +313 -313
- tests/test_distributed.py +65 -65
- tests/test_double_crawlo_fix.py +204 -204
- tests/test_double_crawlo_fix_simple.py +124 -124
- tests/test_download_delay_middleware.py +221 -221
- tests/test_downloader_proxy_compatibility.py +272 -268
- tests/test_edge_cases.py +305 -305
- tests/test_encoding_core.py +56 -56
- tests/test_encoding_detection.py +126 -126
- tests/test_enhanced_error_handler.py +270 -270
- tests/test_enhanced_error_handler_comprehensive.py +245 -245
- tests/test_error_handler_compatibility.py +112 -112
- tests/test_factories.py +252 -252
- tests/test_factory_compatibility.py +196 -196
- tests/test_final_validation.py +153 -153
- tests/test_fingerprint_consistency.py +135 -135
- tests/test_fingerprint_simple.py +51 -51
- tests/test_get_component_logger.py +83 -83
- tests/test_hash_performance.py +99 -99
- tests/test_integration.py +169 -169
- tests/test_item_dedup_redis_key.py +122 -122
- tests/test_large_scale_helper.py +235 -235
- tests/test_logging_enhancements.py +374 -374
- tests/test_logging_final.py +184 -184
- tests/test_logging_integration.py +312 -312
- tests/test_logging_system.py +282 -282
- tests/test_middleware_debug.py +141 -141
- tests/test_mode_consistency.py +51 -51
- tests/test_multi_directory.py +67 -67
- tests/test_multiple_spider_modules.py +80 -80
- tests/test_mysql_pipeline_config.py +165 -0
- tests/test_mysql_pipeline_error.py +99 -0
- tests/test_mysql_pipeline_init_log.py +83 -0
- tests/test_mysql_pipeline_integration.py +133 -0
- tests/test_mysql_pipeline_refactor.py +144 -0
- tests/test_mysql_pipeline_refactor_simple.py +86 -0
- tests/test_mysql_pipeline_robustness.py +196 -0
- tests/test_mysql_pipeline_types.py +89 -0
- tests/test_mysql_update_columns.py +94 -0
- tests/test_offsite_middleware.py +244 -244
- tests/test_offsite_middleware_simple.py +203 -203
- tests/test_optimized_selector_naming.py +100 -100
- tests/test_parsel.py +29 -29
- tests/test_performance.py +327 -327
- tests/test_performance_monitor.py +115 -115
- tests/test_pipeline_fingerprint_consistency.py +86 -86
- tests/test_priority_behavior.py +211 -211
- tests/test_priority_consistency.py +151 -151
- tests/test_priority_consistency_fixed.py +249 -249
- tests/test_proxy_health_check.py +32 -32
- tests/test_proxy_middleware.py +217 -121
- tests/test_proxy_middleware_enhanced.py +212 -216
- tests/test_proxy_middleware_integration.py +142 -137
- tests/test_proxy_middleware_refactored.py +207 -184
- tests/test_proxy_only.py +84 -0
- tests/test_proxy_providers.py +56 -56
- tests/test_proxy_stats.py +19 -19
- tests/test_proxy_strategies.py +59 -59
- tests/test_proxy_with_downloader.py +153 -0
- tests/test_queue_empty_check.py +41 -41
- tests/test_queue_manager_double_crawlo.py +173 -173
- tests/test_queue_manager_redis_key.py +179 -179
- tests/test_queue_naming.py +154 -154
- tests/test_queue_type.py +106 -106
- tests/test_queue_type_redis_config_consistency.py +130 -130
- tests/test_random_headers_default.py +322 -322
- tests/test_random_headers_necessity.py +308 -308
- tests/test_random_user_agent.py +72 -72
- tests/test_redis_config.py +28 -28
- tests/test_redis_connection_pool.py +294 -294
- tests/test_redis_key_naming.py +181 -181
- tests/test_redis_key_validator.py +123 -123
- tests/test_redis_queue.py +224 -224
- tests/test_redis_queue_name_fix.py +175 -175
- tests/test_redis_queue_type_fallback.py +129 -129
- tests/test_request_ignore_middleware.py +182 -182
- tests/test_request_params.py +111 -111
- tests/test_request_serialization.py +70 -70
- tests/test_response_code_middleware.py +349 -349
- tests/test_response_filter_middleware.py +427 -427
- tests/test_response_follow.py +104 -104
- tests/test_response_improvements.py +152 -152
- tests/test_response_selector_methods.py +92 -92
- tests/test_response_url_methods.py +70 -70
- tests/test_response_urljoin.py +86 -86
- tests/test_retry_middleware.py +333 -333
- tests/test_retry_middleware_realistic.py +273 -273
- tests/test_scheduler.py +252 -252
- tests/test_scheduler_config_update.py +133 -133
- tests/test_scrapy_style_encoding.py +112 -112
- tests/test_selector_helper.py +100 -100
- tests/test_selector_optimizations.py +146 -146
- tests/test_simple_response.py +61 -61
- tests/test_spider_loader.py +49 -49
- tests/test_spider_loader_comprehensive.py +69 -69
- tests/test_spider_modules.py +84 -84
- tests/test_spiders/test_spider.py +9 -9
- tests/test_telecom_spider_redis_key.py +205 -205
- tests/test_template_content.py +87 -87
- tests/test_template_redis_key.py +134 -134
- tests/test_tools.py +159 -159
- tests/test_user_agent_randomness.py +176 -176
- tests/test_user_agents.py +96 -96
- tests/untested_features_report.md +138 -138
- tests/verify_debug.py +51 -51
- tests/verify_distributed.py +117 -117
- tests/verify_log_fix.py +111 -111
- tests/verify_mysql_warnings.py +110 -0
- crawlo/logging/async_handler.py +0 -181
- crawlo/logging/monitor.py +0 -153
- crawlo/logging/sampler.py +0 -167
- crawlo/middleware/simple_proxy.py +0 -65
- crawlo/tools/authenticated_proxy.py +0 -241
- crawlo/tools/data_formatter.py +0 -226
- crawlo/tools/data_validator.py +0 -181
- crawlo/tools/encoding_converter.py +0 -127
- crawlo/tools/network_diagnostic.py +0 -365
- crawlo/tools/request_tools.py +0 -83
- crawlo/tools/retry_mechanism.py +0 -224
- crawlo/utils/env_config.py +0 -143
- crawlo/utils/large_scale_config.py +0 -287
- crawlo/utils/system.py +0 -11
- crawlo/utils/tools.py +0 -5
- crawlo-1.4.5.dist-info/METADATA +0 -329
- crawlo-1.4.5.dist-info/RECORD +0 -347
- tests/env_config_example.py +0 -134
- tests/ofweek_scrapy/ofweek_scrapy/spiders/ofweek_spider.py +0 -162
- tests/test_authenticated_proxy.py +0 -142
- tests/test_comprehensive.py +0 -147
- tests/test_dynamic_downloaders_proxy.py +0 -125
- tests/test_dynamic_proxy.py +0 -93
- tests/test_dynamic_proxy_config.py +0 -147
- tests/test_dynamic_proxy_real.py +0 -110
- tests/test_env_config.py +0 -122
- tests/test_framework_env_usage.py +0 -104
- tests/test_large_scale_config.py +0 -113
- tests/test_proxy_api.py +0 -265
- tests/test_real_scenario_proxy.py +0 -196
- tests/tools_example.py +0 -261
- {crawlo-1.4.5.dist-info → crawlo-1.4.7.dist-info}/WHEEL +0 -0
- {crawlo-1.4.5.dist-info → crawlo-1.4.7.dist-info}/entry_points.txt +0 -0
- {crawlo-1.4.5.dist-info → crawlo-1.4.7.dist-info}/top_level.txt +0 -0
|
@@ -1,250 +1,250 @@
|
|
|
1
|
-
#!/usr/bin/env python3
|
|
2
|
-
# -*- coding: utf-8 -*-
|
|
3
|
-
"""
|
|
4
|
-
测试修复后的内存队列和Redis队列优先级行为一致性
|
|
5
|
-
"""
|
|
6
|
-
|
|
7
|
-
import sys
|
|
8
|
-
import os
|
|
9
|
-
import asyncio
|
|
10
|
-
from unittest.mock import Mock, patch
|
|
11
|
-
|
|
12
|
-
# 添加项目根目录到路径
|
|
13
|
-
sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..'))
|
|
14
|
-
|
|
15
|
-
from crawlo.network.request import Request
|
|
16
|
-
from crawlo.queue.pqueue import SpiderPriorityQueue
|
|
17
|
-
from crawlo.queue.redis_priority_queue import RedisPriorityQueue
|
|
18
|
-
|
|
19
|
-
|
|
20
|
-
async def test_priority_consistency_after_fix():
|
|
21
|
-
"""测试修复后内存队列和Redis队列优先级行为的一致性"""
|
|
22
|
-
print("=== 测试修复后内存队列和Redis队列优先级行为一致性 ===")
|
|
23
|
-
|
|
24
|
-
# 创建内存队列
|
|
25
|
-
memory_queue = SpiderPriorityQueue()
|
|
26
|
-
|
|
27
|
-
# 创建Redis队列
|
|
28
|
-
redis_queue = RedisPriorityQueue(
|
|
29
|
-
redis_url="redis://127.0.0.1:6379/15",
|
|
30
|
-
queue_name="test:consistency:fixed:queue"
|
|
31
|
-
)
|
|
32
|
-
|
|
33
|
-
try:
|
|
34
|
-
await redis_queue.connect()
|
|
35
|
-
|
|
36
|
-
# 清理之前的测试数据
|
|
37
|
-
await redis_queue._redis.delete(redis_queue.queue_name)
|
|
38
|
-
await redis_queue._redis.delete(f"{redis_queue.queue_name}:data")
|
|
39
|
-
|
|
40
|
-
# 创建相同优先级的请求(注意Request构造函数会将priority取反存储)
|
|
41
|
-
requests = [
|
|
42
|
-
Request(url="https://priority-100.com", priority=100), # 实际存储为-100(高优先级)
|
|
43
|
-
Request(url="https://priority-0.com", priority=0), # 实际存储为0(正常优先级)
|
|
44
|
-
Request(url="https://priority--100.com", priority=-100) # 实际存储为100(低优先级)
|
|
45
|
-
]
|
|
46
|
-
|
|
47
|
-
# 向内存队列添加请求
|
|
48
|
-
print("向内存队列添加请求...")
|
|
49
|
-
for req in requests:
|
|
50
|
-
# 内存队列直接使用priority值(实际存储的值)
|
|
51
|
-
await memory_queue.put((req.priority, req))
|
|
52
|
-
print(f" 内存队列: {req.url} (stored priority: {req.priority})")
|
|
53
|
-
|
|
54
|
-
# 向Redis队列添加请求
|
|
55
|
-
print("向Redis队列添加请求...")
|
|
56
|
-
for req in requests:
|
|
57
|
-
# Redis队列需要传入priority参数(实际存储的值)
|
|
58
|
-
await redis_queue.put(req, priority=req.priority)
|
|
59
|
-
print(f" Redis队列: {req.url} (stored priority: {req.priority})")
|
|
60
|
-
|
|
61
|
-
print(f" 内存队列大小: {memory_queue.qsize()}")
|
|
62
|
-
print(f" Redis队列大小: {await redis_queue.qsize()}")
|
|
63
|
-
|
|
64
|
-
# 从内存队列获取请求
|
|
65
|
-
print("从内存队列获取请求(应该按priority从小到大,即-100, 0, 100):")
|
|
66
|
-
memory_results = []
|
|
67
|
-
memory_priorities = []
|
|
68
|
-
for i in range(len(requests)):
|
|
69
|
-
item = await memory_queue.get(timeout=1.0)
|
|
70
|
-
if item:
|
|
71
|
-
request = item[1] # 解包(priority, request)元组
|
|
72
|
-
memory_results.append(request.url)
|
|
73
|
-
memory_priorities.append(request.priority)
|
|
74
|
-
print(f" {i+1}. {request.url} (stored priority: {request.priority})")
|
|
75
|
-
|
|
76
|
-
# 从Redis队列获取请求
|
|
77
|
-
print("从Redis队列获取请求(修复后应该与内存队列一致):")
|
|
78
|
-
redis_results = []
|
|
79
|
-
redis_priorities = []
|
|
80
|
-
for i in range(len(requests)):
|
|
81
|
-
request = await redis_queue.get(timeout=2.0)
|
|
82
|
-
if request:
|
|
83
|
-
redis_results.append(request.url)
|
|
84
|
-
redis_priorities.append(request.priority)
|
|
85
|
-
print(f" {i+1}. {request.url} (stored priority: {request.priority})")
|
|
86
|
-
|
|
87
|
-
# 验证一致性
|
|
88
|
-
print("\n一致性检查:")
|
|
89
|
-
print(f" 内存队列出队顺序: {memory_results}")
|
|
90
|
-
print(f" 内存队列优先级顺序: {memory_priorities}")
|
|
91
|
-
print(f" Redis队列出队顺序: {redis_results}")
|
|
92
|
-
print(f" Redis队列优先级顺序: {redis_priorities}")
|
|
93
|
-
|
|
94
|
-
# 验证出队顺序一致性
|
|
95
|
-
if memory_results == redis_results:
|
|
96
|
-
print(" ✅ 出队顺序一致")
|
|
97
|
-
else:
|
|
98
|
-
print(" ❌ 出队顺序不一致")
|
|
99
|
-
return False
|
|
100
|
-
|
|
101
|
-
# 验证优先级顺序一致性(都应该按priority从小到大)
|
|
102
|
-
expected_priority_order = [-100, 0, 100] # 高优先级到低优先级
|
|
103
|
-
if memory_priorities == expected_priority_order and redis_priorities == expected_priority_order:
|
|
104
|
-
print(" ✅ 优先级顺序一致(按priority从小到大)")
|
|
105
|
-
else:
|
|
106
|
-
print(f" ❌ 优先级顺序不一致,期望: {expected_priority_order}")
|
|
107
|
-
return False
|
|
108
|
-
|
|
109
|
-
return True
|
|
110
|
-
|
|
111
|
-
except Exception as e:
|
|
112
|
-
print(f" ❌ 测试失败: {e}")
|
|
113
|
-
import traceback
|
|
114
|
-
traceback.print_exc()
|
|
115
|
-
return False
|
|
116
|
-
finally:
|
|
117
|
-
await redis_queue.close()
|
|
118
|
-
|
|
119
|
-
|
|
120
|
-
async def test_real_world_scenario():
|
|
121
|
-
"""测试真实场景下的优先级行为"""
|
|
122
|
-
print("\n=== 测试真实场景下的优先级行为 ===")
|
|
123
|
-
|
|
124
|
-
# 创建内存队列
|
|
125
|
-
memory_queue = SpiderPriorityQueue()
|
|
126
|
-
|
|
127
|
-
# 创建Redis队列
|
|
128
|
-
redis_queue = RedisPriorityQueue(
|
|
129
|
-
redis_url="redis://127.0.0.1:6379/15",
|
|
130
|
-
queue_name="test:realworld:queue"
|
|
131
|
-
)
|
|
132
|
-
|
|
133
|
-
try:
|
|
134
|
-
await redis_queue.connect()
|
|
135
|
-
|
|
136
|
-
# 清理之前的测试数据
|
|
137
|
-
await redis_queue._redis.delete(redis_queue.queue_name)
|
|
138
|
-
await redis_queue._redis.delete(f"{redis_queue.queue_name}:data")
|
|
139
|
-
|
|
140
|
-
# 创建真实场景的请求
|
|
141
|
-
# 注意:Request构造函数会将传入的priority值取反存储
|
|
142
|
-
# 所以我们要传入负值来获得正值的存储priority
|
|
143
|
-
requests = [
|
|
144
|
-
Request(url="https://urgent-task.com", priority=-200), # 存储为200
|
|
145
|
-
Request(url="https://high-priority-task.com", priority=-100), # 存储为100
|
|
146
|
-
Request(url="https://normal-task.com", priority=0), # 存储为0
|
|
147
|
-
Request(url="https://low-priority-task.com", priority=100), # 存储为-100
|
|
148
|
-
Request(url="https://background-task.com", priority=200) # 存储为-200
|
|
149
|
-
]
|
|
150
|
-
|
|
151
|
-
# 向两个队列添加相同的请求
|
|
152
|
-
print("向队列添加真实场景请求...")
|
|
153
|
-
for req in requests:
|
|
154
|
-
# 内存队列
|
|
155
|
-
await memory_queue.put((req.priority, req))
|
|
156
|
-
# Redis队列
|
|
157
|
-
await redis_queue.put(req, priority=req.priority)
|
|
158
|
-
print(f" {req.url} (stored priority: {req.priority})")
|
|
159
|
-
|
|
160
|
-
# 从两个队列获取请求并比较顺序
|
|
161
|
-
print("\n从内存队列获取请求(应该按stored priority从小到大):")
|
|
162
|
-
memory_results = []
|
|
163
|
-
memory_priorities = []
|
|
164
|
-
for i in range(len(requests)):
|
|
165
|
-
item = await memory_queue.get(timeout=1.0)
|
|
166
|
-
if item:
|
|
167
|
-
request = item[1]
|
|
168
|
-
memory_results.append(request.url)
|
|
169
|
-
memory_priorities.append(request.priority)
|
|
170
|
-
print(f" {i+1}. {request.url} (stored priority: {request.priority})")
|
|
171
|
-
|
|
172
|
-
print("\n从Redis队列获取请求(应该与内存队列一致):")
|
|
173
|
-
redis_results = []
|
|
174
|
-
redis_priorities = []
|
|
175
|
-
for i in range(len(requests)):
|
|
176
|
-
request = await redis_queue.get(timeout=2.0)
|
|
177
|
-
if request:
|
|
178
|
-
redis_results.append(request.url)
|
|
179
|
-
redis_priorities.append(request.priority)
|
|
180
|
-
print(f" {i+1}. {request.url} (stored priority: {request.priority})")
|
|
181
|
-
|
|
182
|
-
# 验证一致性
|
|
183
|
-
print("\n真实场景一致性检查:")
|
|
184
|
-
print(f" 内存队列出队顺序: {memory_results}")
|
|
185
|
-
print(f" 内存队列优先级顺序: {memory_priorities}")
|
|
186
|
-
print(f" Redis队列出队顺序: {redis_results}")
|
|
187
|
-
print(f" Redis队列优先级顺序: {redis_priorities}")
|
|
188
|
-
|
|
189
|
-
# 应该按stored priority从小到大出队(-200, -100, 0, 100, 200)
|
|
190
|
-
expected_order = [
|
|
191
|
-
"https://background-task.com", # stored priority: -200
|
|
192
|
-
"https://low-priority-task.com", # stored priority: -100
|
|
193
|
-
"https://normal-task.com", # stored priority: 0
|
|
194
|
-
"https://high-priority-task.com", # stored priority: 100
|
|
195
|
-
"https://urgent-task.com" # stored priority: 200
|
|
196
|
-
]
|
|
197
|
-
|
|
198
|
-
expected_priority_order = [-200, -100, 0, 100, 200]
|
|
199
|
-
|
|
200
|
-
if (memory_results == expected_order and redis_results == expected_order and
|
|
201
|
-
memory_priorities == expected_priority_order and redis_priorities == expected_priority_order):
|
|
202
|
-
print(" ✅ 真实场景优先级行为一致且正确")
|
|
203
|
-
print(" 出队顺序: 高优先级 -> 低优先级")
|
|
204
|
-
return True
|
|
205
|
-
else:
|
|
206
|
-
print(f" ❌ 真实场景优先级行为不一致或不正确")
|
|
207
|
-
print(f" 期望出队顺序: {expected_order}")
|
|
208
|
-
print(f" 期望优先级顺序: {expected_priority_order}")
|
|
209
|
-
return False
|
|
210
|
-
|
|
211
|
-
except Exception as e:
|
|
212
|
-
print(f" ❌ 测试失败: {e}")
|
|
213
|
-
import traceback
|
|
214
|
-
traceback.print_exc()
|
|
215
|
-
return False
|
|
216
|
-
finally:
|
|
217
|
-
await redis_queue.close()
|
|
218
|
-
|
|
219
|
-
|
|
220
|
-
async def main():
|
|
221
|
-
print("开始测试修复后的队列优先级一致性...")
|
|
222
|
-
|
|
223
|
-
try:
|
|
224
|
-
# 测试基础一致性
|
|
225
|
-
basic_consistent = await test_priority_consistency_after_fix()
|
|
226
|
-
|
|
227
|
-
# 测试真实场景
|
|
228
|
-
realworld_consistent = await test_real_world_scenario()
|
|
229
|
-
|
|
230
|
-
if basic_consistent and realworld_consistent:
|
|
231
|
-
print("\n🎉 修复成功!内存队列和Redis队列优先级行为现在一致")
|
|
232
|
-
print("\n总结:")
|
|
233
|
-
print("1. 修复了Redis队列的score计算方式,从score = -priority改为score = priority")
|
|
234
|
-
print("2. 现在内存队列和Redis队列都遵循'priority数值小优先级高'的原则")
|
|
235
|
-
print("3. 与Request类的优先级设计理念保持一致")
|
|
236
|
-
print("4. 确保了单机模式和分布式模式行为的一致性")
|
|
237
|
-
print("\n注意事项:")
|
|
238
|
-
print(" Request对象构造时会将传入的priority值取反存储")
|
|
239
|
-
print(" 所以Request(url='example.com', priority=-200)实际存储的priority为200")
|
|
240
|
-
else:
|
|
241
|
-
print("\n❌ 修复不完全成功,请检查实现")
|
|
242
|
-
|
|
243
|
-
except Exception as e:
|
|
244
|
-
print(f"\n❌ 测试失败: {e}")
|
|
245
|
-
import traceback
|
|
246
|
-
traceback.print_exc()
|
|
247
|
-
|
|
248
|
-
|
|
249
|
-
if __name__ == "__main__":
|
|
1
|
+
#!/usr/bin/env python3
|
|
2
|
+
# -*- coding: utf-8 -*-
|
|
3
|
+
"""
|
|
4
|
+
测试修复后的内存队列和Redis队列优先级行为一致性
|
|
5
|
+
"""
|
|
6
|
+
|
|
7
|
+
import sys
|
|
8
|
+
import os
|
|
9
|
+
import asyncio
|
|
10
|
+
from unittest.mock import Mock, patch
|
|
11
|
+
|
|
12
|
+
# 添加项目根目录到路径
|
|
13
|
+
sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..'))
|
|
14
|
+
|
|
15
|
+
from crawlo.network.request import Request
|
|
16
|
+
from crawlo.queue.pqueue import SpiderPriorityQueue
|
|
17
|
+
from crawlo.queue.redis_priority_queue import RedisPriorityQueue
|
|
18
|
+
|
|
19
|
+
|
|
20
|
+
async def test_priority_consistency_after_fix():
|
|
21
|
+
"""测试修复后内存队列和Redis队列优先级行为的一致性"""
|
|
22
|
+
print("=== 测试修复后内存队列和Redis队列优先级行为一致性 ===")
|
|
23
|
+
|
|
24
|
+
# 创建内存队列
|
|
25
|
+
memory_queue = SpiderPriorityQueue()
|
|
26
|
+
|
|
27
|
+
# 创建Redis队列
|
|
28
|
+
redis_queue = RedisPriorityQueue(
|
|
29
|
+
redis_url="redis://127.0.0.1:6379/15",
|
|
30
|
+
queue_name="test:consistency:fixed:queue"
|
|
31
|
+
)
|
|
32
|
+
|
|
33
|
+
try:
|
|
34
|
+
await redis_queue.connect()
|
|
35
|
+
|
|
36
|
+
# 清理之前的测试数据
|
|
37
|
+
await redis_queue._redis.delete(redis_queue.queue_name)
|
|
38
|
+
await redis_queue._redis.delete(f"{redis_queue.queue_name}:data")
|
|
39
|
+
|
|
40
|
+
# 创建相同优先级的请求(注意Request构造函数会将priority取反存储)
|
|
41
|
+
requests = [
|
|
42
|
+
Request(url="https://priority-100.com", priority=100), # 实际存储为-100(高优先级)
|
|
43
|
+
Request(url="https://priority-0.com", priority=0), # 实际存储为0(正常优先级)
|
|
44
|
+
Request(url="https://priority--100.com", priority=-100) # 实际存储为100(低优先级)
|
|
45
|
+
]
|
|
46
|
+
|
|
47
|
+
# 向内存队列添加请求
|
|
48
|
+
print("向内存队列添加请求...")
|
|
49
|
+
for req in requests:
|
|
50
|
+
# 内存队列直接使用priority值(实际存储的值)
|
|
51
|
+
await memory_queue.put((req.priority, req))
|
|
52
|
+
print(f" 内存队列: {req.url} (stored priority: {req.priority})")
|
|
53
|
+
|
|
54
|
+
# 向Redis队列添加请求
|
|
55
|
+
print("向Redis队列添加请求...")
|
|
56
|
+
for req in requests:
|
|
57
|
+
# Redis队列需要传入priority参数(实际存储的值)
|
|
58
|
+
await redis_queue.put(req, priority=req.priority)
|
|
59
|
+
print(f" Redis队列: {req.url} (stored priority: {req.priority})")
|
|
60
|
+
|
|
61
|
+
print(f" 内存队列大小: {memory_queue.qsize()}")
|
|
62
|
+
print(f" Redis队列大小: {await redis_queue.qsize()}")
|
|
63
|
+
|
|
64
|
+
# 从内存队列获取请求
|
|
65
|
+
print("从内存队列获取请求(应该按priority从小到大,即-100, 0, 100):")
|
|
66
|
+
memory_results = []
|
|
67
|
+
memory_priorities = []
|
|
68
|
+
for i in range(len(requests)):
|
|
69
|
+
item = await memory_queue.get(timeout=1.0)
|
|
70
|
+
if item:
|
|
71
|
+
request = item[1] # 解包(priority, request)元组
|
|
72
|
+
memory_results.append(request.url)
|
|
73
|
+
memory_priorities.append(request.priority)
|
|
74
|
+
print(f" {i+1}. {request.url} (stored priority: {request.priority})")
|
|
75
|
+
|
|
76
|
+
# 从Redis队列获取请求
|
|
77
|
+
print("从Redis队列获取请求(修复后应该与内存队列一致):")
|
|
78
|
+
redis_results = []
|
|
79
|
+
redis_priorities = []
|
|
80
|
+
for i in range(len(requests)):
|
|
81
|
+
request = await redis_queue.get(timeout=2.0)
|
|
82
|
+
if request:
|
|
83
|
+
redis_results.append(request.url)
|
|
84
|
+
redis_priorities.append(request.priority)
|
|
85
|
+
print(f" {i+1}. {request.url} (stored priority: {request.priority})")
|
|
86
|
+
|
|
87
|
+
# 验证一致性
|
|
88
|
+
print("\n一致性检查:")
|
|
89
|
+
print(f" 内存队列出队顺序: {memory_results}")
|
|
90
|
+
print(f" 内存队列优先级顺序: {memory_priorities}")
|
|
91
|
+
print(f" Redis队列出队顺序: {redis_results}")
|
|
92
|
+
print(f" Redis队列优先级顺序: {redis_priorities}")
|
|
93
|
+
|
|
94
|
+
# 验证出队顺序一致性
|
|
95
|
+
if memory_results == redis_results:
|
|
96
|
+
print(" ✅ 出队顺序一致")
|
|
97
|
+
else:
|
|
98
|
+
print(" ❌ 出队顺序不一致")
|
|
99
|
+
return False
|
|
100
|
+
|
|
101
|
+
# 验证优先级顺序一致性(都应该按priority从小到大)
|
|
102
|
+
expected_priority_order = [-100, 0, 100] # 高优先级到低优先级
|
|
103
|
+
if memory_priorities == expected_priority_order and redis_priorities == expected_priority_order:
|
|
104
|
+
print(" ✅ 优先级顺序一致(按priority从小到大)")
|
|
105
|
+
else:
|
|
106
|
+
print(f" ❌ 优先级顺序不一致,期望: {expected_priority_order}")
|
|
107
|
+
return False
|
|
108
|
+
|
|
109
|
+
return True
|
|
110
|
+
|
|
111
|
+
except Exception as e:
|
|
112
|
+
print(f" ❌ 测试失败: {e}")
|
|
113
|
+
import traceback
|
|
114
|
+
traceback.print_exc()
|
|
115
|
+
return False
|
|
116
|
+
finally:
|
|
117
|
+
await redis_queue.close()
|
|
118
|
+
|
|
119
|
+
|
|
120
|
+
async def test_real_world_scenario():
|
|
121
|
+
"""测试真实场景下的优先级行为"""
|
|
122
|
+
print("\n=== 测试真实场景下的优先级行为 ===")
|
|
123
|
+
|
|
124
|
+
# 创建内存队列
|
|
125
|
+
memory_queue = SpiderPriorityQueue()
|
|
126
|
+
|
|
127
|
+
# 创建Redis队列
|
|
128
|
+
redis_queue = RedisPriorityQueue(
|
|
129
|
+
redis_url="redis://127.0.0.1:6379/15",
|
|
130
|
+
queue_name="test:realworld:queue"
|
|
131
|
+
)
|
|
132
|
+
|
|
133
|
+
try:
|
|
134
|
+
await redis_queue.connect()
|
|
135
|
+
|
|
136
|
+
# 清理之前的测试数据
|
|
137
|
+
await redis_queue._redis.delete(redis_queue.queue_name)
|
|
138
|
+
await redis_queue._redis.delete(f"{redis_queue.queue_name}:data")
|
|
139
|
+
|
|
140
|
+
# 创建真实场景的请求
|
|
141
|
+
# 注意:Request构造函数会将传入的priority值取反存储
|
|
142
|
+
# 所以我们要传入负值来获得正值的存储priority
|
|
143
|
+
requests = [
|
|
144
|
+
Request(url="https://urgent-task.com", priority=-200), # 存储为200
|
|
145
|
+
Request(url="https://high-priority-task.com", priority=-100), # 存储为100
|
|
146
|
+
Request(url="https://normal-task.com", priority=0), # 存储为0
|
|
147
|
+
Request(url="https://low-priority-task.com", priority=100), # 存储为-100
|
|
148
|
+
Request(url="https://background-task.com", priority=200) # 存储为-200
|
|
149
|
+
]
|
|
150
|
+
|
|
151
|
+
# 向两个队列添加相同的请求
|
|
152
|
+
print("向队列添加真实场景请求...")
|
|
153
|
+
for req in requests:
|
|
154
|
+
# 内存队列
|
|
155
|
+
await memory_queue.put((req.priority, req))
|
|
156
|
+
# Redis队列
|
|
157
|
+
await redis_queue.put(req, priority=req.priority)
|
|
158
|
+
print(f" {req.url} (stored priority: {req.priority})")
|
|
159
|
+
|
|
160
|
+
# 从两个队列获取请求并比较顺序
|
|
161
|
+
print("\n从内存队列获取请求(应该按stored priority从小到大):")
|
|
162
|
+
memory_results = []
|
|
163
|
+
memory_priorities = []
|
|
164
|
+
for i in range(len(requests)):
|
|
165
|
+
item = await memory_queue.get(timeout=1.0)
|
|
166
|
+
if item:
|
|
167
|
+
request = item[1]
|
|
168
|
+
memory_results.append(request.url)
|
|
169
|
+
memory_priorities.append(request.priority)
|
|
170
|
+
print(f" {i+1}. {request.url} (stored priority: {request.priority})")
|
|
171
|
+
|
|
172
|
+
print("\n从Redis队列获取请求(应该与内存队列一致):")
|
|
173
|
+
redis_results = []
|
|
174
|
+
redis_priorities = []
|
|
175
|
+
for i in range(len(requests)):
|
|
176
|
+
request = await redis_queue.get(timeout=2.0)
|
|
177
|
+
if request:
|
|
178
|
+
redis_results.append(request.url)
|
|
179
|
+
redis_priorities.append(request.priority)
|
|
180
|
+
print(f" {i+1}. {request.url} (stored priority: {request.priority})")
|
|
181
|
+
|
|
182
|
+
# 验证一致性
|
|
183
|
+
print("\n真实场景一致性检查:")
|
|
184
|
+
print(f" 内存队列出队顺序: {memory_results}")
|
|
185
|
+
print(f" 内存队列优先级顺序: {memory_priorities}")
|
|
186
|
+
print(f" Redis队列出队顺序: {redis_results}")
|
|
187
|
+
print(f" Redis队列优先级顺序: {redis_priorities}")
|
|
188
|
+
|
|
189
|
+
# 应该按stored priority从小到大出队(-200, -100, 0, 100, 200)
|
|
190
|
+
expected_order = [
|
|
191
|
+
"https://background-task.com", # stored priority: -200
|
|
192
|
+
"https://low-priority-task.com", # stored priority: -100
|
|
193
|
+
"https://normal-task.com", # stored priority: 0
|
|
194
|
+
"https://high-priority-task.com", # stored priority: 100
|
|
195
|
+
"https://urgent-task.com" # stored priority: 200
|
|
196
|
+
]
|
|
197
|
+
|
|
198
|
+
expected_priority_order = [-200, -100, 0, 100, 200]
|
|
199
|
+
|
|
200
|
+
if (memory_results == expected_order and redis_results == expected_order and
|
|
201
|
+
memory_priorities == expected_priority_order and redis_priorities == expected_priority_order):
|
|
202
|
+
print(" ✅ 真实场景优先级行为一致且正确")
|
|
203
|
+
print(" 出队顺序: 高优先级 -> 低优先级")
|
|
204
|
+
return True
|
|
205
|
+
else:
|
|
206
|
+
print(f" ❌ 真实场景优先级行为不一致或不正确")
|
|
207
|
+
print(f" 期望出队顺序: {expected_order}")
|
|
208
|
+
print(f" 期望优先级顺序: {expected_priority_order}")
|
|
209
|
+
return False
|
|
210
|
+
|
|
211
|
+
except Exception as e:
|
|
212
|
+
print(f" ❌ 测试失败: {e}")
|
|
213
|
+
import traceback
|
|
214
|
+
traceback.print_exc()
|
|
215
|
+
return False
|
|
216
|
+
finally:
|
|
217
|
+
await redis_queue.close()
|
|
218
|
+
|
|
219
|
+
|
|
220
|
+
async def main():
|
|
221
|
+
print("开始测试修复后的队列优先级一致性...")
|
|
222
|
+
|
|
223
|
+
try:
|
|
224
|
+
# 测试基础一致性
|
|
225
|
+
basic_consistent = await test_priority_consistency_after_fix()
|
|
226
|
+
|
|
227
|
+
# 测试真实场景
|
|
228
|
+
realworld_consistent = await test_real_world_scenario()
|
|
229
|
+
|
|
230
|
+
if basic_consistent and realworld_consistent:
|
|
231
|
+
print("\n🎉 修复成功!内存队列和Redis队列优先级行为现在一致")
|
|
232
|
+
print("\n总结:")
|
|
233
|
+
print("1. 修复了Redis队列的score计算方式,从score = -priority改为score = priority")
|
|
234
|
+
print("2. 现在内存队列和Redis队列都遵循'priority数值小优先级高'的原则")
|
|
235
|
+
print("3. 与Request类的优先级设计理念保持一致")
|
|
236
|
+
print("4. 确保了单机模式和分布式模式行为的一致性")
|
|
237
|
+
print("\n注意事项:")
|
|
238
|
+
print(" Request对象构造时会将传入的priority值取反存储")
|
|
239
|
+
print(" 所以Request(url='example.com', priority=-200)实际存储的priority为200")
|
|
240
|
+
else:
|
|
241
|
+
print("\n❌ 修复不完全成功,请检查实现")
|
|
242
|
+
|
|
243
|
+
except Exception as e:
|
|
244
|
+
print(f"\n❌ 测试失败: {e}")
|
|
245
|
+
import traceback
|
|
246
|
+
traceback.print_exc()
|
|
247
|
+
|
|
248
|
+
|
|
249
|
+
if __name__ == "__main__":
|
|
250
250
|
asyncio.run(main())
|
tests/test_proxy_health_check.py
CHANGED
|
@@ -1,33 +1,33 @@
|
|
|
1
|
-
# tests/test_proxy_health_check.py
|
|
2
|
-
import pytest
|
|
3
|
-
from unittest.mock import AsyncMock, patch
|
|
4
|
-
from crawlo.proxy.health_check import check_single_proxy
|
|
5
|
-
import httpx
|
|
6
|
-
|
|
7
|
-
|
|
8
|
-
@pytest.mark.asyncio
|
|
9
|
-
@patch('httpx.AsyncClient')
|
|
10
|
-
async def test_health_check_success(mock_client_class):
|
|
11
|
-
"""测试健康检查:成功"""
|
|
12
|
-
mock_resp = AsyncMock()
|
|
13
|
-
mock_resp.status_code = 200
|
|
14
|
-
mock_client_class.return_value.__aenter__.return_value.get.return_value = mock_resp
|
|
15
|
-
|
|
16
|
-
proxy_info = {'url': 'http://good:8080', 'healthy': False}
|
|
17
|
-
await check_single_proxy(proxy_info)
|
|
18
|
-
|
|
19
|
-
assert proxy_info['healthy'] is True
|
|
20
|
-
assert proxy_info['failures'] == 0
|
|
21
|
-
|
|
22
|
-
|
|
23
|
-
@pytest.mark.asyncio
|
|
24
|
-
@patch('httpx.AsyncClient')
|
|
25
|
-
async def test_health_check_failure(mock_client_class):
|
|
26
|
-
"""测试健康检查:失败"""
|
|
27
|
-
mock_client_class.return_value.__aenter__.return_value.get.side_effect = httpx.ConnectError("Failed")
|
|
28
|
-
|
|
29
|
-
proxy_info = {'url': 'http://bad:8080', 'healthy': True, 'failures': 0}
|
|
30
|
-
await check_single_proxy(proxy_info)
|
|
31
|
-
|
|
32
|
-
assert proxy_info['healthy'] is False
|
|
1
|
+
# tests/test_proxy_health_check.py
|
|
2
|
+
import pytest
|
|
3
|
+
from unittest.mock import AsyncMock, patch
|
|
4
|
+
from crawlo.proxy.health_check import check_single_proxy
|
|
5
|
+
import httpx
|
|
6
|
+
|
|
7
|
+
|
|
8
|
+
@pytest.mark.asyncio
|
|
9
|
+
@patch('httpx.AsyncClient')
|
|
10
|
+
async def test_health_check_success(mock_client_class):
|
|
11
|
+
"""测试健康检查:成功"""
|
|
12
|
+
mock_resp = AsyncMock()
|
|
13
|
+
mock_resp.status_code = 200
|
|
14
|
+
mock_client_class.return_value.__aenter__.return_value.get.return_value = mock_resp
|
|
15
|
+
|
|
16
|
+
proxy_info = {'url': 'http://good:8080', 'healthy': False}
|
|
17
|
+
await check_single_proxy(proxy_info)
|
|
18
|
+
|
|
19
|
+
assert proxy_info['healthy'] is True
|
|
20
|
+
assert proxy_info['failures'] == 0
|
|
21
|
+
|
|
22
|
+
|
|
23
|
+
@pytest.mark.asyncio
|
|
24
|
+
@patch('httpx.AsyncClient')
|
|
25
|
+
async def test_health_check_failure(mock_client_class):
|
|
26
|
+
"""测试健康检查:失败"""
|
|
27
|
+
mock_client_class.return_value.__aenter__.return_value.get.side_effect = httpx.ConnectError("Failed")
|
|
28
|
+
|
|
29
|
+
proxy_info = {'url': 'http://bad:8080', 'healthy': True, 'failures': 0}
|
|
30
|
+
await check_single_proxy(proxy_info)
|
|
31
|
+
|
|
32
|
+
assert proxy_info['healthy'] is False
|
|
33
33
|
assert proxy_info['failures'] == 1
|