crawlo 1.4.5__py3-none-any.whl → 1.4.7__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of crawlo might be problematic. Click here for more details.
- crawlo/__init__.py +90 -89
- crawlo/__version__.py +1 -1
- crawlo/cli.py +75 -75
- crawlo/commands/__init__.py +14 -14
- crawlo/commands/check.py +594 -594
- crawlo/commands/genspider.py +186 -186
- crawlo/commands/help.py +140 -138
- crawlo/commands/list.py +155 -155
- crawlo/commands/run.py +379 -341
- crawlo/commands/startproject.py +460 -460
- crawlo/commands/stats.py +187 -187
- crawlo/commands/utils.py +196 -196
- crawlo/config.py +320 -312
- crawlo/config_validator.py +277 -277
- crawlo/core/__init__.py +52 -52
- crawlo/core/engine.py +451 -438
- crawlo/core/processor.py +47 -47
- crawlo/core/scheduler.py +290 -291
- crawlo/crawler.py +698 -657
- crawlo/data/__init__.py +5 -5
- crawlo/data/user_agents.py +194 -194
- crawlo/downloader/__init__.py +280 -276
- crawlo/downloader/aiohttp_downloader.py +233 -233
- crawlo/downloader/cffi_downloader.py +250 -245
- crawlo/downloader/httpx_downloader.py +265 -259
- crawlo/downloader/hybrid_downloader.py +212 -212
- crawlo/downloader/playwright_downloader.py +425 -402
- crawlo/downloader/selenium_downloader.py +486 -472
- crawlo/event.py +45 -11
- crawlo/exceptions.py +215 -82
- crawlo/extension/__init__.py +65 -64
- crawlo/extension/health_check.py +141 -141
- crawlo/extension/log_interval.py +94 -94
- crawlo/extension/log_stats.py +70 -70
- crawlo/extension/logging_extension.py +53 -61
- crawlo/extension/memory_monitor.py +104 -104
- crawlo/extension/performance_profiler.py +133 -133
- crawlo/extension/request_recorder.py +107 -107
- crawlo/factories/__init__.py +27 -27
- crawlo/factories/base.py +68 -68
- crawlo/factories/crawler.py +104 -103
- crawlo/factories/registry.py +84 -84
- crawlo/factories/utils.py +135 -0
- crawlo/filters/__init__.py +170 -153
- crawlo/filters/aioredis_filter.py +348 -264
- crawlo/filters/memory_filter.py +261 -276
- crawlo/framework.py +306 -292
- crawlo/initialization/__init__.py +44 -44
- crawlo/initialization/built_in.py +391 -434
- crawlo/initialization/context.py +141 -141
- crawlo/initialization/core.py +240 -194
- crawlo/initialization/phases.py +230 -149
- crawlo/initialization/registry.py +143 -145
- crawlo/initialization/utils.py +49 -0
- crawlo/interfaces.py +23 -23
- crawlo/items/__init__.py +23 -23
- crawlo/items/base.py +23 -23
- crawlo/items/fields.py +52 -52
- crawlo/items/items.py +104 -104
- crawlo/logging/__init__.py +42 -46
- crawlo/logging/config.py +277 -197
- crawlo/logging/factory.py +175 -171
- crawlo/logging/manager.py +104 -112
- crawlo/middleware/__init__.py +87 -24
- crawlo/middleware/default_header.py +132 -132
- crawlo/middleware/download_delay.py +104 -104
- crawlo/middleware/middleware_manager.py +142 -142
- crawlo/middleware/offsite.py +123 -123
- crawlo/middleware/proxy.py +209 -386
- crawlo/middleware/request_ignore.py +86 -86
- crawlo/middleware/response_code.py +150 -150
- crawlo/middleware/response_filter.py +136 -136
- crawlo/middleware/retry.py +124 -124
- crawlo/mode_manager.py +287 -253
- crawlo/network/__init__.py +21 -21
- crawlo/network/request.py +375 -379
- crawlo/network/response.py +569 -664
- crawlo/pipelines/__init__.py +53 -22
- crawlo/pipelines/base_pipeline.py +452 -0
- crawlo/pipelines/bloom_dedup_pipeline.py +146 -146
- crawlo/pipelines/console_pipeline.py +39 -39
- crawlo/pipelines/csv_pipeline.py +316 -316
- crawlo/pipelines/database_dedup_pipeline.py +197 -197
- crawlo/pipelines/json_pipeline.py +218 -218
- crawlo/pipelines/memory_dedup_pipeline.py +105 -105
- crawlo/pipelines/mongo_pipeline.py +140 -132
- crawlo/pipelines/mysql_pipeline.py +470 -326
- crawlo/pipelines/pipeline_manager.py +100 -100
- crawlo/pipelines/redis_dedup_pipeline.py +155 -156
- crawlo/project.py +347 -347
- crawlo/queue/__init__.py +10 -0
- crawlo/queue/pqueue.py +38 -38
- crawlo/queue/queue_manager.py +591 -525
- crawlo/queue/redis_priority_queue.py +519 -370
- crawlo/settings/__init__.py +7 -7
- crawlo/settings/default_settings.py +285 -270
- crawlo/settings/setting_manager.py +219 -219
- crawlo/spider/__init__.py +657 -657
- crawlo/stats_collector.py +82 -73
- crawlo/subscriber.py +129 -129
- crawlo/task_manager.py +138 -138
- crawlo/templates/crawlo.cfg.tmpl +10 -10
- crawlo/templates/project/__init__.py.tmpl +2 -4
- crawlo/templates/project/items.py.tmpl +13 -17
- crawlo/templates/project/middlewares.py.tmpl +38 -38
- crawlo/templates/project/pipelines.py.tmpl +35 -36
- crawlo/templates/project/settings.py.tmpl +110 -157
- crawlo/templates/project/settings_distributed.py.tmpl +156 -161
- crawlo/templates/project/settings_gentle.py.tmpl +170 -171
- crawlo/templates/project/settings_high_performance.py.tmpl +171 -172
- crawlo/templates/project/settings_minimal.py.tmpl +99 -77
- crawlo/templates/project/settings_simple.py.tmpl +168 -169
- crawlo/templates/project/spiders/__init__.py.tmpl +9 -9
- crawlo/templates/run.py.tmpl +23 -30
- crawlo/templates/spider/spider.py.tmpl +33 -144
- crawlo/templates/spiders_init.py.tmpl +5 -10
- crawlo/tools/__init__.py +86 -189
- crawlo/tools/date_tools.py +289 -289
- crawlo/tools/distributed_coordinator.py +384 -384
- crawlo/tools/scenario_adapter.py +262 -262
- crawlo/tools/text_cleaner.py +232 -232
- crawlo/utils/__init__.py +50 -50
- crawlo/utils/batch_processor.py +276 -259
- crawlo/utils/config_manager.py +442 -0
- crawlo/utils/controlled_spider_mixin.py +439 -439
- crawlo/utils/db_helper.py +250 -244
- crawlo/utils/error_handler.py +410 -410
- crawlo/utils/fingerprint.py +121 -121
- crawlo/utils/func_tools.py +82 -82
- crawlo/utils/large_scale_helper.py +344 -344
- crawlo/utils/leak_detector.py +335 -0
- crawlo/utils/log.py +79 -79
- crawlo/utils/misc.py +81 -81
- crawlo/utils/mongo_connection_pool.py +157 -0
- crawlo/utils/mysql_connection_pool.py +197 -0
- crawlo/utils/performance_monitor.py +285 -285
- crawlo/utils/queue_helper.py +175 -175
- crawlo/utils/redis_checker.py +91 -0
- crawlo/utils/redis_connection_pool.py +578 -388
- crawlo/utils/redis_key_validator.py +198 -198
- crawlo/utils/request.py +278 -256
- crawlo/utils/request_serializer.py +225 -225
- crawlo/utils/resource_manager.py +337 -0
- crawlo/utils/selector_helper.py +137 -137
- crawlo/utils/singleton.py +70 -0
- crawlo/utils/spider_loader.py +201 -201
- crawlo/utils/text_helper.py +94 -94
- crawlo/utils/{url.py → url_utils.py} +39 -39
- crawlo-1.4.7.dist-info/METADATA +689 -0
- crawlo-1.4.7.dist-info/RECORD +347 -0
- examples/__init__.py +7 -7
- tests/__init__.py +7 -7
- tests/advanced_tools_example.py +217 -275
- tests/authenticated_proxy_example.py +110 -106
- tests/baidu_performance_test.py +108 -108
- tests/baidu_test.py +59 -59
- tests/bug_check_test.py +250 -250
- tests/cleaners_example.py +160 -160
- tests/comprehensive_framework_test.py +212 -212
- tests/comprehensive_test.py +81 -81
- tests/comprehensive_testing_summary.md +186 -186
- tests/config_validation_demo.py +142 -142
- tests/controlled_spider_example.py +205 -205
- tests/date_tools_example.py +180 -180
- tests/debug_configure.py +69 -69
- tests/debug_framework_logger.py +84 -84
- tests/debug_log_config.py +126 -126
- tests/debug_log_levels.py +63 -63
- tests/debug_pipelines.py +66 -66
- tests/detailed_log_test.py +233 -233
- tests/direct_selector_helper_test.py +96 -96
- tests/distributed_dedup_test.py +467 -0
- tests/distributed_test.py +66 -66
- tests/distributed_test_debug.py +76 -76
- tests/dynamic_loading_example.py +523 -523
- tests/dynamic_loading_test.py +104 -104
- tests/error_handling_example.py +171 -171
- tests/explain_mysql_update_behavior.py +77 -0
- tests/final_comprehensive_test.py +151 -151
- tests/final_log_test.py +260 -260
- tests/final_validation_test.py +182 -182
- tests/fix_log_test.py +142 -142
- tests/framework_performance_test.py +202 -202
- tests/log_buffering_test.py +111 -111
- tests/log_generation_timing_test.py +153 -153
- tests/monitor_redis_dedup.sh +72 -0
- tests/ofweek_scrapy/ofweek_scrapy/items.py +12 -12
- tests/ofweek_scrapy/ofweek_scrapy/middlewares.py +100 -100
- tests/ofweek_scrapy/ofweek_scrapy/pipelines.py +13 -13
- tests/ofweek_scrapy/ofweek_scrapy/settings.py +84 -84
- tests/ofweek_scrapy/scrapy.cfg +11 -11
- tests/optimized_performance_test.py +211 -211
- tests/performance_comparison.py +244 -244
- tests/queue_blocking_test.py +113 -113
- tests/queue_test.py +89 -89
- tests/redis_key_validation_demo.py +130 -130
- tests/request_params_example.py +150 -150
- tests/response_improvements_example.py +144 -144
- tests/scrapy_comparison/ofweek_scrapy.py +138 -138
- tests/scrapy_comparison/scrapy_test.py +133 -133
- tests/simple_cli_test.py +55 -0
- tests/simple_command_test.py +119 -119
- tests/simple_crawlo_test.py +126 -126
- tests/simple_follow_test.py +38 -38
- tests/simple_log_test2.py +137 -137
- tests/simple_optimization_test.py +128 -128
- tests/simple_queue_type_test.py +41 -41
- tests/simple_response_selector_test.py +94 -94
- tests/simple_selector_helper_test.py +154 -154
- tests/simple_selector_test.py +207 -207
- tests/simple_spider_test.py +49 -49
- tests/simple_url_test.py +73 -73
- tests/simulate_mysql_update_test.py +140 -0
- tests/spider_log_timing_test.py +177 -177
- tests/test_advanced_tools.py +148 -148
- tests/test_all_commands.py +230 -230
- tests/test_all_pipeline_fingerprints.py +133 -133
- tests/test_all_redis_key_configs.py +145 -145
- tests/test_asyncmy_usage.py +57 -0
- tests/test_batch_processor.py +178 -178
- tests/test_cleaners.py +54 -54
- tests/test_cli_arguments.py +119 -0
- tests/test_component_factory.py +174 -174
- tests/test_config_consistency.py +80 -80
- tests/test_config_merge.py +152 -152
- tests/test_config_validator.py +182 -182
- tests/test_controlled_spider_mixin.py +79 -79
- tests/test_crawler_process_import.py +38 -38
- tests/test_crawler_process_spider_modules.py +47 -47
- tests/test_crawlo_proxy_integration.py +114 -108
- tests/test_date_tools.py +123 -123
- tests/test_dedup_fix.py +220 -220
- tests/test_dedup_pipeline_consistency.py +124 -124
- tests/test_default_header_middleware.py +313 -313
- tests/test_distributed.py +65 -65
- tests/test_double_crawlo_fix.py +204 -204
- tests/test_double_crawlo_fix_simple.py +124 -124
- tests/test_download_delay_middleware.py +221 -221
- tests/test_downloader_proxy_compatibility.py +272 -268
- tests/test_edge_cases.py +305 -305
- tests/test_encoding_core.py +56 -56
- tests/test_encoding_detection.py +126 -126
- tests/test_enhanced_error_handler.py +270 -270
- tests/test_enhanced_error_handler_comprehensive.py +245 -245
- tests/test_error_handler_compatibility.py +112 -112
- tests/test_factories.py +252 -252
- tests/test_factory_compatibility.py +196 -196
- tests/test_final_validation.py +153 -153
- tests/test_fingerprint_consistency.py +135 -135
- tests/test_fingerprint_simple.py +51 -51
- tests/test_get_component_logger.py +83 -83
- tests/test_hash_performance.py +99 -99
- tests/test_integration.py +169 -169
- tests/test_item_dedup_redis_key.py +122 -122
- tests/test_large_scale_helper.py +235 -235
- tests/test_logging_enhancements.py +374 -374
- tests/test_logging_final.py +184 -184
- tests/test_logging_integration.py +312 -312
- tests/test_logging_system.py +282 -282
- tests/test_middleware_debug.py +141 -141
- tests/test_mode_consistency.py +51 -51
- tests/test_multi_directory.py +67 -67
- tests/test_multiple_spider_modules.py +80 -80
- tests/test_mysql_pipeline_config.py +165 -0
- tests/test_mysql_pipeline_error.py +99 -0
- tests/test_mysql_pipeline_init_log.py +83 -0
- tests/test_mysql_pipeline_integration.py +133 -0
- tests/test_mysql_pipeline_refactor.py +144 -0
- tests/test_mysql_pipeline_refactor_simple.py +86 -0
- tests/test_mysql_pipeline_robustness.py +196 -0
- tests/test_mysql_pipeline_types.py +89 -0
- tests/test_mysql_update_columns.py +94 -0
- tests/test_offsite_middleware.py +244 -244
- tests/test_offsite_middleware_simple.py +203 -203
- tests/test_optimized_selector_naming.py +100 -100
- tests/test_parsel.py +29 -29
- tests/test_performance.py +327 -327
- tests/test_performance_monitor.py +115 -115
- tests/test_pipeline_fingerprint_consistency.py +86 -86
- tests/test_priority_behavior.py +211 -211
- tests/test_priority_consistency.py +151 -151
- tests/test_priority_consistency_fixed.py +249 -249
- tests/test_proxy_health_check.py +32 -32
- tests/test_proxy_middleware.py +217 -121
- tests/test_proxy_middleware_enhanced.py +212 -216
- tests/test_proxy_middleware_integration.py +142 -137
- tests/test_proxy_middleware_refactored.py +207 -184
- tests/test_proxy_only.py +84 -0
- tests/test_proxy_providers.py +56 -56
- tests/test_proxy_stats.py +19 -19
- tests/test_proxy_strategies.py +59 -59
- tests/test_proxy_with_downloader.py +153 -0
- tests/test_queue_empty_check.py +41 -41
- tests/test_queue_manager_double_crawlo.py +173 -173
- tests/test_queue_manager_redis_key.py +179 -179
- tests/test_queue_naming.py +154 -154
- tests/test_queue_type.py +106 -106
- tests/test_queue_type_redis_config_consistency.py +130 -130
- tests/test_random_headers_default.py +322 -322
- tests/test_random_headers_necessity.py +308 -308
- tests/test_random_user_agent.py +72 -72
- tests/test_redis_config.py +28 -28
- tests/test_redis_connection_pool.py +294 -294
- tests/test_redis_key_naming.py +181 -181
- tests/test_redis_key_validator.py +123 -123
- tests/test_redis_queue.py +224 -224
- tests/test_redis_queue_name_fix.py +175 -175
- tests/test_redis_queue_type_fallback.py +129 -129
- tests/test_request_ignore_middleware.py +182 -182
- tests/test_request_params.py +111 -111
- tests/test_request_serialization.py +70 -70
- tests/test_response_code_middleware.py +349 -349
- tests/test_response_filter_middleware.py +427 -427
- tests/test_response_follow.py +104 -104
- tests/test_response_improvements.py +152 -152
- tests/test_response_selector_methods.py +92 -92
- tests/test_response_url_methods.py +70 -70
- tests/test_response_urljoin.py +86 -86
- tests/test_retry_middleware.py +333 -333
- tests/test_retry_middleware_realistic.py +273 -273
- tests/test_scheduler.py +252 -252
- tests/test_scheduler_config_update.py +133 -133
- tests/test_scrapy_style_encoding.py +112 -112
- tests/test_selector_helper.py +100 -100
- tests/test_selector_optimizations.py +146 -146
- tests/test_simple_response.py +61 -61
- tests/test_spider_loader.py +49 -49
- tests/test_spider_loader_comprehensive.py +69 -69
- tests/test_spider_modules.py +84 -84
- tests/test_spiders/test_spider.py +9 -9
- tests/test_telecom_spider_redis_key.py +205 -205
- tests/test_template_content.py +87 -87
- tests/test_template_redis_key.py +134 -134
- tests/test_tools.py +159 -159
- tests/test_user_agent_randomness.py +176 -176
- tests/test_user_agents.py +96 -96
- tests/untested_features_report.md +138 -138
- tests/verify_debug.py +51 -51
- tests/verify_distributed.py +117 -117
- tests/verify_log_fix.py +111 -111
- tests/verify_mysql_warnings.py +110 -0
- crawlo/logging/async_handler.py +0 -181
- crawlo/logging/monitor.py +0 -153
- crawlo/logging/sampler.py +0 -167
- crawlo/middleware/simple_proxy.py +0 -65
- crawlo/tools/authenticated_proxy.py +0 -241
- crawlo/tools/data_formatter.py +0 -226
- crawlo/tools/data_validator.py +0 -181
- crawlo/tools/encoding_converter.py +0 -127
- crawlo/tools/network_diagnostic.py +0 -365
- crawlo/tools/request_tools.py +0 -83
- crawlo/tools/retry_mechanism.py +0 -224
- crawlo/utils/env_config.py +0 -143
- crawlo/utils/large_scale_config.py +0 -287
- crawlo/utils/system.py +0 -11
- crawlo/utils/tools.py +0 -5
- crawlo-1.4.5.dist-info/METADATA +0 -329
- crawlo-1.4.5.dist-info/RECORD +0 -347
- tests/env_config_example.py +0 -134
- tests/ofweek_scrapy/ofweek_scrapy/spiders/ofweek_spider.py +0 -162
- tests/test_authenticated_proxy.py +0 -142
- tests/test_comprehensive.py +0 -147
- tests/test_dynamic_downloaders_proxy.py +0 -125
- tests/test_dynamic_proxy.py +0 -93
- tests/test_dynamic_proxy_config.py +0 -147
- tests/test_dynamic_proxy_real.py +0 -110
- tests/test_env_config.py +0 -122
- tests/test_framework_env_usage.py +0 -104
- tests/test_large_scale_config.py +0 -113
- tests/test_proxy_api.py +0 -265
- tests/test_real_scenario_proxy.py +0 -196
- tests/tools_example.py +0 -261
- {crawlo-1.4.5.dist-info → crawlo-1.4.7.dist-info}/WHEEL +0 -0
- {crawlo-1.4.5.dist-info → crawlo-1.4.7.dist-info}/entry_points.txt +0 -0
- {crawlo-1.4.5.dist-info → crawlo-1.4.7.dist-info}/top_level.txt +0 -0
tests/test_proxy_api.py
DELETED
|
@@ -1,265 +0,0 @@
|
|
|
1
|
-
#!/usr/bin/python
|
|
2
|
-
# -*- coding: UTF-8 -*-
|
|
3
|
-
"""
|
|
4
|
-
代理API测试脚本
|
|
5
|
-
================
|
|
6
|
-
测试指定的代理API接口是否能正常工作
|
|
7
|
-
"""
|
|
8
|
-
|
|
9
|
-
import asyncio
|
|
10
|
-
import aiohttp
|
|
11
|
-
import sys
|
|
12
|
-
import os
|
|
13
|
-
from urllib.parse import urlparse
|
|
14
|
-
|
|
15
|
-
# 添加项目根目录到Python路径
|
|
16
|
-
sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..'))
|
|
17
|
-
|
|
18
|
-
from crawlo.middleware.proxy import ProxyMiddleware
|
|
19
|
-
from crawlo.network.request import Request
|
|
20
|
-
from crawlo.settings.setting_manager import SettingManager
|
|
21
|
-
|
|
22
|
-
|
|
23
|
-
async def test_proxy_api(proxy_api_url):
|
|
24
|
-
"""测试代理API接口"""
|
|
25
|
-
print(f"=== 测试代理API接口 ===")
|
|
26
|
-
print(f"API地址: {proxy_api_url}")
|
|
27
|
-
|
|
28
|
-
try:
|
|
29
|
-
timeout = aiohttp.ClientTimeout(total=10)
|
|
30
|
-
async with aiohttp.ClientSession(timeout=timeout) as session:
|
|
31
|
-
async with session.get(proxy_api_url) as response:
|
|
32
|
-
print(f"状态码: {response.status}")
|
|
33
|
-
print(f"响应头: {response.headers.get('content-type', 'Unknown')}")
|
|
34
|
-
|
|
35
|
-
# 尝试解析JSON响应
|
|
36
|
-
try:
|
|
37
|
-
data = await response.json()
|
|
38
|
-
print(f"响应数据: {data}")
|
|
39
|
-
return data
|
|
40
|
-
except Exception as e:
|
|
41
|
-
# 如果不是JSON,尝试获取文本
|
|
42
|
-
try:
|
|
43
|
-
text = await response.text()
|
|
44
|
-
print(f"响应文本: {text[:200]}{'...' if len(text) > 200 else ''}")
|
|
45
|
-
return text
|
|
46
|
-
except Exception as e2:
|
|
47
|
-
print(f"无法解析响应内容: {e2}")
|
|
48
|
-
return None
|
|
49
|
-
|
|
50
|
-
except asyncio.TimeoutError:
|
|
51
|
-
print("请求超时")
|
|
52
|
-
return None
|
|
53
|
-
except Exception as e:
|
|
54
|
-
print(f"请求失败: {e}")
|
|
55
|
-
return None
|
|
56
|
-
|
|
57
|
-
|
|
58
|
-
def extract_proxy_url(proxy_data):
|
|
59
|
-
"""从API响应中提取代理URL"""
|
|
60
|
-
proxy_url = None
|
|
61
|
-
|
|
62
|
-
if isinstance(proxy_data, dict):
|
|
63
|
-
# 检查是否有status字段且为成功状态
|
|
64
|
-
if proxy_data.get('status') == 0:
|
|
65
|
-
# 获取proxy字段
|
|
66
|
-
proxy_info = proxy_data.get('proxy', {})
|
|
67
|
-
if isinstance(proxy_info, dict):
|
|
68
|
-
# 优先使用https代理,否则使用http代理
|
|
69
|
-
proxy_url = proxy_info.get('https') or proxy_info.get('http')
|
|
70
|
-
elif isinstance(proxy_info, str):
|
|
71
|
-
proxy_url = proxy_info
|
|
72
|
-
else:
|
|
73
|
-
# 直接尝试常见的字段名
|
|
74
|
-
for key in ['proxy', 'data', 'url', 'http', 'https']:
|
|
75
|
-
if key in proxy_data:
|
|
76
|
-
value = proxy_data[key]
|
|
77
|
-
if isinstance(value, str):
|
|
78
|
-
proxy_url = value
|
|
79
|
-
break
|
|
80
|
-
elif isinstance(value, dict):
|
|
81
|
-
proxy_url = value.get('https') or value.get('http')
|
|
82
|
-
break
|
|
83
|
-
|
|
84
|
-
# 如果还是没有找到,尝试更深层的嵌套
|
|
85
|
-
if not proxy_url:
|
|
86
|
-
for key, value in proxy_data.items():
|
|
87
|
-
if isinstance(value, str) and (value.startswith('http://') or value.startswith('https://')):
|
|
88
|
-
proxy_url = value
|
|
89
|
-
break
|
|
90
|
-
elif isinstance(value, dict):
|
|
91
|
-
# 递归查找
|
|
92
|
-
for sub_key, sub_value in value.items():
|
|
93
|
-
if isinstance(sub_value, str) and (sub_value.startswith('http://') or sub_value.startswith('https://')):
|
|
94
|
-
proxy_url = sub_value
|
|
95
|
-
break
|
|
96
|
-
if proxy_url:
|
|
97
|
-
break
|
|
98
|
-
|
|
99
|
-
elif isinstance(proxy_data, str):
|
|
100
|
-
# 如果响应是字符串,直接使用
|
|
101
|
-
if proxy_data.startswith('http://') or proxy_data.startswith('https://'):
|
|
102
|
-
proxy_url = proxy_data
|
|
103
|
-
|
|
104
|
-
return proxy_url
|
|
105
|
-
|
|
106
|
-
|
|
107
|
-
async def test_target_url_without_proxy(target_url):
|
|
108
|
-
"""不使用代理直接测试访问目标URL"""
|
|
109
|
-
print(f"\n=== 直接访问目标URL(不使用代理) ===")
|
|
110
|
-
print(f"目标URL: {target_url}")
|
|
111
|
-
|
|
112
|
-
try:
|
|
113
|
-
timeout = aiohttp.ClientTimeout(total=15)
|
|
114
|
-
async with aiohttp.ClientSession(timeout=timeout) as session:
|
|
115
|
-
# 添加用户代理头,避免被反爬虫机制拦截
|
|
116
|
-
headers = {
|
|
117
|
-
'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/124.0.0.0 Safari/537.36'
|
|
118
|
-
}
|
|
119
|
-
async with session.get(target_url, headers=headers) as response:
|
|
120
|
-
print(f"状态码: {response.status}")
|
|
121
|
-
print(f"响应头: {response.headers.get('content-type', 'Unknown')}")
|
|
122
|
-
|
|
123
|
-
# 只读取响应状态,不尝试解码内容
|
|
124
|
-
return response.status == 200
|
|
125
|
-
|
|
126
|
-
except asyncio.TimeoutError:
|
|
127
|
-
print("请求超时")
|
|
128
|
-
return False
|
|
129
|
-
except Exception as e:
|
|
130
|
-
print(f"请求失败: {e}")
|
|
131
|
-
return False
|
|
132
|
-
|
|
133
|
-
|
|
134
|
-
async def test_target_url_with_proxy(proxy_url, target_url, max_retries=3):
|
|
135
|
-
"""使用代理测试访问目标URL"""
|
|
136
|
-
print(f"\n=== 使用代理测试访问目标URL ===")
|
|
137
|
-
print(f"代理地址: {proxy_url}")
|
|
138
|
-
print(f"目标URL: {target_url}")
|
|
139
|
-
|
|
140
|
-
# 添加用户代理头,避免被反爬虫机制拦截
|
|
141
|
-
headers = {
|
|
142
|
-
'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/124.0.0.0 Safari/537.36'
|
|
143
|
-
}
|
|
144
|
-
|
|
145
|
-
for attempt in range(max_retries):
|
|
146
|
-
if attempt > 0:
|
|
147
|
-
print(f"\n第 {attempt + 1} 次重试...")
|
|
148
|
-
|
|
149
|
-
try:
|
|
150
|
-
# 创建aiohttp客户端会话
|
|
151
|
-
timeout = aiohttp.ClientTimeout(total=15)
|
|
152
|
-
async with aiohttp.ClientSession(timeout=timeout, headers=headers) as session:
|
|
153
|
-
# 处理代理URL,支持带认证的代理
|
|
154
|
-
if isinstance(proxy_url, str) and "@" in proxy_url and "://" in proxy_url:
|
|
155
|
-
parsed = urlparse(proxy_url)
|
|
156
|
-
if parsed.username and parsed.password:
|
|
157
|
-
# 提取认证信息
|
|
158
|
-
auth = aiohttp.BasicAuth(parsed.username, parsed.password)
|
|
159
|
-
# 清理代理URL,移除认证信息
|
|
160
|
-
clean_proxy = f"{parsed.scheme}://{parsed.hostname}"
|
|
161
|
-
if parsed.port:
|
|
162
|
-
clean_proxy += f":{parsed.port}"
|
|
163
|
-
|
|
164
|
-
print(f"使用带认证的代理: {clean_proxy}")
|
|
165
|
-
async with session.get(target_url, proxy=clean_proxy, proxy_auth=auth) as response:
|
|
166
|
-
print(f"状态码: {response.status}")
|
|
167
|
-
print(f"响应头: {response.headers.get('content-type', 'Unknown')}")
|
|
168
|
-
return response.status == 200
|
|
169
|
-
else:
|
|
170
|
-
# 没有认证信息的代理
|
|
171
|
-
print(f"使用普通代理: {proxy_url}")
|
|
172
|
-
async with session.get(target_url, proxy=proxy_url) as response:
|
|
173
|
-
print(f"状态码: {response.status}")
|
|
174
|
-
print(f"响应头: {response.headers.get('content-type', 'Unknown')}")
|
|
175
|
-
return response.status == 200
|
|
176
|
-
else:
|
|
177
|
-
# 直接使用代理URL
|
|
178
|
-
print(f"使用代理: {proxy_url}")
|
|
179
|
-
async with session.get(target_url, proxy=proxy_url) as response:
|
|
180
|
-
print(f"状态码: {response.status}")
|
|
181
|
-
print(f"响应头: {response.headers.get('content-type', 'Unknown')}")
|
|
182
|
-
return response.status == 200
|
|
183
|
-
|
|
184
|
-
except asyncio.TimeoutError:
|
|
185
|
-
print("请求超时")
|
|
186
|
-
if attempt < max_retries - 1:
|
|
187
|
-
await asyncio.sleep(2) # 等待2秒后重试
|
|
188
|
-
continue
|
|
189
|
-
except aiohttp.ClientConnectorError as e:
|
|
190
|
-
print(f"连接错误: {e}")
|
|
191
|
-
if attempt < max_retries - 1:
|
|
192
|
-
await asyncio.sleep(2) # 等待2秒后重试
|
|
193
|
-
continue
|
|
194
|
-
except aiohttp.ClientHttpProxyError as e:
|
|
195
|
-
print(f"代理HTTP错误: {e}")
|
|
196
|
-
if attempt < max_retries - 1:
|
|
197
|
-
await asyncio.sleep(2) # 等待2秒后重试
|
|
198
|
-
continue
|
|
199
|
-
except aiohttp.ServerDisconnectedError as e:
|
|
200
|
-
print(f"服务器断开连接: {e}")
|
|
201
|
-
if attempt < max_retries - 1:
|
|
202
|
-
await asyncio.sleep(2) # 等待2秒后重试
|
|
203
|
-
continue
|
|
204
|
-
except Exception as e:
|
|
205
|
-
print(f"请求失败: {e}")
|
|
206
|
-
if attempt < max_retries - 1:
|
|
207
|
-
await asyncio.sleep(2) # 等待2秒后重试
|
|
208
|
-
continue
|
|
209
|
-
|
|
210
|
-
return False
|
|
211
|
-
|
|
212
|
-
|
|
213
|
-
async def main():
|
|
214
|
-
"""主测试函数"""
|
|
215
|
-
# 指定的代理API和测试链接
|
|
216
|
-
proxy_api = 'http://test.proxy.api:8080/proxy/getitem/'
|
|
217
|
-
target_url = 'https://stock.10jqka.com.cn/20240315/c655957791.shtml'
|
|
218
|
-
|
|
219
|
-
print("开始测试代理接口和目标链接访问...\n")
|
|
220
|
-
|
|
221
|
-
# 1. 测试代理API接口
|
|
222
|
-
proxy_data = await test_proxy_api(proxy_api)
|
|
223
|
-
|
|
224
|
-
if not proxy_data:
|
|
225
|
-
print("代理API测试失败,无法获取代理信息")
|
|
226
|
-
return
|
|
227
|
-
|
|
228
|
-
# 2. 从API响应中提取代理URL
|
|
229
|
-
proxy_url = extract_proxy_url(proxy_data)
|
|
230
|
-
|
|
231
|
-
if not proxy_url:
|
|
232
|
-
print("无法从API响应中提取代理URL")
|
|
233
|
-
print(f"API响应内容: {proxy_data}")
|
|
234
|
-
return
|
|
235
|
-
|
|
236
|
-
print(f"\n提取到的代理URL: {proxy_url}")
|
|
237
|
-
|
|
238
|
-
# 3. 首先尝试直接访问,确认目标URL是否可访问
|
|
239
|
-
print("\n=== 测试直接访问目标URL ===")
|
|
240
|
-
direct_success = await test_target_url_without_proxy(target_url)
|
|
241
|
-
|
|
242
|
-
if direct_success:
|
|
243
|
-
print("直接访问目标URL成功")
|
|
244
|
-
else:
|
|
245
|
-
print("直接访问目标URL失败")
|
|
246
|
-
|
|
247
|
-
# 4. 使用代理访问目标URL
|
|
248
|
-
print("\n=== 测试使用代理访问目标URL ===")
|
|
249
|
-
proxy_success = await test_target_url_with_proxy(proxy_url, target_url)
|
|
250
|
-
|
|
251
|
-
if proxy_success:
|
|
252
|
-
print(f"代理测试成功!代理 {proxy_url} 可以正常访问目标链接")
|
|
253
|
-
else:
|
|
254
|
-
print(f"代理测试失败!代理 {proxy_url} 无法访问目标链接")
|
|
255
|
-
|
|
256
|
-
# 5. 总结
|
|
257
|
-
print(f"\n=== 测试总结 ===")
|
|
258
|
-
print(f"代理API访问: {'成功' if proxy_data else '失败'}")
|
|
259
|
-
print(f"代理提取: {'成功' if proxy_url else '失败'}")
|
|
260
|
-
print(f"直接访问: {'成功' if direct_success else '失败'}")
|
|
261
|
-
print(f"代理访问: {'成功' if proxy_success else '失败'}")
|
|
262
|
-
|
|
263
|
-
|
|
264
|
-
if __name__ == "__main__":
|
|
265
|
-
asyncio.run(main())
|
|
@@ -1,196 +0,0 @@
|
|
|
1
|
-
#!/usr/bin/python
|
|
2
|
-
# -*- coding: UTF-8 -*-
|
|
3
|
-
"""
|
|
4
|
-
真实场景代理测试
|
|
5
|
-
================
|
|
6
|
-
使用用户提供的headers、cookies和URL测试代理功能
|
|
7
|
-
"""
|
|
8
|
-
|
|
9
|
-
import asyncio
|
|
10
|
-
import aiohttp
|
|
11
|
-
import sys
|
|
12
|
-
import os
|
|
13
|
-
from urllib.parse import urlparse
|
|
14
|
-
|
|
15
|
-
# 添加项目根目录到Python路径
|
|
16
|
-
sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..'))
|
|
17
|
-
|
|
18
|
-
# 用户提供的请求头
|
|
19
|
-
HEADERS = {
|
|
20
|
-
"accept": "text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,image/apng,*/*;q=0.8,application/signed-exchange;v=b3;q=0.7",
|
|
21
|
-
"accept-language": "zh-CN,zh;q=0.9,en;q=0.8",
|
|
22
|
-
"cache-control": "no-cache",
|
|
23
|
-
"pragma": "no-cache",
|
|
24
|
-
"priority": "u=0, i",
|
|
25
|
-
"sec-ch-ua": "\"Chromium\";v=\"140\", \"Not=A?Brand\";v=\"24\", \"Google Chrome\";v=\"140\"",
|
|
26
|
-
"sec-ch-ua-mobile": "?0",
|
|
27
|
-
"sec-ch-ua-platform": "\"Windows\"",
|
|
28
|
-
"sec-fetch-dest": "document",
|
|
29
|
-
"sec-fetch-mode": "navigate",
|
|
30
|
-
"sec-fetch-site": "none",
|
|
31
|
-
"sec-fetch-user": "?1",
|
|
32
|
-
"upgrade-insecure-requests": "1",
|
|
33
|
-
"user-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/140.0.0.0 Safari/537.36"
|
|
34
|
-
}
|
|
35
|
-
|
|
36
|
-
# 用户提供的cookies
|
|
37
|
-
COOKIES = {
|
|
38
|
-
"Hm_lvt_722143063e4892925903024537075d0d": "1758071793",
|
|
39
|
-
"Hm_lvt_929f8b362150b1f77b477230541dbbc2": "1758071793",
|
|
40
|
-
"historystock": "600699",
|
|
41
|
-
"spversion": "20130314",
|
|
42
|
-
"cid": "f9bc812da2c3a7ddf6d5df1fa2d497091758076438",
|
|
43
|
-
"u_ukey": "A10702B8689642C6BE607730E11E6E4A",
|
|
44
|
-
"u_uver": "1.0.0",
|
|
45
|
-
"u_dpass": "Qk3U07X7SHGKa0AcRUg1R1DVWbPioD9Eg270bdikvlwWWXexbsXnRsQNt%2B04iXwdHi80LrSsTFH9a%2B6rtRvqGg%3D%3D",
|
|
46
|
-
"u_did": "E3ED337393E1429DA56E380DD00B3CCD",
|
|
47
|
-
"u_ttype": "WEB",
|
|
48
|
-
"user_status": "0",
|
|
49
|
-
"ttype": "WEB",
|
|
50
|
-
"log": "",
|
|
51
|
-
"Hm_lvt_69929b9dce4c22a060bd22d703b2a280": "1758079404,1758113068,1758157144",
|
|
52
|
-
"HMACCOUNT": "08DF0D235A291EAA",
|
|
53
|
-
"Hm_lvt_78c58f01938e4d85eaf619eae71b4ed1": "1758071793,1758113068,1758157144",
|
|
54
|
-
"user": "MDpteF9lNXRkY3RpdHo6Ok5vbmU6NTAwOjgxNzYyOTAwNDo3LDExMTExMTExMTExLDQwOzQ0LDExLDQwOzYsMSw0MDs1LDEsNDA7MSwxMDEsNDA7MiwxLDQwOzMsMSw0MDs1LDEsNDA7OCwwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMSw0MDsxMDIsMSw0MDoxNjo6OjgwNzYyOTAwNDoxNzU4MTYxNTE0Ojo6MTc1ODA3MjA2MDo2MDQ4MDA6MDoxYTQ0NmFlNDY4M2VmZWY3YmNjYTczY2U3ODZmZTNiODg6ZGVmYXVsdF81OjA%3D",
|
|
55
|
-
"userid": "807629004",
|
|
56
|
-
"u_name": "mx_e5tdctitz",
|
|
57
|
-
"escapename": "mx_e5tdctitz",
|
|
58
|
-
"ticket": "85eea709becdd924d7eb975351de629e",
|
|
59
|
-
"utk": "8959c4c6b6f5fb7628864feab15473f4",
|
|
60
|
-
"sess_tk": "eyJ0eXAiOiJKV1QiLCJhbGciOiJFUzI1NiIsImtpZCI6InNlc3NfdGtfMSIsImJ0eSI6InNlc3NfdGsifQ.eyJqdGkiOiI4ODNiZmU4NmU3M2NhN2NjN2JlZmVmODM0NmFlNDZhNDEiLCJpYXQiOjE3NTgxNjE1MTQsImV4cCI6MTc1ODc2NjMxNCwic3ViIjoiODA3NjI5MDA0IiwiaXNzIjoidXBhc3MuaXdlbmNhaS5jb20iLCJhdWQiOiIyMDIwMTExODUyODg5MDcyIiwiYWN0Ijoib2ZjIiwiY3VocyI6ImIwNTcyZDVjOWNlNDg0MGFlOWYxYTlhYjU3NGZkNjQyYjgzNmExN2E3Y2NhZjk4ZWRiNzI5ZmJkOWFjOGVkYmYifQ.UBNIzxGvQQtXSiIcB_1JJl-EuAc1S9j2LcTLXjwy4ImhDDbh1oJvyRdDUrXdUpwBpIyx5zgYqgt_3FEhY_iayw",
|
|
61
|
-
"cuc": "ap2eap3gg99g",
|
|
62
|
-
"Hm_lvt_f79b64788a4e377c608617fba4c736e2": "1758161692",
|
|
63
|
-
"v": "A1glI4rWhPCQGqh0MvA0ioufKY3vQbzLHqWQT5JJpBNGLfazOlGMW261YNrh",
|
|
64
|
-
"Hm_lpvt_78c58f01938e4d85eaf619eae71b4ed1": "1758163145",
|
|
65
|
-
"Hm_lpvt_f79b64788a4e377c608617fba4c736e2": "1758163145",
|
|
66
|
-
"Hm_lpvt_69929b9dce4c22a060bd22d703b2a280": "1758163145"
|
|
67
|
-
}
|
|
68
|
-
|
|
69
|
-
# 用户提供的URL
|
|
70
|
-
URL = "https://stock.10jqka.com.cn/20240315/c655957791.shtml"
|
|
71
|
-
|
|
72
|
-
|
|
73
|
-
async def test_without_proxy():
|
|
74
|
-
"""不使用代理直接测试访问"""
|
|
75
|
-
print("=== 不使用代理直接访问 ===")
|
|
76
|
-
print(f"URL: {URL}")
|
|
77
|
-
|
|
78
|
-
try:
|
|
79
|
-
timeout = aiohttp.ClientTimeout(total=15)
|
|
80
|
-
async with aiohttp.ClientSession(timeout=timeout, headers=HEADERS, cookies=COOKIES) as session:
|
|
81
|
-
async with session.get(URL) as response:
|
|
82
|
-
print(f"状态码: {response.status}")
|
|
83
|
-
if response.status == 200:
|
|
84
|
-
print("直接访问成功")
|
|
85
|
-
return True
|
|
86
|
-
else:
|
|
87
|
-
print(f"直接访问失败,状态码: {response.status}")
|
|
88
|
-
return False
|
|
89
|
-
except Exception as e:
|
|
90
|
-
print(f"直接访问出错: {e}")
|
|
91
|
-
return False
|
|
92
|
-
|
|
93
|
-
|
|
94
|
-
async def test_with_proxy(proxy_url):
|
|
95
|
-
"""使用代理测试访问"""
|
|
96
|
-
print(f"\n=== 使用代理访问 ===")
|
|
97
|
-
print(f"代理地址: {proxy_url}")
|
|
98
|
-
print(f"URL: {URL}")
|
|
99
|
-
|
|
100
|
-
try:
|
|
101
|
-
timeout = aiohttp.ClientTimeout(total=15)
|
|
102
|
-
async with aiohttp.ClientSession(timeout=timeout, headers=HEADERS, cookies=COOKIES) as session:
|
|
103
|
-
# 处理带认证的代理
|
|
104
|
-
if "@" in proxy_url and "://" in proxy_url:
|
|
105
|
-
parsed = urlparse(proxy_url)
|
|
106
|
-
if parsed.username and parsed.password:
|
|
107
|
-
# 提取认证信息
|
|
108
|
-
auth = aiohttp.BasicAuth(parsed.username, parsed.password)
|
|
109
|
-
# 清理代理URL
|
|
110
|
-
clean_proxy = f"{parsed.scheme}://{parsed.hostname}"
|
|
111
|
-
if parsed.port:
|
|
112
|
-
clean_proxy += f":{parsed.port}"
|
|
113
|
-
|
|
114
|
-
print(f"使用带认证的代理: {clean_proxy}")
|
|
115
|
-
async with session.get(URL, proxy=clean_proxy, proxy_auth=auth) as response:
|
|
116
|
-
print(f"状态码: {response.status}")
|
|
117
|
-
if response.status == 200:
|
|
118
|
-
print("代理访问成功")
|
|
119
|
-
return True
|
|
120
|
-
else:
|
|
121
|
-
print(f"代理访问失败,状态码: {response.status}")
|
|
122
|
-
return False
|
|
123
|
-
else:
|
|
124
|
-
# 直接使用代理URL
|
|
125
|
-
print(f"使用代理: {proxy_url}")
|
|
126
|
-
async with session.get(URL, proxy=proxy_url) as response:
|
|
127
|
-
print(f"状态码: {response.status}")
|
|
128
|
-
if response.status == 200:
|
|
129
|
-
print("代理访问成功")
|
|
130
|
-
return True
|
|
131
|
-
else:
|
|
132
|
-
print(f"代理访问失败,状态码: {response.status}")
|
|
133
|
-
return False
|
|
134
|
-
except Exception as e:
|
|
135
|
-
print(f"代理访问出错: {e}")
|
|
136
|
-
return False
|
|
137
|
-
|
|
138
|
-
|
|
139
|
-
async def get_proxy_from_api():
|
|
140
|
-
"""从代理API获取代理"""
|
|
141
|
-
proxy_api = 'http://test.proxy.api:8080/proxy/getitem/'
|
|
142
|
-
print(f"\n=== 从代理API获取代理 ===")
|
|
143
|
-
print(f"API地址: {proxy_api}")
|
|
144
|
-
|
|
145
|
-
try:
|
|
146
|
-
timeout = aiohttp.ClientTimeout(total=10)
|
|
147
|
-
async with aiohttp.ClientSession(timeout=timeout) as session:
|
|
148
|
-
async with session.get(proxy_api) as response:
|
|
149
|
-
print(f"状态码: {response.status}")
|
|
150
|
-
|
|
151
|
-
if response.status == 200:
|
|
152
|
-
data = await response.json()
|
|
153
|
-
print(f"响应数据: {data}")
|
|
154
|
-
|
|
155
|
-
# 提取代理URL
|
|
156
|
-
if isinstance(data, dict) and data.get('status') == 0:
|
|
157
|
-
proxy_info = data.get('proxy', {})
|
|
158
|
-
if isinstance(proxy_info, dict):
|
|
159
|
-
proxy_url = proxy_info.get('https') or proxy_info.get('http')
|
|
160
|
-
if proxy_url:
|
|
161
|
-
print(f"提取到的代理URL: {proxy_url}")
|
|
162
|
-
return proxy_url
|
|
163
|
-
print("无法获取代理URL")
|
|
164
|
-
return None
|
|
165
|
-
except Exception as e:
|
|
166
|
-
print(f"API请求出错: {e}")
|
|
167
|
-
return None
|
|
168
|
-
|
|
169
|
-
|
|
170
|
-
async def main():
|
|
171
|
-
"""主测试函数"""
|
|
172
|
-
print("开始真实场景代理测试...")
|
|
173
|
-
print("=" * 50)
|
|
174
|
-
|
|
175
|
-
# 1. 首先测试不使用代理直接访问
|
|
176
|
-
direct_success = await test_without_proxy()
|
|
177
|
-
|
|
178
|
-
# 2. 从代理API获取代理
|
|
179
|
-
proxy_url = await get_proxy_from_api()
|
|
180
|
-
|
|
181
|
-
if not proxy_url:
|
|
182
|
-
print("\n无法获取代理,测试结束")
|
|
183
|
-
return
|
|
184
|
-
|
|
185
|
-
# 3. 使用代理访问
|
|
186
|
-
proxy_success = await test_with_proxy(proxy_url)
|
|
187
|
-
|
|
188
|
-
# 4. 测试结果总结
|
|
189
|
-
print(f"\n{'='*30}")
|
|
190
|
-
print("测试结果:")
|
|
191
|
-
print(f"直接访问: {'成功' if direct_success else '失败'}")
|
|
192
|
-
print(f"代理访问: {'成功' if proxy_success else '失败'}")
|
|
193
|
-
|
|
194
|
-
|
|
195
|
-
if __name__ == "__main__":
|
|
196
|
-
asyncio.run(main())
|