crawlo 1.4.5__py3-none-any.whl → 1.4.7__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of crawlo might be problematic. Click here for more details.
- crawlo/__init__.py +90 -89
- crawlo/__version__.py +1 -1
- crawlo/cli.py +75 -75
- crawlo/commands/__init__.py +14 -14
- crawlo/commands/check.py +594 -594
- crawlo/commands/genspider.py +186 -186
- crawlo/commands/help.py +140 -138
- crawlo/commands/list.py +155 -155
- crawlo/commands/run.py +379 -341
- crawlo/commands/startproject.py +460 -460
- crawlo/commands/stats.py +187 -187
- crawlo/commands/utils.py +196 -196
- crawlo/config.py +320 -312
- crawlo/config_validator.py +277 -277
- crawlo/core/__init__.py +52 -52
- crawlo/core/engine.py +451 -438
- crawlo/core/processor.py +47 -47
- crawlo/core/scheduler.py +290 -291
- crawlo/crawler.py +698 -657
- crawlo/data/__init__.py +5 -5
- crawlo/data/user_agents.py +194 -194
- crawlo/downloader/__init__.py +280 -276
- crawlo/downloader/aiohttp_downloader.py +233 -233
- crawlo/downloader/cffi_downloader.py +250 -245
- crawlo/downloader/httpx_downloader.py +265 -259
- crawlo/downloader/hybrid_downloader.py +212 -212
- crawlo/downloader/playwright_downloader.py +425 -402
- crawlo/downloader/selenium_downloader.py +486 -472
- crawlo/event.py +45 -11
- crawlo/exceptions.py +215 -82
- crawlo/extension/__init__.py +65 -64
- crawlo/extension/health_check.py +141 -141
- crawlo/extension/log_interval.py +94 -94
- crawlo/extension/log_stats.py +70 -70
- crawlo/extension/logging_extension.py +53 -61
- crawlo/extension/memory_monitor.py +104 -104
- crawlo/extension/performance_profiler.py +133 -133
- crawlo/extension/request_recorder.py +107 -107
- crawlo/factories/__init__.py +27 -27
- crawlo/factories/base.py +68 -68
- crawlo/factories/crawler.py +104 -103
- crawlo/factories/registry.py +84 -84
- crawlo/factories/utils.py +135 -0
- crawlo/filters/__init__.py +170 -153
- crawlo/filters/aioredis_filter.py +348 -264
- crawlo/filters/memory_filter.py +261 -276
- crawlo/framework.py +306 -292
- crawlo/initialization/__init__.py +44 -44
- crawlo/initialization/built_in.py +391 -434
- crawlo/initialization/context.py +141 -141
- crawlo/initialization/core.py +240 -194
- crawlo/initialization/phases.py +230 -149
- crawlo/initialization/registry.py +143 -145
- crawlo/initialization/utils.py +49 -0
- crawlo/interfaces.py +23 -23
- crawlo/items/__init__.py +23 -23
- crawlo/items/base.py +23 -23
- crawlo/items/fields.py +52 -52
- crawlo/items/items.py +104 -104
- crawlo/logging/__init__.py +42 -46
- crawlo/logging/config.py +277 -197
- crawlo/logging/factory.py +175 -171
- crawlo/logging/manager.py +104 -112
- crawlo/middleware/__init__.py +87 -24
- crawlo/middleware/default_header.py +132 -132
- crawlo/middleware/download_delay.py +104 -104
- crawlo/middleware/middleware_manager.py +142 -142
- crawlo/middleware/offsite.py +123 -123
- crawlo/middleware/proxy.py +209 -386
- crawlo/middleware/request_ignore.py +86 -86
- crawlo/middleware/response_code.py +150 -150
- crawlo/middleware/response_filter.py +136 -136
- crawlo/middleware/retry.py +124 -124
- crawlo/mode_manager.py +287 -253
- crawlo/network/__init__.py +21 -21
- crawlo/network/request.py +375 -379
- crawlo/network/response.py +569 -664
- crawlo/pipelines/__init__.py +53 -22
- crawlo/pipelines/base_pipeline.py +452 -0
- crawlo/pipelines/bloom_dedup_pipeline.py +146 -146
- crawlo/pipelines/console_pipeline.py +39 -39
- crawlo/pipelines/csv_pipeline.py +316 -316
- crawlo/pipelines/database_dedup_pipeline.py +197 -197
- crawlo/pipelines/json_pipeline.py +218 -218
- crawlo/pipelines/memory_dedup_pipeline.py +105 -105
- crawlo/pipelines/mongo_pipeline.py +140 -132
- crawlo/pipelines/mysql_pipeline.py +470 -326
- crawlo/pipelines/pipeline_manager.py +100 -100
- crawlo/pipelines/redis_dedup_pipeline.py +155 -156
- crawlo/project.py +347 -347
- crawlo/queue/__init__.py +10 -0
- crawlo/queue/pqueue.py +38 -38
- crawlo/queue/queue_manager.py +591 -525
- crawlo/queue/redis_priority_queue.py +519 -370
- crawlo/settings/__init__.py +7 -7
- crawlo/settings/default_settings.py +285 -270
- crawlo/settings/setting_manager.py +219 -219
- crawlo/spider/__init__.py +657 -657
- crawlo/stats_collector.py +82 -73
- crawlo/subscriber.py +129 -129
- crawlo/task_manager.py +138 -138
- crawlo/templates/crawlo.cfg.tmpl +10 -10
- crawlo/templates/project/__init__.py.tmpl +2 -4
- crawlo/templates/project/items.py.tmpl +13 -17
- crawlo/templates/project/middlewares.py.tmpl +38 -38
- crawlo/templates/project/pipelines.py.tmpl +35 -36
- crawlo/templates/project/settings.py.tmpl +110 -157
- crawlo/templates/project/settings_distributed.py.tmpl +156 -161
- crawlo/templates/project/settings_gentle.py.tmpl +170 -171
- crawlo/templates/project/settings_high_performance.py.tmpl +171 -172
- crawlo/templates/project/settings_minimal.py.tmpl +99 -77
- crawlo/templates/project/settings_simple.py.tmpl +168 -169
- crawlo/templates/project/spiders/__init__.py.tmpl +9 -9
- crawlo/templates/run.py.tmpl +23 -30
- crawlo/templates/spider/spider.py.tmpl +33 -144
- crawlo/templates/spiders_init.py.tmpl +5 -10
- crawlo/tools/__init__.py +86 -189
- crawlo/tools/date_tools.py +289 -289
- crawlo/tools/distributed_coordinator.py +384 -384
- crawlo/tools/scenario_adapter.py +262 -262
- crawlo/tools/text_cleaner.py +232 -232
- crawlo/utils/__init__.py +50 -50
- crawlo/utils/batch_processor.py +276 -259
- crawlo/utils/config_manager.py +442 -0
- crawlo/utils/controlled_spider_mixin.py +439 -439
- crawlo/utils/db_helper.py +250 -244
- crawlo/utils/error_handler.py +410 -410
- crawlo/utils/fingerprint.py +121 -121
- crawlo/utils/func_tools.py +82 -82
- crawlo/utils/large_scale_helper.py +344 -344
- crawlo/utils/leak_detector.py +335 -0
- crawlo/utils/log.py +79 -79
- crawlo/utils/misc.py +81 -81
- crawlo/utils/mongo_connection_pool.py +157 -0
- crawlo/utils/mysql_connection_pool.py +197 -0
- crawlo/utils/performance_monitor.py +285 -285
- crawlo/utils/queue_helper.py +175 -175
- crawlo/utils/redis_checker.py +91 -0
- crawlo/utils/redis_connection_pool.py +578 -388
- crawlo/utils/redis_key_validator.py +198 -198
- crawlo/utils/request.py +278 -256
- crawlo/utils/request_serializer.py +225 -225
- crawlo/utils/resource_manager.py +337 -0
- crawlo/utils/selector_helper.py +137 -137
- crawlo/utils/singleton.py +70 -0
- crawlo/utils/spider_loader.py +201 -201
- crawlo/utils/text_helper.py +94 -94
- crawlo/utils/{url.py → url_utils.py} +39 -39
- crawlo-1.4.7.dist-info/METADATA +689 -0
- crawlo-1.4.7.dist-info/RECORD +347 -0
- examples/__init__.py +7 -7
- tests/__init__.py +7 -7
- tests/advanced_tools_example.py +217 -275
- tests/authenticated_proxy_example.py +110 -106
- tests/baidu_performance_test.py +108 -108
- tests/baidu_test.py +59 -59
- tests/bug_check_test.py +250 -250
- tests/cleaners_example.py +160 -160
- tests/comprehensive_framework_test.py +212 -212
- tests/comprehensive_test.py +81 -81
- tests/comprehensive_testing_summary.md +186 -186
- tests/config_validation_demo.py +142 -142
- tests/controlled_spider_example.py +205 -205
- tests/date_tools_example.py +180 -180
- tests/debug_configure.py +69 -69
- tests/debug_framework_logger.py +84 -84
- tests/debug_log_config.py +126 -126
- tests/debug_log_levels.py +63 -63
- tests/debug_pipelines.py +66 -66
- tests/detailed_log_test.py +233 -233
- tests/direct_selector_helper_test.py +96 -96
- tests/distributed_dedup_test.py +467 -0
- tests/distributed_test.py +66 -66
- tests/distributed_test_debug.py +76 -76
- tests/dynamic_loading_example.py +523 -523
- tests/dynamic_loading_test.py +104 -104
- tests/error_handling_example.py +171 -171
- tests/explain_mysql_update_behavior.py +77 -0
- tests/final_comprehensive_test.py +151 -151
- tests/final_log_test.py +260 -260
- tests/final_validation_test.py +182 -182
- tests/fix_log_test.py +142 -142
- tests/framework_performance_test.py +202 -202
- tests/log_buffering_test.py +111 -111
- tests/log_generation_timing_test.py +153 -153
- tests/monitor_redis_dedup.sh +72 -0
- tests/ofweek_scrapy/ofweek_scrapy/items.py +12 -12
- tests/ofweek_scrapy/ofweek_scrapy/middlewares.py +100 -100
- tests/ofweek_scrapy/ofweek_scrapy/pipelines.py +13 -13
- tests/ofweek_scrapy/ofweek_scrapy/settings.py +84 -84
- tests/ofweek_scrapy/scrapy.cfg +11 -11
- tests/optimized_performance_test.py +211 -211
- tests/performance_comparison.py +244 -244
- tests/queue_blocking_test.py +113 -113
- tests/queue_test.py +89 -89
- tests/redis_key_validation_demo.py +130 -130
- tests/request_params_example.py +150 -150
- tests/response_improvements_example.py +144 -144
- tests/scrapy_comparison/ofweek_scrapy.py +138 -138
- tests/scrapy_comparison/scrapy_test.py +133 -133
- tests/simple_cli_test.py +55 -0
- tests/simple_command_test.py +119 -119
- tests/simple_crawlo_test.py +126 -126
- tests/simple_follow_test.py +38 -38
- tests/simple_log_test2.py +137 -137
- tests/simple_optimization_test.py +128 -128
- tests/simple_queue_type_test.py +41 -41
- tests/simple_response_selector_test.py +94 -94
- tests/simple_selector_helper_test.py +154 -154
- tests/simple_selector_test.py +207 -207
- tests/simple_spider_test.py +49 -49
- tests/simple_url_test.py +73 -73
- tests/simulate_mysql_update_test.py +140 -0
- tests/spider_log_timing_test.py +177 -177
- tests/test_advanced_tools.py +148 -148
- tests/test_all_commands.py +230 -230
- tests/test_all_pipeline_fingerprints.py +133 -133
- tests/test_all_redis_key_configs.py +145 -145
- tests/test_asyncmy_usage.py +57 -0
- tests/test_batch_processor.py +178 -178
- tests/test_cleaners.py +54 -54
- tests/test_cli_arguments.py +119 -0
- tests/test_component_factory.py +174 -174
- tests/test_config_consistency.py +80 -80
- tests/test_config_merge.py +152 -152
- tests/test_config_validator.py +182 -182
- tests/test_controlled_spider_mixin.py +79 -79
- tests/test_crawler_process_import.py +38 -38
- tests/test_crawler_process_spider_modules.py +47 -47
- tests/test_crawlo_proxy_integration.py +114 -108
- tests/test_date_tools.py +123 -123
- tests/test_dedup_fix.py +220 -220
- tests/test_dedup_pipeline_consistency.py +124 -124
- tests/test_default_header_middleware.py +313 -313
- tests/test_distributed.py +65 -65
- tests/test_double_crawlo_fix.py +204 -204
- tests/test_double_crawlo_fix_simple.py +124 -124
- tests/test_download_delay_middleware.py +221 -221
- tests/test_downloader_proxy_compatibility.py +272 -268
- tests/test_edge_cases.py +305 -305
- tests/test_encoding_core.py +56 -56
- tests/test_encoding_detection.py +126 -126
- tests/test_enhanced_error_handler.py +270 -270
- tests/test_enhanced_error_handler_comprehensive.py +245 -245
- tests/test_error_handler_compatibility.py +112 -112
- tests/test_factories.py +252 -252
- tests/test_factory_compatibility.py +196 -196
- tests/test_final_validation.py +153 -153
- tests/test_fingerprint_consistency.py +135 -135
- tests/test_fingerprint_simple.py +51 -51
- tests/test_get_component_logger.py +83 -83
- tests/test_hash_performance.py +99 -99
- tests/test_integration.py +169 -169
- tests/test_item_dedup_redis_key.py +122 -122
- tests/test_large_scale_helper.py +235 -235
- tests/test_logging_enhancements.py +374 -374
- tests/test_logging_final.py +184 -184
- tests/test_logging_integration.py +312 -312
- tests/test_logging_system.py +282 -282
- tests/test_middleware_debug.py +141 -141
- tests/test_mode_consistency.py +51 -51
- tests/test_multi_directory.py +67 -67
- tests/test_multiple_spider_modules.py +80 -80
- tests/test_mysql_pipeline_config.py +165 -0
- tests/test_mysql_pipeline_error.py +99 -0
- tests/test_mysql_pipeline_init_log.py +83 -0
- tests/test_mysql_pipeline_integration.py +133 -0
- tests/test_mysql_pipeline_refactor.py +144 -0
- tests/test_mysql_pipeline_refactor_simple.py +86 -0
- tests/test_mysql_pipeline_robustness.py +196 -0
- tests/test_mysql_pipeline_types.py +89 -0
- tests/test_mysql_update_columns.py +94 -0
- tests/test_offsite_middleware.py +244 -244
- tests/test_offsite_middleware_simple.py +203 -203
- tests/test_optimized_selector_naming.py +100 -100
- tests/test_parsel.py +29 -29
- tests/test_performance.py +327 -327
- tests/test_performance_monitor.py +115 -115
- tests/test_pipeline_fingerprint_consistency.py +86 -86
- tests/test_priority_behavior.py +211 -211
- tests/test_priority_consistency.py +151 -151
- tests/test_priority_consistency_fixed.py +249 -249
- tests/test_proxy_health_check.py +32 -32
- tests/test_proxy_middleware.py +217 -121
- tests/test_proxy_middleware_enhanced.py +212 -216
- tests/test_proxy_middleware_integration.py +142 -137
- tests/test_proxy_middleware_refactored.py +207 -184
- tests/test_proxy_only.py +84 -0
- tests/test_proxy_providers.py +56 -56
- tests/test_proxy_stats.py +19 -19
- tests/test_proxy_strategies.py +59 -59
- tests/test_proxy_with_downloader.py +153 -0
- tests/test_queue_empty_check.py +41 -41
- tests/test_queue_manager_double_crawlo.py +173 -173
- tests/test_queue_manager_redis_key.py +179 -179
- tests/test_queue_naming.py +154 -154
- tests/test_queue_type.py +106 -106
- tests/test_queue_type_redis_config_consistency.py +130 -130
- tests/test_random_headers_default.py +322 -322
- tests/test_random_headers_necessity.py +308 -308
- tests/test_random_user_agent.py +72 -72
- tests/test_redis_config.py +28 -28
- tests/test_redis_connection_pool.py +294 -294
- tests/test_redis_key_naming.py +181 -181
- tests/test_redis_key_validator.py +123 -123
- tests/test_redis_queue.py +224 -224
- tests/test_redis_queue_name_fix.py +175 -175
- tests/test_redis_queue_type_fallback.py +129 -129
- tests/test_request_ignore_middleware.py +182 -182
- tests/test_request_params.py +111 -111
- tests/test_request_serialization.py +70 -70
- tests/test_response_code_middleware.py +349 -349
- tests/test_response_filter_middleware.py +427 -427
- tests/test_response_follow.py +104 -104
- tests/test_response_improvements.py +152 -152
- tests/test_response_selector_methods.py +92 -92
- tests/test_response_url_methods.py +70 -70
- tests/test_response_urljoin.py +86 -86
- tests/test_retry_middleware.py +333 -333
- tests/test_retry_middleware_realistic.py +273 -273
- tests/test_scheduler.py +252 -252
- tests/test_scheduler_config_update.py +133 -133
- tests/test_scrapy_style_encoding.py +112 -112
- tests/test_selector_helper.py +100 -100
- tests/test_selector_optimizations.py +146 -146
- tests/test_simple_response.py +61 -61
- tests/test_spider_loader.py +49 -49
- tests/test_spider_loader_comprehensive.py +69 -69
- tests/test_spider_modules.py +84 -84
- tests/test_spiders/test_spider.py +9 -9
- tests/test_telecom_spider_redis_key.py +205 -205
- tests/test_template_content.py +87 -87
- tests/test_template_redis_key.py +134 -134
- tests/test_tools.py +159 -159
- tests/test_user_agent_randomness.py +176 -176
- tests/test_user_agents.py +96 -96
- tests/untested_features_report.md +138 -138
- tests/verify_debug.py +51 -51
- tests/verify_distributed.py +117 -117
- tests/verify_log_fix.py +111 -111
- tests/verify_mysql_warnings.py +110 -0
- crawlo/logging/async_handler.py +0 -181
- crawlo/logging/monitor.py +0 -153
- crawlo/logging/sampler.py +0 -167
- crawlo/middleware/simple_proxy.py +0 -65
- crawlo/tools/authenticated_proxy.py +0 -241
- crawlo/tools/data_formatter.py +0 -226
- crawlo/tools/data_validator.py +0 -181
- crawlo/tools/encoding_converter.py +0 -127
- crawlo/tools/network_diagnostic.py +0 -365
- crawlo/tools/request_tools.py +0 -83
- crawlo/tools/retry_mechanism.py +0 -224
- crawlo/utils/env_config.py +0 -143
- crawlo/utils/large_scale_config.py +0 -287
- crawlo/utils/system.py +0 -11
- crawlo/utils/tools.py +0 -5
- crawlo-1.4.5.dist-info/METADATA +0 -329
- crawlo-1.4.5.dist-info/RECORD +0 -347
- tests/env_config_example.py +0 -134
- tests/ofweek_scrapy/ofweek_scrapy/spiders/ofweek_spider.py +0 -162
- tests/test_authenticated_proxy.py +0 -142
- tests/test_comprehensive.py +0 -147
- tests/test_dynamic_downloaders_proxy.py +0 -125
- tests/test_dynamic_proxy.py +0 -93
- tests/test_dynamic_proxy_config.py +0 -147
- tests/test_dynamic_proxy_real.py +0 -110
- tests/test_env_config.py +0 -122
- tests/test_framework_env_usage.py +0 -104
- tests/test_large_scale_config.py +0 -113
- tests/test_proxy_api.py +0 -265
- tests/test_real_scenario_proxy.py +0 -196
- tests/tools_example.py +0 -261
- {crawlo-1.4.5.dist-info → crawlo-1.4.7.dist-info}/WHEEL +0 -0
- {crawlo-1.4.5.dist-info → crawlo-1.4.7.dist-info}/entry_points.txt +0 -0
- {crawlo-1.4.5.dist-info → crawlo-1.4.7.dist-info}/top_level.txt +0 -0
tests/date_tools_example.py
CHANGED
|
@@ -1,181 +1,181 @@
|
|
|
1
|
-
#!/usr/bin/python
|
|
2
|
-
# -*- coding: UTF-8 -*-
|
|
3
|
-
"""
|
|
4
|
-
Crawlo框架日期工具使用示例
|
|
5
|
-
"""
|
|
6
|
-
from crawlo.tools import (
|
|
7
|
-
TimeUtils,
|
|
8
|
-
parse_time,
|
|
9
|
-
format_time,
|
|
10
|
-
time_diff,
|
|
11
|
-
to_timestamp,
|
|
12
|
-
to_datetime,
|
|
13
|
-
now,
|
|
14
|
-
to_timezone,
|
|
15
|
-
to_utc,
|
|
16
|
-
to_local
|
|
17
|
-
)
|
|
18
|
-
|
|
19
|
-
|
|
20
|
-
def demo_date_parsing():
|
|
21
|
-
"""演示日期解析功能"""
|
|
22
|
-
print("=== 日期解析演示 ===\n")
|
|
23
|
-
|
|
24
|
-
# 解析各种格式的日期字符串
|
|
25
|
-
date_strings = [
|
|
26
|
-
"2025-09-10 14:30:00",
|
|
27
|
-
"September 10, 2025 2:30 PM",
|
|
28
|
-
"2025/09/10 14:30:00",
|
|
29
|
-
"10-09-2025 14:30:00",
|
|
30
|
-
"2025年9月10日 14时30分00秒"
|
|
31
|
-
]
|
|
32
|
-
|
|
33
|
-
for date_str in date_strings:
|
|
34
|
-
parsed = parse_time(date_str)
|
|
35
|
-
print(f"解析 '{date_str}' -> {parsed}")
|
|
36
|
-
|
|
37
|
-
print()
|
|
38
|
-
|
|
39
|
-
|
|
40
|
-
def demo_date_formatting():
|
|
41
|
-
"""演示日期格式化功能"""
|
|
42
|
-
print("=== 日期格式化演示 ===\n")
|
|
43
|
-
|
|
44
|
-
# 获取当前时间
|
|
45
|
-
current_time = now()
|
|
46
|
-
print(f"当前时间: {current_time}")
|
|
47
|
-
|
|
48
|
-
# 使用不同格式进行格式化
|
|
49
|
-
formats = [
|
|
50
|
-
"%Y-%m-%d",
|
|
51
|
-
"%Y年%m月%d日",
|
|
52
|
-
"%Y-%m-%d %H:%M:%S",
|
|
53
|
-
"%B %d, %Y at %I:%M %p",
|
|
54
|
-
"%A, %B %d, %Y"
|
|
55
|
-
]
|
|
56
|
-
|
|
57
|
-
for fmt in formats:
|
|
58
|
-
formatted = format_time(current_time, fmt)
|
|
59
|
-
print(f"格式化为 '{fmt}' -> {formatted}")
|
|
60
|
-
|
|
61
|
-
print()
|
|
62
|
-
|
|
63
|
-
|
|
64
|
-
def demo_time_calculations():
|
|
65
|
-
"""演示时间计算功能"""
|
|
66
|
-
print("=== 时间计算演示 ===\n")
|
|
67
|
-
|
|
68
|
-
# 计算时间差
|
|
69
|
-
start_time = "2025-09-10 10:00:00"
|
|
70
|
-
end_time = "2025-09-10 14:30:45"
|
|
71
|
-
|
|
72
|
-
diff_seconds = time_diff(start_time, end_time, "seconds")
|
|
73
|
-
diff_minutes = time_diff(start_time, end_time, "minutes")
|
|
74
|
-
diff_hours = time_diff(start_time, end_time, "hours")
|
|
75
|
-
|
|
76
|
-
print(f"起始时间: {start_time}")
|
|
77
|
-
print(f"结束时间: {end_time}")
|
|
78
|
-
print(f"时间差 - 秒: {diff_seconds}")
|
|
79
|
-
print(f"时间差 - 分钟: {diff_minutes}")
|
|
80
|
-
print(f"时间差 - 小时: {diff_hours}")
|
|
81
|
-
|
|
82
|
-
print()
|
|
83
|
-
|
|
84
|
-
# 日期加减
|
|
85
|
-
base_date = "2025-09-10"
|
|
86
|
-
plus_30_days = TimeUtils.add_days(base_date, 30)
|
|
87
|
-
minus_15_days = TimeUtils.add_days(base_date, -15)
|
|
88
|
-
plus_3_months = TimeUtils.add_months(base_date, 3)
|
|
89
|
-
|
|
90
|
-
print(f"基础日期: {base_date}")
|
|
91
|
-
print(f"加30天: {plus_30_days}")
|
|
92
|
-
print(f"减15天: {minus_15_days}")
|
|
93
|
-
print(f"加3个月: {plus_3_months}")
|
|
94
|
-
|
|
95
|
-
print()
|
|
96
|
-
|
|
97
|
-
|
|
98
|
-
def demo_timezone_conversion():
|
|
99
|
-
"""演示时区转换功能"""
|
|
100
|
-
print("=== 时区转换演示 ===\n")
|
|
101
|
-
|
|
102
|
-
# 创建一个时间
|
|
103
|
-
dt = parse_time("2025-09-10 14:30:00")
|
|
104
|
-
print(f"原始时间: {dt}")
|
|
105
|
-
|
|
106
|
-
# 转换为UTC时区
|
|
107
|
-
utc_time = to_utc(dt)
|
|
108
|
-
print(f"UTC时间: {utc_time}")
|
|
109
|
-
|
|
110
|
-
# 转换为本地时区
|
|
111
|
-
local_time = to_local(dt)
|
|
112
|
-
print(f"本地时间: {local_time}")
|
|
113
|
-
|
|
114
|
-
# 转换为纽约时区
|
|
115
|
-
ny_time = to_timezone(dt, "America/New_York")
|
|
116
|
-
print(f"纽约时间: {ny_time}")
|
|
117
|
-
|
|
118
|
-
# 转换为伦敦时区
|
|
119
|
-
london_time = to_timezone(dt, "Europe/London")
|
|
120
|
-
print(f"伦敦时间: {london_time}")
|
|
121
|
-
|
|
122
|
-
print()
|
|
123
|
-
|
|
124
|
-
|
|
125
|
-
def demo_timestamp_conversion():
|
|
126
|
-
"""演示时间戳转换功能"""
|
|
127
|
-
print("=== 时间戳转换演示 ===\n")
|
|
128
|
-
|
|
129
|
-
# 获取当前时间戳
|
|
130
|
-
current_timestamp = to_timestamp(now())
|
|
131
|
-
print(f"当前时间戳: {current_timestamp}")
|
|
132
|
-
|
|
133
|
-
# 从时间戳转换为datetime
|
|
134
|
-
dt_from_timestamp = to_datetime(current_timestamp)
|
|
135
|
-
print(f"从时间戳转换: {dt_from_timestamp}")
|
|
136
|
-
|
|
137
|
-
# 从时间戳创建带时区的datetime
|
|
138
|
-
dt_with_tz = to_timezone(to_datetime(current_timestamp), "Asia/Shanghai")
|
|
139
|
-
print(f"带时区的时间: {dt_with_tz}")
|
|
140
|
-
|
|
141
|
-
print()
|
|
142
|
-
|
|
143
|
-
|
|
144
|
-
def demo_in_spider():
|
|
145
|
-
"""演示在爬虫中使用日期工具"""
|
|
146
|
-
print("=== 在爬虫中使用日期工具 ===\n")
|
|
147
|
-
print("在爬虫项目中,您可以这样使用日期工具:")
|
|
148
|
-
print("""
|
|
149
|
-
from crawlo import Spider
|
|
150
|
-
from crawlo.tools import parse_time, format_time
|
|
151
|
-
|
|
152
|
-
class NewsSpider(Spider):
|
|
153
|
-
def parse(self, response):
|
|
154
|
-
# 提取文章发布时间
|
|
155
|
-
publish_time_str = response.css('.publish-time::text').get()
|
|
156
|
-
|
|
157
|
-
# 解析发布时间
|
|
158
|
-
publish_time = parse_time(publish_time_str)
|
|
159
|
-
|
|
160
|
-
# 格式化时间为标准格式
|
|
161
|
-
formatted_time = format_time(publish_time, "%Y-%m-%d %H:%M:%S")
|
|
162
|
-
|
|
163
|
-
# 计算文章发布多久了
|
|
164
|
-
current_time = self.tools.now()
|
|
165
|
-
hours_since_publish = self.tools.time_diff(publish_time, current_time, "hours")
|
|
166
|
-
|
|
167
|
-
# 根据发布时间决定是否继续处理
|
|
168
|
-
if hours_since_publish < 24: # 只处理24小时内的文章
|
|
169
|
-
# 处理文章...
|
|
170
|
-
pass
|
|
171
|
-
""")
|
|
172
|
-
|
|
173
|
-
|
|
174
|
-
if __name__ == '__main__':
|
|
175
|
-
# 运行演示
|
|
176
|
-
demo_date_parsing()
|
|
177
|
-
demo_date_formatting()
|
|
178
|
-
demo_time_calculations()
|
|
179
|
-
demo_timezone_conversion()
|
|
180
|
-
demo_timestamp_conversion()
|
|
1
|
+
#!/usr/bin/python
|
|
2
|
+
# -*- coding: UTF-8 -*-
|
|
3
|
+
"""
|
|
4
|
+
Crawlo框架日期工具使用示例
|
|
5
|
+
"""
|
|
6
|
+
from crawlo.tools import (
|
|
7
|
+
TimeUtils,
|
|
8
|
+
parse_time,
|
|
9
|
+
format_time,
|
|
10
|
+
time_diff,
|
|
11
|
+
to_timestamp,
|
|
12
|
+
to_datetime,
|
|
13
|
+
now,
|
|
14
|
+
to_timezone,
|
|
15
|
+
to_utc,
|
|
16
|
+
to_local
|
|
17
|
+
)
|
|
18
|
+
|
|
19
|
+
|
|
20
|
+
def demo_date_parsing():
|
|
21
|
+
"""演示日期解析功能"""
|
|
22
|
+
print("=== 日期解析演示 ===\n")
|
|
23
|
+
|
|
24
|
+
# 解析各种格式的日期字符串
|
|
25
|
+
date_strings = [
|
|
26
|
+
"2025-09-10 14:30:00",
|
|
27
|
+
"September 10, 2025 2:30 PM",
|
|
28
|
+
"2025/09/10 14:30:00",
|
|
29
|
+
"10-09-2025 14:30:00",
|
|
30
|
+
"2025年9月10日 14时30分00秒"
|
|
31
|
+
]
|
|
32
|
+
|
|
33
|
+
for date_str in date_strings:
|
|
34
|
+
parsed = parse_time(date_str)
|
|
35
|
+
print(f"解析 '{date_str}' -> {parsed}")
|
|
36
|
+
|
|
37
|
+
print()
|
|
38
|
+
|
|
39
|
+
|
|
40
|
+
def demo_date_formatting():
|
|
41
|
+
"""演示日期格式化功能"""
|
|
42
|
+
print("=== 日期格式化演示 ===\n")
|
|
43
|
+
|
|
44
|
+
# 获取当前时间
|
|
45
|
+
current_time = now()
|
|
46
|
+
print(f"当前时间: {current_time}")
|
|
47
|
+
|
|
48
|
+
# 使用不同格式进行格式化
|
|
49
|
+
formats = [
|
|
50
|
+
"%Y-%m-%d",
|
|
51
|
+
"%Y年%m月%d日",
|
|
52
|
+
"%Y-%m-%d %H:%M:%S",
|
|
53
|
+
"%B %d, %Y at %I:%M %p",
|
|
54
|
+
"%A, %B %d, %Y"
|
|
55
|
+
]
|
|
56
|
+
|
|
57
|
+
for fmt in formats:
|
|
58
|
+
formatted = format_time(current_time, fmt)
|
|
59
|
+
print(f"格式化为 '{fmt}' -> {formatted}")
|
|
60
|
+
|
|
61
|
+
print()
|
|
62
|
+
|
|
63
|
+
|
|
64
|
+
def demo_time_calculations():
|
|
65
|
+
"""演示时间计算功能"""
|
|
66
|
+
print("=== 时间计算演示 ===\n")
|
|
67
|
+
|
|
68
|
+
# 计算时间差
|
|
69
|
+
start_time = "2025-09-10 10:00:00"
|
|
70
|
+
end_time = "2025-09-10 14:30:45"
|
|
71
|
+
|
|
72
|
+
diff_seconds = time_diff(start_time, end_time, "seconds")
|
|
73
|
+
diff_minutes = time_diff(start_time, end_time, "minutes")
|
|
74
|
+
diff_hours = time_diff(start_time, end_time, "hours")
|
|
75
|
+
|
|
76
|
+
print(f"起始时间: {start_time}")
|
|
77
|
+
print(f"结束时间: {end_time}")
|
|
78
|
+
print(f"时间差 - 秒: {diff_seconds}")
|
|
79
|
+
print(f"时间差 - 分钟: {diff_minutes}")
|
|
80
|
+
print(f"时间差 - 小时: {diff_hours}")
|
|
81
|
+
|
|
82
|
+
print()
|
|
83
|
+
|
|
84
|
+
# 日期加减
|
|
85
|
+
base_date = "2025-09-10"
|
|
86
|
+
plus_30_days = TimeUtils.add_days(base_date, 30)
|
|
87
|
+
minus_15_days = TimeUtils.add_days(base_date, -15)
|
|
88
|
+
plus_3_months = TimeUtils.add_months(base_date, 3)
|
|
89
|
+
|
|
90
|
+
print(f"基础日期: {base_date}")
|
|
91
|
+
print(f"加30天: {plus_30_days}")
|
|
92
|
+
print(f"减15天: {minus_15_days}")
|
|
93
|
+
print(f"加3个月: {plus_3_months}")
|
|
94
|
+
|
|
95
|
+
print()
|
|
96
|
+
|
|
97
|
+
|
|
98
|
+
def demo_timezone_conversion():
|
|
99
|
+
"""演示时区转换功能"""
|
|
100
|
+
print("=== 时区转换演示 ===\n")
|
|
101
|
+
|
|
102
|
+
# 创建一个时间
|
|
103
|
+
dt = parse_time("2025-09-10 14:30:00")
|
|
104
|
+
print(f"原始时间: {dt}")
|
|
105
|
+
|
|
106
|
+
# 转换为UTC时区
|
|
107
|
+
utc_time = to_utc(dt)
|
|
108
|
+
print(f"UTC时间: {utc_time}")
|
|
109
|
+
|
|
110
|
+
# 转换为本地时区
|
|
111
|
+
local_time = to_local(dt)
|
|
112
|
+
print(f"本地时间: {local_time}")
|
|
113
|
+
|
|
114
|
+
# 转换为纽约时区
|
|
115
|
+
ny_time = to_timezone(dt, "America/New_York")
|
|
116
|
+
print(f"纽约时间: {ny_time}")
|
|
117
|
+
|
|
118
|
+
# 转换为伦敦时区
|
|
119
|
+
london_time = to_timezone(dt, "Europe/London")
|
|
120
|
+
print(f"伦敦时间: {london_time}")
|
|
121
|
+
|
|
122
|
+
print()
|
|
123
|
+
|
|
124
|
+
|
|
125
|
+
def demo_timestamp_conversion():
|
|
126
|
+
"""演示时间戳转换功能"""
|
|
127
|
+
print("=== 时间戳转换演示 ===\n")
|
|
128
|
+
|
|
129
|
+
# 获取当前时间戳
|
|
130
|
+
current_timestamp = to_timestamp(now())
|
|
131
|
+
print(f"当前时间戳: {current_timestamp}")
|
|
132
|
+
|
|
133
|
+
# 从时间戳转换为datetime
|
|
134
|
+
dt_from_timestamp = to_datetime(current_timestamp)
|
|
135
|
+
print(f"从时间戳转换: {dt_from_timestamp}")
|
|
136
|
+
|
|
137
|
+
# 从时间戳创建带时区的datetime
|
|
138
|
+
dt_with_tz = to_timezone(to_datetime(current_timestamp), "Asia/Shanghai")
|
|
139
|
+
print(f"带时区的时间: {dt_with_tz}")
|
|
140
|
+
|
|
141
|
+
print()
|
|
142
|
+
|
|
143
|
+
|
|
144
|
+
def demo_in_spider():
|
|
145
|
+
"""演示在爬虫中使用日期工具"""
|
|
146
|
+
print("=== 在爬虫中使用日期工具 ===\n")
|
|
147
|
+
print("在爬虫项目中,您可以这样使用日期工具:")
|
|
148
|
+
print("""
|
|
149
|
+
from crawlo import Spider
|
|
150
|
+
from crawlo.tools import parse_time, format_time
|
|
151
|
+
|
|
152
|
+
class NewsSpider(Spider):
|
|
153
|
+
def parse(self, response):
|
|
154
|
+
# 提取文章发布时间
|
|
155
|
+
publish_time_str = response.css('.publish-time::text').get()
|
|
156
|
+
|
|
157
|
+
# 解析发布时间
|
|
158
|
+
publish_time = parse_time(publish_time_str)
|
|
159
|
+
|
|
160
|
+
# 格式化时间为标准格式
|
|
161
|
+
formatted_time = format_time(publish_time, "%Y-%m-%d %H:%M:%S")
|
|
162
|
+
|
|
163
|
+
# 计算文章发布多久了
|
|
164
|
+
current_time = self.tools.now()
|
|
165
|
+
hours_since_publish = self.tools.time_diff(publish_time, current_time, "hours")
|
|
166
|
+
|
|
167
|
+
# 根据发布时间决定是否继续处理
|
|
168
|
+
if hours_since_publish < 24: # 只处理24小时内的文章
|
|
169
|
+
# 处理文章...
|
|
170
|
+
pass
|
|
171
|
+
""")
|
|
172
|
+
|
|
173
|
+
|
|
174
|
+
if __name__ == '__main__':
|
|
175
|
+
# 运行演示
|
|
176
|
+
demo_date_parsing()
|
|
177
|
+
demo_date_formatting()
|
|
178
|
+
demo_time_calculations()
|
|
179
|
+
demo_timezone_conversion()
|
|
180
|
+
demo_timestamp_conversion()
|
|
181
181
|
demo_in_spider()
|
tests/debug_configure.py
CHANGED
|
@@ -1,70 +1,70 @@
|
|
|
1
|
-
#!/usr/bin/env python
|
|
2
|
-
# -*- coding: UTF-8 -*-
|
|
3
|
-
"""
|
|
4
|
-
逐步调试LoggerManager.configure
|
|
5
|
-
"""
|
|
6
|
-
import sys
|
|
7
|
-
import os
|
|
8
|
-
sys.path.insert(0, '/')
|
|
9
|
-
|
|
10
|
-
from crawlo.utils.log import LoggerManager
|
|
11
|
-
|
|
12
|
-
print("=== 逐步调试LoggerManager.configure ===")
|
|
13
|
-
|
|
14
|
-
print("1. 检查初始状态")
|
|
15
|
-
print(f" _early_initialized: {LoggerManager._early_initialized}")
|
|
16
|
-
print(f" _configured: {LoggerManager._configured}")
|
|
17
|
-
|
|
18
|
-
print("2. 调用early_initialize")
|
|
19
|
-
LoggerManager.early_initialize()
|
|
20
|
-
print(f" _early_initialized: {LoggerManager._early_initialized}")
|
|
21
|
-
|
|
22
|
-
print("3. 进入configure - 开始")
|
|
23
|
-
# 模拟configure方法的各个步骤
|
|
24
|
-
with LoggerManager._config_lock:
|
|
25
|
-
print(" 获得锁")
|
|
26
|
-
|
|
27
|
-
if LoggerManager._configured:
|
|
28
|
-
print(" 已配置,直接返回")
|
|
29
|
-
else:
|
|
30
|
-
print(" 开始配置...")
|
|
31
|
-
|
|
32
|
-
# 更新状态
|
|
33
|
-
print(" 设置_log_state")
|
|
34
|
-
from crawlo.utils.log import _log_state
|
|
35
|
-
_log_state['current_step'] = 'basic_setup'
|
|
36
|
-
|
|
37
|
-
print(" 处理参数")
|
|
38
|
-
kwargs = {'LOG_LEVEL': 'INFO', 'LOG_FILE': 'test.log'}
|
|
39
|
-
get_val = lambda k, d=None: kwargs.get(k, d)
|
|
40
|
-
|
|
41
|
-
filename = get_val('LOG_FILE')
|
|
42
|
-
level = get_val('LOG_LEVEL', None)
|
|
43
|
-
if level is None:
|
|
44
|
-
level = 'INFO'
|
|
45
|
-
|
|
46
|
-
print(f" filename: {filename}")
|
|
47
|
-
print(f" level: {level}")
|
|
48
|
-
|
|
49
|
-
print(" 设置默认值")
|
|
50
|
-
LoggerManager._default_filename = filename
|
|
51
|
-
LoggerManager._default_level = LoggerManager._to_level(level)
|
|
52
|
-
LoggerManager._default_file_level = LoggerManager._to_level(level)
|
|
53
|
-
LoggerManager._default_console_level = LoggerManager._default_level
|
|
54
|
-
|
|
55
|
-
print(" 清空缓存")
|
|
56
|
-
LoggerManager.logger_cache.clear()
|
|
57
|
-
|
|
58
|
-
print(" 设置已配置状态")
|
|
59
|
-
LoggerManager._configured = True
|
|
60
|
-
_log_state['current_step'] = 'full_config'
|
|
61
|
-
|
|
62
|
-
print(" 配置完成")
|
|
63
|
-
|
|
64
|
-
print("4. 测试创建logger")
|
|
65
|
-
from crawlo.utils.log import get_logger
|
|
66
|
-
logger = get_logger('test')
|
|
67
|
-
print(f" Logger: {logger}")
|
|
68
|
-
print(f" Handlers: {len(logger.handlers)}")
|
|
69
|
-
|
|
1
|
+
#!/usr/bin/env python
|
|
2
|
+
# -*- coding: UTF-8 -*-
|
|
3
|
+
"""
|
|
4
|
+
逐步调试LoggerManager.configure
|
|
5
|
+
"""
|
|
6
|
+
import sys
|
|
7
|
+
import os
|
|
8
|
+
sys.path.insert(0, '/')
|
|
9
|
+
|
|
10
|
+
from crawlo.utils.log import LoggerManager
|
|
11
|
+
|
|
12
|
+
print("=== 逐步调试LoggerManager.configure ===")
|
|
13
|
+
|
|
14
|
+
print("1. 检查初始状态")
|
|
15
|
+
print(f" _early_initialized: {LoggerManager._early_initialized}")
|
|
16
|
+
print(f" _configured: {LoggerManager._configured}")
|
|
17
|
+
|
|
18
|
+
print("2. 调用early_initialize")
|
|
19
|
+
LoggerManager.early_initialize()
|
|
20
|
+
print(f" _early_initialized: {LoggerManager._early_initialized}")
|
|
21
|
+
|
|
22
|
+
print("3. 进入configure - 开始")
|
|
23
|
+
# 模拟configure方法的各个步骤
|
|
24
|
+
with LoggerManager._config_lock:
|
|
25
|
+
print(" 获得锁")
|
|
26
|
+
|
|
27
|
+
if LoggerManager._configured:
|
|
28
|
+
print(" 已配置,直接返回")
|
|
29
|
+
else:
|
|
30
|
+
print(" 开始配置...")
|
|
31
|
+
|
|
32
|
+
# 更新状态
|
|
33
|
+
print(" 设置_log_state")
|
|
34
|
+
from crawlo.utils.log import _log_state
|
|
35
|
+
_log_state['current_step'] = 'basic_setup'
|
|
36
|
+
|
|
37
|
+
print(" 处理参数")
|
|
38
|
+
kwargs = {'LOG_LEVEL': 'INFO', 'LOG_FILE': 'test.log'}
|
|
39
|
+
get_val = lambda k, d=None: kwargs.get(k, d)
|
|
40
|
+
|
|
41
|
+
filename = get_val('LOG_FILE')
|
|
42
|
+
level = get_val('LOG_LEVEL', None)
|
|
43
|
+
if level is None:
|
|
44
|
+
level = 'INFO'
|
|
45
|
+
|
|
46
|
+
print(f" filename: {filename}")
|
|
47
|
+
print(f" level: {level}")
|
|
48
|
+
|
|
49
|
+
print(" 设置默认值")
|
|
50
|
+
LoggerManager._default_filename = filename
|
|
51
|
+
LoggerManager._default_level = LoggerManager._to_level(level)
|
|
52
|
+
LoggerManager._default_file_level = LoggerManager._to_level(level)
|
|
53
|
+
LoggerManager._default_console_level = LoggerManager._default_level
|
|
54
|
+
|
|
55
|
+
print(" 清空缓存")
|
|
56
|
+
LoggerManager.logger_cache.clear()
|
|
57
|
+
|
|
58
|
+
print(" 设置已配置状态")
|
|
59
|
+
LoggerManager._configured = True
|
|
60
|
+
_log_state['current_step'] = 'full_config'
|
|
61
|
+
|
|
62
|
+
print(" 配置完成")
|
|
63
|
+
|
|
64
|
+
print("4. 测试创建logger")
|
|
65
|
+
from crawlo.utils.log import get_logger
|
|
66
|
+
logger = get_logger('test')
|
|
67
|
+
print(f" Logger: {logger}")
|
|
68
|
+
print(f" Handlers: {len(logger.handlers)}")
|
|
69
|
+
|
|
70
70
|
print("=== 调试完成 ===")
|
tests/debug_framework_logger.py
CHANGED
|
@@ -1,85 +1,85 @@
|
|
|
1
|
-
#!/usr/bin/env python
|
|
2
|
-
# -*- coding: UTF-8 -*-
|
|
3
|
-
"""
|
|
4
|
-
调试框架logger配置
|
|
5
|
-
"""
|
|
6
|
-
import sys
|
|
7
|
-
import os
|
|
8
|
-
sys.path.insert(0, '/')
|
|
9
|
-
|
|
10
|
-
from crawlo.initialization import initialize_framework, get_framework_initializer
|
|
11
|
-
from crawlo.utils.log import get_logger, LoggerManager
|
|
12
|
-
import logging
|
|
13
|
-
|
|
14
|
-
def debug_framework_logger():
|
|
15
|
-
print("=== 调试框架logger配置 ===")
|
|
16
|
-
|
|
17
|
-
# 1. 初始化框架,模拟ofweek_standalone的配置
|
|
18
|
-
print("1. 初始化框架...")
|
|
19
|
-
custom_settings = {
|
|
20
|
-
'LOG_LEVEL': 'INFO',
|
|
21
|
-
'LOG_FILE': 'logs/debug_framework.log',
|
|
22
|
-
'PROJECT_NAME': 'debug_test',
|
|
23
|
-
'RUN_MODE': 'standalone'
|
|
24
|
-
}
|
|
25
|
-
|
|
26
|
-
# 确保日志目录存在
|
|
27
|
-
os.makedirs('../logs', exist_ok=True)
|
|
28
|
-
|
|
29
|
-
settings = initialize_framework(custom_settings)
|
|
30
|
-
print(f" LOG_LEVEL: {settings.get('LOG_LEVEL')}")
|
|
31
|
-
print(f" LOG_FILE: {settings.get('LOG_FILE')}")
|
|
32
|
-
|
|
33
|
-
# 2. 获取框架初始化管理器
|
|
34
|
-
init_manager = get_framework_initializer()
|
|
35
|
-
print(f" 框架是否就绪: {init_manager.is_ready}")
|
|
36
|
-
print(f" 初始化阶段: {init_manager.phase}")
|
|
37
|
-
|
|
38
|
-
# 3. 测试框架logger
|
|
39
|
-
framework_logger = init_manager.logger
|
|
40
|
-
print(f" 框架logger: {framework_logger}")
|
|
41
|
-
if framework_logger:
|
|
42
|
-
print(f" 名称: {framework_logger.name}")
|
|
43
|
-
print(f" 级别: {framework_logger.level} ({logging.getLevelName(framework_logger.level)})")
|
|
44
|
-
print(f" 处理器数量: {len(framework_logger.handlers)}")
|
|
45
|
-
|
|
46
|
-
for i, handler in enumerate(framework_logger.handlers):
|
|
47
|
-
handler_type = type(handler).__name__
|
|
48
|
-
handler_level = handler.level
|
|
49
|
-
print(f" 处理器{i}: {handler_type}, 级别: {handler_level} ({logging.getLevelName(handler_level)})")
|
|
50
|
-
if hasattr(handler, 'baseFilename'):
|
|
51
|
-
print(f" 文件: {handler.baseFilename}")
|
|
52
|
-
|
|
53
|
-
# 4. 手动创建一个crawlo.framework logger对比
|
|
54
|
-
manual_logger = get_logger('crawlo.framework')
|
|
55
|
-
print(f" 手动创建的logger: {manual_logger}")
|
|
56
|
-
if manual_logger:
|
|
57
|
-
print(f" 名称: {manual_logger.name}")
|
|
58
|
-
print(f" 级别: {manual_logger.level} ({logging.getLevelName(manual_logger.level)})")
|
|
59
|
-
print(f" 处理器数量: {len(manual_logger.handlers)}")
|
|
60
|
-
|
|
61
|
-
for i, handler in enumerate(manual_logger.handlers):
|
|
62
|
-
handler_type = type(handler).__name__
|
|
63
|
-
handler_level = handler.level
|
|
64
|
-
print(f" 处理器{i}: {handler_type}, 级别: {handler_level} ({logging.getLevelName(handler_level)})")
|
|
65
|
-
if hasattr(handler, 'baseFilename'):
|
|
66
|
-
print(f" 文件: {handler.baseFilename}")
|
|
67
|
-
|
|
68
|
-
# 5. 测试日志输出
|
|
69
|
-
print("2. 测试日志输出...")
|
|
70
|
-
|
|
71
|
-
if framework_logger:
|
|
72
|
-
framework_logger.info("这是框架logger测试消息 - INFO级别")
|
|
73
|
-
framework_logger.debug("这是框架logger测试消息 - DEBUG级别")
|
|
74
|
-
|
|
75
|
-
if manual_logger:
|
|
76
|
-
manual_logger.info("这是手动logger测试消息 - INFO级别")
|
|
77
|
-
manual_logger.debug("这是手动logger测试消息 - DEBUG级别")
|
|
78
|
-
|
|
79
|
-
# 6. 检查是否同一个实例
|
|
80
|
-
print(f"3. 是否同一实例: {framework_logger is manual_logger}")
|
|
81
|
-
|
|
82
|
-
print("=== 调试完成 ===")
|
|
83
|
-
|
|
84
|
-
if __name__ == "__main__":
|
|
1
|
+
#!/usr/bin/env python
|
|
2
|
+
# -*- coding: UTF-8 -*-
|
|
3
|
+
"""
|
|
4
|
+
调试框架logger配置
|
|
5
|
+
"""
|
|
6
|
+
import sys
|
|
7
|
+
import os
|
|
8
|
+
sys.path.insert(0, '/')
|
|
9
|
+
|
|
10
|
+
from crawlo.initialization import initialize_framework, get_framework_initializer
|
|
11
|
+
from crawlo.utils.log import get_logger, LoggerManager
|
|
12
|
+
import logging
|
|
13
|
+
|
|
14
|
+
def debug_framework_logger():
|
|
15
|
+
print("=== 调试框架logger配置 ===")
|
|
16
|
+
|
|
17
|
+
# 1. 初始化框架,模拟ofweek_standalone的配置
|
|
18
|
+
print("1. 初始化框架...")
|
|
19
|
+
custom_settings = {
|
|
20
|
+
'LOG_LEVEL': 'INFO',
|
|
21
|
+
'LOG_FILE': 'logs/debug_framework.log',
|
|
22
|
+
'PROJECT_NAME': 'debug_test',
|
|
23
|
+
'RUN_MODE': 'standalone'
|
|
24
|
+
}
|
|
25
|
+
|
|
26
|
+
# 确保日志目录存在
|
|
27
|
+
os.makedirs('../logs', exist_ok=True)
|
|
28
|
+
|
|
29
|
+
settings = initialize_framework(custom_settings)
|
|
30
|
+
print(f" LOG_LEVEL: {settings.get('LOG_LEVEL')}")
|
|
31
|
+
print(f" LOG_FILE: {settings.get('LOG_FILE')}")
|
|
32
|
+
|
|
33
|
+
# 2. 获取框架初始化管理器
|
|
34
|
+
init_manager = get_framework_initializer()
|
|
35
|
+
print(f" 框架是否就绪: {init_manager.is_ready}")
|
|
36
|
+
print(f" 初始化阶段: {init_manager.phase}")
|
|
37
|
+
|
|
38
|
+
# 3. 测试框架logger
|
|
39
|
+
framework_logger = init_manager.logger
|
|
40
|
+
print(f" 框架logger: {framework_logger}")
|
|
41
|
+
if framework_logger:
|
|
42
|
+
print(f" 名称: {framework_logger.name}")
|
|
43
|
+
print(f" 级别: {framework_logger.level} ({logging.getLevelName(framework_logger.level)})")
|
|
44
|
+
print(f" 处理器数量: {len(framework_logger.handlers)}")
|
|
45
|
+
|
|
46
|
+
for i, handler in enumerate(framework_logger.handlers):
|
|
47
|
+
handler_type = type(handler).__name__
|
|
48
|
+
handler_level = handler.level
|
|
49
|
+
print(f" 处理器{i}: {handler_type}, 级别: {handler_level} ({logging.getLevelName(handler_level)})")
|
|
50
|
+
if hasattr(handler, 'baseFilename'):
|
|
51
|
+
print(f" 文件: {handler.baseFilename}")
|
|
52
|
+
|
|
53
|
+
# 4. 手动创建一个crawlo.framework logger对比
|
|
54
|
+
manual_logger = get_logger('crawlo.framework')
|
|
55
|
+
print(f" 手动创建的logger: {manual_logger}")
|
|
56
|
+
if manual_logger:
|
|
57
|
+
print(f" 名称: {manual_logger.name}")
|
|
58
|
+
print(f" 级别: {manual_logger.level} ({logging.getLevelName(manual_logger.level)})")
|
|
59
|
+
print(f" 处理器数量: {len(manual_logger.handlers)}")
|
|
60
|
+
|
|
61
|
+
for i, handler in enumerate(manual_logger.handlers):
|
|
62
|
+
handler_type = type(handler).__name__
|
|
63
|
+
handler_level = handler.level
|
|
64
|
+
print(f" 处理器{i}: {handler_type}, 级别: {handler_level} ({logging.getLevelName(handler_level)})")
|
|
65
|
+
if hasattr(handler, 'baseFilename'):
|
|
66
|
+
print(f" 文件: {handler.baseFilename}")
|
|
67
|
+
|
|
68
|
+
# 5. 测试日志输出
|
|
69
|
+
print("2. 测试日志输出...")
|
|
70
|
+
|
|
71
|
+
if framework_logger:
|
|
72
|
+
framework_logger.info("这是框架logger测试消息 - INFO级别")
|
|
73
|
+
framework_logger.debug("这是框架logger测试消息 - DEBUG级别")
|
|
74
|
+
|
|
75
|
+
if manual_logger:
|
|
76
|
+
manual_logger.info("这是手动logger测试消息 - INFO级别")
|
|
77
|
+
manual_logger.debug("这是手动logger测试消息 - DEBUG级别")
|
|
78
|
+
|
|
79
|
+
# 6. 检查是否同一个实例
|
|
80
|
+
print(f"3. 是否同一实例: {framework_logger is manual_logger}")
|
|
81
|
+
|
|
82
|
+
print("=== 调试完成 ===")
|
|
83
|
+
|
|
84
|
+
if __name__ == "__main__":
|
|
85
85
|
debug_framework_logger()
|