aio-scrapy 2.1.4__tar.gz → 2.1.7__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/LICENSE +1 -1
- aio_scrapy-2.1.7/PKG-INFO +147 -0
- aio_scrapy-2.1.7/README.md +69 -0
- aio_scrapy-2.1.7/aio_scrapy.egg-info/PKG-INFO +147 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aio_scrapy.egg-info/SOURCES.txt +6 -5
- aio_scrapy-2.1.7/aioscrapy/VERSION +1 -0
- aio_scrapy-2.1.7/aioscrapy/cmdline.py +603 -0
- aio_scrapy-2.1.7/aioscrapy/core/downloader/__init__.py +758 -0
- aio_scrapy-2.1.7/aioscrapy/core/downloader/handlers/__init__.py +267 -0
- aio_scrapy-2.1.7/aioscrapy/core/downloader/handlers/aiohttp.py +289 -0
- aio_scrapy-2.1.7/aioscrapy/core/downloader/handlers/curl_cffi.py +188 -0
- aio_scrapy-2.1.7/aioscrapy/core/downloader/handlers/httpx.py +214 -0
- aio_scrapy-2.1.7/aioscrapy/core/downloader/handlers/pyhttpx.py +200 -0
- aio_scrapy-2.1.7/aioscrapy/core/downloader/handlers/requests.py +181 -0
- aio_scrapy-2.1.7/aioscrapy/core/downloader/handlers/webdriver/__init__.py +2 -0
- aio_scrapy-2.1.7/aioscrapy/core/downloader/handlers/webdriver/drissionpage.py +493 -0
- aio_scrapy-2.1.7/aioscrapy/core/downloader/handlers/webdriver/driverpool.py +234 -0
- aio_scrapy-2.1.7/aioscrapy/core/downloader/handlers/webdriver/playwright.py +498 -0
- aio_scrapy-2.1.7/aioscrapy/core/engine.py +618 -0
- aio_scrapy-2.1.7/aioscrapy/core/scheduler.py +510 -0
- aio_scrapy-2.1.7/aioscrapy/core/scraper.py +714 -0
- aio_scrapy-2.1.7/aioscrapy/crawler.py +642 -0
- aio_scrapy-2.1.7/aioscrapy/db/__init__.py +221 -0
- aio_scrapy-2.1.7/aioscrapy/db/absmanager.py +253 -0
- aio_scrapy-2.1.7/aioscrapy/db/aiomongo.py +374 -0
- aio_scrapy-2.1.7/aioscrapy/db/aiomysql.py +477 -0
- aio_scrapy-2.1.7/aioscrapy/db/aiopg.py +407 -0
- aio_scrapy-2.1.7/aioscrapy/db/aiorabbitmq.py +611 -0
- aio_scrapy-2.1.7/aioscrapy/db/aioredis.py +343 -0
- aio_scrapy-2.1.7/aioscrapy/dupefilters/__init__.py +152 -0
- aio_scrapy-2.1.7/aioscrapy/dupefilters/disk.py +166 -0
- aio_scrapy-2.1.7/aioscrapy/dupefilters/redis.py +763 -0
- aio_scrapy-2.1.7/aioscrapy/exceptions.py +254 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/http/__init__.py +1 -1
- aio_scrapy-2.1.7/aioscrapy/http/headers.py +291 -0
- aio_scrapy-2.1.7/aioscrapy/http/request/__init__.py +444 -0
- aio_scrapy-2.1.7/aioscrapy/http/request/form.py +119 -0
- aio_scrapy-2.1.7/aioscrapy/http/request/json_request.py +175 -0
- aio_scrapy-2.1.7/aioscrapy/http/response/__init__.py +481 -0
- aio_scrapy-2.1.7/aioscrapy/http/response/html.py +51 -0
- aio_scrapy-2.1.7/aioscrapy/http/response/text.py +712 -0
- aio_scrapy-2.1.7/aioscrapy/http/response/web_driver.py +144 -0
- aio_scrapy-2.1.7/aioscrapy/http/response/xml.py +54 -0
- aio_scrapy-2.1.7/aioscrapy/libs/downloader/defaultheaders.py +86 -0
- aio_scrapy-2.1.7/aioscrapy/libs/downloader/downloadtimeout.py +115 -0
- aio_scrapy-2.1.7/aioscrapy/libs/downloader/ja3fingerprint.py +129 -0
- aio_scrapy-2.1.7/aioscrapy/libs/downloader/retry.py +304 -0
- aio_scrapy-2.1.7/aioscrapy/libs/downloader/stats.py +173 -0
- aio_scrapy-2.1.7/aioscrapy/libs/downloader/useragent.py +114 -0
- aio_scrapy-2.1.7/aioscrapy/libs/extensions/closespider.py +233 -0
- aio_scrapy-2.1.7/aioscrapy/libs/extensions/corestats.py +196 -0
- aio_scrapy-2.1.7/aioscrapy/libs/extensions/logstats.py +194 -0
- aio_scrapy-2.1.7/aioscrapy/libs/extensions/metric.py +516 -0
- aio_scrapy-2.1.7/aioscrapy/libs/extensions/throttle.py +322 -0
- aio_scrapy-2.1.7/aioscrapy/libs/pipelines/__init__.py +480 -0
- aio_scrapy-2.1.7/aioscrapy/libs/pipelines/csv.py +328 -0
- aio_scrapy-2.1.7/aioscrapy/libs/pipelines/excel.py +545 -0
- aio_scrapy-2.1.7/aioscrapy/libs/pipelines/mongo.py +188 -0
- aio_scrapy-2.1.7/aioscrapy/libs/pipelines/mysql.py +93 -0
- aio_scrapy-2.1.7/aioscrapy/libs/pipelines/pg.py +93 -0
- aio_scrapy-2.1.7/aioscrapy/libs/spider/depth.py +192 -0
- aio_scrapy-2.1.7/aioscrapy/libs/spider/httperror.py +194 -0
- aio_scrapy-2.1.7/aioscrapy/libs/spider/offsite.py +283 -0
- aio_scrapy-2.1.7/aioscrapy/libs/spider/referer.py +739 -0
- aio_scrapy-2.1.7/aioscrapy/libs/spider/urllength.py +133 -0
- aio_scrapy-2.1.7/aioscrapy/link.py +160 -0
- aio_scrapy-2.1.7/aioscrapy/logformatter.py +285 -0
- aio_scrapy-2.1.7/aioscrapy/middleware/absmanager.py +416 -0
- aio_scrapy-2.1.7/aioscrapy/middleware/downloader.py +287 -0
- aio_scrapy-2.1.7/aioscrapy/middleware/extension.py +65 -0
- aio_scrapy-2.1.7/aioscrapy/middleware/itempipeline.py +114 -0
- aio_scrapy-2.1.7/aioscrapy/middleware/spider.py +485 -0
- aio_scrapy-2.1.7/aioscrapy/process.py +252 -0
- aio_scrapy-2.1.7/aioscrapy/proxy/__init__.py +189 -0
- aio_scrapy-2.1.7/aioscrapy/proxy/redis.py +206 -0
- aio_scrapy-2.1.7/aioscrapy/queue/__init__.py +219 -0
- aio_scrapy-2.1.7/aioscrapy/scrapyd/runner.py +185 -0
- aio_scrapy-2.1.7/aioscrapy/serializer.py +219 -0
- aio_scrapy-2.1.7/aioscrapy/settings/__init__.py +948 -0
- aio_scrapy-2.1.7/aioscrapy/settings/default_settings.py +483 -0
- aio_scrapy-2.1.7/aioscrapy/signalmanager.py +199 -0
- aio_scrapy-2.1.7/aioscrapy/signals.py +206 -0
- aio_scrapy-2.1.7/aioscrapy/spiderloader.py +261 -0
- aio_scrapy-2.1.7/aioscrapy/spiders/__init__.py +350 -0
- aio_scrapy-2.1.7/aioscrapy/statscollectors.py +391 -0
- aio_scrapy-2.1.7/aioscrapy/utils/conf.py +523 -0
- aio_scrapy-2.1.7/aioscrapy/utils/curl.py +258 -0
- aio_scrapy-2.1.7/aioscrapy/utils/decorators.py +95 -0
- aio_scrapy-2.1.7/aioscrapy/utils/deprecate.py +334 -0
- aio_scrapy-2.1.7/aioscrapy/utils/httpobj.py +71 -0
- aio_scrapy-2.1.7/aioscrapy/utils/log.py +130 -0
- aio_scrapy-2.1.7/aioscrapy/utils/misc.py +270 -0
- aio_scrapy-2.1.7/aioscrapy/utils/ossignal.py +85 -0
- aio_scrapy-2.1.7/aioscrapy/utils/project.py +251 -0
- aio_scrapy-2.1.7/aioscrapy/utils/python.py +358 -0
- aio_scrapy-2.1.7/aioscrapy/utils/reqser.py +89 -0
- aio_scrapy-2.1.7/aioscrapy/utils/request.py +228 -0
- aio_scrapy-2.1.7/aioscrapy/utils/response.py +123 -0
- aio_scrapy-2.1.7/aioscrapy/utils/signal.py +243 -0
- aio_scrapy-2.1.7/aioscrapy/utils/spider.py +67 -0
- aio_scrapy-2.1.7/aioscrapy/utils/template.py +123 -0
- aio_scrapy-2.1.7/aioscrapy/utils/tools.py +252 -0
- aio_scrapy-2.1.7/aioscrapy/utils/trackref.py +253 -0
- aio_scrapy-2.1.7/aioscrapy/utils/url.py +469 -0
- aio-scrapy-2.1.4/PKG-INFO +0 -239
- aio-scrapy-2.1.4/README.md +0 -173
- aio-scrapy-2.1.4/aio_scrapy.egg-info/PKG-INFO +0 -239
- aio-scrapy-2.1.4/aioscrapy/VERSION +0 -1
- aio-scrapy-2.1.4/aioscrapy/cmdline.py +0 -170
- aio-scrapy-2.1.4/aioscrapy/core/downloader/__init__.py +0 -253
- aio-scrapy-2.1.4/aioscrapy/core/downloader/handlers/__init__.py +0 -85
- aio-scrapy-2.1.4/aioscrapy/core/downloader/handlers/aiohttp.py +0 -105
- aio-scrapy-2.1.4/aioscrapy/core/downloader/handlers/curl_cffi.py +0 -67
- aio-scrapy-2.1.4/aioscrapy/core/downloader/handlers/httpx.py +0 -84
- aio-scrapy-2.1.4/aioscrapy/core/downloader/handlers/playwright/__init__.py +0 -115
- aio-scrapy-2.1.4/aioscrapy/core/downloader/handlers/playwright/driverpool.py +0 -59
- aio-scrapy-2.1.4/aioscrapy/core/downloader/handlers/playwright/webdriver.py +0 -96
- aio-scrapy-2.1.4/aioscrapy/core/downloader/handlers/pyhttpx.py +0 -68
- aio-scrapy-2.1.4/aioscrapy/core/downloader/handlers/requests.py +0 -63
- aio-scrapy-2.1.4/aioscrapy/core/engine.py +0 -257
- aio-scrapy-2.1.4/aioscrapy/core/scheduler.py +0 -196
- aio-scrapy-2.1.4/aioscrapy/core/scraper.py +0 -238
- aio-scrapy-2.1.4/aioscrapy/crawler.py +0 -260
- aio-scrapy-2.1.4/aioscrapy/db/__init__.py +0 -72
- aio-scrapy-2.1.4/aioscrapy/db/absmanager.py +0 -47
- aio-scrapy-2.1.4/aioscrapy/db/aiomongo.py +0 -92
- aio-scrapy-2.1.4/aioscrapy/db/aiomysql.py +0 -124
- aio-scrapy-2.1.4/aioscrapy/db/aiopg.py +0 -110
- aio-scrapy-2.1.4/aioscrapy/db/aiorabbitmq.py +0 -171
- aio-scrapy-2.1.4/aioscrapy/db/aioredis.py +0 -94
- aio-scrapy-2.1.4/aioscrapy/dupefilters/__init__.py +0 -47
- aio-scrapy-2.1.4/aioscrapy/dupefilters/disk.py +0 -44
- aio-scrapy-2.1.4/aioscrapy/dupefilters/redis.py +0 -197
- aio-scrapy-2.1.4/aioscrapy/exceptions.py +0 -116
- aio-scrapy-2.1.4/aioscrapy/http/headers.py +0 -57
- aio-scrapy-2.1.4/aioscrapy/http/request/__init__.py +0 -198
- aio-scrapy-2.1.4/aioscrapy/http/request/form.py +0 -39
- aio-scrapy-2.1.4/aioscrapy/http/request/json_request.py +0 -63
- aio-scrapy-2.1.4/aioscrapy/http/response/__init__.py +0 -208
- aio-scrapy-2.1.4/aioscrapy/http/response/html.py +0 -12
- aio-scrapy-2.1.4/aioscrapy/http/response/playwright.py +0 -36
- aio-scrapy-2.1.4/aioscrapy/http/response/text.py +0 -265
- aio-scrapy-2.1.4/aioscrapy/http/response/xml.py +0 -12
- aio-scrapy-2.1.4/aioscrapy/libs/downloader/defaultheaders.py +0 -22
- aio-scrapy-2.1.4/aioscrapy/libs/downloader/downloadtimeout.py +0 -26
- aio-scrapy-2.1.4/aioscrapy/libs/downloader/ja3fingerprint.py +0 -36
- aio-scrapy-2.1.4/aioscrapy/libs/downloader/retry.py +0 -118
- aio-scrapy-2.1.4/aioscrapy/libs/downloader/stats.py +0 -31
- aio-scrapy-2.1.4/aioscrapy/libs/downloader/useragent.py +0 -23
- aio-scrapy-2.1.4/aioscrapy/libs/extensions/closespider.py +0 -71
- aio-scrapy-2.1.4/aioscrapy/libs/extensions/corestats.py +0 -46
- aio-scrapy-2.1.4/aioscrapy/libs/extensions/logstats.py +0 -50
- aio-scrapy-2.1.4/aioscrapy/libs/extensions/metric.py +0 -147
- aio-scrapy-2.1.4/aioscrapy/libs/extensions/throttle.py +0 -88
- aio-scrapy-2.1.4/aioscrapy/libs/pipelines/__init__.py +0 -137
- aio-scrapy-2.1.4/aioscrapy/libs/pipelines/csv.py +0 -86
- aio-scrapy-2.1.4/aioscrapy/libs/pipelines/execl.py +0 -169
- aio-scrapy-2.1.4/aioscrapy/libs/pipelines/mongo.py +0 -56
- aio-scrapy-2.1.4/aioscrapy/libs/pipelines/mysql.py +0 -26
- aio-scrapy-2.1.4/aioscrapy/libs/pipelines/pg.py +0 -26
- aio-scrapy-2.1.4/aioscrapy/libs/spider/depth.py +0 -54
- aio-scrapy-2.1.4/aioscrapy/libs/spider/httperror.py +0 -54
- aio-scrapy-2.1.4/aioscrapy/libs/spider/offsite.py +0 -83
- aio-scrapy-2.1.4/aioscrapy/libs/spider/referer.py +0 -364
- aio-scrapy-2.1.4/aioscrapy/libs/spider/urllength.py +0 -37
- aio-scrapy-2.1.4/aioscrapy/link.py +0 -53
- aio-scrapy-2.1.4/aioscrapy/logformatter.py +0 -94
- aio-scrapy-2.1.4/aioscrapy/middleware/absmanager.py +0 -90
- aio-scrapy-2.1.4/aioscrapy/middleware/downloader.py +0 -69
- aio-scrapy-2.1.4/aioscrapy/middleware/extension.py +0 -16
- aio-scrapy-2.1.4/aioscrapy/middleware/itempipeline.py +0 -18
- aio-scrapy-2.1.4/aioscrapy/middleware/spider.py +0 -132
- aio-scrapy-2.1.4/aioscrapy/process.py +0 -52
- aio-scrapy-2.1.4/aioscrapy/proxy/__init__.py +0 -50
- aio-scrapy-2.1.4/aioscrapy/proxy/redis.py +0 -72
- aio-scrapy-2.1.4/aioscrapy/queue/__init__.py +0 -67
- aio-scrapy-2.1.4/aioscrapy/scrapyd/runner.py +0 -64
- aio-scrapy-2.1.4/aioscrapy/serializer.py +0 -39
- aio-scrapy-2.1.4/aioscrapy/settings/__init__.py +0 -466
- aio-scrapy-2.1.4/aioscrapy/settings/default_settings.py +0 -183
- aio-scrapy-2.1.4/aioscrapy/signalmanager.py +0 -68
- aio-scrapy-2.1.4/aioscrapy/signals.py +0 -24
- aio-scrapy-2.1.4/aioscrapy/spiderloader.py +0 -108
- aio-scrapy-2.1.4/aioscrapy/spiders/__init__.py +0 -123
- aio-scrapy-2.1.4/aioscrapy/statscollectors.py +0 -80
- aio-scrapy-2.1.4/aioscrapy/utils/conf.py +0 -195
- aio-scrapy-2.1.4/aioscrapy/utils/curl.py +0 -106
- aio-scrapy-2.1.4/aioscrapy/utils/decorators.py +0 -25
- aio-scrapy-2.1.4/aioscrapy/utils/deprecate.py +0 -141
- aio-scrapy-2.1.4/aioscrapy/utils/httpobj.py +0 -19
- aio-scrapy-2.1.4/aioscrapy/utils/log.py +0 -51
- aio-scrapy-2.1.4/aioscrapy/utils/misc.py +0 -102
- aio-scrapy-2.1.4/aioscrapy/utils/ossignal.py +0 -23
- aio-scrapy-2.1.4/aioscrapy/utils/project.py +0 -89
- aio-scrapy-2.1.4/aioscrapy/utils/python.py +0 -148
- aio-scrapy-2.1.4/aioscrapy/utils/reqser.py +0 -15
- aio-scrapy-2.1.4/aioscrapy/utils/request.py +0 -67
- aio-scrapy-2.1.4/aioscrapy/utils/response.py +0 -38
- aio-scrapy-2.1.4/aioscrapy/utils/signal.py +0 -61
- aio-scrapy-2.1.4/aioscrapy/utils/spider.py +0 -20
- aio-scrapy-2.1.4/aioscrapy/utils/template.py +0 -36
- aio-scrapy-2.1.4/aioscrapy/utils/tools.py +0 -78
- aio-scrapy-2.1.4/aioscrapy/utils/trackref.py +0 -67
- aio-scrapy-2.1.4/aioscrapy/utils/url.py +0 -164
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/MANIFEST.in +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aio_scrapy.egg-info/dependency_links.txt +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aio_scrapy.egg-info/entry_points.txt +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aio_scrapy.egg-info/not-zip-safe +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aio_scrapy.egg-info/requires.txt +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aio_scrapy.egg-info/top_level.txt +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/__init__.py +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/__main__.py +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/commands/__init__.py +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/commands/crawl.py +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/commands/genspider.py +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/commands/list.py +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/commands/runspider.py +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/commands/settings.py +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/commands/startproject.py +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/commands/version.py +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/core/__init__.py +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/libs/__init__.py +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/libs/downloader/__init__.py +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/libs/extensions/__init__.py +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/libs/spider/__init__.py +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/middleware/__init__.py +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/queue/memory.py +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/queue/rabbitmq.py +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/queue/redis.py +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/scrapyd/__init__.py +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/templates/project/aioscrapy.cfg +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/templates/project/module/__init__.py +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/templates/project/module/middlewares.py.tmpl +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/templates/project/module/pipelines.py.tmpl +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/templates/project/module/settings.py.tmpl +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/templates/project/module/spiders/__init__.py +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/templates/spiders/basic.tmpl +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/templates/spiders/single.tmpl +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/utils/__init__.py +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/setup.cfg +0 -0
- {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/setup.py +0 -0
|
@@ -0,0 +1,147 @@
|
|
|
1
|
+
Metadata-Version: 2.2
|
|
2
|
+
Name: aio-scrapy
|
|
3
|
+
Version: 2.1.7
|
|
4
|
+
Summary: A high-level Web Crawling and Web Scraping framework based on Asyncio
|
|
5
|
+
Home-page: https://github.com/conlin-huang/aio-scrapy.git
|
|
6
|
+
Author: conlin
|
|
7
|
+
Author-email: 995018884@qq.com
|
|
8
|
+
License: MIT
|
|
9
|
+
Keywords: aio-scrapy,scrapy,aioscrapy,scrapy redis,asyncio,spider
|
|
10
|
+
Classifier: License :: OSI Approved :: MIT License
|
|
11
|
+
Classifier: Intended Audience :: Developers
|
|
12
|
+
Classifier: Operating System :: OS Independent
|
|
13
|
+
Classifier: Programming Language :: Python :: 3.9
|
|
14
|
+
Classifier: Programming Language :: Python :: 3.10
|
|
15
|
+
Classifier: Programming Language :: Python :: 3.11
|
|
16
|
+
Classifier: Topic :: Internet :: WWW/HTTP
|
|
17
|
+
Classifier: Topic :: Software Development :: Libraries :: Application Frameworks
|
|
18
|
+
Classifier: Topic :: Software Development :: Libraries :: Python Modules
|
|
19
|
+
Requires-Python: >=3.9
|
|
20
|
+
Description-Content-Type: text/markdown
|
|
21
|
+
License-File: LICENSE
|
|
22
|
+
Requires-Dist: aiohttp
|
|
23
|
+
Requires-Dist: ujson
|
|
24
|
+
Requires-Dist: w3lib>=1.17.0
|
|
25
|
+
Requires-Dist: parsel>=1.5.0
|
|
26
|
+
Requires-Dist: PyDispatcher>=2.0.5
|
|
27
|
+
Requires-Dist: zope.interface>=5.1.0
|
|
28
|
+
Requires-Dist: redis>=4.3.1
|
|
29
|
+
Requires-Dist: aiomultiprocess>=0.9.0
|
|
30
|
+
Requires-Dist: loguru>=0.7.0
|
|
31
|
+
Requires-Dist: anyio>=3.6.2
|
|
32
|
+
Provides-Extra: all
|
|
33
|
+
Requires-Dist: aiomysql>=0.1.1; extra == "all"
|
|
34
|
+
Requires-Dist: httpx[http2]>=0.23.0; extra == "all"
|
|
35
|
+
Requires-Dist: aio-pika>=8.1.1; extra == "all"
|
|
36
|
+
Requires-Dist: cryptography; extra == "all"
|
|
37
|
+
Requires-Dist: motor>=2.1.0; extra == "all"
|
|
38
|
+
Requires-Dist: pyhttpx>=2.10.1; extra == "all"
|
|
39
|
+
Requires-Dist: asyncpg>=0.27.0; extra == "all"
|
|
40
|
+
Requires-Dist: XlsxWriter>=3.1.2; extra == "all"
|
|
41
|
+
Requires-Dist: pillow>=9.4.0; extra == "all"
|
|
42
|
+
Requires-Dist: requests>=2.28.2; extra == "all"
|
|
43
|
+
Requires-Dist: curl_cffi; extra == "all"
|
|
44
|
+
Provides-Extra: aiomysql
|
|
45
|
+
Requires-Dist: aiomysql>=0.1.1; extra == "aiomysql"
|
|
46
|
+
Requires-Dist: cryptography; extra == "aiomysql"
|
|
47
|
+
Provides-Extra: httpx
|
|
48
|
+
Requires-Dist: httpx[http2]>=0.23.0; extra == "httpx"
|
|
49
|
+
Provides-Extra: aio-pika
|
|
50
|
+
Requires-Dist: aio-pika>=8.1.1; extra == "aio-pika"
|
|
51
|
+
Provides-Extra: mongo
|
|
52
|
+
Requires-Dist: motor>=2.1.0; extra == "mongo"
|
|
53
|
+
Provides-Extra: playwright
|
|
54
|
+
Requires-Dist: playwright>=1.31.1; extra == "playwright"
|
|
55
|
+
Provides-Extra: pyhttpx
|
|
56
|
+
Requires-Dist: pyhttpx>=2.10.4; extra == "pyhttpx"
|
|
57
|
+
Provides-Extra: curl-cffi
|
|
58
|
+
Requires-Dist: curl_cffi>=0.6.1; extra == "curl-cffi"
|
|
59
|
+
Provides-Extra: requests
|
|
60
|
+
Requires-Dist: requests>=2.28.2; extra == "requests"
|
|
61
|
+
Provides-Extra: pg
|
|
62
|
+
Requires-Dist: asyncpg>=0.27.0; extra == "pg"
|
|
63
|
+
Provides-Extra: execl
|
|
64
|
+
Requires-Dist: XlsxWriter>=3.1.2; extra == "execl"
|
|
65
|
+
Requires-Dist: pillow>=9.4.0; extra == "execl"
|
|
66
|
+
Dynamic: author
|
|
67
|
+
Dynamic: author-email
|
|
68
|
+
Dynamic: classifier
|
|
69
|
+
Dynamic: description
|
|
70
|
+
Dynamic: description-content-type
|
|
71
|
+
Dynamic: home-page
|
|
72
|
+
Dynamic: keywords
|
|
73
|
+
Dynamic: license
|
|
74
|
+
Dynamic: provides-extra
|
|
75
|
+
Dynamic: requires-dist
|
|
76
|
+
Dynamic: requires-python
|
|
77
|
+
Dynamic: summary
|
|
78
|
+
|
|
79
|
+
# AioScrapy
|
|
80
|
+
|
|
81
|
+
AioScrapy是一个基于Python异步IO的强大网络爬虫框架。它的设计理念源自Scrapy,但完全基于异步IO实现,提供更高的性能和更灵活的配置选项。</br>
|
|
82
|
+
AioScrapy is a powerful asynchronous web crawling framework built on Python's asyncio library. It is inspired by Scrapy but completely reimplemented with asynchronous IO, offering higher performance and more flexible configuration options.
|
|
83
|
+
|
|
84
|
+
## 特性 | Features
|
|
85
|
+
|
|
86
|
+
- **完全异步**:基于Python的asyncio库,实现高效的并发爬取
|
|
87
|
+
- **多种下载处理程序**:支持多种HTTP客户端,包括aiohttp、httpx、requests、pyhttpx、curl_cffi、DrissionPage和playwright
|
|
88
|
+
- **灵活的中间件系统**:轻松添加自定义功能和处理逻辑
|
|
89
|
+
- **强大的数据处理管道**:支持多种数据库存储选项
|
|
90
|
+
- **内置信号系统**:方便的事件处理机制
|
|
91
|
+
- **丰富的配置选项**:高度可定制的爬虫行为
|
|
92
|
+
- **分布式爬取**:支持使用Redis和RabbitMQ进行分布式爬取
|
|
93
|
+
- **数据库集成**:内置支持Redis、MySQL、MongoDB、PostgreSQL和RabbitMQ
|
|
94
|
+
|
|
95
|
+
|
|
96
|
+
- **Fully Asynchronous**: Built on Python's asyncio for efficient concurrent crawling
|
|
97
|
+
- **Multiple Download Handlers**: Support for various HTTP clients including aiohttp, httpx, requests, pyhttpx, curl_cffi, DrissionPage and playwright
|
|
98
|
+
- **Flexible Middleware System**: Easily add custom functionality and processing logic
|
|
99
|
+
- **Powerful Data Processing Pipelines**: Support for various database storage options
|
|
100
|
+
- **Built-in Signal System**: Convenient event handling mechanism
|
|
101
|
+
- **Rich Configuration Options**: Highly customizable crawler behavior
|
|
102
|
+
- **Distributed Crawling**: Support for distributed crawling using Redis and RabbitMQ
|
|
103
|
+
- **Database Integration**: Built-in support for Redis, MySQL, MongoDB, PostgreSQL, and RabbitMQ
|
|
104
|
+
|
|
105
|
+
## 安装 | Installation
|
|
106
|
+
|
|
107
|
+
### 要求 | Requirements
|
|
108
|
+
|
|
109
|
+
- Python 3.9+
|
|
110
|
+
|
|
111
|
+
### 使用pip安装 | Install with pip
|
|
112
|
+
|
|
113
|
+
```bash
|
|
114
|
+
pip install aio-scrapy
|
|
115
|
+
|
|
116
|
+
# Install the latest aio-scrapy
|
|
117
|
+
# pip install git+https://github.com/ConlinH/aio-scrapy
|
|
118
|
+
```
|
|
119
|
+
|
|
120
|
+
## 文档 | Documentation
|
|
121
|
+
|
|
122
|
+
## 文档目录 | Documentation Contents
|
|
123
|
+
- [安装指南 | Installation Guide](docs/installation.md)
|
|
124
|
+
- [快速入门 | Quick Start](docs/quickstart.md)
|
|
125
|
+
- [核心概念 | Core Concepts](docs/concepts.md)
|
|
126
|
+
- [爬虫指南 | Spider Guide](docs/spiders.md)
|
|
127
|
+
- [下载器 | Downloaders](docs/downloaders.md)
|
|
128
|
+
- [中间件 | Middlewares](docs/middlewares.md)
|
|
129
|
+
- [管道 | Pipelines](docs/pipelines.md)
|
|
130
|
+
- [队列 | Queues](docs/queues.md)
|
|
131
|
+
- [请求过滤器 | Request Filters](docs/dupefilters.md)
|
|
132
|
+
- [代理 | Proxy](docs/proxy.md)
|
|
133
|
+
- [数据库连接 | Database Connections](docs/databases.md)
|
|
134
|
+
- [分布式部署 | Distributed Deployment](docs/distributed.md)
|
|
135
|
+
- [配置参考 | Settings Reference](docs/settings.md)
|
|
136
|
+
- [API参考 | API Reference](docs/api.md)
|
|
137
|
+
- [示例 | Example](example)
|
|
138
|
+
|
|
139
|
+
## 许可证 | License
|
|
140
|
+
|
|
141
|
+
本项目采用MIT许可证 - 详情请查看LICENSE文件。</br>
|
|
142
|
+
This project is licensed under the MIT License - see the LICENSE file for details.
|
|
143
|
+
|
|
144
|
+
|
|
145
|
+
## 联系
|
|
146
|
+
QQ: 995018884 </br>
|
|
147
|
+
WeChat: h995018884
|
|
@@ -0,0 +1,69 @@
|
|
|
1
|
+
# AioScrapy
|
|
2
|
+
|
|
3
|
+
AioScrapy是一个基于Python异步IO的强大网络爬虫框架。它的设计理念源自Scrapy,但完全基于异步IO实现,提供更高的性能和更灵活的配置选项。</br>
|
|
4
|
+
AioScrapy is a powerful asynchronous web crawling framework built on Python's asyncio library. It is inspired by Scrapy but completely reimplemented with asynchronous IO, offering higher performance and more flexible configuration options.
|
|
5
|
+
|
|
6
|
+
## 特性 | Features
|
|
7
|
+
|
|
8
|
+
- **完全异步**:基于Python的asyncio库,实现高效的并发爬取
|
|
9
|
+
- **多种下载处理程序**:支持多种HTTP客户端,包括aiohttp、httpx、requests、pyhttpx、curl_cffi、DrissionPage和playwright
|
|
10
|
+
- **灵活的中间件系统**:轻松添加自定义功能和处理逻辑
|
|
11
|
+
- **强大的数据处理管道**:支持多种数据库存储选项
|
|
12
|
+
- **内置信号系统**:方便的事件处理机制
|
|
13
|
+
- **丰富的配置选项**:高度可定制的爬虫行为
|
|
14
|
+
- **分布式爬取**:支持使用Redis和RabbitMQ进行分布式爬取
|
|
15
|
+
- **数据库集成**:内置支持Redis、MySQL、MongoDB、PostgreSQL和RabbitMQ
|
|
16
|
+
|
|
17
|
+
|
|
18
|
+
- **Fully Asynchronous**: Built on Python's asyncio for efficient concurrent crawling
|
|
19
|
+
- **Multiple Download Handlers**: Support for various HTTP clients including aiohttp, httpx, requests, pyhttpx, curl_cffi, DrissionPage and playwright
|
|
20
|
+
- **Flexible Middleware System**: Easily add custom functionality and processing logic
|
|
21
|
+
- **Powerful Data Processing Pipelines**: Support for various database storage options
|
|
22
|
+
- **Built-in Signal System**: Convenient event handling mechanism
|
|
23
|
+
- **Rich Configuration Options**: Highly customizable crawler behavior
|
|
24
|
+
- **Distributed Crawling**: Support for distributed crawling using Redis and RabbitMQ
|
|
25
|
+
- **Database Integration**: Built-in support for Redis, MySQL, MongoDB, PostgreSQL, and RabbitMQ
|
|
26
|
+
|
|
27
|
+
## 安装 | Installation
|
|
28
|
+
|
|
29
|
+
### 要求 | Requirements
|
|
30
|
+
|
|
31
|
+
- Python 3.9+
|
|
32
|
+
|
|
33
|
+
### 使用pip安装 | Install with pip
|
|
34
|
+
|
|
35
|
+
```bash
|
|
36
|
+
pip install aio-scrapy
|
|
37
|
+
|
|
38
|
+
# Install the latest aio-scrapy
|
|
39
|
+
# pip install git+https://github.com/ConlinH/aio-scrapy
|
|
40
|
+
```
|
|
41
|
+
|
|
42
|
+
## 文档 | Documentation
|
|
43
|
+
|
|
44
|
+
## 文档目录 | Documentation Contents
|
|
45
|
+
- [安装指南 | Installation Guide](docs/installation.md)
|
|
46
|
+
- [快速入门 | Quick Start](docs/quickstart.md)
|
|
47
|
+
- [核心概念 | Core Concepts](docs/concepts.md)
|
|
48
|
+
- [爬虫指南 | Spider Guide](docs/spiders.md)
|
|
49
|
+
- [下载器 | Downloaders](docs/downloaders.md)
|
|
50
|
+
- [中间件 | Middlewares](docs/middlewares.md)
|
|
51
|
+
- [管道 | Pipelines](docs/pipelines.md)
|
|
52
|
+
- [队列 | Queues](docs/queues.md)
|
|
53
|
+
- [请求过滤器 | Request Filters](docs/dupefilters.md)
|
|
54
|
+
- [代理 | Proxy](docs/proxy.md)
|
|
55
|
+
- [数据库连接 | Database Connections](docs/databases.md)
|
|
56
|
+
- [分布式部署 | Distributed Deployment](docs/distributed.md)
|
|
57
|
+
- [配置参考 | Settings Reference](docs/settings.md)
|
|
58
|
+
- [API参考 | API Reference](docs/api.md)
|
|
59
|
+
- [示例 | Example](example)
|
|
60
|
+
|
|
61
|
+
## 许可证 | License
|
|
62
|
+
|
|
63
|
+
本项目采用MIT许可证 - 详情请查看LICENSE文件。</br>
|
|
64
|
+
This project is licensed under the MIT License - see the LICENSE file for details.
|
|
65
|
+
|
|
66
|
+
|
|
67
|
+
## 联系
|
|
68
|
+
QQ: 995018884 </br>
|
|
69
|
+
WeChat: h995018884
|
|
@@ -0,0 +1,147 @@
|
|
|
1
|
+
Metadata-Version: 2.2
|
|
2
|
+
Name: aio-scrapy
|
|
3
|
+
Version: 2.1.7
|
|
4
|
+
Summary: A high-level Web Crawling and Web Scraping framework based on Asyncio
|
|
5
|
+
Home-page: https://github.com/conlin-huang/aio-scrapy.git
|
|
6
|
+
Author: conlin
|
|
7
|
+
Author-email: 995018884@qq.com
|
|
8
|
+
License: MIT
|
|
9
|
+
Keywords: aio-scrapy,scrapy,aioscrapy,scrapy redis,asyncio,spider
|
|
10
|
+
Classifier: License :: OSI Approved :: MIT License
|
|
11
|
+
Classifier: Intended Audience :: Developers
|
|
12
|
+
Classifier: Operating System :: OS Independent
|
|
13
|
+
Classifier: Programming Language :: Python :: 3.9
|
|
14
|
+
Classifier: Programming Language :: Python :: 3.10
|
|
15
|
+
Classifier: Programming Language :: Python :: 3.11
|
|
16
|
+
Classifier: Topic :: Internet :: WWW/HTTP
|
|
17
|
+
Classifier: Topic :: Software Development :: Libraries :: Application Frameworks
|
|
18
|
+
Classifier: Topic :: Software Development :: Libraries :: Python Modules
|
|
19
|
+
Requires-Python: >=3.9
|
|
20
|
+
Description-Content-Type: text/markdown
|
|
21
|
+
License-File: LICENSE
|
|
22
|
+
Requires-Dist: aiohttp
|
|
23
|
+
Requires-Dist: ujson
|
|
24
|
+
Requires-Dist: w3lib>=1.17.0
|
|
25
|
+
Requires-Dist: parsel>=1.5.0
|
|
26
|
+
Requires-Dist: PyDispatcher>=2.0.5
|
|
27
|
+
Requires-Dist: zope.interface>=5.1.0
|
|
28
|
+
Requires-Dist: redis>=4.3.1
|
|
29
|
+
Requires-Dist: aiomultiprocess>=0.9.0
|
|
30
|
+
Requires-Dist: loguru>=0.7.0
|
|
31
|
+
Requires-Dist: anyio>=3.6.2
|
|
32
|
+
Provides-Extra: all
|
|
33
|
+
Requires-Dist: aiomysql>=0.1.1; extra == "all"
|
|
34
|
+
Requires-Dist: httpx[http2]>=0.23.0; extra == "all"
|
|
35
|
+
Requires-Dist: aio-pika>=8.1.1; extra == "all"
|
|
36
|
+
Requires-Dist: cryptography; extra == "all"
|
|
37
|
+
Requires-Dist: motor>=2.1.0; extra == "all"
|
|
38
|
+
Requires-Dist: pyhttpx>=2.10.1; extra == "all"
|
|
39
|
+
Requires-Dist: asyncpg>=0.27.0; extra == "all"
|
|
40
|
+
Requires-Dist: XlsxWriter>=3.1.2; extra == "all"
|
|
41
|
+
Requires-Dist: pillow>=9.4.0; extra == "all"
|
|
42
|
+
Requires-Dist: requests>=2.28.2; extra == "all"
|
|
43
|
+
Requires-Dist: curl_cffi; extra == "all"
|
|
44
|
+
Provides-Extra: aiomysql
|
|
45
|
+
Requires-Dist: aiomysql>=0.1.1; extra == "aiomysql"
|
|
46
|
+
Requires-Dist: cryptography; extra == "aiomysql"
|
|
47
|
+
Provides-Extra: httpx
|
|
48
|
+
Requires-Dist: httpx[http2]>=0.23.0; extra == "httpx"
|
|
49
|
+
Provides-Extra: aio-pika
|
|
50
|
+
Requires-Dist: aio-pika>=8.1.1; extra == "aio-pika"
|
|
51
|
+
Provides-Extra: mongo
|
|
52
|
+
Requires-Dist: motor>=2.1.0; extra == "mongo"
|
|
53
|
+
Provides-Extra: playwright
|
|
54
|
+
Requires-Dist: playwright>=1.31.1; extra == "playwright"
|
|
55
|
+
Provides-Extra: pyhttpx
|
|
56
|
+
Requires-Dist: pyhttpx>=2.10.4; extra == "pyhttpx"
|
|
57
|
+
Provides-Extra: curl-cffi
|
|
58
|
+
Requires-Dist: curl_cffi>=0.6.1; extra == "curl-cffi"
|
|
59
|
+
Provides-Extra: requests
|
|
60
|
+
Requires-Dist: requests>=2.28.2; extra == "requests"
|
|
61
|
+
Provides-Extra: pg
|
|
62
|
+
Requires-Dist: asyncpg>=0.27.0; extra == "pg"
|
|
63
|
+
Provides-Extra: execl
|
|
64
|
+
Requires-Dist: XlsxWriter>=3.1.2; extra == "execl"
|
|
65
|
+
Requires-Dist: pillow>=9.4.0; extra == "execl"
|
|
66
|
+
Dynamic: author
|
|
67
|
+
Dynamic: author-email
|
|
68
|
+
Dynamic: classifier
|
|
69
|
+
Dynamic: description
|
|
70
|
+
Dynamic: description-content-type
|
|
71
|
+
Dynamic: home-page
|
|
72
|
+
Dynamic: keywords
|
|
73
|
+
Dynamic: license
|
|
74
|
+
Dynamic: provides-extra
|
|
75
|
+
Dynamic: requires-dist
|
|
76
|
+
Dynamic: requires-python
|
|
77
|
+
Dynamic: summary
|
|
78
|
+
|
|
79
|
+
# AioScrapy
|
|
80
|
+
|
|
81
|
+
AioScrapy是一个基于Python异步IO的强大网络爬虫框架。它的设计理念源自Scrapy,但完全基于异步IO实现,提供更高的性能和更灵活的配置选项。</br>
|
|
82
|
+
AioScrapy is a powerful asynchronous web crawling framework built on Python's asyncio library. It is inspired by Scrapy but completely reimplemented with asynchronous IO, offering higher performance and more flexible configuration options.
|
|
83
|
+
|
|
84
|
+
## 特性 | Features
|
|
85
|
+
|
|
86
|
+
- **完全异步**:基于Python的asyncio库,实现高效的并发爬取
|
|
87
|
+
- **多种下载处理程序**:支持多种HTTP客户端,包括aiohttp、httpx、requests、pyhttpx、curl_cffi、DrissionPage和playwright
|
|
88
|
+
- **灵活的中间件系统**:轻松添加自定义功能和处理逻辑
|
|
89
|
+
- **强大的数据处理管道**:支持多种数据库存储选项
|
|
90
|
+
- **内置信号系统**:方便的事件处理机制
|
|
91
|
+
- **丰富的配置选项**:高度可定制的爬虫行为
|
|
92
|
+
- **分布式爬取**:支持使用Redis和RabbitMQ进行分布式爬取
|
|
93
|
+
- **数据库集成**:内置支持Redis、MySQL、MongoDB、PostgreSQL和RabbitMQ
|
|
94
|
+
|
|
95
|
+
|
|
96
|
+
- **Fully Asynchronous**: Built on Python's asyncio for efficient concurrent crawling
|
|
97
|
+
- **Multiple Download Handlers**: Support for various HTTP clients including aiohttp, httpx, requests, pyhttpx, curl_cffi, DrissionPage and playwright
|
|
98
|
+
- **Flexible Middleware System**: Easily add custom functionality and processing logic
|
|
99
|
+
- **Powerful Data Processing Pipelines**: Support for various database storage options
|
|
100
|
+
- **Built-in Signal System**: Convenient event handling mechanism
|
|
101
|
+
- **Rich Configuration Options**: Highly customizable crawler behavior
|
|
102
|
+
- **Distributed Crawling**: Support for distributed crawling using Redis and RabbitMQ
|
|
103
|
+
- **Database Integration**: Built-in support for Redis, MySQL, MongoDB, PostgreSQL, and RabbitMQ
|
|
104
|
+
|
|
105
|
+
## 安装 | Installation
|
|
106
|
+
|
|
107
|
+
### 要求 | Requirements
|
|
108
|
+
|
|
109
|
+
- Python 3.9+
|
|
110
|
+
|
|
111
|
+
### 使用pip安装 | Install with pip
|
|
112
|
+
|
|
113
|
+
```bash
|
|
114
|
+
pip install aio-scrapy
|
|
115
|
+
|
|
116
|
+
# Install the latest aio-scrapy
|
|
117
|
+
# pip install git+https://github.com/ConlinH/aio-scrapy
|
|
118
|
+
```
|
|
119
|
+
|
|
120
|
+
## 文档 | Documentation
|
|
121
|
+
|
|
122
|
+
## 文档目录 | Documentation Contents
|
|
123
|
+
- [安装指南 | Installation Guide](docs/installation.md)
|
|
124
|
+
- [快速入门 | Quick Start](docs/quickstart.md)
|
|
125
|
+
- [核心概念 | Core Concepts](docs/concepts.md)
|
|
126
|
+
- [爬虫指南 | Spider Guide](docs/spiders.md)
|
|
127
|
+
- [下载器 | Downloaders](docs/downloaders.md)
|
|
128
|
+
- [中间件 | Middlewares](docs/middlewares.md)
|
|
129
|
+
- [管道 | Pipelines](docs/pipelines.md)
|
|
130
|
+
- [队列 | Queues](docs/queues.md)
|
|
131
|
+
- [请求过滤器 | Request Filters](docs/dupefilters.md)
|
|
132
|
+
- [代理 | Proxy](docs/proxy.md)
|
|
133
|
+
- [数据库连接 | Database Connections](docs/databases.md)
|
|
134
|
+
- [分布式部署 | Distributed Deployment](docs/distributed.md)
|
|
135
|
+
- [配置参考 | Settings Reference](docs/settings.md)
|
|
136
|
+
- [API参考 | API Reference](docs/api.md)
|
|
137
|
+
- [示例 | Example](example)
|
|
138
|
+
|
|
139
|
+
## 许可证 | License
|
|
140
|
+
|
|
141
|
+
本项目采用MIT许可证 - 详情请查看LICENSE文件。</br>
|
|
142
|
+
This project is licensed under the MIT License - see the LICENSE file for details.
|
|
143
|
+
|
|
144
|
+
|
|
145
|
+
## 联系
|
|
146
|
+
QQ: 995018884 </br>
|
|
147
|
+
WeChat: h995018884
|
|
@@ -42,9 +42,10 @@ aioscrapy/core/downloader/handlers/curl_cffi.py
|
|
|
42
42
|
aioscrapy/core/downloader/handlers/httpx.py
|
|
43
43
|
aioscrapy/core/downloader/handlers/pyhttpx.py
|
|
44
44
|
aioscrapy/core/downloader/handlers/requests.py
|
|
45
|
-
aioscrapy/core/downloader/handlers/
|
|
46
|
-
aioscrapy/core/downloader/handlers/
|
|
47
|
-
aioscrapy/core/downloader/handlers/
|
|
45
|
+
aioscrapy/core/downloader/handlers/webdriver/__init__.py
|
|
46
|
+
aioscrapy/core/downloader/handlers/webdriver/drissionpage.py
|
|
47
|
+
aioscrapy/core/downloader/handlers/webdriver/driverpool.py
|
|
48
|
+
aioscrapy/core/downloader/handlers/webdriver/playwright.py
|
|
48
49
|
aioscrapy/db/__init__.py
|
|
49
50
|
aioscrapy/db/absmanager.py
|
|
50
51
|
aioscrapy/db/aiomongo.py
|
|
@@ -62,8 +63,8 @@ aioscrapy/http/request/form.py
|
|
|
62
63
|
aioscrapy/http/request/json_request.py
|
|
63
64
|
aioscrapy/http/response/__init__.py
|
|
64
65
|
aioscrapy/http/response/html.py
|
|
65
|
-
aioscrapy/http/response/playwright.py
|
|
66
66
|
aioscrapy/http/response/text.py
|
|
67
|
+
aioscrapy/http/response/web_driver.py
|
|
67
68
|
aioscrapy/http/response/xml.py
|
|
68
69
|
aioscrapy/libs/__init__.py
|
|
69
70
|
aioscrapy/libs/downloader/__init__.py
|
|
@@ -81,7 +82,7 @@ aioscrapy/libs/extensions/metric.py
|
|
|
81
82
|
aioscrapy/libs/extensions/throttle.py
|
|
82
83
|
aioscrapy/libs/pipelines/__init__.py
|
|
83
84
|
aioscrapy/libs/pipelines/csv.py
|
|
84
|
-
aioscrapy/libs/pipelines/
|
|
85
|
+
aioscrapy/libs/pipelines/excel.py
|
|
85
86
|
aioscrapy/libs/pipelines/mongo.py
|
|
86
87
|
aioscrapy/libs/pipelines/mysql.py
|
|
87
88
|
aioscrapy/libs/pipelines/pg.py
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
2.1.7
|