aio-scrapy 2.1.4__tar.gz → 2.1.7__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (241) hide show
  1. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/LICENSE +1 -1
  2. aio_scrapy-2.1.7/PKG-INFO +147 -0
  3. aio_scrapy-2.1.7/README.md +69 -0
  4. aio_scrapy-2.1.7/aio_scrapy.egg-info/PKG-INFO +147 -0
  5. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aio_scrapy.egg-info/SOURCES.txt +6 -5
  6. aio_scrapy-2.1.7/aioscrapy/VERSION +1 -0
  7. aio_scrapy-2.1.7/aioscrapy/cmdline.py +603 -0
  8. aio_scrapy-2.1.7/aioscrapy/core/downloader/__init__.py +758 -0
  9. aio_scrapy-2.1.7/aioscrapy/core/downloader/handlers/__init__.py +267 -0
  10. aio_scrapy-2.1.7/aioscrapy/core/downloader/handlers/aiohttp.py +289 -0
  11. aio_scrapy-2.1.7/aioscrapy/core/downloader/handlers/curl_cffi.py +188 -0
  12. aio_scrapy-2.1.7/aioscrapy/core/downloader/handlers/httpx.py +214 -0
  13. aio_scrapy-2.1.7/aioscrapy/core/downloader/handlers/pyhttpx.py +200 -0
  14. aio_scrapy-2.1.7/aioscrapy/core/downloader/handlers/requests.py +181 -0
  15. aio_scrapy-2.1.7/aioscrapy/core/downloader/handlers/webdriver/__init__.py +2 -0
  16. aio_scrapy-2.1.7/aioscrapy/core/downloader/handlers/webdriver/drissionpage.py +493 -0
  17. aio_scrapy-2.1.7/aioscrapy/core/downloader/handlers/webdriver/driverpool.py +234 -0
  18. aio_scrapy-2.1.7/aioscrapy/core/downloader/handlers/webdriver/playwright.py +498 -0
  19. aio_scrapy-2.1.7/aioscrapy/core/engine.py +618 -0
  20. aio_scrapy-2.1.7/aioscrapy/core/scheduler.py +510 -0
  21. aio_scrapy-2.1.7/aioscrapy/core/scraper.py +714 -0
  22. aio_scrapy-2.1.7/aioscrapy/crawler.py +642 -0
  23. aio_scrapy-2.1.7/aioscrapy/db/__init__.py +221 -0
  24. aio_scrapy-2.1.7/aioscrapy/db/absmanager.py +253 -0
  25. aio_scrapy-2.1.7/aioscrapy/db/aiomongo.py +374 -0
  26. aio_scrapy-2.1.7/aioscrapy/db/aiomysql.py +477 -0
  27. aio_scrapy-2.1.7/aioscrapy/db/aiopg.py +407 -0
  28. aio_scrapy-2.1.7/aioscrapy/db/aiorabbitmq.py +611 -0
  29. aio_scrapy-2.1.7/aioscrapy/db/aioredis.py +343 -0
  30. aio_scrapy-2.1.7/aioscrapy/dupefilters/__init__.py +152 -0
  31. aio_scrapy-2.1.7/aioscrapy/dupefilters/disk.py +166 -0
  32. aio_scrapy-2.1.7/aioscrapy/dupefilters/redis.py +763 -0
  33. aio_scrapy-2.1.7/aioscrapy/exceptions.py +254 -0
  34. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/http/__init__.py +1 -1
  35. aio_scrapy-2.1.7/aioscrapy/http/headers.py +291 -0
  36. aio_scrapy-2.1.7/aioscrapy/http/request/__init__.py +444 -0
  37. aio_scrapy-2.1.7/aioscrapy/http/request/form.py +119 -0
  38. aio_scrapy-2.1.7/aioscrapy/http/request/json_request.py +175 -0
  39. aio_scrapy-2.1.7/aioscrapy/http/response/__init__.py +481 -0
  40. aio_scrapy-2.1.7/aioscrapy/http/response/html.py +51 -0
  41. aio_scrapy-2.1.7/aioscrapy/http/response/text.py +712 -0
  42. aio_scrapy-2.1.7/aioscrapy/http/response/web_driver.py +144 -0
  43. aio_scrapy-2.1.7/aioscrapy/http/response/xml.py +54 -0
  44. aio_scrapy-2.1.7/aioscrapy/libs/downloader/defaultheaders.py +86 -0
  45. aio_scrapy-2.1.7/aioscrapy/libs/downloader/downloadtimeout.py +115 -0
  46. aio_scrapy-2.1.7/aioscrapy/libs/downloader/ja3fingerprint.py +129 -0
  47. aio_scrapy-2.1.7/aioscrapy/libs/downloader/retry.py +304 -0
  48. aio_scrapy-2.1.7/aioscrapy/libs/downloader/stats.py +173 -0
  49. aio_scrapy-2.1.7/aioscrapy/libs/downloader/useragent.py +114 -0
  50. aio_scrapy-2.1.7/aioscrapy/libs/extensions/closespider.py +233 -0
  51. aio_scrapy-2.1.7/aioscrapy/libs/extensions/corestats.py +196 -0
  52. aio_scrapy-2.1.7/aioscrapy/libs/extensions/logstats.py +194 -0
  53. aio_scrapy-2.1.7/aioscrapy/libs/extensions/metric.py +516 -0
  54. aio_scrapy-2.1.7/aioscrapy/libs/extensions/throttle.py +322 -0
  55. aio_scrapy-2.1.7/aioscrapy/libs/pipelines/__init__.py +480 -0
  56. aio_scrapy-2.1.7/aioscrapy/libs/pipelines/csv.py +328 -0
  57. aio_scrapy-2.1.7/aioscrapy/libs/pipelines/excel.py +545 -0
  58. aio_scrapy-2.1.7/aioscrapy/libs/pipelines/mongo.py +188 -0
  59. aio_scrapy-2.1.7/aioscrapy/libs/pipelines/mysql.py +93 -0
  60. aio_scrapy-2.1.7/aioscrapy/libs/pipelines/pg.py +93 -0
  61. aio_scrapy-2.1.7/aioscrapy/libs/spider/depth.py +192 -0
  62. aio_scrapy-2.1.7/aioscrapy/libs/spider/httperror.py +194 -0
  63. aio_scrapy-2.1.7/aioscrapy/libs/spider/offsite.py +283 -0
  64. aio_scrapy-2.1.7/aioscrapy/libs/spider/referer.py +739 -0
  65. aio_scrapy-2.1.7/aioscrapy/libs/spider/urllength.py +133 -0
  66. aio_scrapy-2.1.7/aioscrapy/link.py +160 -0
  67. aio_scrapy-2.1.7/aioscrapy/logformatter.py +285 -0
  68. aio_scrapy-2.1.7/aioscrapy/middleware/absmanager.py +416 -0
  69. aio_scrapy-2.1.7/aioscrapy/middleware/downloader.py +287 -0
  70. aio_scrapy-2.1.7/aioscrapy/middleware/extension.py +65 -0
  71. aio_scrapy-2.1.7/aioscrapy/middleware/itempipeline.py +114 -0
  72. aio_scrapy-2.1.7/aioscrapy/middleware/spider.py +485 -0
  73. aio_scrapy-2.1.7/aioscrapy/process.py +252 -0
  74. aio_scrapy-2.1.7/aioscrapy/proxy/__init__.py +189 -0
  75. aio_scrapy-2.1.7/aioscrapy/proxy/redis.py +206 -0
  76. aio_scrapy-2.1.7/aioscrapy/queue/__init__.py +219 -0
  77. aio_scrapy-2.1.7/aioscrapy/scrapyd/runner.py +185 -0
  78. aio_scrapy-2.1.7/aioscrapy/serializer.py +219 -0
  79. aio_scrapy-2.1.7/aioscrapy/settings/__init__.py +948 -0
  80. aio_scrapy-2.1.7/aioscrapy/settings/default_settings.py +483 -0
  81. aio_scrapy-2.1.7/aioscrapy/signalmanager.py +199 -0
  82. aio_scrapy-2.1.7/aioscrapy/signals.py +206 -0
  83. aio_scrapy-2.1.7/aioscrapy/spiderloader.py +261 -0
  84. aio_scrapy-2.1.7/aioscrapy/spiders/__init__.py +350 -0
  85. aio_scrapy-2.1.7/aioscrapy/statscollectors.py +391 -0
  86. aio_scrapy-2.1.7/aioscrapy/utils/conf.py +523 -0
  87. aio_scrapy-2.1.7/aioscrapy/utils/curl.py +258 -0
  88. aio_scrapy-2.1.7/aioscrapy/utils/decorators.py +95 -0
  89. aio_scrapy-2.1.7/aioscrapy/utils/deprecate.py +334 -0
  90. aio_scrapy-2.1.7/aioscrapy/utils/httpobj.py +71 -0
  91. aio_scrapy-2.1.7/aioscrapy/utils/log.py +130 -0
  92. aio_scrapy-2.1.7/aioscrapy/utils/misc.py +270 -0
  93. aio_scrapy-2.1.7/aioscrapy/utils/ossignal.py +85 -0
  94. aio_scrapy-2.1.7/aioscrapy/utils/project.py +251 -0
  95. aio_scrapy-2.1.7/aioscrapy/utils/python.py +358 -0
  96. aio_scrapy-2.1.7/aioscrapy/utils/reqser.py +89 -0
  97. aio_scrapy-2.1.7/aioscrapy/utils/request.py +228 -0
  98. aio_scrapy-2.1.7/aioscrapy/utils/response.py +123 -0
  99. aio_scrapy-2.1.7/aioscrapy/utils/signal.py +243 -0
  100. aio_scrapy-2.1.7/aioscrapy/utils/spider.py +67 -0
  101. aio_scrapy-2.1.7/aioscrapy/utils/template.py +123 -0
  102. aio_scrapy-2.1.7/aioscrapy/utils/tools.py +252 -0
  103. aio_scrapy-2.1.7/aioscrapy/utils/trackref.py +253 -0
  104. aio_scrapy-2.1.7/aioscrapy/utils/url.py +469 -0
  105. aio-scrapy-2.1.4/PKG-INFO +0 -239
  106. aio-scrapy-2.1.4/README.md +0 -173
  107. aio-scrapy-2.1.4/aio_scrapy.egg-info/PKG-INFO +0 -239
  108. aio-scrapy-2.1.4/aioscrapy/VERSION +0 -1
  109. aio-scrapy-2.1.4/aioscrapy/cmdline.py +0 -170
  110. aio-scrapy-2.1.4/aioscrapy/core/downloader/__init__.py +0 -253
  111. aio-scrapy-2.1.4/aioscrapy/core/downloader/handlers/__init__.py +0 -85
  112. aio-scrapy-2.1.4/aioscrapy/core/downloader/handlers/aiohttp.py +0 -105
  113. aio-scrapy-2.1.4/aioscrapy/core/downloader/handlers/curl_cffi.py +0 -67
  114. aio-scrapy-2.1.4/aioscrapy/core/downloader/handlers/httpx.py +0 -84
  115. aio-scrapy-2.1.4/aioscrapy/core/downloader/handlers/playwright/__init__.py +0 -115
  116. aio-scrapy-2.1.4/aioscrapy/core/downloader/handlers/playwright/driverpool.py +0 -59
  117. aio-scrapy-2.1.4/aioscrapy/core/downloader/handlers/playwright/webdriver.py +0 -96
  118. aio-scrapy-2.1.4/aioscrapy/core/downloader/handlers/pyhttpx.py +0 -68
  119. aio-scrapy-2.1.4/aioscrapy/core/downloader/handlers/requests.py +0 -63
  120. aio-scrapy-2.1.4/aioscrapy/core/engine.py +0 -257
  121. aio-scrapy-2.1.4/aioscrapy/core/scheduler.py +0 -196
  122. aio-scrapy-2.1.4/aioscrapy/core/scraper.py +0 -238
  123. aio-scrapy-2.1.4/aioscrapy/crawler.py +0 -260
  124. aio-scrapy-2.1.4/aioscrapy/db/__init__.py +0 -72
  125. aio-scrapy-2.1.4/aioscrapy/db/absmanager.py +0 -47
  126. aio-scrapy-2.1.4/aioscrapy/db/aiomongo.py +0 -92
  127. aio-scrapy-2.1.4/aioscrapy/db/aiomysql.py +0 -124
  128. aio-scrapy-2.1.4/aioscrapy/db/aiopg.py +0 -110
  129. aio-scrapy-2.1.4/aioscrapy/db/aiorabbitmq.py +0 -171
  130. aio-scrapy-2.1.4/aioscrapy/db/aioredis.py +0 -94
  131. aio-scrapy-2.1.4/aioscrapy/dupefilters/__init__.py +0 -47
  132. aio-scrapy-2.1.4/aioscrapy/dupefilters/disk.py +0 -44
  133. aio-scrapy-2.1.4/aioscrapy/dupefilters/redis.py +0 -197
  134. aio-scrapy-2.1.4/aioscrapy/exceptions.py +0 -116
  135. aio-scrapy-2.1.4/aioscrapy/http/headers.py +0 -57
  136. aio-scrapy-2.1.4/aioscrapy/http/request/__init__.py +0 -198
  137. aio-scrapy-2.1.4/aioscrapy/http/request/form.py +0 -39
  138. aio-scrapy-2.1.4/aioscrapy/http/request/json_request.py +0 -63
  139. aio-scrapy-2.1.4/aioscrapy/http/response/__init__.py +0 -208
  140. aio-scrapy-2.1.4/aioscrapy/http/response/html.py +0 -12
  141. aio-scrapy-2.1.4/aioscrapy/http/response/playwright.py +0 -36
  142. aio-scrapy-2.1.4/aioscrapy/http/response/text.py +0 -265
  143. aio-scrapy-2.1.4/aioscrapy/http/response/xml.py +0 -12
  144. aio-scrapy-2.1.4/aioscrapy/libs/downloader/defaultheaders.py +0 -22
  145. aio-scrapy-2.1.4/aioscrapy/libs/downloader/downloadtimeout.py +0 -26
  146. aio-scrapy-2.1.4/aioscrapy/libs/downloader/ja3fingerprint.py +0 -36
  147. aio-scrapy-2.1.4/aioscrapy/libs/downloader/retry.py +0 -118
  148. aio-scrapy-2.1.4/aioscrapy/libs/downloader/stats.py +0 -31
  149. aio-scrapy-2.1.4/aioscrapy/libs/downloader/useragent.py +0 -23
  150. aio-scrapy-2.1.4/aioscrapy/libs/extensions/closespider.py +0 -71
  151. aio-scrapy-2.1.4/aioscrapy/libs/extensions/corestats.py +0 -46
  152. aio-scrapy-2.1.4/aioscrapy/libs/extensions/logstats.py +0 -50
  153. aio-scrapy-2.1.4/aioscrapy/libs/extensions/metric.py +0 -147
  154. aio-scrapy-2.1.4/aioscrapy/libs/extensions/throttle.py +0 -88
  155. aio-scrapy-2.1.4/aioscrapy/libs/pipelines/__init__.py +0 -137
  156. aio-scrapy-2.1.4/aioscrapy/libs/pipelines/csv.py +0 -86
  157. aio-scrapy-2.1.4/aioscrapy/libs/pipelines/execl.py +0 -169
  158. aio-scrapy-2.1.4/aioscrapy/libs/pipelines/mongo.py +0 -56
  159. aio-scrapy-2.1.4/aioscrapy/libs/pipelines/mysql.py +0 -26
  160. aio-scrapy-2.1.4/aioscrapy/libs/pipelines/pg.py +0 -26
  161. aio-scrapy-2.1.4/aioscrapy/libs/spider/depth.py +0 -54
  162. aio-scrapy-2.1.4/aioscrapy/libs/spider/httperror.py +0 -54
  163. aio-scrapy-2.1.4/aioscrapy/libs/spider/offsite.py +0 -83
  164. aio-scrapy-2.1.4/aioscrapy/libs/spider/referer.py +0 -364
  165. aio-scrapy-2.1.4/aioscrapy/libs/spider/urllength.py +0 -37
  166. aio-scrapy-2.1.4/aioscrapy/link.py +0 -53
  167. aio-scrapy-2.1.4/aioscrapy/logformatter.py +0 -94
  168. aio-scrapy-2.1.4/aioscrapy/middleware/absmanager.py +0 -90
  169. aio-scrapy-2.1.4/aioscrapy/middleware/downloader.py +0 -69
  170. aio-scrapy-2.1.4/aioscrapy/middleware/extension.py +0 -16
  171. aio-scrapy-2.1.4/aioscrapy/middleware/itempipeline.py +0 -18
  172. aio-scrapy-2.1.4/aioscrapy/middleware/spider.py +0 -132
  173. aio-scrapy-2.1.4/aioscrapy/process.py +0 -52
  174. aio-scrapy-2.1.4/aioscrapy/proxy/__init__.py +0 -50
  175. aio-scrapy-2.1.4/aioscrapy/proxy/redis.py +0 -72
  176. aio-scrapy-2.1.4/aioscrapy/queue/__init__.py +0 -67
  177. aio-scrapy-2.1.4/aioscrapy/scrapyd/runner.py +0 -64
  178. aio-scrapy-2.1.4/aioscrapy/serializer.py +0 -39
  179. aio-scrapy-2.1.4/aioscrapy/settings/__init__.py +0 -466
  180. aio-scrapy-2.1.4/aioscrapy/settings/default_settings.py +0 -183
  181. aio-scrapy-2.1.4/aioscrapy/signalmanager.py +0 -68
  182. aio-scrapy-2.1.4/aioscrapy/signals.py +0 -24
  183. aio-scrapy-2.1.4/aioscrapy/spiderloader.py +0 -108
  184. aio-scrapy-2.1.4/aioscrapy/spiders/__init__.py +0 -123
  185. aio-scrapy-2.1.4/aioscrapy/statscollectors.py +0 -80
  186. aio-scrapy-2.1.4/aioscrapy/utils/conf.py +0 -195
  187. aio-scrapy-2.1.4/aioscrapy/utils/curl.py +0 -106
  188. aio-scrapy-2.1.4/aioscrapy/utils/decorators.py +0 -25
  189. aio-scrapy-2.1.4/aioscrapy/utils/deprecate.py +0 -141
  190. aio-scrapy-2.1.4/aioscrapy/utils/httpobj.py +0 -19
  191. aio-scrapy-2.1.4/aioscrapy/utils/log.py +0 -51
  192. aio-scrapy-2.1.4/aioscrapy/utils/misc.py +0 -102
  193. aio-scrapy-2.1.4/aioscrapy/utils/ossignal.py +0 -23
  194. aio-scrapy-2.1.4/aioscrapy/utils/project.py +0 -89
  195. aio-scrapy-2.1.4/aioscrapy/utils/python.py +0 -148
  196. aio-scrapy-2.1.4/aioscrapy/utils/reqser.py +0 -15
  197. aio-scrapy-2.1.4/aioscrapy/utils/request.py +0 -67
  198. aio-scrapy-2.1.4/aioscrapy/utils/response.py +0 -38
  199. aio-scrapy-2.1.4/aioscrapy/utils/signal.py +0 -61
  200. aio-scrapy-2.1.4/aioscrapy/utils/spider.py +0 -20
  201. aio-scrapy-2.1.4/aioscrapy/utils/template.py +0 -36
  202. aio-scrapy-2.1.4/aioscrapy/utils/tools.py +0 -78
  203. aio-scrapy-2.1.4/aioscrapy/utils/trackref.py +0 -67
  204. aio-scrapy-2.1.4/aioscrapy/utils/url.py +0 -164
  205. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/MANIFEST.in +0 -0
  206. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aio_scrapy.egg-info/dependency_links.txt +0 -0
  207. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aio_scrapy.egg-info/entry_points.txt +0 -0
  208. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aio_scrapy.egg-info/not-zip-safe +0 -0
  209. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aio_scrapy.egg-info/requires.txt +0 -0
  210. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aio_scrapy.egg-info/top_level.txt +0 -0
  211. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/__init__.py +0 -0
  212. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/__main__.py +0 -0
  213. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/commands/__init__.py +0 -0
  214. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/commands/crawl.py +0 -0
  215. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/commands/genspider.py +0 -0
  216. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/commands/list.py +0 -0
  217. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/commands/runspider.py +0 -0
  218. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/commands/settings.py +0 -0
  219. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/commands/startproject.py +0 -0
  220. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/commands/version.py +0 -0
  221. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/core/__init__.py +0 -0
  222. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/libs/__init__.py +0 -0
  223. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/libs/downloader/__init__.py +0 -0
  224. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/libs/extensions/__init__.py +0 -0
  225. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/libs/spider/__init__.py +0 -0
  226. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/middleware/__init__.py +0 -0
  227. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/queue/memory.py +0 -0
  228. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/queue/rabbitmq.py +0 -0
  229. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/queue/redis.py +0 -0
  230. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/scrapyd/__init__.py +0 -0
  231. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/templates/project/aioscrapy.cfg +0 -0
  232. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/templates/project/module/__init__.py +0 -0
  233. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/templates/project/module/middlewares.py.tmpl +0 -0
  234. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/templates/project/module/pipelines.py.tmpl +0 -0
  235. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/templates/project/module/settings.py.tmpl +0 -0
  236. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/templates/project/module/spiders/__init__.py +0 -0
  237. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/templates/spiders/basic.tmpl +0 -0
  238. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/templates/spiders/single.tmpl +0 -0
  239. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/aioscrapy/utils/__init__.py +0 -0
  240. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/setup.cfg +0 -0
  241. {aio-scrapy-2.1.4 → aio_scrapy-2.1.7}/setup.py +0 -0
@@ -1,6 +1,6 @@
1
1
  MIT License
2
2
 
3
- Copyright (c) 2021 conlin-huang
3
+ Copyright (c) 2021 ConlinH
4
4
 
5
5
  Permission is hereby granted, free of charge, to any person obtaining a copy
6
6
  of this software and associated documentation files (the "Software"), to deal
@@ -0,0 +1,147 @@
1
+ Metadata-Version: 2.2
2
+ Name: aio-scrapy
3
+ Version: 2.1.7
4
+ Summary: A high-level Web Crawling and Web Scraping framework based on Asyncio
5
+ Home-page: https://github.com/conlin-huang/aio-scrapy.git
6
+ Author: conlin
7
+ Author-email: 995018884@qq.com
8
+ License: MIT
9
+ Keywords: aio-scrapy,scrapy,aioscrapy,scrapy redis,asyncio,spider
10
+ Classifier: License :: OSI Approved :: MIT License
11
+ Classifier: Intended Audience :: Developers
12
+ Classifier: Operating System :: OS Independent
13
+ Classifier: Programming Language :: Python :: 3.9
14
+ Classifier: Programming Language :: Python :: 3.10
15
+ Classifier: Programming Language :: Python :: 3.11
16
+ Classifier: Topic :: Internet :: WWW/HTTP
17
+ Classifier: Topic :: Software Development :: Libraries :: Application Frameworks
18
+ Classifier: Topic :: Software Development :: Libraries :: Python Modules
19
+ Requires-Python: >=3.9
20
+ Description-Content-Type: text/markdown
21
+ License-File: LICENSE
22
+ Requires-Dist: aiohttp
23
+ Requires-Dist: ujson
24
+ Requires-Dist: w3lib>=1.17.0
25
+ Requires-Dist: parsel>=1.5.0
26
+ Requires-Dist: PyDispatcher>=2.0.5
27
+ Requires-Dist: zope.interface>=5.1.0
28
+ Requires-Dist: redis>=4.3.1
29
+ Requires-Dist: aiomultiprocess>=0.9.0
30
+ Requires-Dist: loguru>=0.7.0
31
+ Requires-Dist: anyio>=3.6.2
32
+ Provides-Extra: all
33
+ Requires-Dist: aiomysql>=0.1.1; extra == "all"
34
+ Requires-Dist: httpx[http2]>=0.23.0; extra == "all"
35
+ Requires-Dist: aio-pika>=8.1.1; extra == "all"
36
+ Requires-Dist: cryptography; extra == "all"
37
+ Requires-Dist: motor>=2.1.0; extra == "all"
38
+ Requires-Dist: pyhttpx>=2.10.1; extra == "all"
39
+ Requires-Dist: asyncpg>=0.27.0; extra == "all"
40
+ Requires-Dist: XlsxWriter>=3.1.2; extra == "all"
41
+ Requires-Dist: pillow>=9.4.0; extra == "all"
42
+ Requires-Dist: requests>=2.28.2; extra == "all"
43
+ Requires-Dist: curl_cffi; extra == "all"
44
+ Provides-Extra: aiomysql
45
+ Requires-Dist: aiomysql>=0.1.1; extra == "aiomysql"
46
+ Requires-Dist: cryptography; extra == "aiomysql"
47
+ Provides-Extra: httpx
48
+ Requires-Dist: httpx[http2]>=0.23.0; extra == "httpx"
49
+ Provides-Extra: aio-pika
50
+ Requires-Dist: aio-pika>=8.1.1; extra == "aio-pika"
51
+ Provides-Extra: mongo
52
+ Requires-Dist: motor>=2.1.0; extra == "mongo"
53
+ Provides-Extra: playwright
54
+ Requires-Dist: playwright>=1.31.1; extra == "playwright"
55
+ Provides-Extra: pyhttpx
56
+ Requires-Dist: pyhttpx>=2.10.4; extra == "pyhttpx"
57
+ Provides-Extra: curl-cffi
58
+ Requires-Dist: curl_cffi>=0.6.1; extra == "curl-cffi"
59
+ Provides-Extra: requests
60
+ Requires-Dist: requests>=2.28.2; extra == "requests"
61
+ Provides-Extra: pg
62
+ Requires-Dist: asyncpg>=0.27.0; extra == "pg"
63
+ Provides-Extra: execl
64
+ Requires-Dist: XlsxWriter>=3.1.2; extra == "execl"
65
+ Requires-Dist: pillow>=9.4.0; extra == "execl"
66
+ Dynamic: author
67
+ Dynamic: author-email
68
+ Dynamic: classifier
69
+ Dynamic: description
70
+ Dynamic: description-content-type
71
+ Dynamic: home-page
72
+ Dynamic: keywords
73
+ Dynamic: license
74
+ Dynamic: provides-extra
75
+ Dynamic: requires-dist
76
+ Dynamic: requires-python
77
+ Dynamic: summary
78
+
79
+ # AioScrapy
80
+
81
+ AioScrapy是一个基于Python异步IO的强大网络爬虫框架。它的设计理念源自Scrapy,但完全基于异步IO实现,提供更高的性能和更灵活的配置选项。</br>
82
+ AioScrapy is a powerful asynchronous web crawling framework built on Python's asyncio library. It is inspired by Scrapy but completely reimplemented with asynchronous IO, offering higher performance and more flexible configuration options.
83
+
84
+ ## 特性 | Features
85
+
86
+ - **完全异步**:基于Python的asyncio库,实现高效的并发爬取
87
+ - **多种下载处理程序**:支持多种HTTP客户端,包括aiohttp、httpx、requests、pyhttpx、curl_cffi、DrissionPage和playwright
88
+ - **灵活的中间件系统**:轻松添加自定义功能和处理逻辑
89
+ - **强大的数据处理管道**:支持多种数据库存储选项
90
+ - **内置信号系统**:方便的事件处理机制
91
+ - **丰富的配置选项**:高度可定制的爬虫行为
92
+ - **分布式爬取**:支持使用Redis和RabbitMQ进行分布式爬取
93
+ - **数据库集成**:内置支持Redis、MySQL、MongoDB、PostgreSQL和RabbitMQ
94
+
95
+
96
+ - **Fully Asynchronous**: Built on Python's asyncio for efficient concurrent crawling
97
+ - **Multiple Download Handlers**: Support for various HTTP clients including aiohttp, httpx, requests, pyhttpx, curl_cffi, DrissionPage and playwright
98
+ - **Flexible Middleware System**: Easily add custom functionality and processing logic
99
+ - **Powerful Data Processing Pipelines**: Support for various database storage options
100
+ - **Built-in Signal System**: Convenient event handling mechanism
101
+ - **Rich Configuration Options**: Highly customizable crawler behavior
102
+ - **Distributed Crawling**: Support for distributed crawling using Redis and RabbitMQ
103
+ - **Database Integration**: Built-in support for Redis, MySQL, MongoDB, PostgreSQL, and RabbitMQ
104
+
105
+ ## 安装 | Installation
106
+
107
+ ### 要求 | Requirements
108
+
109
+ - Python 3.9+
110
+
111
+ ### 使用pip安装 | Install with pip
112
+
113
+ ```bash
114
+ pip install aio-scrapy
115
+
116
+ # Install the latest aio-scrapy
117
+ # pip install git+https://github.com/ConlinH/aio-scrapy
118
+ ```
119
+
120
+ ## 文档 | Documentation
121
+
122
+ ## 文档目录 | Documentation Contents
123
+ - [安装指南 | Installation Guide](docs/installation.md)
124
+ - [快速入门 | Quick Start](docs/quickstart.md)
125
+ - [核心概念 | Core Concepts](docs/concepts.md)
126
+ - [爬虫指南 | Spider Guide](docs/spiders.md)
127
+ - [下载器 | Downloaders](docs/downloaders.md)
128
+ - [中间件 | Middlewares](docs/middlewares.md)
129
+ - [管道 | Pipelines](docs/pipelines.md)
130
+ - [队列 | Queues](docs/queues.md)
131
+ - [请求过滤器 | Request Filters](docs/dupefilters.md)
132
+ - [代理 | Proxy](docs/proxy.md)
133
+ - [数据库连接 | Database Connections](docs/databases.md)
134
+ - [分布式部署 | Distributed Deployment](docs/distributed.md)
135
+ - [配置参考 | Settings Reference](docs/settings.md)
136
+ - [API参考 | API Reference](docs/api.md)
137
+ - [示例 | Example](example)
138
+
139
+ ## 许可证 | License
140
+
141
+ 本项目采用MIT许可证 - 详情请查看LICENSE文件。</br>
142
+ This project is licensed under the MIT License - see the LICENSE file for details.
143
+
144
+
145
+ ## 联系
146
+ QQ: 995018884 </br>
147
+ WeChat: h995018884
@@ -0,0 +1,69 @@
1
+ # AioScrapy
2
+
3
+ AioScrapy是一个基于Python异步IO的强大网络爬虫框架。它的设计理念源自Scrapy,但完全基于异步IO实现,提供更高的性能和更灵活的配置选项。</br>
4
+ AioScrapy is a powerful asynchronous web crawling framework built on Python's asyncio library. It is inspired by Scrapy but completely reimplemented with asynchronous IO, offering higher performance and more flexible configuration options.
5
+
6
+ ## 特性 | Features
7
+
8
+ - **完全异步**:基于Python的asyncio库,实现高效的并发爬取
9
+ - **多种下载处理程序**:支持多种HTTP客户端,包括aiohttp、httpx、requests、pyhttpx、curl_cffi、DrissionPage和playwright
10
+ - **灵活的中间件系统**:轻松添加自定义功能和处理逻辑
11
+ - **强大的数据处理管道**:支持多种数据库存储选项
12
+ - **内置信号系统**:方便的事件处理机制
13
+ - **丰富的配置选项**:高度可定制的爬虫行为
14
+ - **分布式爬取**:支持使用Redis和RabbitMQ进行分布式爬取
15
+ - **数据库集成**:内置支持Redis、MySQL、MongoDB、PostgreSQL和RabbitMQ
16
+
17
+
18
+ - **Fully Asynchronous**: Built on Python's asyncio for efficient concurrent crawling
19
+ - **Multiple Download Handlers**: Support for various HTTP clients including aiohttp, httpx, requests, pyhttpx, curl_cffi, DrissionPage and playwright
20
+ - **Flexible Middleware System**: Easily add custom functionality and processing logic
21
+ - **Powerful Data Processing Pipelines**: Support for various database storage options
22
+ - **Built-in Signal System**: Convenient event handling mechanism
23
+ - **Rich Configuration Options**: Highly customizable crawler behavior
24
+ - **Distributed Crawling**: Support for distributed crawling using Redis and RabbitMQ
25
+ - **Database Integration**: Built-in support for Redis, MySQL, MongoDB, PostgreSQL, and RabbitMQ
26
+
27
+ ## 安装 | Installation
28
+
29
+ ### 要求 | Requirements
30
+
31
+ - Python 3.9+
32
+
33
+ ### 使用pip安装 | Install with pip
34
+
35
+ ```bash
36
+ pip install aio-scrapy
37
+
38
+ # Install the latest aio-scrapy
39
+ # pip install git+https://github.com/ConlinH/aio-scrapy
40
+ ```
41
+
42
+ ## 文档 | Documentation
43
+
44
+ ## 文档目录 | Documentation Contents
45
+ - [安装指南 | Installation Guide](docs/installation.md)
46
+ - [快速入门 | Quick Start](docs/quickstart.md)
47
+ - [核心概念 | Core Concepts](docs/concepts.md)
48
+ - [爬虫指南 | Spider Guide](docs/spiders.md)
49
+ - [下载器 | Downloaders](docs/downloaders.md)
50
+ - [中间件 | Middlewares](docs/middlewares.md)
51
+ - [管道 | Pipelines](docs/pipelines.md)
52
+ - [队列 | Queues](docs/queues.md)
53
+ - [请求过滤器 | Request Filters](docs/dupefilters.md)
54
+ - [代理 | Proxy](docs/proxy.md)
55
+ - [数据库连接 | Database Connections](docs/databases.md)
56
+ - [分布式部署 | Distributed Deployment](docs/distributed.md)
57
+ - [配置参考 | Settings Reference](docs/settings.md)
58
+ - [API参考 | API Reference](docs/api.md)
59
+ - [示例 | Example](example)
60
+
61
+ ## 许可证 | License
62
+
63
+ 本项目采用MIT许可证 - 详情请查看LICENSE文件。</br>
64
+ This project is licensed under the MIT License - see the LICENSE file for details.
65
+
66
+
67
+ ## 联系
68
+ QQ: 995018884 </br>
69
+ WeChat: h995018884
@@ -0,0 +1,147 @@
1
+ Metadata-Version: 2.2
2
+ Name: aio-scrapy
3
+ Version: 2.1.7
4
+ Summary: A high-level Web Crawling and Web Scraping framework based on Asyncio
5
+ Home-page: https://github.com/conlin-huang/aio-scrapy.git
6
+ Author: conlin
7
+ Author-email: 995018884@qq.com
8
+ License: MIT
9
+ Keywords: aio-scrapy,scrapy,aioscrapy,scrapy redis,asyncio,spider
10
+ Classifier: License :: OSI Approved :: MIT License
11
+ Classifier: Intended Audience :: Developers
12
+ Classifier: Operating System :: OS Independent
13
+ Classifier: Programming Language :: Python :: 3.9
14
+ Classifier: Programming Language :: Python :: 3.10
15
+ Classifier: Programming Language :: Python :: 3.11
16
+ Classifier: Topic :: Internet :: WWW/HTTP
17
+ Classifier: Topic :: Software Development :: Libraries :: Application Frameworks
18
+ Classifier: Topic :: Software Development :: Libraries :: Python Modules
19
+ Requires-Python: >=3.9
20
+ Description-Content-Type: text/markdown
21
+ License-File: LICENSE
22
+ Requires-Dist: aiohttp
23
+ Requires-Dist: ujson
24
+ Requires-Dist: w3lib>=1.17.0
25
+ Requires-Dist: parsel>=1.5.0
26
+ Requires-Dist: PyDispatcher>=2.0.5
27
+ Requires-Dist: zope.interface>=5.1.0
28
+ Requires-Dist: redis>=4.3.1
29
+ Requires-Dist: aiomultiprocess>=0.9.0
30
+ Requires-Dist: loguru>=0.7.0
31
+ Requires-Dist: anyio>=3.6.2
32
+ Provides-Extra: all
33
+ Requires-Dist: aiomysql>=0.1.1; extra == "all"
34
+ Requires-Dist: httpx[http2]>=0.23.0; extra == "all"
35
+ Requires-Dist: aio-pika>=8.1.1; extra == "all"
36
+ Requires-Dist: cryptography; extra == "all"
37
+ Requires-Dist: motor>=2.1.0; extra == "all"
38
+ Requires-Dist: pyhttpx>=2.10.1; extra == "all"
39
+ Requires-Dist: asyncpg>=0.27.0; extra == "all"
40
+ Requires-Dist: XlsxWriter>=3.1.2; extra == "all"
41
+ Requires-Dist: pillow>=9.4.0; extra == "all"
42
+ Requires-Dist: requests>=2.28.2; extra == "all"
43
+ Requires-Dist: curl_cffi; extra == "all"
44
+ Provides-Extra: aiomysql
45
+ Requires-Dist: aiomysql>=0.1.1; extra == "aiomysql"
46
+ Requires-Dist: cryptography; extra == "aiomysql"
47
+ Provides-Extra: httpx
48
+ Requires-Dist: httpx[http2]>=0.23.0; extra == "httpx"
49
+ Provides-Extra: aio-pika
50
+ Requires-Dist: aio-pika>=8.1.1; extra == "aio-pika"
51
+ Provides-Extra: mongo
52
+ Requires-Dist: motor>=2.1.0; extra == "mongo"
53
+ Provides-Extra: playwright
54
+ Requires-Dist: playwright>=1.31.1; extra == "playwright"
55
+ Provides-Extra: pyhttpx
56
+ Requires-Dist: pyhttpx>=2.10.4; extra == "pyhttpx"
57
+ Provides-Extra: curl-cffi
58
+ Requires-Dist: curl_cffi>=0.6.1; extra == "curl-cffi"
59
+ Provides-Extra: requests
60
+ Requires-Dist: requests>=2.28.2; extra == "requests"
61
+ Provides-Extra: pg
62
+ Requires-Dist: asyncpg>=0.27.0; extra == "pg"
63
+ Provides-Extra: execl
64
+ Requires-Dist: XlsxWriter>=3.1.2; extra == "execl"
65
+ Requires-Dist: pillow>=9.4.0; extra == "execl"
66
+ Dynamic: author
67
+ Dynamic: author-email
68
+ Dynamic: classifier
69
+ Dynamic: description
70
+ Dynamic: description-content-type
71
+ Dynamic: home-page
72
+ Dynamic: keywords
73
+ Dynamic: license
74
+ Dynamic: provides-extra
75
+ Dynamic: requires-dist
76
+ Dynamic: requires-python
77
+ Dynamic: summary
78
+
79
+ # AioScrapy
80
+
81
+ AioScrapy是一个基于Python异步IO的强大网络爬虫框架。它的设计理念源自Scrapy,但完全基于异步IO实现,提供更高的性能和更灵活的配置选项。</br>
82
+ AioScrapy is a powerful asynchronous web crawling framework built on Python's asyncio library. It is inspired by Scrapy but completely reimplemented with asynchronous IO, offering higher performance and more flexible configuration options.
83
+
84
+ ## 特性 | Features
85
+
86
+ - **完全异步**:基于Python的asyncio库,实现高效的并发爬取
87
+ - **多种下载处理程序**:支持多种HTTP客户端,包括aiohttp、httpx、requests、pyhttpx、curl_cffi、DrissionPage和playwright
88
+ - **灵活的中间件系统**:轻松添加自定义功能和处理逻辑
89
+ - **强大的数据处理管道**:支持多种数据库存储选项
90
+ - **内置信号系统**:方便的事件处理机制
91
+ - **丰富的配置选项**:高度可定制的爬虫行为
92
+ - **分布式爬取**:支持使用Redis和RabbitMQ进行分布式爬取
93
+ - **数据库集成**:内置支持Redis、MySQL、MongoDB、PostgreSQL和RabbitMQ
94
+
95
+
96
+ - **Fully Asynchronous**: Built on Python's asyncio for efficient concurrent crawling
97
+ - **Multiple Download Handlers**: Support for various HTTP clients including aiohttp, httpx, requests, pyhttpx, curl_cffi, DrissionPage and playwright
98
+ - **Flexible Middleware System**: Easily add custom functionality and processing logic
99
+ - **Powerful Data Processing Pipelines**: Support for various database storage options
100
+ - **Built-in Signal System**: Convenient event handling mechanism
101
+ - **Rich Configuration Options**: Highly customizable crawler behavior
102
+ - **Distributed Crawling**: Support for distributed crawling using Redis and RabbitMQ
103
+ - **Database Integration**: Built-in support for Redis, MySQL, MongoDB, PostgreSQL, and RabbitMQ
104
+
105
+ ## 安装 | Installation
106
+
107
+ ### 要求 | Requirements
108
+
109
+ - Python 3.9+
110
+
111
+ ### 使用pip安装 | Install with pip
112
+
113
+ ```bash
114
+ pip install aio-scrapy
115
+
116
+ # Install the latest aio-scrapy
117
+ # pip install git+https://github.com/ConlinH/aio-scrapy
118
+ ```
119
+
120
+ ## 文档 | Documentation
121
+
122
+ ## 文档目录 | Documentation Contents
123
+ - [安装指南 | Installation Guide](docs/installation.md)
124
+ - [快速入门 | Quick Start](docs/quickstart.md)
125
+ - [核心概念 | Core Concepts](docs/concepts.md)
126
+ - [爬虫指南 | Spider Guide](docs/spiders.md)
127
+ - [下载器 | Downloaders](docs/downloaders.md)
128
+ - [中间件 | Middlewares](docs/middlewares.md)
129
+ - [管道 | Pipelines](docs/pipelines.md)
130
+ - [队列 | Queues](docs/queues.md)
131
+ - [请求过滤器 | Request Filters](docs/dupefilters.md)
132
+ - [代理 | Proxy](docs/proxy.md)
133
+ - [数据库连接 | Database Connections](docs/databases.md)
134
+ - [分布式部署 | Distributed Deployment](docs/distributed.md)
135
+ - [配置参考 | Settings Reference](docs/settings.md)
136
+ - [API参考 | API Reference](docs/api.md)
137
+ - [示例 | Example](example)
138
+
139
+ ## 许可证 | License
140
+
141
+ 本项目采用MIT许可证 - 详情请查看LICENSE文件。</br>
142
+ This project is licensed under the MIT License - see the LICENSE file for details.
143
+
144
+
145
+ ## 联系
146
+ QQ: 995018884 </br>
147
+ WeChat: h995018884
@@ -42,9 +42,10 @@ aioscrapy/core/downloader/handlers/curl_cffi.py
42
42
  aioscrapy/core/downloader/handlers/httpx.py
43
43
  aioscrapy/core/downloader/handlers/pyhttpx.py
44
44
  aioscrapy/core/downloader/handlers/requests.py
45
- aioscrapy/core/downloader/handlers/playwright/__init__.py
46
- aioscrapy/core/downloader/handlers/playwright/driverpool.py
47
- aioscrapy/core/downloader/handlers/playwright/webdriver.py
45
+ aioscrapy/core/downloader/handlers/webdriver/__init__.py
46
+ aioscrapy/core/downloader/handlers/webdriver/drissionpage.py
47
+ aioscrapy/core/downloader/handlers/webdriver/driverpool.py
48
+ aioscrapy/core/downloader/handlers/webdriver/playwright.py
48
49
  aioscrapy/db/__init__.py
49
50
  aioscrapy/db/absmanager.py
50
51
  aioscrapy/db/aiomongo.py
@@ -62,8 +63,8 @@ aioscrapy/http/request/form.py
62
63
  aioscrapy/http/request/json_request.py
63
64
  aioscrapy/http/response/__init__.py
64
65
  aioscrapy/http/response/html.py
65
- aioscrapy/http/response/playwright.py
66
66
  aioscrapy/http/response/text.py
67
+ aioscrapy/http/response/web_driver.py
67
68
  aioscrapy/http/response/xml.py
68
69
  aioscrapy/libs/__init__.py
69
70
  aioscrapy/libs/downloader/__init__.py
@@ -81,7 +82,7 @@ aioscrapy/libs/extensions/metric.py
81
82
  aioscrapy/libs/extensions/throttle.py
82
83
  aioscrapy/libs/pipelines/__init__.py
83
84
  aioscrapy/libs/pipelines/csv.py
84
- aioscrapy/libs/pipelines/execl.py
85
+ aioscrapy/libs/pipelines/excel.py
85
86
  aioscrapy/libs/pipelines/mongo.py
86
87
  aioscrapy/libs/pipelines/mysql.py
87
88
  aioscrapy/libs/pipelines/pg.py
@@ -0,0 +1 @@
1
+ 2.1.7