scrapling 0.1__tar.gz → 0.1.1__tar.gz
Sign up to get free protection for your applications and to get access to all the features.
- {scrapling-0.1/scrapling.egg-info → scrapling-0.1.1}/PKG-INFO +4 -5
- {scrapling-0.1 → scrapling-0.1.1}/README.md +3 -3
- {scrapling-0.1 → scrapling-0.1.1}/scrapling/__init__.py +1 -1
- {scrapling-0.1 → scrapling-0.1.1/scrapling.egg-info}/PKG-INFO +4 -5
- {scrapling-0.1 → scrapling-0.1.1}/setup.cfg +1 -1
- {scrapling-0.1 → scrapling-0.1.1}/setup.py +2 -3
- {scrapling-0.1 → scrapling-0.1.1}/LICENSE +0 -0
- {scrapling-0.1 → scrapling-0.1.1}/MANIFEST.in +0 -0
- {scrapling-0.1 → scrapling-0.1.1}/scrapling/custom_types.py +0 -0
- {scrapling-0.1 → scrapling-0.1.1}/scrapling/mixins.py +0 -0
- {scrapling-0.1 → scrapling-0.1.1}/scrapling/parser.py +0 -0
- {scrapling-0.1 → scrapling-0.1.1}/scrapling/storage_adaptors.py +0 -0
- {scrapling-0.1 → scrapling-0.1.1}/scrapling/translator.py +0 -0
- {scrapling-0.1 → scrapling-0.1.1}/scrapling/utils.py +0 -0
- {scrapling-0.1 → scrapling-0.1.1}/scrapling.egg-info/SOURCES.txt +0 -0
- {scrapling-0.1 → scrapling-0.1.1}/scrapling.egg-info/dependency_links.txt +0 -0
- {scrapling-0.1 → scrapling-0.1.1}/scrapling.egg-info/not-zip-safe +0 -0
- {scrapling-0.1 → scrapling-0.1.1}/scrapling.egg-info/requires.txt +0 -0
- {scrapling-0.1 → scrapling-0.1.1}/scrapling.egg-info/top_level.txt +0 -0
- {scrapling-0.1 → scrapling-0.1.1}/tests/test_all_functions.py +0 -0
@@ -1,12 +1,12 @@
|
|
1
1
|
Metadata-Version: 2.1
|
2
2
|
Name: scrapling
|
3
|
-
Version: 0.1
|
3
|
+
Version: 0.1.1
|
4
4
|
Summary: Scrapling is a powerful, flexible, and high-performance web scraping library for Python. It
|
5
5
|
Home-page: https://github.com/D4Vinci/Scrapling
|
6
6
|
Author: Karim Shoair
|
7
7
|
Author-email: karim.shoair@pm.me
|
8
8
|
License: BSD
|
9
|
-
Project-URL: Documentation, https://github.com/D4Vinci/Scrapling/
|
9
|
+
Project-URL: Documentation, https://github.com/D4Vinci/Scrapling/tree/main/docs
|
10
10
|
Project-URL: Source, https://github.com/D4Vinci/Scrapling
|
11
11
|
Project-URL: Tracker, https://github.com/D4Vinci/Scrapling/issues
|
12
12
|
Classifier: Operating System :: OS Independent
|
@@ -20,7 +20,6 @@ Classifier: Topic :: Text Processing :: Markup :: HTML
|
|
20
20
|
Classifier: Topic :: Software Development :: Libraries :: Python Modules
|
21
21
|
Classifier: Programming Language :: Python :: 3
|
22
22
|
Classifier: Programming Language :: Python :: 3 :: Only
|
23
|
-
Classifier: Programming Language :: Python :: 3.6
|
24
23
|
Classifier: Programming Language :: Python :: 3.7
|
25
24
|
Classifier: Programming Language :: Python :: 3.8
|
26
25
|
Classifier: Programming Language :: Python :: 3.9
|
@@ -40,7 +39,7 @@ Requires-Dist: orjson>=3
|
|
40
39
|
Requires-Dist: tldextract
|
41
40
|
|
42
41
|
# 🕷️ Scrapling: Lightning-Fast, Adaptive Web Scraping for Python
|
43
|
-
[](https://github.com/D4Vinci/Scrapling/actions/workflows/tests.yml) [](https://badge.fury.io/py/Scrapling) [](https://pypi.org/project/scrapling/) [](https://opensource.org/licenses/BSD-3-Clause)
|
44
43
|
|
45
44
|
Dealing with failing web scrapers due to website changes? Meet Scrapling.
|
46
45
|
|
@@ -78,7 +77,7 @@ products = page.css('.product', auto_match=True) # Still finds them!
|
|
78
77
|
|
79
78
|
## Getting Started
|
80
79
|
|
81
|
-
Let's walk through a basic example that demonstrates small group of Scrapling's core features:
|
80
|
+
Let's walk through a basic example that demonstrates a small group of Scrapling's core features:
|
82
81
|
|
83
82
|
```python
|
84
83
|
import requests
|
@@ -1,5 +1,5 @@
|
|
1
1
|
# 🕷️ Scrapling: Lightning-Fast, Adaptive Web Scraping for Python
|
2
|
-
[](https://github.com/D4Vinci/Scrapling/actions/workflows/tests.yml) [](https://badge.fury.io/py/Scrapling) [](https://pypi.org/project/scrapling/) [](https://opensource.org/licenses/BSD-3-Clause)
|
3
3
|
|
4
4
|
Dealing with failing web scrapers due to website changes? Meet Scrapling.
|
5
5
|
|
@@ -37,7 +37,7 @@ products = page.css('.product', auto_match=True) # Still finds them!
|
|
37
37
|
|
38
38
|
## Getting Started
|
39
39
|
|
40
|
-
Let's walk through a basic example that demonstrates small group of Scrapling's core features:
|
40
|
+
Let's walk through a basic example that demonstrates a small group of Scrapling's core features:
|
41
41
|
|
42
42
|
```python
|
43
43
|
import requests
|
@@ -431,4 +431,4 @@ This project includes code adapted from:
|
|
431
431
|
- In the auto-matching save process, the unique properties of the first element from the selection results are the only ones that get saved. So if the selector you are using selects different elements on the page that are in different locations, auto-matching will probably return to you the first element only when you relocate it later. This doesn't include combined CSS selectors (Using commas to combine more than one selector for example) as these selectors get separated and each selector gets executed alone.
|
432
432
|
- Currently, Scrapling is not compatible with async/await.
|
433
433
|
|
434
|
-
<div align="center"><small>Made with ❤️ by Karim Shoair</small></div><br>
|
434
|
+
<div align="center"><small>Made with ❤️ by Karim Shoair</small></div><br>
|
@@ -3,7 +3,7 @@ from scrapling.parser import Adaptor, Adaptors
|
|
3
3
|
from scrapling.custom_types import TextHandler, AttributesHandler
|
4
4
|
|
5
5
|
__author__ = "Karim Shoair (karim.shoair@pm.me)"
|
6
|
-
__version__ = "0.1"
|
6
|
+
__version__ = "0.1.1"
|
7
7
|
__copyright__ = "Copyright (c) 2024 Karim Shoair"
|
8
8
|
|
9
9
|
|
@@ -1,12 +1,12 @@
|
|
1
1
|
Metadata-Version: 2.1
|
2
2
|
Name: scrapling
|
3
|
-
Version: 0.1
|
3
|
+
Version: 0.1.1
|
4
4
|
Summary: Scrapling is a powerful, flexible, and high-performance web scraping library for Python. It
|
5
5
|
Home-page: https://github.com/D4Vinci/Scrapling
|
6
6
|
Author: Karim Shoair
|
7
7
|
Author-email: karim.shoair@pm.me
|
8
8
|
License: BSD
|
9
|
-
Project-URL: Documentation, https://github.com/D4Vinci/Scrapling/
|
9
|
+
Project-URL: Documentation, https://github.com/D4Vinci/Scrapling/tree/main/docs
|
10
10
|
Project-URL: Source, https://github.com/D4Vinci/Scrapling
|
11
11
|
Project-URL: Tracker, https://github.com/D4Vinci/Scrapling/issues
|
12
12
|
Classifier: Operating System :: OS Independent
|
@@ -20,7 +20,6 @@ Classifier: Topic :: Text Processing :: Markup :: HTML
|
|
20
20
|
Classifier: Topic :: Software Development :: Libraries :: Python Modules
|
21
21
|
Classifier: Programming Language :: Python :: 3
|
22
22
|
Classifier: Programming Language :: Python :: 3 :: Only
|
23
|
-
Classifier: Programming Language :: Python :: 3.6
|
24
23
|
Classifier: Programming Language :: Python :: 3.7
|
25
24
|
Classifier: Programming Language :: Python :: 3.8
|
26
25
|
Classifier: Programming Language :: Python :: 3.9
|
@@ -40,7 +39,7 @@ Requires-Dist: orjson>=3
|
|
40
39
|
Requires-Dist: tldextract
|
41
40
|
|
42
41
|
# 🕷️ Scrapling: Lightning-Fast, Adaptive Web Scraping for Python
|
43
|
-
[](https://github.com/D4Vinci/Scrapling/actions/workflows/tests.yml) [](https://badge.fury.io/py/Scrapling) [](https://pypi.org/project/scrapling/) [](https://opensource.org/licenses/BSD-3-Clause)
|
44
43
|
|
45
44
|
Dealing with failing web scrapers due to website changes? Meet Scrapling.
|
46
45
|
|
@@ -78,7 +77,7 @@ products = page.css('.product', auto_match=True) # Still finds them!
|
|
78
77
|
|
79
78
|
## Getting Started
|
80
79
|
|
81
|
-
Let's walk through a basic example that demonstrates small group of Scrapling's core features:
|
80
|
+
Let's walk through a basic example that demonstrates a small group of Scrapling's core features:
|
82
81
|
|
83
82
|
```python
|
84
83
|
import requests
|
@@ -6,7 +6,7 @@ with open("README.md", "r", encoding="utf-8") as fh:
|
|
6
6
|
|
7
7
|
setup(
|
8
8
|
name="scrapling",
|
9
|
-
version="0.1",
|
9
|
+
version="0.1.1",
|
10
10
|
description="""Scrapling is a powerful, flexible, and high-performance web scraping library for Python. It
|
11
11
|
simplifies the process of extracting data from websites, even when they undergo structural changes, and offers
|
12
12
|
impressive speed improvements over many popular scraping tools.""",
|
@@ -36,7 +36,6 @@ setup(
|
|
36
36
|
"Topic :: Software Development :: Libraries :: Python Modules",
|
37
37
|
"Programming Language :: Python :: 3",
|
38
38
|
"Programming Language :: Python :: 3 :: Only",
|
39
|
-
"Programming Language :: Python :: 3.6",
|
40
39
|
"Programming Language :: Python :: 3.7",
|
41
40
|
"Programming Language :: Python :: 3.8",
|
42
41
|
"Programming Language :: Python :: 3.9",
|
@@ -58,7 +57,7 @@ setup(
|
|
58
57
|
python_requires=">=3.7",
|
59
58
|
url="https://github.com/D4Vinci/Scrapling",
|
60
59
|
project_urls={
|
61
|
-
"Documentation": "https://github.com/D4Vinci/Scrapling/
|
60
|
+
"Documentation": "https://github.com/D4Vinci/Scrapling/tree/main/docs", # For now
|
62
61
|
"Source": "https://github.com/D4Vinci/Scrapling",
|
63
62
|
"Tracker": "https://github.com/D4Vinci/Scrapling/issues",
|
64
63
|
}
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|