spiderforce4ai 1.8__tar.gz → 1.9__tar.gz

Sign up to get free protection for your applications and to get access to all the features.
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.2
2
2
  Name: spiderforce4ai
3
- Version: 1.8
3
+ Version: 1.9
4
4
  Summary: Python wrapper for SpiderForce4AI HTML-to-Markdown conversion service
5
5
  Home-page: https://petertam.pro
6
6
  Author: Piotr Tamulewicz
@@ -4,7 +4,7 @@ build-backend = "setuptools.build_meta"
4
4
 
5
5
  [project]
6
6
  name = "spiderforce4ai"
7
- version = "1.8"
7
+ version = "1.9"
8
8
  description = "Python wrapper for SpiderForce4AI HTML-to-Markdown conversion service"
9
9
  readme = "README.md"
10
10
  authors = [{name = "Piotr Tamulewicz", email = "pt@petertam.pro"}]
@@ -3,7 +3,7 @@ from setuptools import setup, find_packages
3
3
 
4
4
  setup(
5
5
  name="spiderforce4ai",
6
- version="1.8",
6
+ version="1.9",
7
7
  author="Piotr Tamulewicz",
8
8
  author_email="pt@petertam.pro",
9
9
  description="Python wrapper for SpiderForce4AI HTML-to-Markdown conversion service",
@@ -22,7 +22,7 @@ from multiprocessing import Pool
22
22
 
23
23
  console = Console()
24
24
 
25
- def extract_metadata_headers(markdown: str) -> str:
25
+ def extract_metadata_headers(markdown: str, url: str = '') -> str:
26
26
  """Extract metadata and headers from markdown content with SEO formatting."""
27
27
  lines = markdown.split('\n')
28
28
  extracted = []
@@ -62,10 +62,10 @@ def extract_metadata_headers(markdown: str) -> str:
62
62
  metadata['language'] = value
63
63
 
64
64
  # Add formatted metadata section with URL first
65
- extracted.append(f"URL: {metadata.get('url', '')}")
66
- extracted.append(f"Title: {metadata['title']}")
65
+ extracted.append(f"URL: {url}") # Use the actual crawled URL
66
+ extracted.append(f"Title: {metadata['title'] or url.split('/')[-2].replace('-', ' ').title()}") # Fallback to URL segment
67
67
  extracted.append(f"Description: {metadata['description']}")
68
- extracted.append(f"CanonicalUrl: {metadata['canonical_url']}")
68
+ extracted.append(f"CanonicalUrl: {metadata['canonical_url'] or url}") # Fallback to crawled URL
69
69
  extracted.append(f"Language: {metadata['language'] or 'en'}") # Default to 'en' if not specified
70
70
  extracted.append("") # Empty line after metadata
71
71
 
@@ -247,7 +247,7 @@ def _process_url_parallel(args: Tuple[str, str, CrawlConfig]) -> CrawlResult:
247
247
 
248
248
  # Handle combined markdown file
249
249
  if config.combine_to_one_markdown:
250
- content = markdown if config.combine_to_one_markdown == 'full' else extract_metadata_headers(markdown)
250
+ content = markdown if config.combine_to_one_markdown == 'full' else extract_metadata_headers(markdown, url)
251
251
  combined_content = f"\n----PAGE----\n{url}\n\n{content}\n----PAGE END----\n"
252
252
 
253
253
  with open(config.combined_markdown_file, 'a', encoding='utf-8') as f:
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.2
2
2
  Name: spiderforce4ai
3
- Version: 1.8
3
+ Version: 1.9
4
4
  Summary: Python wrapper for SpiderForce4AI HTML-to-Markdown conversion service
5
5
  Home-page: https://petertam.pro
6
6
  Author: Piotr Tamulewicz
File without changes
File without changes