firecrawl 2.5.4__tar.gz → 2.6.0__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of firecrawl might be problematic. Click here for more details.

File without changes
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: firecrawl
3
- Version: 2.5.4
3
+ Version: 2.6.0
4
4
  Summary: Python SDK for Firecrawl API
5
5
  Home-page: https://github.com/mendableai/firecrawl
6
6
  Author: Mendable.ai
File without changes
@@ -13,7 +13,7 @@ import os
13
13
 
14
14
  from .firecrawl import FirecrawlApp, AsyncFirecrawlApp, JsonConfig, ScrapeOptions, ChangeTrackingOptions # noqa
15
15
 
16
- __version__ = "2.5.4"
16
+ __version__ = "2.6.0"
17
17
 
18
18
  # Define the logger for the Firecrawl project
19
19
  logger: logging.Logger = logging.getLogger("firecrawl")
@@ -347,6 +347,7 @@ class GenerateLLMsTextParams(pydantic.BaseModel):
347
347
  """
348
348
  maxUrls: Optional[int] = 10
349
349
  showFullText: Optional[bool] = False
350
+ cache: Optional[bool] = True
350
351
  __experimental_stream: Optional[bool] = None
351
352
 
352
353
  class DeepResearchParams(pydantic.BaseModel):
@@ -1870,6 +1871,7 @@ class FirecrawlApp:
1870
1871
  *,
1871
1872
  max_urls: Optional[int] = None,
1872
1873
  show_full_text: Optional[bool] = None,
1874
+ cache: Optional[bool] = None,
1873
1875
  experimental_stream: Optional[bool] = None) -> GenerateLLMsTextStatusResponse:
1874
1876
  """
1875
1877
  Generate LLMs.txt for a given URL and poll until completion.
@@ -1878,6 +1880,7 @@ class FirecrawlApp:
1878
1880
  url (str): Target URL to generate LLMs.txt from
1879
1881
  max_urls (Optional[int]): Maximum URLs to process (default: 10)
1880
1882
  show_full_text (Optional[bool]): Include full text in output (default: False)
1883
+ cache (Optional[bool]): Whether to use cached content if available (default: True)
1881
1884
  experimental_stream (Optional[bool]): Enable experimental streaming
1882
1885
 
1883
1886
  Returns:
@@ -1893,6 +1896,7 @@ class FirecrawlApp:
1893
1896
  params = GenerateLLMsTextParams(
1894
1897
  maxUrls=max_urls,
1895
1898
  showFullText=show_full_text,
1899
+ cache=cache,
1896
1900
  __experimental_stream=experimental_stream
1897
1901
  )
1898
1902
 
@@ -1900,6 +1904,7 @@ class FirecrawlApp:
1900
1904
  url,
1901
1905
  max_urls=max_urls,
1902
1906
  show_full_text=show_full_text,
1907
+ cache=cache,
1903
1908
  experimental_stream=experimental_stream
1904
1909
  )
1905
1910
 
@@ -1935,6 +1940,7 @@ class FirecrawlApp:
1935
1940
  *,
1936
1941
  max_urls: Optional[int] = None,
1937
1942
  show_full_text: Optional[bool] = None,
1943
+ cache: Optional[bool] = None,
1938
1944
  experimental_stream: Optional[bool] = None) -> GenerateLLMsTextResponse:
1939
1945
  """
1940
1946
  Initiate an asynchronous LLMs.txt generation operation.
@@ -1943,6 +1949,7 @@ class FirecrawlApp:
1943
1949
  url (str): The target URL to generate LLMs.txt from. Must be a valid HTTP/HTTPS URL.
1944
1950
  max_urls (Optional[int]): Maximum URLs to process (default: 10)
1945
1951
  show_full_text (Optional[bool]): Include full text in output (default: False)
1952
+ cache (Optional[bool]): Whether to use cached content if available (default: True)
1946
1953
  experimental_stream (Optional[bool]): Enable experimental streaming
1947
1954
 
1948
1955
  Returns:
@@ -1957,6 +1964,7 @@ class FirecrawlApp:
1957
1964
  params = GenerateLLMsTextParams(
1958
1965
  maxUrls=max_urls,
1959
1966
  showFullText=show_full_text,
1967
+ cache=cache,
1960
1968
  __experimental_stream=experimental_stream
1961
1969
  )
1962
1970
 
@@ -4001,6 +4009,7 @@ class AsyncFirecrawlApp(FirecrawlApp):
4001
4009
  url,
4002
4010
  max_urls=max_urls,
4003
4011
  show_full_text=show_full_text,
4012
+ cache=cache,
4004
4013
  experimental_stream=experimental_stream
4005
4014
  )
4006
4015
  if not response.get('success') or 'id' not in response:
@@ -4027,6 +4036,7 @@ class AsyncFirecrawlApp(FirecrawlApp):
4027
4036
  *,
4028
4037
  max_urls: Optional[int] = None,
4029
4038
  show_full_text: Optional[bool] = None,
4039
+ cache: Optional[bool] = None,
4030
4040
  experimental_stream: Optional[bool] = None) -> GenerateLLMsTextResponse:
4031
4041
  """
4032
4042
  Initiate an asynchronous LLMs.txt generation job without waiting for completion.
@@ -4035,6 +4045,7 @@ class AsyncFirecrawlApp(FirecrawlApp):
4035
4045
  url (str): Target URL to generate LLMs.txt from
4036
4046
  max_urls (Optional[int]): Maximum URLs to process (default: 10)
4037
4047
  show_full_text (Optional[bool]): Include full text in output (default: False)
4048
+ cache (Optional[bool]): Whether to use cached content if available (default: True)
4038
4049
  experimental_stream (Optional[bool]): Enable experimental streaming
4039
4050
 
4040
4051
  Returns:
@@ -4057,6 +4068,7 @@ class AsyncFirecrawlApp(FirecrawlApp):
4057
4068
  params = GenerateLLMsTextParams(
4058
4069
  maxUrls=max_urls,
4059
4070
  showFullText=show_full_text,
4071
+ cache=cache,
4060
4072
  __experimental_stream=experimental_stream
4061
4073
  )
4062
4074
 
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: firecrawl
3
- Version: 2.5.4
3
+ Version: 2.6.0
4
4
  Summary: Python SDK for Firecrawl API
5
5
  Home-page: https://github.com/mendableai/firecrawl
6
6
  Author: Mendable.ai
@@ -1,4 +1,2 @@
1
- build
2
- dist
3
1
  firecrawl
4
2
  tests
File without changes
File without changes
File without changes