firecrawl 4.18.2 → 4.18.4

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/README.md CHANGED
@@ -1,6 +1,6 @@
1
1
  # Firecrawl Node SDK
2
2
 
3
- The Firecrawl Node SDK is a library that allows you to easily scrape and crawl websites, and output the data in a format ready for use with language models (LLMs). It provides a simple and intuitive interface for interacting with the Firecrawl API.
3
+ The Firecrawl Node SDK is a library that allows you to easily search, scrape, and interact with the web, and output the data in a format ready for use with language models (LLMs). It provides a simple and intuitive interface for the Firecrawl API.
4
4
 
5
5
  ## Installation
6
6
 
@@ -8,7 +8,7 @@ var require_package = __commonJS({
8
8
  "package.json"(exports, module) {
9
9
  module.exports = {
10
10
  name: "@mendable/firecrawl-js",
11
- version: "4.18.2",
11
+ version: "4.18.4",
12
12
  description: "JavaScript SDK for Firecrawl API",
13
13
  main: "dist/index.js",
14
14
  types: "dist/index.d.ts",
@@ -78,7 +78,8 @@ var require_package = __commonJS({
78
78
  "picomatch@<4.0.4": ">=4.0.4",
79
79
  handlebars: ">=4.7.9",
80
80
  "brace-expansion": ">=5.0.5",
81
- "axios@<1.15.0": "1.15.0"
81
+ "axios@<1.15.0": "1.15.0",
82
+ "follow-redirects@<1.16.0": ">=1.16.0 <2.0.0"
82
83
  }
83
84
  }
84
85
  };
package/dist/index.cjs CHANGED
@@ -35,7 +35,7 @@ var require_package = __commonJS({
35
35
  "package.json"(exports2, module2) {
36
36
  module2.exports = {
37
37
  name: "@mendable/firecrawl-js",
38
- version: "4.18.2",
38
+ version: "4.18.4",
39
39
  description: "JavaScript SDK for Firecrawl API",
40
40
  main: "dist/index.js",
41
41
  types: "dist/index.d.ts",
@@ -105,7 +105,8 @@ var require_package = __commonJS({
105
105
  "picomatch@<4.0.4": ">=4.0.4",
106
106
  handlebars: ">=4.7.9",
107
107
  "brace-expansion": ">=5.0.5",
108
- "axios@<1.15.0": "1.15.0"
108
+ "axios@<1.15.0": "1.15.0",
109
+ "follow-redirects@<1.16.0": ">=1.16.0 <2.0.0"
109
110
  }
110
111
  }
111
112
  };
@@ -589,6 +590,7 @@ function prepareCrawlPayload(request) {
589
590
  if (request.includePaths) data.includePaths = request.includePaths;
590
591
  if (request.maxDiscoveryDepth != null) data.maxDiscoveryDepth = request.maxDiscoveryDepth;
591
592
  if (request.sitemap != null) data.sitemap = request.sitemap;
593
+ if (request.robotsUserAgent != null) data.robotsUserAgent = request.robotsUserAgent;
592
594
  if (request.ignoreQueryParameters != null) data.ignoreQueryParameters = request.ignoreQueryParameters;
593
595
  if (request.deduplicateSimilarURLs != null) data.deduplicateSimilarURLs = request.deduplicateSimilarURLs;
594
596
  if (request.limit != null) data.limit = request.limit;
package/dist/index.d.cts CHANGED
@@ -400,6 +400,8 @@ interface CrawlOptions {
400
400
  crawlEntireDomain?: boolean;
401
401
  allowExternalLinks?: boolean;
402
402
  allowSubdomains?: boolean;
403
+ ignoreRobotsTxt?: boolean;
404
+ robotsUserAgent?: string | null;
403
405
  delay?: number | null;
404
406
  maxConcurrency?: number | null;
405
407
  webhook?: string | WebhookConfig | null;
package/dist/index.d.ts CHANGED
@@ -400,6 +400,8 @@ interface CrawlOptions {
400
400
  crawlEntireDomain?: boolean;
401
401
  allowExternalLinks?: boolean;
402
402
  allowSubdomains?: boolean;
403
+ ignoreRobotsTxt?: boolean;
404
+ robotsUserAgent?: string | null;
403
405
  delay?: number | null;
404
406
  maxConcurrency?: number | null;
405
407
  webhook?: string | WebhookConfig | null;
package/dist/index.js CHANGED
@@ -1,6 +1,6 @@
1
1
  import {
2
2
  require_package
3
- } from "./chunk-QU5TGKGP.js";
3
+ } from "./chunk-ZUJQPZTD.js";
4
4
 
5
5
  // src/v2/utils/httpClient.ts
6
6
  import axios from "axios";
@@ -466,6 +466,7 @@ function prepareCrawlPayload(request) {
466
466
  if (request.includePaths) data.includePaths = request.includePaths;
467
467
  if (request.maxDiscoveryDepth != null) data.maxDiscoveryDepth = request.maxDiscoveryDepth;
468
468
  if (request.sitemap != null) data.sitemap = request.sitemap;
469
+ if (request.robotsUserAgent != null) data.robotsUserAgent = request.robotsUserAgent;
469
470
  if (request.ignoreQueryParameters != null) data.ignoreQueryParameters = request.ignoreQueryParameters;
470
471
  if (request.deduplicateSimilarURLs != null) data.deduplicateSimilarURLs = request.deduplicateSimilarURLs;
471
472
  if (request.limit != null) data.limit = request.limit;
@@ -1559,7 +1560,7 @@ var FirecrawlApp = class {
1559
1560
  if (typeof process !== "undefined" && process.env && process.env.npm_package_version) {
1560
1561
  return process.env.npm_package_version;
1561
1562
  }
1562
- const packageJson = await import("./package-ZEW57A2P.js");
1563
+ const packageJson = await import("./package-43GY3VT3.js");
1563
1564
  return packageJson.default.version;
1564
1565
  } catch (error) {
1565
1566
  const isTest = typeof process !== "undefined" && (process.env.JEST_WORKER_ID != null || false);
@@ -1,4 +1,4 @@
1
1
  import {
2
2
  require_package
3
- } from "./chunk-QU5TGKGP.js";
3
+ } from "./chunk-ZUJQPZTD.js";
4
4
  export default require_package();
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "firecrawl",
3
- "version": "4.18.2",
3
+ "version": "4.18.4",
4
4
  "description": "JavaScript SDK for Firecrawl API",
5
5
  "main": "dist/index.js",
6
6
  "types": "dist/index.d.ts",
@@ -27,6 +27,7 @@ function prepareCrawlPayload(request: CrawlRequest): Record<string, unknown> {
27
27
  if (request.includePaths) data.includePaths = request.includePaths;
28
28
  if (request.maxDiscoveryDepth != null) data.maxDiscoveryDepth = request.maxDiscoveryDepth;
29
29
  if (request.sitemap != null) data.sitemap = request.sitemap;
30
+ if (request.robotsUserAgent != null) data.robotsUserAgent = request.robotsUserAgent;
30
31
  if (request.ignoreQueryParameters != null) data.ignoreQueryParameters = request.ignoreQueryParameters;
31
32
  if (request.deduplicateSimilarURLs != null) data.deduplicateSimilarURLs = request.deduplicateSimilarURLs;
32
33
  if (request.limit != null) data.limit = request.limit;
package/src/v2/types.ts CHANGED
@@ -482,6 +482,8 @@ export interface CrawlOptions {
482
482
  crawlEntireDomain?: boolean;
483
483
  allowExternalLinks?: boolean;
484
484
  allowSubdomains?: boolean;
485
+ ignoreRobotsTxt?: boolean;
486
+ robotsUserAgent?: string | null;
485
487
  delay?: number | null;
486
488
  maxConcurrency?: number | null;
487
489
  webhook?: string | WebhookConfig | null;