firecrawl 4.18.2 → 4.18.4
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +1 -1
- package/dist/{chunk-QU5TGKGP.js → chunk-ZUJQPZTD.js} +3 -2
- package/dist/index.cjs +4 -2
- package/dist/index.d.cts +2 -0
- package/dist/index.d.ts +2 -0
- package/dist/index.js +3 -2
- package/dist/{package-ZEW57A2P.js → package-43GY3VT3.js} +1 -1
- package/package.json +1 -1
- package/src/v2/methods/crawl.ts +1 -0
- package/src/v2/types.ts +2 -0
package/README.md
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Firecrawl Node SDK
|
|
2
2
|
|
|
3
|
-
The Firecrawl Node SDK is a library that allows you to easily scrape and
|
|
3
|
+
The Firecrawl Node SDK is a library that allows you to easily search, scrape, and interact with the web, and output the data in a format ready for use with language models (LLMs). It provides a simple and intuitive interface for the Firecrawl API.
|
|
4
4
|
|
|
5
5
|
## Installation
|
|
6
6
|
|
|
@@ -8,7 +8,7 @@ var require_package = __commonJS({
|
|
|
8
8
|
"package.json"(exports, module) {
|
|
9
9
|
module.exports = {
|
|
10
10
|
name: "@mendable/firecrawl-js",
|
|
11
|
-
version: "4.18.
|
|
11
|
+
version: "4.18.4",
|
|
12
12
|
description: "JavaScript SDK for Firecrawl API",
|
|
13
13
|
main: "dist/index.js",
|
|
14
14
|
types: "dist/index.d.ts",
|
|
@@ -78,7 +78,8 @@ var require_package = __commonJS({
|
|
|
78
78
|
"picomatch@<4.0.4": ">=4.0.4",
|
|
79
79
|
handlebars: ">=4.7.9",
|
|
80
80
|
"brace-expansion": ">=5.0.5",
|
|
81
|
-
"axios@<1.15.0": "1.15.0"
|
|
81
|
+
"axios@<1.15.0": "1.15.0",
|
|
82
|
+
"follow-redirects@<1.16.0": ">=1.16.0 <2.0.0"
|
|
82
83
|
}
|
|
83
84
|
}
|
|
84
85
|
};
|
package/dist/index.cjs
CHANGED
|
@@ -35,7 +35,7 @@ var require_package = __commonJS({
|
|
|
35
35
|
"package.json"(exports2, module2) {
|
|
36
36
|
module2.exports = {
|
|
37
37
|
name: "@mendable/firecrawl-js",
|
|
38
|
-
version: "4.18.
|
|
38
|
+
version: "4.18.4",
|
|
39
39
|
description: "JavaScript SDK for Firecrawl API",
|
|
40
40
|
main: "dist/index.js",
|
|
41
41
|
types: "dist/index.d.ts",
|
|
@@ -105,7 +105,8 @@ var require_package = __commonJS({
|
|
|
105
105
|
"picomatch@<4.0.4": ">=4.0.4",
|
|
106
106
|
handlebars: ">=4.7.9",
|
|
107
107
|
"brace-expansion": ">=5.0.5",
|
|
108
|
-
"axios@<1.15.0": "1.15.0"
|
|
108
|
+
"axios@<1.15.0": "1.15.0",
|
|
109
|
+
"follow-redirects@<1.16.0": ">=1.16.0 <2.0.0"
|
|
109
110
|
}
|
|
110
111
|
}
|
|
111
112
|
};
|
|
@@ -589,6 +590,7 @@ function prepareCrawlPayload(request) {
|
|
|
589
590
|
if (request.includePaths) data.includePaths = request.includePaths;
|
|
590
591
|
if (request.maxDiscoveryDepth != null) data.maxDiscoveryDepth = request.maxDiscoveryDepth;
|
|
591
592
|
if (request.sitemap != null) data.sitemap = request.sitemap;
|
|
593
|
+
if (request.robotsUserAgent != null) data.robotsUserAgent = request.robotsUserAgent;
|
|
592
594
|
if (request.ignoreQueryParameters != null) data.ignoreQueryParameters = request.ignoreQueryParameters;
|
|
593
595
|
if (request.deduplicateSimilarURLs != null) data.deduplicateSimilarURLs = request.deduplicateSimilarURLs;
|
|
594
596
|
if (request.limit != null) data.limit = request.limit;
|
package/dist/index.d.cts
CHANGED
|
@@ -400,6 +400,8 @@ interface CrawlOptions {
|
|
|
400
400
|
crawlEntireDomain?: boolean;
|
|
401
401
|
allowExternalLinks?: boolean;
|
|
402
402
|
allowSubdomains?: boolean;
|
|
403
|
+
ignoreRobotsTxt?: boolean;
|
|
404
|
+
robotsUserAgent?: string | null;
|
|
403
405
|
delay?: number | null;
|
|
404
406
|
maxConcurrency?: number | null;
|
|
405
407
|
webhook?: string | WebhookConfig | null;
|
package/dist/index.d.ts
CHANGED
|
@@ -400,6 +400,8 @@ interface CrawlOptions {
|
|
|
400
400
|
crawlEntireDomain?: boolean;
|
|
401
401
|
allowExternalLinks?: boolean;
|
|
402
402
|
allowSubdomains?: boolean;
|
|
403
|
+
ignoreRobotsTxt?: boolean;
|
|
404
|
+
robotsUserAgent?: string | null;
|
|
403
405
|
delay?: number | null;
|
|
404
406
|
maxConcurrency?: number | null;
|
|
405
407
|
webhook?: string | WebhookConfig | null;
|
package/dist/index.js
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
import {
|
|
2
2
|
require_package
|
|
3
|
-
} from "./chunk-
|
|
3
|
+
} from "./chunk-ZUJQPZTD.js";
|
|
4
4
|
|
|
5
5
|
// src/v2/utils/httpClient.ts
|
|
6
6
|
import axios from "axios";
|
|
@@ -466,6 +466,7 @@ function prepareCrawlPayload(request) {
|
|
|
466
466
|
if (request.includePaths) data.includePaths = request.includePaths;
|
|
467
467
|
if (request.maxDiscoveryDepth != null) data.maxDiscoveryDepth = request.maxDiscoveryDepth;
|
|
468
468
|
if (request.sitemap != null) data.sitemap = request.sitemap;
|
|
469
|
+
if (request.robotsUserAgent != null) data.robotsUserAgent = request.robotsUserAgent;
|
|
469
470
|
if (request.ignoreQueryParameters != null) data.ignoreQueryParameters = request.ignoreQueryParameters;
|
|
470
471
|
if (request.deduplicateSimilarURLs != null) data.deduplicateSimilarURLs = request.deduplicateSimilarURLs;
|
|
471
472
|
if (request.limit != null) data.limit = request.limit;
|
|
@@ -1559,7 +1560,7 @@ var FirecrawlApp = class {
|
|
|
1559
1560
|
if (typeof process !== "undefined" && process.env && process.env.npm_package_version) {
|
|
1560
1561
|
return process.env.npm_package_version;
|
|
1561
1562
|
}
|
|
1562
|
-
const packageJson = await import("./package-
|
|
1563
|
+
const packageJson = await import("./package-43GY3VT3.js");
|
|
1563
1564
|
return packageJson.default.version;
|
|
1564
1565
|
} catch (error) {
|
|
1565
1566
|
const isTest = typeof process !== "undefined" && (process.env.JEST_WORKER_ID != null || false);
|
package/package.json
CHANGED
package/src/v2/methods/crawl.ts
CHANGED
|
@@ -27,6 +27,7 @@ function prepareCrawlPayload(request: CrawlRequest): Record<string, unknown> {
|
|
|
27
27
|
if (request.includePaths) data.includePaths = request.includePaths;
|
|
28
28
|
if (request.maxDiscoveryDepth != null) data.maxDiscoveryDepth = request.maxDiscoveryDepth;
|
|
29
29
|
if (request.sitemap != null) data.sitemap = request.sitemap;
|
|
30
|
+
if (request.robotsUserAgent != null) data.robotsUserAgent = request.robotsUserAgent;
|
|
30
31
|
if (request.ignoreQueryParameters != null) data.ignoreQueryParameters = request.ignoreQueryParameters;
|
|
31
32
|
if (request.deduplicateSimilarURLs != null) data.deduplicateSimilarURLs = request.deduplicateSimilarURLs;
|
|
32
33
|
if (request.limit != null) data.limit = request.limit;
|
package/src/v2/types.ts
CHANGED
|
@@ -482,6 +482,8 @@ export interface CrawlOptions {
|
|
|
482
482
|
crawlEntireDomain?: boolean;
|
|
483
483
|
allowExternalLinks?: boolean;
|
|
484
484
|
allowSubdomains?: boolean;
|
|
485
|
+
ignoreRobotsTxt?: boolean;
|
|
486
|
+
robotsUserAgent?: string | null;
|
|
485
487
|
delay?: number | null;
|
|
486
488
|
maxConcurrency?: number | null;
|
|
487
489
|
webhook?: string | WebhookConfig | null;
|