@crawlee/core 4.0.0-beta.5 → 4.0.0-beta.50
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +9 -5
- package/autoscaling/autoscaled_pool.d.ts +3 -5
- package/autoscaling/autoscaled_pool.d.ts.map +1 -1
- package/autoscaling/autoscaled_pool.js +3 -9
- package/autoscaling/autoscaled_pool.js.map +1 -1
- package/autoscaling/snapshotter.d.ts +3 -13
- package/autoscaling/snapshotter.d.ts.map +1 -1
- package/autoscaling/snapshotter.js +18 -29
- package/autoscaling/snapshotter.js.map +1 -1
- package/autoscaling/system_status.d.ts +0 -3
- package/autoscaling/system_status.d.ts.map +1 -1
- package/autoscaling/system_status.js +2 -3
- package/autoscaling/system_status.js.map +1 -1
- package/configuration.d.ts +85 -227
- package/configuration.d.ts.map +1 -1
- package/configuration.js +159 -223
- package/configuration.js.map +1 -1
- package/cookie_utils.d.ts +1 -1
- package/cookie_utils.d.ts.map +1 -1
- package/cookie_utils.js +8 -8
- package/cookie_utils.js.map +1 -1
- package/crawlers/context_pipeline.d.ts +71 -0
- package/crawlers/context_pipeline.d.ts.map +1 -0
- package/crawlers/context_pipeline.js +123 -0
- package/crawlers/context_pipeline.js.map +1 -0
- package/crawlers/crawler_commons.d.ts +19 -27
- package/crawlers/crawler_commons.d.ts.map +1 -1
- package/crawlers/crawler_commons.js +12 -20
- package/crawlers/crawler_commons.js.map +1 -1
- package/crawlers/crawler_utils.js +1 -1
- package/crawlers/crawler_utils.js.map +1 -1
- package/crawlers/error_snapshotter.d.ts +3 -2
- package/crawlers/error_snapshotter.d.ts.map +1 -1
- package/crawlers/error_snapshotter.js +2 -2
- package/crawlers/error_snapshotter.js.map +1 -1
- package/crawlers/error_tracker.d.ts +2 -1
- package/crawlers/error_tracker.d.ts.map +1 -1
- package/crawlers/error_tracker.js.map +1 -1
- package/crawlers/index.d.ts +1 -1
- package/crawlers/index.d.ts.map +1 -1
- package/crawlers/index.js +1 -1
- package/crawlers/index.js.map +1 -1
- package/crawlers/internals/types.d.ts +8 -0
- package/crawlers/internals/types.d.ts.map +1 -0
- package/crawlers/internals/types.js +2 -0
- package/crawlers/internals/types.js.map +1 -0
- package/crawlers/statistics.d.ts +15 -15
- package/crawlers/statistics.d.ts.map +1 -1
- package/crawlers/statistics.js +21 -18
- package/crawlers/statistics.js.map +1 -1
- package/enqueue_links/enqueue_links.d.ts +32 -18
- package/enqueue_links/enqueue_links.d.ts.map +1 -1
- package/enqueue_links/enqueue_links.js +45 -24
- package/enqueue_links/enqueue_links.js.map +1 -1
- package/enqueue_links/shared.d.ts +25 -8
- package/enqueue_links/shared.d.ts.map +1 -1
- package/enqueue_links/shared.js +69 -37
- package/enqueue_links/shared.js.map +1 -1
- package/errors.d.ts +30 -0
- package/errors.d.ts.map +1 -1
- package/errors.js +44 -0
- package/errors.js.map +1 -1
- package/events/event_manager.d.ts +8 -5
- package/events/event_manager.d.ts.map +1 -1
- package/events/event_manager.js +7 -9
- package/events/event_manager.js.map +1 -1
- package/events/local_event_manager.d.ts +14 -4
- package/events/local_event_manager.d.ts.map +1 -1
- package/events/local_event_manager.js +33 -39
- package/events/local_event_manager.js.map +1 -1
- package/index.d.ts +3 -2
- package/index.d.ts.map +1 -1
- package/index.js +2 -1
- package/index.js.map +1 -1
- package/log.d.ts +82 -2
- package/log.d.ts.map +1 -1
- package/log.js +102 -0
- package/log.js.map +1 -1
- package/package.json +9 -10
- package/proxy_configuration.d.ts +14 -148
- package/proxy_configuration.d.ts.map +1 -1
- package/proxy_configuration.js +19 -167
- package/proxy_configuration.js.map +1 -1
- package/recoverable_state.d.ts +121 -0
- package/recoverable_state.d.ts.map +1 -0
- package/recoverable_state.js +142 -0
- package/recoverable_state.js.map +1 -0
- package/request.d.ts +74 -8
- package/request.d.ts.map +1 -1
- package/request.js +87 -18
- package/request.js.map +1 -1
- package/router.d.ts.map +1 -1
- package/router.js.map +1 -1
- package/serialization.js +1 -1
- package/serialization.js.map +1 -1
- package/service_locator.d.ts +162 -0
- package/service_locator.d.ts.map +1 -0
- package/service_locator.js +247 -0
- package/service_locator.js.map +1 -0
- package/session_pool/session.d.ts +9 -31
- package/session_pool/session.d.ts.map +1 -1
- package/session_pool/session.js +17 -21
- package/session_pool/session.js.map +1 -1
- package/session_pool/session_pool.d.ts +51 -57
- package/session_pool/session_pool.d.ts.map +1 -1
- package/session_pool/session_pool.js +86 -79
- package/session_pool/session_pool.js.map +1 -1
- package/storages/dataset.d.ts +63 -19
- package/storages/dataset.d.ts.map +1 -1
- package/storages/dataset.js +86 -22
- package/storages/dataset.js.map +1 -1
- package/storages/index.d.ts +2 -0
- package/storages/index.d.ts.map +1 -1
- package/storages/index.js +2 -0
- package/storages/index.js.map +1 -1
- package/storages/key_value_store.d.ts +79 -10
- package/storages/key_value_store.d.ts.map +1 -1
- package/storages/key_value_store.js +104 -23
- package/storages/key_value_store.js.map +1 -1
- package/storages/request_list.d.ts +9 -9
- package/storages/request_list.d.ts.map +1 -1
- package/storages/request_list.js +13 -8
- package/storages/request_list.js.map +1 -1
- package/storages/request_list_adapter.d.ts +58 -0
- package/storages/request_list_adapter.d.ts.map +1 -0
- package/storages/request_list_adapter.js +81 -0
- package/storages/request_list_adapter.js.map +1 -0
- package/storages/request_manager_tandem.d.ts +68 -0
- package/storages/request_manager_tandem.d.ts.map +1 -0
- package/storages/request_manager_tandem.js +124 -0
- package/storages/request_manager_tandem.js.map +1 -0
- package/storages/request_provider.d.ts +86 -23
- package/storages/request_provider.d.ts.map +1 -1
- package/storages/request_provider.js +114 -74
- package/storages/request_provider.js.map +1 -1
- package/storages/request_queue.d.ts +1 -3
- package/storages/request_queue.d.ts.map +1 -1
- package/storages/request_queue.js +2 -4
- package/storages/request_queue.js.map +1 -1
- package/storages/request_queue_v2.d.ts +3 -3
- package/storages/request_queue_v2.d.ts.map +1 -1
- package/storages/request_queue_v2.js +4 -5
- package/storages/request_queue_v2.js.map +1 -1
- package/storages/sitemap_request_list.d.ts +5 -5
- package/storages/sitemap_request_list.d.ts.map +1 -1
- package/storages/sitemap_request_list.js +10 -7
- package/storages/sitemap_request_list.js.map +1 -1
- package/storages/storage_manager.d.ts +22 -17
- package/storages/storage_manager.d.ts.map +1 -1
- package/storages/storage_manager.js +58 -52
- package/storages/storage_manager.js.map +1 -1
- package/storages/utils.d.ts.map +1 -1
- package/storages/utils.js +6 -5
- package/storages/utils.js.map +1 -1
- package/typedefs.d.ts +1 -1
- package/typedefs.d.ts.map +1 -1
- package/crawlers/crawler_extension.d.ts +0 -12
- package/crawlers/crawler_extension.d.ts.map +0 -1
- package/crawlers/crawler_extension.js +0 -14
- package/crawlers/crawler_extension.js.map +0 -1
- package/http_clients/base-http-client.d.ts +0 -134
- package/http_clients/base-http-client.d.ts.map +0 -1
- package/http_clients/base-http-client.js +0 -33
- package/http_clients/base-http-client.js.map +0 -1
- package/http_clients/form-data-like.d.ts +0 -67
- package/http_clients/form-data-like.d.ts.map +0 -1
- package/http_clients/form-data-like.js +0 -5
- package/http_clients/form-data-like.js.map +0 -1
- package/http_clients/got-scraping-http-client.d.ts +0 -15
- package/http_clients/got-scraping-http-client.d.ts.map +0 -1
- package/http_clients/got-scraping-http-client.js +0 -69
- package/http_clients/got-scraping-http-client.js.map +0 -1
- package/http_clients/index.d.ts +0 -3
- package/http_clients/index.d.ts.map +0 -1
- package/http_clients/index.js +0 -3
- package/http_clients/index.js.map +0 -1
- package/tsconfig.build.tsbuildinfo +0 -1
package/log.d.ts
CHANGED
|
@@ -1,3 +1,83 @@
|
|
|
1
|
-
import
|
|
2
|
-
|
|
1
|
+
import type { CrawleeLogger, CrawleeLoggerOptions } from '@crawlee/types';
|
|
2
|
+
import type { LoggerOptions } from '@apify/log';
|
|
3
|
+
import log, { Log, Logger, LoggerJson, LoggerText, LogLevel } from '@apify/log';
|
|
4
|
+
export type { CrawleeLogger, CrawleeLoggerOptions };
|
|
5
|
+
/**
|
|
6
|
+
* Abstract base class for custom Crawlee logger implementations.
|
|
7
|
+
*
|
|
8
|
+
* Subclasses must implement two methods:
|
|
9
|
+
* - {@link BaseCrawleeLogger.logWithLevel} — the core logging dispatch
|
|
10
|
+
* - {@link BaseCrawleeLogger.createChild} — how to create a child logger instance
|
|
11
|
+
*
|
|
12
|
+
* All other `CrawleeLogger` methods (`error`, `warning`, `info`, `debug`, etc.)
|
|
13
|
+
* are derived automatically. Level filtering is entirely the responsibility of the
|
|
14
|
+
* underlying library — `logWithLevel()` is called for every message.
|
|
15
|
+
*
|
|
16
|
+
* **Example — Winston adapter:**
|
|
17
|
+
* ```typescript
|
|
18
|
+
* const CRAWLEE_TO_WINSTON = { 1: 'error', 2: 'warn', 3: 'warn', 4: 'info', 5: 'debug', 6: 'debug' };
|
|
19
|
+
*
|
|
20
|
+
* class WinstonAdapter extends BaseCrawleeLogger {
|
|
21
|
+
* constructor(private logger: winston.Logger, options?: Partial<CrawleeLoggerOptions>) {
|
|
22
|
+
* super(options);
|
|
23
|
+
* }
|
|
24
|
+
*
|
|
25
|
+
* logWithLevel(level: number, message: string, data?: Record<string, unknown>): void {
|
|
26
|
+
* this.logger.log(CRAWLEE_TO_WINSTON[level] ?? 'info', message, data);
|
|
27
|
+
* }
|
|
28
|
+
*
|
|
29
|
+
* protected createChild(options: Partial<CrawleeLoggerOptions>): CrawleeLogger {
|
|
30
|
+
* return new WinstonAdapter(this.logger.child({ prefix: options.prefix }), { ...this.getOptions(), ...options });
|
|
31
|
+
* }
|
|
32
|
+
* }
|
|
33
|
+
* ```
|
|
34
|
+
*/
|
|
35
|
+
export declare abstract class BaseCrawleeLogger implements CrawleeLogger {
|
|
36
|
+
private options;
|
|
37
|
+
private readonly warningsLogged;
|
|
38
|
+
constructor(options?: Partial<CrawleeLoggerOptions>);
|
|
39
|
+
/**
|
|
40
|
+
* Core logging method. Subclasses must implement this to dispatch log messages
|
|
41
|
+
* to the underlying logger (Winston, Pino, console, etc.).
|
|
42
|
+
*
|
|
43
|
+
* Level filtering is the responsibility of the underlying library — this method
|
|
44
|
+
* is called for every message regardless of the current level.
|
|
45
|
+
*
|
|
46
|
+
* @param level Crawlee log level (use {@link LogLevel} constants)
|
|
47
|
+
* @param message The log message
|
|
48
|
+
* @param data Optional structured data to attach to the log entry
|
|
49
|
+
*/
|
|
50
|
+
abstract logWithLevel(level: number, message: string, data?: Record<string, unknown>): void;
|
|
51
|
+
/**
|
|
52
|
+
* Creates a child logger instance. Subclasses must implement this to define
|
|
53
|
+
* how child loggers are created for the underlying logger.
|
|
54
|
+
*/
|
|
55
|
+
protected abstract createChild(options: Partial<CrawleeLoggerOptions>): CrawleeLogger;
|
|
56
|
+
getOptions(): CrawleeLoggerOptions;
|
|
57
|
+
setOptions(options: Partial<CrawleeLoggerOptions>): void;
|
|
58
|
+
child(options: Partial<CrawleeLoggerOptions>): CrawleeLogger;
|
|
59
|
+
error(message: string, data?: Record<string, unknown>): void;
|
|
60
|
+
exception(exception: Error, message: string, data?: Record<string, unknown>): void;
|
|
61
|
+
softFail(message: string, data?: Record<string, unknown>): void;
|
|
62
|
+
warning(message: string, data?: Record<string, unknown>): void;
|
|
63
|
+
warningOnce(message: string): void;
|
|
64
|
+
info(message: string, data?: Record<string, unknown>): void;
|
|
65
|
+
debug(message: string, data?: Record<string, unknown>): void;
|
|
66
|
+
perf(message: string, data?: Record<string, unknown>): void;
|
|
67
|
+
deprecated(message: string): void;
|
|
68
|
+
}
|
|
69
|
+
/**
|
|
70
|
+
* Adapter that wraps `@apify/log`'s {@link Log} instance to implement the {@link CrawleeLogger} interface.
|
|
71
|
+
*
|
|
72
|
+
* This is the default logger used by Crawlee when no custom logger is configured.
|
|
73
|
+
* Users who want to use a different logging library should implement {@link BaseCrawleeLogger} directly.
|
|
74
|
+
*/
|
|
75
|
+
export declare class ApifyLogAdapter extends BaseCrawleeLogger {
|
|
76
|
+
private readonly apifyLog;
|
|
77
|
+
constructor(apifyLog: Log, options?: Partial<CrawleeLoggerOptions>);
|
|
78
|
+
logWithLevel(level: number, message: string, data?: Record<string, unknown>): void;
|
|
79
|
+
protected createChild(options: Partial<CrawleeLoggerOptions>): CrawleeLogger;
|
|
80
|
+
}
|
|
81
|
+
export { log, Log, LogLevel, Logger, LoggerJson, LoggerText };
|
|
82
|
+
export type { LoggerOptions };
|
|
3
83
|
//# sourceMappingURL=log.d.ts.map
|
package/log.d.ts.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"log.d.ts","sourceRoot":"","sources":["../src/log.ts"],"names":[],"mappings":"AAAA,OAAO,GAAG,EAAE,EAAE,GAAG,EAAE,MAAM,EAAE,UAAU,EAAE,aAAa,EAAE,UAAU,EAAE,QAAQ,EAAE,MAAM,
|
|
1
|
+
{"version":3,"file":"log.d.ts","sourceRoot":"","sources":["../src/log.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,aAAa,EAAE,oBAAoB,EAAE,MAAM,gBAAgB,CAAC;AAE1E,OAAO,KAAK,EAAE,aAAa,EAAE,MAAM,YAAY,CAAC;AAChD,OAAO,GAAG,EAAE,EAAE,GAAG,EAAE,MAAM,EAAE,UAAU,EAAE,UAAU,EAAE,QAAQ,EAAE,MAAM,YAAY,CAAC;AAEhF,YAAY,EAAE,aAAa,EAAE,oBAAoB,EAAE,CAAC;AAEpD;;;;;;;;;;;;;;;;;;;;;;;;;;;;;GA6BG;AACH,8BAAsB,iBAAkB,YAAW,aAAa;IAC5D,OAAO,CAAC,OAAO,CAAuB;IACtC,OAAO,CAAC,QAAQ,CAAC,cAAc,CAAqB;gBAExC,OAAO,GAAE,OAAO,CAAC,oBAAoB,CAAM;IAIvD;;;;;;;;;;OAUG;IACH,QAAQ,CAAC,YAAY,CAAC,KAAK,EAAE,MAAM,EAAE,OAAO,EAAE,MAAM,EAAE,IAAI,CAAC,EAAE,MAAM,CAAC,MAAM,EAAE,OAAO,CAAC,GAAG,IAAI;IAE3F;;;OAGG;IACH,SAAS,CAAC,QAAQ,CAAC,WAAW,CAAC,OAAO,EAAE,OAAO,CAAC,oBAAoB,CAAC,GAAG,aAAa;IAErF,UAAU,IAAI,oBAAoB;IAIlC,UAAU,CAAC,OAAO,EAAE,OAAO,CAAC,oBAAoB,CAAC,GAAG,IAAI;IAIxD,KAAK,CAAC,OAAO,EAAE,OAAO,CAAC,oBAAoB,CAAC,GAAG,aAAa;IAI5D,KAAK,CAAC,OAAO,EAAE,MAAM,EAAE,IAAI,CAAC,EAAE,MAAM,CAAC,MAAM,EAAE,OAAO,CAAC,GAAG,IAAI;IAI5D,SAAS,CAAC,SAAS,EAAE,KAAK,EAAE,OAAO,EAAE,MAAM,EAAE,IAAI,CAAC,EAAE,MAAM,CAAC,MAAM,EAAE,OAAO,CAAC,GAAG,IAAI;IAQlF,QAAQ,CAAC,OAAO,EAAE,MAAM,EAAE,IAAI,CAAC,EAAE,MAAM,CAAC,MAAM,EAAE,OAAO,CAAC,GAAG,IAAI;IAI/D,OAAO,CAAC,OAAO,EAAE,MAAM,EAAE,IAAI,CAAC,EAAE,MAAM,CAAC,MAAM,EAAE,OAAO,CAAC,GAAG,IAAI;IAI9D,WAAW,CAAC,OAAO,EAAE,MAAM,GAAG,IAAI;IAOlC,IAAI,CAAC,OAAO,EAAE,MAAM,EAAE,IAAI,CAAC,EAAE,MAAM,CAAC,MAAM,EAAE,OAAO,CAAC,GAAG,IAAI;IAI3D,KAAK,CAAC,OAAO,EAAE,MAAM,EAAE,IAAI,CAAC,EAAE,MAAM,CAAC,MAAM,EAAE,OAAO,CAAC,GAAG,IAAI;IAI5D,IAAI,CAAC,OAAO,EAAE,MAAM,EAAE,IAAI,CAAC,EAAE,MAAM,CAAC,MAAM,EAAE,OAAO,CAAC,GAAG,IAAI;IAI3D,UAAU,CAAC,OAAO,EAAE,MAAM,GAAG,IAAI;CAGpC;AAED;;;;;GAKG;AACH,qBAAa,eAAgB,SAAQ,iBAAiB;IAE9C,OAAO,CAAC,QAAQ,CAAC,QAAQ;gBAAR,QAAQ,EAAE,GAAG,EAC9B,OAAO,CAAC,EAAE,OAAO,CAAC,oBAAoB,CAAC;IAK3C,YAAY,CAAC,KAAK,EAAE,MAAM,EAAE,OAAO,EAAE,MAAM,EAAE,IAAI,CAAC,EAAE,MAAM,CAAC,MAAM,EAAE,OAAO,CAAC,GAAG,IAAI;IAIlF,SAAS,CAAC,WAAW,CAAC,OAAO,EAAE,OAAO,CAAC,oBAAoB,CAAC,GAAG,aAAa;CAM/E;AAED,OAAO,EAAE,GAAG,EAAE,GAAG,EAAE,QAAQ,EAAE,MAAM,EAAE,UAAU,EAAE,UAAU,EAAE,CAAC;AAC9D,YAAY,EAAE,aAAa,EAAE,CAAC"}
|
package/log.js
CHANGED
|
@@ -1,3 +1,105 @@
|
|
|
1
1
|
import log, { Log, Logger, LoggerJson, LoggerText, LogLevel } from '@apify/log';
|
|
2
|
+
/**
|
|
3
|
+
* Abstract base class for custom Crawlee logger implementations.
|
|
4
|
+
*
|
|
5
|
+
* Subclasses must implement two methods:
|
|
6
|
+
* - {@link BaseCrawleeLogger.logWithLevel} — the core logging dispatch
|
|
7
|
+
* - {@link BaseCrawleeLogger.createChild} — how to create a child logger instance
|
|
8
|
+
*
|
|
9
|
+
* All other `CrawleeLogger` methods (`error`, `warning`, `info`, `debug`, etc.)
|
|
10
|
+
* are derived automatically. Level filtering is entirely the responsibility of the
|
|
11
|
+
* underlying library — `logWithLevel()` is called for every message.
|
|
12
|
+
*
|
|
13
|
+
* **Example — Winston adapter:**
|
|
14
|
+
* ```typescript
|
|
15
|
+
* const CRAWLEE_TO_WINSTON = { 1: 'error', 2: 'warn', 3: 'warn', 4: 'info', 5: 'debug', 6: 'debug' };
|
|
16
|
+
*
|
|
17
|
+
* class WinstonAdapter extends BaseCrawleeLogger {
|
|
18
|
+
* constructor(private logger: winston.Logger, options?: Partial<CrawleeLoggerOptions>) {
|
|
19
|
+
* super(options);
|
|
20
|
+
* }
|
|
21
|
+
*
|
|
22
|
+
* logWithLevel(level: number, message: string, data?: Record<string, unknown>): void {
|
|
23
|
+
* this.logger.log(CRAWLEE_TO_WINSTON[level] ?? 'info', message, data);
|
|
24
|
+
* }
|
|
25
|
+
*
|
|
26
|
+
* protected createChild(options: Partial<CrawleeLoggerOptions>): CrawleeLogger {
|
|
27
|
+
* return new WinstonAdapter(this.logger.child({ prefix: options.prefix }), { ...this.getOptions(), ...options });
|
|
28
|
+
* }
|
|
29
|
+
* }
|
|
30
|
+
* ```
|
|
31
|
+
*/
|
|
32
|
+
export class BaseCrawleeLogger {
|
|
33
|
+
options;
|
|
34
|
+
warningsLogged = new Set();
|
|
35
|
+
constructor(options = {}) {
|
|
36
|
+
this.options = options;
|
|
37
|
+
}
|
|
38
|
+
getOptions() {
|
|
39
|
+
return this.options;
|
|
40
|
+
}
|
|
41
|
+
setOptions(options) {
|
|
42
|
+
this.options = { ...this.options, ...options };
|
|
43
|
+
}
|
|
44
|
+
child(options) {
|
|
45
|
+
return this.createChild(options);
|
|
46
|
+
}
|
|
47
|
+
error(message, data) {
|
|
48
|
+
this.logWithLevel(LogLevel.ERROR, message, data);
|
|
49
|
+
}
|
|
50
|
+
exception(exception, message, data) {
|
|
51
|
+
this.logWithLevel(LogLevel.ERROR, `${message}: ${exception.message}`, {
|
|
52
|
+
...data,
|
|
53
|
+
stack: exception.stack,
|
|
54
|
+
exception,
|
|
55
|
+
});
|
|
56
|
+
}
|
|
57
|
+
softFail(message, data) {
|
|
58
|
+
this.logWithLevel(LogLevel.SOFT_FAIL, message, data);
|
|
59
|
+
}
|
|
60
|
+
warning(message, data) {
|
|
61
|
+
this.logWithLevel(LogLevel.WARNING, message, data);
|
|
62
|
+
}
|
|
63
|
+
warningOnce(message) {
|
|
64
|
+
if (!this.warningsLogged.has(message)) {
|
|
65
|
+
this.warningsLogged.add(message);
|
|
66
|
+
this.warning(message);
|
|
67
|
+
}
|
|
68
|
+
}
|
|
69
|
+
info(message, data) {
|
|
70
|
+
this.logWithLevel(LogLevel.INFO, message, data);
|
|
71
|
+
}
|
|
72
|
+
debug(message, data) {
|
|
73
|
+
this.logWithLevel(LogLevel.DEBUG, message, data);
|
|
74
|
+
}
|
|
75
|
+
perf(message, data) {
|
|
76
|
+
this.logWithLevel(LogLevel.PERF, `[PERF] ${message}`, data);
|
|
77
|
+
}
|
|
78
|
+
deprecated(message) {
|
|
79
|
+
this.warningOnce(`[DEPRECATED] ${message}`);
|
|
80
|
+
}
|
|
81
|
+
}
|
|
82
|
+
/**
|
|
83
|
+
* Adapter that wraps `@apify/log`'s {@link Log} instance to implement the {@link CrawleeLogger} interface.
|
|
84
|
+
*
|
|
85
|
+
* This is the default logger used by Crawlee when no custom logger is configured.
|
|
86
|
+
* Users who want to use a different logging library should implement {@link BaseCrawleeLogger} directly.
|
|
87
|
+
*/
|
|
88
|
+
export class ApifyLogAdapter extends BaseCrawleeLogger {
|
|
89
|
+
apifyLog;
|
|
90
|
+
constructor(apifyLog, options) {
|
|
91
|
+
super(options ?? {});
|
|
92
|
+
this.apifyLog = apifyLog;
|
|
93
|
+
}
|
|
94
|
+
logWithLevel(level, message, data) {
|
|
95
|
+
this.apifyLog.internal(level, message, data);
|
|
96
|
+
}
|
|
97
|
+
createChild(options) {
|
|
98
|
+
return new ApifyLogAdapter(this.apifyLog.child({ prefix: options.prefix ?? null }), {
|
|
99
|
+
...this.getOptions(),
|
|
100
|
+
...options,
|
|
101
|
+
});
|
|
102
|
+
}
|
|
103
|
+
}
|
|
2
104
|
export { log, Log, LogLevel, Logger, LoggerJson, LoggerText };
|
|
3
105
|
//# sourceMappingURL=log.js.map
|
package/log.js.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"log.js","sourceRoot":"","sources":["../src/log.ts"],"names":[],"mappings":"
|
|
1
|
+
{"version":3,"file":"log.js","sourceRoot":"","sources":["../src/log.ts"],"names":[],"mappings":"AAGA,OAAO,GAAG,EAAE,EAAE,GAAG,EAAE,MAAM,EAAE,UAAU,EAAE,UAAU,EAAE,QAAQ,EAAE,MAAM,YAAY,CAAC;AAIhF;;;;;;;;;;;;;;;;;;;;;;;;;;;;;GA6BG;AACH,MAAM,OAAgB,iBAAiB;IAC3B,OAAO,CAAuB;IACrB,cAAc,GAAG,IAAI,GAAG,EAAU,CAAC;IAEpD,YAAY,UAAyC,EAAE;QACnD,IAAI,CAAC,OAAO,GAAG,OAAO,CAAC;IAC3B,CAAC;IAqBD,UAAU;QACN,OAAO,IAAI,CAAC,OAAO,CAAC;IACxB,CAAC;IAED,UAAU,CAAC,OAAsC;QAC7C,IAAI,CAAC,OAAO,GAAG,EAAE,GAAG,IAAI,CAAC,OAAO,EAAE,GAAG,OAAO,EAAE,CAAC;IACnD,CAAC;IAED,KAAK,CAAC,OAAsC;QACxC,OAAO,IAAI,CAAC,WAAW,CAAC,OAAO,CAAC,CAAC;IACrC,CAAC;IAED,KAAK,CAAC,OAAe,EAAE,IAA8B;QACjD,IAAI,CAAC,YAAY,CAAC,QAAQ,CAAC,KAAK,EAAE,OAAO,EAAE,IAAI,CAAC,CAAC;IACrD,CAAC;IAED,SAAS,CAAC,SAAgB,EAAE,OAAe,EAAE,IAA8B;QACvE,IAAI,CAAC,YAAY,CAAC,QAAQ,CAAC,KAAK,EAAE,GAAG,OAAO,KAAK,SAAS,CAAC,OAAO,EAAE,EAAE;YAClE,GAAG,IAAI;YACP,KAAK,EAAE,SAAS,CAAC,KAAK;YACtB,SAAS;SACZ,CAAC,CAAC;IACP,CAAC;IAED,QAAQ,CAAC,OAAe,EAAE,IAA8B;QACpD,IAAI,CAAC,YAAY,CAAC,QAAQ,CAAC,SAAS,EAAE,OAAO,EAAE,IAAI,CAAC,CAAC;IACzD,CAAC;IAED,OAAO,CAAC,OAAe,EAAE,IAA8B;QACnD,IAAI,CAAC,YAAY,CAAC,QAAQ,CAAC,OAAO,EAAE,OAAO,EAAE,IAAI,CAAC,CAAC;IACvD,CAAC;IAED,WAAW,CAAC,OAAe;QACvB,IAAI,CAAC,IAAI,CAAC,cAAc,CAAC,GAAG,CAAC,OAAO,CAAC,EAAE,CAAC;YACpC,IAAI,CAAC,cAAc,CAAC,GAAG,CAAC,OAAO,CAAC,CAAC;YACjC,IAAI,CAAC,OAAO,CAAC,OAAO,CAAC,CAAC;QAC1B,CAAC;IACL,CAAC;IAED,IAAI,CAAC,OAAe,EAAE,IAA8B;QAChD,IAAI,CAAC,YAAY,CAAC,QAAQ,CAAC,IAAI,EAAE,OAAO,EAAE,IAAI,CAAC,CAAC;IACpD,CAAC;IAED,KAAK,CAAC,OAAe,EAAE,IAA8B;QACjD,IAAI,CAAC,YAAY,CAAC,QAAQ,CAAC,KAAK,EAAE,OAAO,EAAE,IAAI,CAAC,CAAC;IACrD,CAAC;IAED,IAAI,CAAC,OAAe,EAAE,IAA8B;QAChD,IAAI,CAAC,YAAY,CAAC,QAAQ,CAAC,IAAI,EAAE,UAAU,OAAO,EAAE,EAAE,IAAI,CAAC,CAAC;IAChE,CAAC;IAED,UAAU,CAAC,OAAe;QACtB,IAAI,CAAC,WAAW,CAAC,gBAAgB,OAAO,EAAE,CAAC,CAAC;IAChD,CAAC;CACJ;AAED;;;;;GAKG;AACH,MAAM,OAAO,eAAgB,SAAQ,iBAAiB;IAE7B;IADrB,YACqB,QAAa,EAC9B,OAAuC;QAEvC,KAAK,CAAC,OAAO,IAAI,EAAE,CAAC,CAAC;QAHJ,aAAQ,GAAR,QAAQ,CAAK;IAIlC,CAAC;IAED,YAAY,CAAC,KAAa,EAAE,OAAe,EAAE,IAA8B;QACvE,IAAI,CAAC,QAAQ,CAAC,QAAQ,CAAC,KAAiB,EAAE,OAAO,EAAE,IAAI,CAAC,CAAC;IAC7D,CAAC;IAES,WAAW,CAAC,OAAsC;QACxD,OAAO,IAAI,eAAe,CAAC,IAAI,CAAC,QAAQ,CAAC,KAAK,CAAC,EAAE,MAAM,EAAE,OAAO,CAAC,MAAM,IAAI,IAAI,EAAE,CAAC,EAAE;YAChF,GAAG,IAAI,CAAC,UAAU,EAAE;YACpB,GAAG,OAAO;SACb,CAAC,CAAC;IACP,CAAC;CACJ;AAED,OAAO,EAAE,GAAG,EAAE,GAAG,EAAE,QAAQ,EAAE,MAAM,EAAE,UAAU,EAAE,UAAU,EAAE,CAAC"}
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@crawlee/core",
|
|
3
|
-
"version": "4.0.0-beta.
|
|
3
|
+
"version": "4.0.0-beta.50",
|
|
4
4
|
"description": "The scalable web crawling and scraping library for JavaScript/Node.js. Enables development of data extraction and web automation jobs (not only) with headless Chrome and Puppeteer.",
|
|
5
5
|
"engines": {
|
|
6
6
|
"node": ">=22.0.0"
|
|
@@ -38,7 +38,7 @@
|
|
|
38
38
|
},
|
|
39
39
|
"homepage": "https://crawlee.dev",
|
|
40
40
|
"scripts": {
|
|
41
|
-
"build": "
|
|
41
|
+
"build": "pnpm clean && pnpm compile && pnpm copy",
|
|
42
42
|
"clean": "rimraf ./dist",
|
|
43
43
|
"compile": "tsc -p tsconfig.build.json",
|
|
44
44
|
"copy": "tsx ../../scripts/copy.ts"
|
|
@@ -53,22 +53,21 @@
|
|
|
53
53
|
"@apify/pseudo_url": "^2.0.59",
|
|
54
54
|
"@apify/timeout": "^0.3.2",
|
|
55
55
|
"@apify/utilities": "^2.15.5",
|
|
56
|
-
"@crawlee/memory-storage": "4.0.0-beta.
|
|
57
|
-
"@crawlee/types": "4.0.0-beta.
|
|
58
|
-
"@crawlee/utils": "4.0.0-beta.
|
|
56
|
+
"@crawlee/memory-storage": "4.0.0-beta.50",
|
|
57
|
+
"@crawlee/types": "4.0.0-beta.50",
|
|
58
|
+
"@crawlee/utils": "4.0.0-beta.50",
|
|
59
59
|
"@sapphire/async-queue": "^1.5.5",
|
|
60
60
|
"@vladfrangu/async_event_emitter": "^2.4.6",
|
|
61
61
|
"csv-stringify": "^6.5.2",
|
|
62
|
-
"fs-extra": "^11.3.0",
|
|
63
|
-
"got-scraping": "^4.1.1",
|
|
64
62
|
"json5": "^2.2.3",
|
|
65
63
|
"minimatch": "^10.0.1",
|
|
66
64
|
"ow": "^2.0.0",
|
|
67
65
|
"stream-json": "^1.9.1",
|
|
68
66
|
"tldts": "^7.0.6",
|
|
69
|
-
"tough-cookie": "^
|
|
67
|
+
"tough-cookie": "^6.0.0",
|
|
70
68
|
"tslib": "^2.8.1",
|
|
71
|
-
"type-fest": "^4.41.0"
|
|
69
|
+
"type-fest": "^4.41.0",
|
|
70
|
+
"zod": "^3.24.0 || ^4.0.0"
|
|
72
71
|
},
|
|
73
72
|
"lerna": {
|
|
74
73
|
"command": {
|
|
@@ -77,5 +76,5 @@
|
|
|
77
76
|
}
|
|
78
77
|
}
|
|
79
78
|
},
|
|
80
|
-
"gitHead": "
|
|
79
|
+
"gitHead": "cdd082a1e9d106e6ea46fbe21928cc2f1978f4f7"
|
|
81
80
|
}
|
package/proxy_configuration.d.ts
CHANGED
|
@@ -1,129 +1,28 @@
|
|
|
1
|
+
import type { ProxyInfo } from '@crawlee/types';
|
|
1
2
|
import type { Request } from './request.js';
|
|
2
3
|
export interface ProxyConfigurationFunction {
|
|
3
|
-
(
|
|
4
|
+
(options?: {
|
|
4
5
|
request?: Request;
|
|
5
6
|
}): string | null | Promise<string | null>;
|
|
6
7
|
}
|
|
8
|
+
type UrlList = (string | null)[];
|
|
7
9
|
export interface ProxyConfigurationOptions {
|
|
8
10
|
/**
|
|
9
11
|
* An array of custom proxy URLs to be rotated.
|
|
10
12
|
* Custom proxies are not compatible with Apify Proxy and an attempt to use both
|
|
11
13
|
* configuration options will cause an error to be thrown on initialize.
|
|
12
14
|
*/
|
|
13
|
-
proxyUrls?:
|
|
15
|
+
proxyUrls?: UrlList;
|
|
14
16
|
/**
|
|
15
|
-
* Custom function that allows you to generate the new proxy URL dynamically. It gets
|
|
17
|
+
* Custom function that allows you to generate the new proxy URL dynamically. It gets an optional parameter with the `Request` object when applicable.
|
|
16
18
|
* Can return either stringified proxy URL or `null` if the proxy should not be used. Can be asynchronous.
|
|
17
19
|
*
|
|
18
20
|
* This function is used to generate the URL when {@link ProxyConfiguration.newUrl} or {@link ProxyConfiguration.newProxyInfo} is called.
|
|
19
21
|
*/
|
|
20
22
|
newUrlFunction?: ProxyConfigurationFunction;
|
|
21
|
-
/**
|
|
22
|
-
* An array of custom proxy URLs to be rotated stratified in tiers.
|
|
23
|
-
* This is a more advanced version of `proxyUrls` that allows you to define a hierarchy of proxy URLs
|
|
24
|
-
* If everything goes well, all the requests will be sent through the first proxy URL in the list.
|
|
25
|
-
* Whenever the crawler encounters a problem with the current proxy on the given domain, it will switch to the higher tier for this domain.
|
|
26
|
-
* The crawler probes lower-level proxies at intervals to check if it can make the tier downshift.
|
|
27
|
-
*
|
|
28
|
-
* This feature is useful when you have a set of proxies with different performance characteristics (speed, price, antibot performance etc.) and you want to use the best one for each domain.
|
|
29
|
-
*
|
|
30
|
-
* Use `null` as a proxy URL to disable the proxy for the given tier.
|
|
31
|
-
*/
|
|
32
|
-
tieredProxyUrls?: (string | null)[][];
|
|
33
23
|
}
|
|
34
|
-
|
|
35
|
-
proxyUrl: string | null;
|
|
36
|
-
proxyTier?: number;
|
|
37
|
-
}
|
|
38
|
-
/**
|
|
39
|
-
* The main purpose of the ProxyInfo object is to provide information
|
|
40
|
-
* about the current proxy connection used by the crawler for the request.
|
|
41
|
-
* Outside of crawlers, you can get this object by calling {@link ProxyConfiguration.newProxyInfo}.
|
|
42
|
-
*
|
|
43
|
-
* **Example usage:**
|
|
44
|
-
*
|
|
45
|
-
* ```javascript
|
|
46
|
-
* const proxyConfiguration = new ProxyConfiguration({
|
|
47
|
-
* proxyUrls: ['...', '...'] // List of Proxy URLs to rotate
|
|
48
|
-
* });
|
|
49
|
-
*
|
|
50
|
-
* // Getting proxyInfo object by calling class method directly
|
|
51
|
-
* const proxyInfo = await proxyConfiguration.newProxyInfo();
|
|
52
|
-
*
|
|
53
|
-
* // In crawler
|
|
54
|
-
* const crawler = new CheerioCrawler({
|
|
55
|
-
* // ...
|
|
56
|
-
* proxyConfiguration,
|
|
57
|
-
* requestHandler({ proxyInfo }) {
|
|
58
|
-
* // Getting used proxy URL
|
|
59
|
-
* const proxyUrl = proxyInfo.url;
|
|
60
|
-
*
|
|
61
|
-
* // Getting ID of used Session
|
|
62
|
-
* const sessionIdentifier = proxyInfo.sessionId;
|
|
63
|
-
* }
|
|
64
|
-
* })
|
|
65
|
-
*
|
|
66
|
-
* ```
|
|
67
|
-
*/
|
|
68
|
-
export interface ProxyInfo {
|
|
69
|
-
/**
|
|
70
|
-
* The identifier of used {@link Session}, if used.
|
|
71
|
-
*/
|
|
72
|
-
sessionId?: string;
|
|
73
|
-
/**
|
|
74
|
-
* The URL of the proxy.
|
|
75
|
-
*/
|
|
76
|
-
url: string;
|
|
77
|
-
/**
|
|
78
|
-
* Username for the proxy.
|
|
79
|
-
*/
|
|
80
|
-
username?: string;
|
|
81
|
-
/**
|
|
82
|
-
* User's password for the proxy.
|
|
83
|
-
*/
|
|
84
|
-
password: string;
|
|
85
|
-
/**
|
|
86
|
-
* Hostname of your proxy.
|
|
87
|
-
*/
|
|
88
|
-
hostname: string;
|
|
89
|
-
/**
|
|
90
|
-
* Proxy port.
|
|
91
|
-
*/
|
|
92
|
-
port: number | string;
|
|
93
|
-
/**
|
|
94
|
-
* Proxy tier for the current proxy, if applicable (only for `tieredProxyUrls`).
|
|
95
|
-
*/
|
|
96
|
-
proxyTier?: number;
|
|
97
|
-
}
|
|
98
|
-
interface TieredProxyOptions {
|
|
24
|
+
interface NewUrlOptions {
|
|
99
25
|
request?: Request;
|
|
100
|
-
proxyTier?: number;
|
|
101
|
-
}
|
|
102
|
-
/**
|
|
103
|
-
* Internal class for tracking the proxy tier history for a specific domain.
|
|
104
|
-
*
|
|
105
|
-
* Predicts the best proxy tier for the next request based on the error history for different proxy tiers.
|
|
106
|
-
*/
|
|
107
|
-
declare class ProxyTierTracker {
|
|
108
|
-
private histogram;
|
|
109
|
-
private currentTier;
|
|
110
|
-
constructor(tieredProxyUrls: (string | null)[][]);
|
|
111
|
-
/**
|
|
112
|
-
* Processes a single step of the algorithm and updates the current tier prediction based on the error history.
|
|
113
|
-
*/
|
|
114
|
-
private processStep;
|
|
115
|
-
/**
|
|
116
|
-
* Increases the error score for the given proxy tier. This raises the chance of picking a different proxy tier for the subsequent requests.
|
|
117
|
-
*
|
|
118
|
-
* The error score is increased by 10 for the given tier. This means that this tier will be disadvantaged for the next 10 requests (every new request prediction decreases the error score by 1).
|
|
119
|
-
* @param tier The proxy tier to mark as problematic.
|
|
120
|
-
*/
|
|
121
|
-
addError(tier: number): void;
|
|
122
|
-
/**
|
|
123
|
-
* Returns the best proxy tier for the next request based on the error history for different proxy tiers.
|
|
124
|
-
* @returns The proxy tier prediction
|
|
125
|
-
*/
|
|
126
|
-
predictTier(): number;
|
|
127
26
|
}
|
|
128
27
|
/**
|
|
129
28
|
* Configures connection to a proxy server with the provided options. Proxy servers are used to prevent target websites from blocking
|
|
@@ -156,13 +55,11 @@ declare class ProxyTierTracker {
|
|
|
156
55
|
export declare class ProxyConfiguration {
|
|
157
56
|
isManInTheMiddle: boolean;
|
|
158
57
|
protected nextCustomUrlIndex: number;
|
|
159
|
-
protected proxyUrls?:
|
|
160
|
-
protected
|
|
161
|
-
protected usedProxyUrls: Map<string, string>;
|
|
58
|
+
protected proxyUrls?: UrlList;
|
|
59
|
+
protected usedProxyUrls: Map<string, string | null>;
|
|
162
60
|
protected newUrlFunction?: ProxyConfigurationFunction;
|
|
163
61
|
// @ts-ignore optional peer dependency or compatibility with es2022
|
|
164
|
-
protected log: import("@
|
|
165
|
-
protected domainTiers: Map<string, ProxyTierTracker>;
|
|
62
|
+
protected log: import("@crawlee/types").CrawleeLogger;
|
|
166
63
|
/**
|
|
167
64
|
* Creates a {@link ProxyConfiguration} instance based on the provided options. Proxy servers are used to prevent target websites from
|
|
168
65
|
* blocking your crawlers based on IP address rate limits or blacklists. Setting proxy configuration in your crawlers automatically configures
|
|
@@ -190,53 +87,22 @@ export declare class ProxyConfiguration {
|
|
|
190
87
|
* the currently used proxy via the requestHandler parameter `proxyInfo`.
|
|
191
88
|
* Use it if you want to work with a rich representation of a proxy URL.
|
|
192
89
|
* If you need the URL string only, use {@link ProxyConfiguration.newUrl}.
|
|
193
|
-
* @param [sessionId]
|
|
194
|
-
* Represents the identifier of user {@link Session} that can be managed by the {@link SessionPool} or
|
|
195
|
-
* you can use the Apify Proxy [Session](https://docs.apify.com/proxy#sessions) identifier.
|
|
196
|
-
* When the provided sessionId is a number, it's converted to a string. Property sessionId of
|
|
197
|
-
* {@link ProxyInfo} is always returned as a type string.
|
|
198
90
|
*
|
|
199
|
-
* All the HTTP requests going through the proxy with the same session identifier
|
|
200
|
-
* will use the same target proxy server (i.e. the same IP address).
|
|
201
|
-
* The identifier must not be longer than 50 characters and include only the following: `0-9`, `a-z`, `A-Z`, `"."`, `"_"` and `"~"`.
|
|
202
91
|
* @return Represents information about used proxy and its configuration.
|
|
203
92
|
*/
|
|
204
|
-
newProxyInfo(
|
|
93
|
+
newProxyInfo(options?: NewUrlOptions): Promise<ProxyInfo | undefined>;
|
|
205
94
|
/**
|
|
206
|
-
*
|
|
207
|
-
* @param _sessionId Session identifier
|
|
208
|
-
* @param options Options for the tiered proxy rotation
|
|
209
|
-
* @returns An object with the proxy URL and the proxy tier used.
|
|
210
|
-
*/
|
|
211
|
-
protected _handleTieredUrl(_sessionId: string, options?: TieredProxyOptions): TieredProxy;
|
|
212
|
-
/**
|
|
213
|
-
* Given a `Request` object, this function returns the tier of the proxy that should be used for the request.
|
|
95
|
+
* Returns a new proxy URL based on provided configuration options.
|
|
214
96
|
*
|
|
215
|
-
* This returns `null` if `tieredProxyUrls` option is not set.
|
|
216
|
-
*/
|
|
217
|
-
protected predictProxyTier(request: Request): number | null;
|
|
218
|
-
/**
|
|
219
|
-
* Returns a new proxy URL based on provided configuration options and the `sessionId` parameter.
|
|
220
|
-
* @param [sessionId]
|
|
221
|
-
* Represents the identifier of user {@link Session} that can be managed by the {@link SessionPool} or
|
|
222
|
-
* you can use the Apify Proxy [Session](https://docs.apify.com/proxy#sessions) identifier.
|
|
223
|
-
* When the provided sessionId is a number, it's converted to a string.
|
|
224
|
-
*
|
|
225
|
-
* All the HTTP requests going through the proxy with the same session identifier
|
|
226
|
-
* will use the same target proxy server (i.e. the same IP address).
|
|
227
|
-
* The identifier must not be longer than 50 characters and include only the following: `0-9`, `a-z`, `A-Z`, `"."`, `"_"` and `"~"`.
|
|
228
97
|
* @return A string with a proxy URL, including authentication credentials and port number.
|
|
229
98
|
* For example, `http://bob:password123@proxy.example.com:8000`
|
|
230
99
|
*/
|
|
231
|
-
newUrl(
|
|
232
|
-
|
|
233
|
-
* Handles custom url rotation with session
|
|
234
|
-
*/
|
|
235
|
-
protected _handleCustomUrl(sessionId?: string): string;
|
|
100
|
+
newUrl(options?: NewUrlOptions): Promise<string | undefined>;
|
|
101
|
+
protected _handleProxyUrlsList(): string | null;
|
|
236
102
|
/**
|
|
237
103
|
* Calls the custom newUrlFunction and checks format of its return value
|
|
238
104
|
*/
|
|
239
|
-
protected _callNewUrlFunction(
|
|
105
|
+
protected _callNewUrlFunction(options?: {
|
|
240
106
|
request?: Request;
|
|
241
107
|
}): Promise<string | null>;
|
|
242
108
|
protected _throwCannotCombineCustomMethods(): never;
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"proxy_configuration.d.ts","sourceRoot":"","sources":["../src/proxy_configuration.ts"],"names":[],"mappings":"
|
|
1
|
+
{"version":3,"file":"proxy_configuration.d.ts","sourceRoot":"","sources":["../src/proxy_configuration.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAc,SAAS,EAAE,MAAM,gBAAgB,CAAC;AAG5D,OAAO,KAAK,EAAE,OAAO,EAAE,MAAM,cAAc,CAAC;AAG5C,MAAM,WAAW,0BAA0B;IACvC,CAAC,OAAO,CAAC,EAAE;QAAE,OAAO,CAAC,EAAE,OAAO,CAAA;KAAE,GAAG,MAAM,GAAG,IAAI,GAAG,OAAO,CAAC,MAAM,GAAG,IAAI,CAAC,CAAC;CAC7E;AAED,KAAK,OAAO,GAAG,CAAC,MAAM,GAAG,IAAI,CAAC,EAAE,CAAC;AAEjC,MAAM,WAAW,yBAAyB;IACtC;;;;OAIG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IAEpB;;;;;OAKG;IACH,cAAc,CAAC,EAAE,0BAA0B,CAAC;CAC/C;AAED,UAAU,aAAa;IACnB,OAAO,CAAC,EAAE,OAAO,CAAC;CACrB;AAED;;;;;;;;;;;;;;;;;;;;;;;;;;;GA2BG;AACH,qBAAa,kBAAkB;IAC3B,gBAAgB,UAAS;IACzB,SAAS,CAAC,kBAAkB,SAAK;IACjC,SAAS,CAAC,SAAS,CAAC,EAAE,OAAO,CAAC;IAC9B,SAAS,CAAC,aAAa,6BAAoC;IAC3D,SAAS,CAAC,cAAc,CAAC,EAAE,0BAA0B,CAAC;IACtD,SAAS,CAAC,GAAG,yCAAsE;IAEnF;;;;;;;;;;;;;;;;;;;OAmBG;gBACS,OAAO,GAAE,yBAA8B;IA2BnD;;;;;;;;OAQG;IACG,YAAY,CAAC,OAAO,CAAC,EAAE,aAAa,GAAG,OAAO,CAAC,SAAS,GAAG,SAAS,CAAC;IAe3E;;;;;OAKG;IACG,MAAM,CAAC,OAAO,CAAC,EAAE,aAAa,GAAG,OAAO,CAAC,MAAM,GAAG,SAAS,CAAC;IAQlE,SAAS,CAAC,oBAAoB,IAAI,MAAM,GAAG,IAAI;IAI/C;;OAEG;cACa,mBAAmB,CAAC,OAAO,CAAC,EAAE;QAAE,OAAO,CAAC,EAAE,OAAO,CAAA;KAAE;IAcnE,SAAS,CAAC,gCAAgC,IAAI,KAAK;IAMnD,SAAS,CAAC,uBAAuB,IAAI,KAAK;CAG7C"}
|