@crawlee/core 4.0.0-beta.5 → 4.0.0-beta.51

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (177) hide show
  1. package/README.md +9 -5
  2. package/autoscaling/autoscaled_pool.d.ts +3 -5
  3. package/autoscaling/autoscaled_pool.d.ts.map +1 -1
  4. package/autoscaling/autoscaled_pool.js +3 -9
  5. package/autoscaling/autoscaled_pool.js.map +1 -1
  6. package/autoscaling/snapshotter.d.ts +3 -13
  7. package/autoscaling/snapshotter.d.ts.map +1 -1
  8. package/autoscaling/snapshotter.js +18 -29
  9. package/autoscaling/snapshotter.js.map +1 -1
  10. package/autoscaling/system_status.d.ts +0 -3
  11. package/autoscaling/system_status.d.ts.map +1 -1
  12. package/autoscaling/system_status.js +2 -3
  13. package/autoscaling/system_status.js.map +1 -1
  14. package/configuration.d.ts +85 -227
  15. package/configuration.d.ts.map +1 -1
  16. package/configuration.js +159 -223
  17. package/configuration.js.map +1 -1
  18. package/cookie_utils.d.ts +1 -1
  19. package/cookie_utils.d.ts.map +1 -1
  20. package/cookie_utils.js +8 -8
  21. package/cookie_utils.js.map +1 -1
  22. package/crawlers/context_pipeline.d.ts +71 -0
  23. package/crawlers/context_pipeline.d.ts.map +1 -0
  24. package/crawlers/context_pipeline.js +123 -0
  25. package/crawlers/context_pipeline.js.map +1 -0
  26. package/crawlers/crawler_commons.d.ts +19 -27
  27. package/crawlers/crawler_commons.d.ts.map +1 -1
  28. package/crawlers/crawler_commons.js +12 -20
  29. package/crawlers/crawler_commons.js.map +1 -1
  30. package/crawlers/crawler_utils.js +1 -1
  31. package/crawlers/crawler_utils.js.map +1 -1
  32. package/crawlers/error_snapshotter.d.ts +3 -2
  33. package/crawlers/error_snapshotter.d.ts.map +1 -1
  34. package/crawlers/error_snapshotter.js +2 -2
  35. package/crawlers/error_snapshotter.js.map +1 -1
  36. package/crawlers/error_tracker.d.ts +2 -1
  37. package/crawlers/error_tracker.d.ts.map +1 -1
  38. package/crawlers/error_tracker.js.map +1 -1
  39. package/crawlers/index.d.ts +1 -1
  40. package/crawlers/index.d.ts.map +1 -1
  41. package/crawlers/index.js +1 -1
  42. package/crawlers/index.js.map +1 -1
  43. package/crawlers/internals/types.d.ts +8 -0
  44. package/crawlers/internals/types.d.ts.map +1 -0
  45. package/crawlers/internals/types.js +2 -0
  46. package/crawlers/internals/types.js.map +1 -0
  47. package/crawlers/statistics.d.ts +15 -15
  48. package/crawlers/statistics.d.ts.map +1 -1
  49. package/crawlers/statistics.js +21 -18
  50. package/crawlers/statistics.js.map +1 -1
  51. package/enqueue_links/enqueue_links.d.ts +32 -18
  52. package/enqueue_links/enqueue_links.d.ts.map +1 -1
  53. package/enqueue_links/enqueue_links.js +45 -24
  54. package/enqueue_links/enqueue_links.js.map +1 -1
  55. package/enqueue_links/shared.d.ts +25 -8
  56. package/enqueue_links/shared.d.ts.map +1 -1
  57. package/enqueue_links/shared.js +69 -37
  58. package/enqueue_links/shared.js.map +1 -1
  59. package/errors.d.ts +30 -0
  60. package/errors.d.ts.map +1 -1
  61. package/errors.js +44 -0
  62. package/errors.js.map +1 -1
  63. package/events/event_manager.d.ts +8 -5
  64. package/events/event_manager.d.ts.map +1 -1
  65. package/events/event_manager.js +7 -9
  66. package/events/event_manager.js.map +1 -1
  67. package/events/local_event_manager.d.ts +14 -4
  68. package/events/local_event_manager.d.ts.map +1 -1
  69. package/events/local_event_manager.js +33 -39
  70. package/events/local_event_manager.js.map +1 -1
  71. package/index.d.ts +3 -2
  72. package/index.d.ts.map +1 -1
  73. package/index.js +2 -1
  74. package/index.js.map +1 -1
  75. package/log.d.ts +82 -2
  76. package/log.d.ts.map +1 -1
  77. package/log.js +102 -0
  78. package/log.js.map +1 -1
  79. package/package.json +9 -10
  80. package/proxy_configuration.d.ts +14 -148
  81. package/proxy_configuration.d.ts.map +1 -1
  82. package/proxy_configuration.js +19 -167
  83. package/proxy_configuration.js.map +1 -1
  84. package/recoverable_state.d.ts +121 -0
  85. package/recoverable_state.d.ts.map +1 -0
  86. package/recoverable_state.js +142 -0
  87. package/recoverable_state.js.map +1 -0
  88. package/request.d.ts +74 -8
  89. package/request.d.ts.map +1 -1
  90. package/request.js +87 -18
  91. package/request.js.map +1 -1
  92. package/router.d.ts.map +1 -1
  93. package/router.js.map +1 -1
  94. package/serialization.js +1 -1
  95. package/serialization.js.map +1 -1
  96. package/service_locator.d.ts +162 -0
  97. package/service_locator.d.ts.map +1 -0
  98. package/service_locator.js +247 -0
  99. package/service_locator.js.map +1 -0
  100. package/session_pool/session.d.ts +9 -31
  101. package/session_pool/session.d.ts.map +1 -1
  102. package/session_pool/session.js +17 -21
  103. package/session_pool/session.js.map +1 -1
  104. package/session_pool/session_pool.d.ts +51 -57
  105. package/session_pool/session_pool.d.ts.map +1 -1
  106. package/session_pool/session_pool.js +86 -79
  107. package/session_pool/session_pool.js.map +1 -1
  108. package/storages/dataset.d.ts +63 -19
  109. package/storages/dataset.d.ts.map +1 -1
  110. package/storages/dataset.js +86 -22
  111. package/storages/dataset.js.map +1 -1
  112. package/storages/index.d.ts +2 -0
  113. package/storages/index.d.ts.map +1 -1
  114. package/storages/index.js +2 -0
  115. package/storages/index.js.map +1 -1
  116. package/storages/key_value_store.d.ts +79 -10
  117. package/storages/key_value_store.d.ts.map +1 -1
  118. package/storages/key_value_store.js +104 -23
  119. package/storages/key_value_store.js.map +1 -1
  120. package/storages/request_list.d.ts +9 -9
  121. package/storages/request_list.d.ts.map +1 -1
  122. package/storages/request_list.js +13 -8
  123. package/storages/request_list.js.map +1 -1
  124. package/storages/request_list_adapter.d.ts +58 -0
  125. package/storages/request_list_adapter.d.ts.map +1 -0
  126. package/storages/request_list_adapter.js +81 -0
  127. package/storages/request_list_adapter.js.map +1 -0
  128. package/storages/request_manager_tandem.d.ts +68 -0
  129. package/storages/request_manager_tandem.d.ts.map +1 -0
  130. package/storages/request_manager_tandem.js +124 -0
  131. package/storages/request_manager_tandem.js.map +1 -0
  132. package/storages/request_provider.d.ts +86 -23
  133. package/storages/request_provider.d.ts.map +1 -1
  134. package/storages/request_provider.js +114 -74
  135. package/storages/request_provider.js.map +1 -1
  136. package/storages/request_queue.d.ts +1 -3
  137. package/storages/request_queue.d.ts.map +1 -1
  138. package/storages/request_queue.js +2 -4
  139. package/storages/request_queue.js.map +1 -1
  140. package/storages/request_queue_v2.d.ts +3 -3
  141. package/storages/request_queue_v2.d.ts.map +1 -1
  142. package/storages/request_queue_v2.js +4 -5
  143. package/storages/request_queue_v2.js.map +1 -1
  144. package/storages/sitemap_request_list.d.ts +5 -5
  145. package/storages/sitemap_request_list.d.ts.map +1 -1
  146. package/storages/sitemap_request_list.js +10 -7
  147. package/storages/sitemap_request_list.js.map +1 -1
  148. package/storages/storage_manager.d.ts +22 -17
  149. package/storages/storage_manager.d.ts.map +1 -1
  150. package/storages/storage_manager.js +58 -52
  151. package/storages/storage_manager.js.map +1 -1
  152. package/storages/utils.d.ts.map +1 -1
  153. package/storages/utils.js +6 -5
  154. package/storages/utils.js.map +1 -1
  155. package/typedefs.d.ts +1 -1
  156. package/typedefs.d.ts.map +1 -1
  157. package/crawlers/crawler_extension.d.ts +0 -12
  158. package/crawlers/crawler_extension.d.ts.map +0 -1
  159. package/crawlers/crawler_extension.js +0 -14
  160. package/crawlers/crawler_extension.js.map +0 -1
  161. package/http_clients/base-http-client.d.ts +0 -134
  162. package/http_clients/base-http-client.d.ts.map +0 -1
  163. package/http_clients/base-http-client.js +0 -33
  164. package/http_clients/base-http-client.js.map +0 -1
  165. package/http_clients/form-data-like.d.ts +0 -67
  166. package/http_clients/form-data-like.d.ts.map +0 -1
  167. package/http_clients/form-data-like.js +0 -5
  168. package/http_clients/form-data-like.js.map +0 -1
  169. package/http_clients/got-scraping-http-client.d.ts +0 -15
  170. package/http_clients/got-scraping-http-client.d.ts.map +0 -1
  171. package/http_clients/got-scraping-http-client.js +0 -69
  172. package/http_clients/got-scraping-http-client.js.map +0 -1
  173. package/http_clients/index.d.ts +0 -3
  174. package/http_clients/index.d.ts.map +0 -1
  175. package/http_clients/index.js +0 -3
  176. package/http_clients/index.js.map +0 -1
  177. package/tsconfig.build.tsbuildinfo +0 -1
package/log.d.ts CHANGED
@@ -1,3 +1,83 @@
1
- import log, { Log, Logger, LoggerJson, LoggerOptions, LoggerText, LogLevel } from '@apify/log';
2
- export { log, Log, LoggerOptions, LogLevel, Logger, LoggerJson, LoggerText };
1
+ import type { CrawleeLogger, CrawleeLoggerOptions } from '@crawlee/types';
2
+ import type { LoggerOptions } from '@apify/log';
3
+ import log, { Log, Logger, LoggerJson, LoggerText, LogLevel } from '@apify/log';
4
+ export type { CrawleeLogger, CrawleeLoggerOptions };
5
+ /**
6
+ * Abstract base class for custom Crawlee logger implementations.
7
+ *
8
+ * Subclasses must implement two methods:
9
+ * - {@link BaseCrawleeLogger.logWithLevel} — the core logging dispatch
10
+ * - {@link BaseCrawleeLogger.createChild} — how to create a child logger instance
11
+ *
12
+ * All other `CrawleeLogger` methods (`error`, `warning`, `info`, `debug`, etc.)
13
+ * are derived automatically. Level filtering is entirely the responsibility of the
14
+ * underlying library — `logWithLevel()` is called for every message.
15
+ *
16
+ * **Example — Winston adapter:**
17
+ * ```typescript
18
+ * const CRAWLEE_TO_WINSTON = { 1: 'error', 2: 'warn', 3: 'warn', 4: 'info', 5: 'debug', 6: 'debug' };
19
+ *
20
+ * class WinstonAdapter extends BaseCrawleeLogger {
21
+ * constructor(private logger: winston.Logger, options?: Partial<CrawleeLoggerOptions>) {
22
+ * super(options);
23
+ * }
24
+ *
25
+ * logWithLevel(level: number, message: string, data?: Record<string, unknown>): void {
26
+ * this.logger.log(CRAWLEE_TO_WINSTON[level] ?? 'info', message, data);
27
+ * }
28
+ *
29
+ * protected createChild(options: Partial<CrawleeLoggerOptions>): CrawleeLogger {
30
+ * return new WinstonAdapter(this.logger.child({ prefix: options.prefix }), { ...this.getOptions(), ...options });
31
+ * }
32
+ * }
33
+ * ```
34
+ */
35
+ export declare abstract class BaseCrawleeLogger implements CrawleeLogger {
36
+ private options;
37
+ private readonly warningsLogged;
38
+ constructor(options?: Partial<CrawleeLoggerOptions>);
39
+ /**
40
+ * Core logging method. Subclasses must implement this to dispatch log messages
41
+ * to the underlying logger (Winston, Pino, console, etc.).
42
+ *
43
+ * Level filtering is the responsibility of the underlying library — this method
44
+ * is called for every message regardless of the current level.
45
+ *
46
+ * @param level Crawlee log level (use {@link LogLevel} constants)
47
+ * @param message The log message
48
+ * @param data Optional structured data to attach to the log entry
49
+ */
50
+ abstract logWithLevel(level: number, message: string, data?: Record<string, unknown>): void;
51
+ /**
52
+ * Creates a child logger instance. Subclasses must implement this to define
53
+ * how child loggers are created for the underlying logger.
54
+ */
55
+ protected abstract createChild(options: Partial<CrawleeLoggerOptions>): CrawleeLogger;
56
+ getOptions(): CrawleeLoggerOptions;
57
+ setOptions(options: Partial<CrawleeLoggerOptions>): void;
58
+ child(options: Partial<CrawleeLoggerOptions>): CrawleeLogger;
59
+ error(message: string, data?: Record<string, unknown>): void;
60
+ exception(exception: Error, message: string, data?: Record<string, unknown>): void;
61
+ softFail(message: string, data?: Record<string, unknown>): void;
62
+ warning(message: string, data?: Record<string, unknown>): void;
63
+ warningOnce(message: string): void;
64
+ info(message: string, data?: Record<string, unknown>): void;
65
+ debug(message: string, data?: Record<string, unknown>): void;
66
+ perf(message: string, data?: Record<string, unknown>): void;
67
+ deprecated(message: string): void;
68
+ }
69
+ /**
70
+ * Adapter that wraps `@apify/log`'s {@link Log} instance to implement the {@link CrawleeLogger} interface.
71
+ *
72
+ * This is the default logger used by Crawlee when no custom logger is configured.
73
+ * Users who want to use a different logging library should implement {@link BaseCrawleeLogger} directly.
74
+ */
75
+ export declare class ApifyLogAdapter extends BaseCrawleeLogger {
76
+ private readonly apifyLog;
77
+ constructor(apifyLog: Log, options?: Partial<CrawleeLoggerOptions>);
78
+ logWithLevel(level: number, message: string, data?: Record<string, unknown>): void;
79
+ protected createChild(options: Partial<CrawleeLoggerOptions>): CrawleeLogger;
80
+ }
81
+ export { log, Log, LogLevel, Logger, LoggerJson, LoggerText };
82
+ export type { LoggerOptions };
3
83
  //# sourceMappingURL=log.d.ts.map
package/log.d.ts.map CHANGED
@@ -1 +1 @@
1
- {"version":3,"file":"log.d.ts","sourceRoot":"","sources":["../src/log.ts"],"names":[],"mappings":"AAAA,OAAO,GAAG,EAAE,EAAE,GAAG,EAAE,MAAM,EAAE,UAAU,EAAE,aAAa,EAAE,UAAU,EAAE,QAAQ,EAAE,MAAM,YAAY,CAAC;AAE/F,OAAO,EAAE,GAAG,EAAE,GAAG,EAAE,aAAa,EAAE,QAAQ,EAAE,MAAM,EAAE,UAAU,EAAE,UAAU,EAAE,CAAC"}
1
+ {"version":3,"file":"log.d.ts","sourceRoot":"","sources":["../src/log.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,aAAa,EAAE,oBAAoB,EAAE,MAAM,gBAAgB,CAAC;AAE1E,OAAO,KAAK,EAAE,aAAa,EAAE,MAAM,YAAY,CAAC;AAChD,OAAO,GAAG,EAAE,EAAE,GAAG,EAAE,MAAM,EAAE,UAAU,EAAE,UAAU,EAAE,QAAQ,EAAE,MAAM,YAAY,CAAC;AAEhF,YAAY,EAAE,aAAa,EAAE,oBAAoB,EAAE,CAAC;AAEpD;;;;;;;;;;;;;;;;;;;;;;;;;;;;;GA6BG;AACH,8BAAsB,iBAAkB,YAAW,aAAa;IAC5D,OAAO,CAAC,OAAO,CAAuB;IACtC,OAAO,CAAC,QAAQ,CAAC,cAAc,CAAqB;gBAExC,OAAO,GAAE,OAAO,CAAC,oBAAoB,CAAM;IAIvD;;;;;;;;;;OAUG;IACH,QAAQ,CAAC,YAAY,CAAC,KAAK,EAAE,MAAM,EAAE,OAAO,EAAE,MAAM,EAAE,IAAI,CAAC,EAAE,MAAM,CAAC,MAAM,EAAE,OAAO,CAAC,GAAG,IAAI;IAE3F;;;OAGG;IACH,SAAS,CAAC,QAAQ,CAAC,WAAW,CAAC,OAAO,EAAE,OAAO,CAAC,oBAAoB,CAAC,GAAG,aAAa;IAErF,UAAU,IAAI,oBAAoB;IAIlC,UAAU,CAAC,OAAO,EAAE,OAAO,CAAC,oBAAoB,CAAC,GAAG,IAAI;IAIxD,KAAK,CAAC,OAAO,EAAE,OAAO,CAAC,oBAAoB,CAAC,GAAG,aAAa;IAI5D,KAAK,CAAC,OAAO,EAAE,MAAM,EAAE,IAAI,CAAC,EAAE,MAAM,CAAC,MAAM,EAAE,OAAO,CAAC,GAAG,IAAI;IAI5D,SAAS,CAAC,SAAS,EAAE,KAAK,EAAE,OAAO,EAAE,MAAM,EAAE,IAAI,CAAC,EAAE,MAAM,CAAC,MAAM,EAAE,OAAO,CAAC,GAAG,IAAI;IAQlF,QAAQ,CAAC,OAAO,EAAE,MAAM,EAAE,IAAI,CAAC,EAAE,MAAM,CAAC,MAAM,EAAE,OAAO,CAAC,GAAG,IAAI;IAI/D,OAAO,CAAC,OAAO,EAAE,MAAM,EAAE,IAAI,CAAC,EAAE,MAAM,CAAC,MAAM,EAAE,OAAO,CAAC,GAAG,IAAI;IAI9D,WAAW,CAAC,OAAO,EAAE,MAAM,GAAG,IAAI;IAOlC,IAAI,CAAC,OAAO,EAAE,MAAM,EAAE,IAAI,CAAC,EAAE,MAAM,CAAC,MAAM,EAAE,OAAO,CAAC,GAAG,IAAI;IAI3D,KAAK,CAAC,OAAO,EAAE,MAAM,EAAE,IAAI,CAAC,EAAE,MAAM,CAAC,MAAM,EAAE,OAAO,CAAC,GAAG,IAAI;IAI5D,IAAI,CAAC,OAAO,EAAE,MAAM,EAAE,IAAI,CAAC,EAAE,MAAM,CAAC,MAAM,EAAE,OAAO,CAAC,GAAG,IAAI;IAI3D,UAAU,CAAC,OAAO,EAAE,MAAM,GAAG,IAAI;CAGpC;AAED;;;;;GAKG;AACH,qBAAa,eAAgB,SAAQ,iBAAiB;IAE9C,OAAO,CAAC,QAAQ,CAAC,QAAQ;gBAAR,QAAQ,EAAE,GAAG,EAC9B,OAAO,CAAC,EAAE,OAAO,CAAC,oBAAoB,CAAC;IAK3C,YAAY,CAAC,KAAK,EAAE,MAAM,EAAE,OAAO,EAAE,MAAM,EAAE,IAAI,CAAC,EAAE,MAAM,CAAC,MAAM,EAAE,OAAO,CAAC,GAAG,IAAI;IAIlF,SAAS,CAAC,WAAW,CAAC,OAAO,EAAE,OAAO,CAAC,oBAAoB,CAAC,GAAG,aAAa;CAM/E;AAED,OAAO,EAAE,GAAG,EAAE,GAAG,EAAE,QAAQ,EAAE,MAAM,EAAE,UAAU,EAAE,UAAU,EAAE,CAAC;AAC9D,YAAY,EAAE,aAAa,EAAE,CAAC"}
package/log.js CHANGED
@@ -1,3 +1,105 @@
1
1
  import log, { Log, Logger, LoggerJson, LoggerText, LogLevel } from '@apify/log';
2
+ /**
3
+ * Abstract base class for custom Crawlee logger implementations.
4
+ *
5
+ * Subclasses must implement two methods:
6
+ * - {@link BaseCrawleeLogger.logWithLevel} — the core logging dispatch
7
+ * - {@link BaseCrawleeLogger.createChild} — how to create a child logger instance
8
+ *
9
+ * All other `CrawleeLogger` methods (`error`, `warning`, `info`, `debug`, etc.)
10
+ * are derived automatically. Level filtering is entirely the responsibility of the
11
+ * underlying library — `logWithLevel()` is called for every message.
12
+ *
13
+ * **Example — Winston adapter:**
14
+ * ```typescript
15
+ * const CRAWLEE_TO_WINSTON = { 1: 'error', 2: 'warn', 3: 'warn', 4: 'info', 5: 'debug', 6: 'debug' };
16
+ *
17
+ * class WinstonAdapter extends BaseCrawleeLogger {
18
+ * constructor(private logger: winston.Logger, options?: Partial<CrawleeLoggerOptions>) {
19
+ * super(options);
20
+ * }
21
+ *
22
+ * logWithLevel(level: number, message: string, data?: Record<string, unknown>): void {
23
+ * this.logger.log(CRAWLEE_TO_WINSTON[level] ?? 'info', message, data);
24
+ * }
25
+ *
26
+ * protected createChild(options: Partial<CrawleeLoggerOptions>): CrawleeLogger {
27
+ * return new WinstonAdapter(this.logger.child({ prefix: options.prefix }), { ...this.getOptions(), ...options });
28
+ * }
29
+ * }
30
+ * ```
31
+ */
32
+ export class BaseCrawleeLogger {
33
+ options;
34
+ warningsLogged = new Set();
35
+ constructor(options = {}) {
36
+ this.options = options;
37
+ }
38
+ getOptions() {
39
+ return this.options;
40
+ }
41
+ setOptions(options) {
42
+ this.options = { ...this.options, ...options };
43
+ }
44
+ child(options) {
45
+ return this.createChild(options);
46
+ }
47
+ error(message, data) {
48
+ this.logWithLevel(LogLevel.ERROR, message, data);
49
+ }
50
+ exception(exception, message, data) {
51
+ this.logWithLevel(LogLevel.ERROR, `${message}: ${exception.message}`, {
52
+ ...data,
53
+ stack: exception.stack,
54
+ exception,
55
+ });
56
+ }
57
+ softFail(message, data) {
58
+ this.logWithLevel(LogLevel.SOFT_FAIL, message, data);
59
+ }
60
+ warning(message, data) {
61
+ this.logWithLevel(LogLevel.WARNING, message, data);
62
+ }
63
+ warningOnce(message) {
64
+ if (!this.warningsLogged.has(message)) {
65
+ this.warningsLogged.add(message);
66
+ this.warning(message);
67
+ }
68
+ }
69
+ info(message, data) {
70
+ this.logWithLevel(LogLevel.INFO, message, data);
71
+ }
72
+ debug(message, data) {
73
+ this.logWithLevel(LogLevel.DEBUG, message, data);
74
+ }
75
+ perf(message, data) {
76
+ this.logWithLevel(LogLevel.PERF, `[PERF] ${message}`, data);
77
+ }
78
+ deprecated(message) {
79
+ this.warningOnce(`[DEPRECATED] ${message}`);
80
+ }
81
+ }
82
+ /**
83
+ * Adapter that wraps `@apify/log`'s {@link Log} instance to implement the {@link CrawleeLogger} interface.
84
+ *
85
+ * This is the default logger used by Crawlee when no custom logger is configured.
86
+ * Users who want to use a different logging library should implement {@link BaseCrawleeLogger} directly.
87
+ */
88
+ export class ApifyLogAdapter extends BaseCrawleeLogger {
89
+ apifyLog;
90
+ constructor(apifyLog, options) {
91
+ super(options ?? {});
92
+ this.apifyLog = apifyLog;
93
+ }
94
+ logWithLevel(level, message, data) {
95
+ this.apifyLog.internal(level, message, data);
96
+ }
97
+ createChild(options) {
98
+ return new ApifyLogAdapter(this.apifyLog.child({ prefix: options.prefix ?? null }), {
99
+ ...this.getOptions(),
100
+ ...options,
101
+ });
102
+ }
103
+ }
2
104
  export { log, Log, LogLevel, Logger, LoggerJson, LoggerText };
3
105
  //# sourceMappingURL=log.js.map
package/log.js.map CHANGED
@@ -1 +1 @@
1
- {"version":3,"file":"log.js","sourceRoot":"","sources":["../src/log.ts"],"names":[],"mappings":"AAAA,OAAO,GAAG,EAAE,EAAE,GAAG,EAAE,MAAM,EAAE,UAAU,EAAiB,UAAU,EAAE,QAAQ,EAAE,MAAM,YAAY,CAAC;AAE/F,OAAO,EAAE,GAAG,EAAE,GAAG,EAAiB,QAAQ,EAAE,MAAM,EAAE,UAAU,EAAE,UAAU,EAAE,CAAC"}
1
+ {"version":3,"file":"log.js","sourceRoot":"","sources":["../src/log.ts"],"names":[],"mappings":"AAGA,OAAO,GAAG,EAAE,EAAE,GAAG,EAAE,MAAM,EAAE,UAAU,EAAE,UAAU,EAAE,QAAQ,EAAE,MAAM,YAAY,CAAC;AAIhF;;;;;;;;;;;;;;;;;;;;;;;;;;;;;GA6BG;AACH,MAAM,OAAgB,iBAAiB;IAC3B,OAAO,CAAuB;IACrB,cAAc,GAAG,IAAI,GAAG,EAAU,CAAC;IAEpD,YAAY,UAAyC,EAAE;QACnD,IAAI,CAAC,OAAO,GAAG,OAAO,CAAC;IAC3B,CAAC;IAqBD,UAAU;QACN,OAAO,IAAI,CAAC,OAAO,CAAC;IACxB,CAAC;IAED,UAAU,CAAC,OAAsC;QAC7C,IAAI,CAAC,OAAO,GAAG,EAAE,GAAG,IAAI,CAAC,OAAO,EAAE,GAAG,OAAO,EAAE,CAAC;IACnD,CAAC;IAED,KAAK,CAAC,OAAsC;QACxC,OAAO,IAAI,CAAC,WAAW,CAAC,OAAO,CAAC,CAAC;IACrC,CAAC;IAED,KAAK,CAAC,OAAe,EAAE,IAA8B;QACjD,IAAI,CAAC,YAAY,CAAC,QAAQ,CAAC,KAAK,EAAE,OAAO,EAAE,IAAI,CAAC,CAAC;IACrD,CAAC;IAED,SAAS,CAAC,SAAgB,EAAE,OAAe,EAAE,IAA8B;QACvE,IAAI,CAAC,YAAY,CAAC,QAAQ,CAAC,KAAK,EAAE,GAAG,OAAO,KAAK,SAAS,CAAC,OAAO,EAAE,EAAE;YAClE,GAAG,IAAI;YACP,KAAK,EAAE,SAAS,CAAC,KAAK;YACtB,SAAS;SACZ,CAAC,CAAC;IACP,CAAC;IAED,QAAQ,CAAC,OAAe,EAAE,IAA8B;QACpD,IAAI,CAAC,YAAY,CAAC,QAAQ,CAAC,SAAS,EAAE,OAAO,EAAE,IAAI,CAAC,CAAC;IACzD,CAAC;IAED,OAAO,CAAC,OAAe,EAAE,IAA8B;QACnD,IAAI,CAAC,YAAY,CAAC,QAAQ,CAAC,OAAO,EAAE,OAAO,EAAE,IAAI,CAAC,CAAC;IACvD,CAAC;IAED,WAAW,CAAC,OAAe;QACvB,IAAI,CAAC,IAAI,CAAC,cAAc,CAAC,GAAG,CAAC,OAAO,CAAC,EAAE,CAAC;YACpC,IAAI,CAAC,cAAc,CAAC,GAAG,CAAC,OAAO,CAAC,CAAC;YACjC,IAAI,CAAC,OAAO,CAAC,OAAO,CAAC,CAAC;QAC1B,CAAC;IACL,CAAC;IAED,IAAI,CAAC,OAAe,EAAE,IAA8B;QAChD,IAAI,CAAC,YAAY,CAAC,QAAQ,CAAC,IAAI,EAAE,OAAO,EAAE,IAAI,CAAC,CAAC;IACpD,CAAC;IAED,KAAK,CAAC,OAAe,EAAE,IAA8B;QACjD,IAAI,CAAC,YAAY,CAAC,QAAQ,CAAC,KAAK,EAAE,OAAO,EAAE,IAAI,CAAC,CAAC;IACrD,CAAC;IAED,IAAI,CAAC,OAAe,EAAE,IAA8B;QAChD,IAAI,CAAC,YAAY,CAAC,QAAQ,CAAC,IAAI,EAAE,UAAU,OAAO,EAAE,EAAE,IAAI,CAAC,CAAC;IAChE,CAAC;IAED,UAAU,CAAC,OAAe;QACtB,IAAI,CAAC,WAAW,CAAC,gBAAgB,OAAO,EAAE,CAAC,CAAC;IAChD,CAAC;CACJ;AAED;;;;;GAKG;AACH,MAAM,OAAO,eAAgB,SAAQ,iBAAiB;IAE7B;IADrB,YACqB,QAAa,EAC9B,OAAuC;QAEvC,KAAK,CAAC,OAAO,IAAI,EAAE,CAAC,CAAC;QAHJ,aAAQ,GAAR,QAAQ,CAAK;IAIlC,CAAC;IAED,YAAY,CAAC,KAAa,EAAE,OAAe,EAAE,IAA8B;QACvE,IAAI,CAAC,QAAQ,CAAC,QAAQ,CAAC,KAAiB,EAAE,OAAO,EAAE,IAAI,CAAC,CAAC;IAC7D,CAAC;IAES,WAAW,CAAC,OAAsC;QACxD,OAAO,IAAI,eAAe,CAAC,IAAI,CAAC,QAAQ,CAAC,KAAK,CAAC,EAAE,MAAM,EAAE,OAAO,CAAC,MAAM,IAAI,IAAI,EAAE,CAAC,EAAE;YAChF,GAAG,IAAI,CAAC,UAAU,EAAE;YACpB,GAAG,OAAO;SACb,CAAC,CAAC;IACP,CAAC;CACJ;AAED,OAAO,EAAE,GAAG,EAAE,GAAG,EAAE,QAAQ,EAAE,MAAM,EAAE,UAAU,EAAE,UAAU,EAAE,CAAC"}
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@crawlee/core",
3
- "version": "4.0.0-beta.5",
3
+ "version": "4.0.0-beta.51",
4
4
  "description": "The scalable web crawling and scraping library for JavaScript/Node.js. Enables development of data extraction and web automation jobs (not only) with headless Chrome and Puppeteer.",
5
5
  "engines": {
6
6
  "node": ">=22.0.0"
@@ -38,7 +38,7 @@
38
38
  },
39
39
  "homepage": "https://crawlee.dev",
40
40
  "scripts": {
41
- "build": "yarn clean && yarn compile && yarn copy",
41
+ "build": "pnpm clean && pnpm compile && pnpm copy",
42
42
  "clean": "rimraf ./dist",
43
43
  "compile": "tsc -p tsconfig.build.json",
44
44
  "copy": "tsx ../../scripts/copy.ts"
@@ -53,22 +53,21 @@
53
53
  "@apify/pseudo_url": "^2.0.59",
54
54
  "@apify/timeout": "^0.3.2",
55
55
  "@apify/utilities": "^2.15.5",
56
- "@crawlee/memory-storage": "4.0.0-beta.5",
57
- "@crawlee/types": "4.0.0-beta.5",
58
- "@crawlee/utils": "4.0.0-beta.5",
56
+ "@crawlee/memory-storage": "4.0.0-beta.51",
57
+ "@crawlee/types": "4.0.0-beta.51",
58
+ "@crawlee/utils": "4.0.0-beta.51",
59
59
  "@sapphire/async-queue": "^1.5.5",
60
60
  "@vladfrangu/async_event_emitter": "^2.4.6",
61
61
  "csv-stringify": "^6.5.2",
62
- "fs-extra": "^11.3.0",
63
- "got-scraping": "^4.1.1",
64
62
  "json5": "^2.2.3",
65
63
  "minimatch": "^10.0.1",
66
64
  "ow": "^2.0.0",
67
65
  "stream-json": "^1.9.1",
68
66
  "tldts": "^7.0.6",
69
- "tough-cookie": "^5.1.2",
67
+ "tough-cookie": "^6.0.0",
70
68
  "tslib": "^2.8.1",
71
- "type-fest": "^4.41.0"
69
+ "type-fest": "^4.41.0",
70
+ "zod": "^3.24.0 || ^4.0.0"
72
71
  },
73
72
  "lerna": {
74
73
  "command": {
@@ -77,5 +76,5 @@
77
76
  }
78
77
  }
79
78
  },
80
- "gitHead": "9db9799f5eb88ab29f6156ee1574140b438aa199"
79
+ "gitHead": "321493aea4dfb8073bb5e5d112c331645cac3d11"
81
80
  }
@@ -1,129 +1,28 @@
1
+ import type { ProxyInfo } from '@crawlee/types';
1
2
  import type { Request } from './request.js';
2
3
  export interface ProxyConfigurationFunction {
3
- (sessionId: string | number, options?: {
4
+ (options?: {
4
5
  request?: Request;
5
6
  }): string | null | Promise<string | null>;
6
7
  }
8
+ type UrlList = (string | null)[];
7
9
  export interface ProxyConfigurationOptions {
8
10
  /**
9
11
  * An array of custom proxy URLs to be rotated.
10
12
  * Custom proxies are not compatible with Apify Proxy and an attempt to use both
11
13
  * configuration options will cause an error to be thrown on initialize.
12
14
  */
13
- proxyUrls?: string[];
15
+ proxyUrls?: UrlList;
14
16
  /**
15
- * Custom function that allows you to generate the new proxy URL dynamically. It gets the `sessionId` as a parameter and an optional parameter with the `Request` object when applicable.
17
+ * Custom function that allows you to generate the new proxy URL dynamically. It gets an optional parameter with the `Request` object when applicable.
16
18
  * Can return either stringified proxy URL or `null` if the proxy should not be used. Can be asynchronous.
17
19
  *
18
20
  * This function is used to generate the URL when {@link ProxyConfiguration.newUrl} or {@link ProxyConfiguration.newProxyInfo} is called.
19
21
  */
20
22
  newUrlFunction?: ProxyConfigurationFunction;
21
- /**
22
- * An array of custom proxy URLs to be rotated stratified in tiers.
23
- * This is a more advanced version of `proxyUrls` that allows you to define a hierarchy of proxy URLs
24
- * If everything goes well, all the requests will be sent through the first proxy URL in the list.
25
- * Whenever the crawler encounters a problem with the current proxy on the given domain, it will switch to the higher tier for this domain.
26
- * The crawler probes lower-level proxies at intervals to check if it can make the tier downshift.
27
- *
28
- * This feature is useful when you have a set of proxies with different performance characteristics (speed, price, antibot performance etc.) and you want to use the best one for each domain.
29
- *
30
- * Use `null` as a proxy URL to disable the proxy for the given tier.
31
- */
32
- tieredProxyUrls?: (string | null)[][];
33
23
  }
34
- export interface TieredProxy {
35
- proxyUrl: string | null;
36
- proxyTier?: number;
37
- }
38
- /**
39
- * The main purpose of the ProxyInfo object is to provide information
40
- * about the current proxy connection used by the crawler for the request.
41
- * Outside of crawlers, you can get this object by calling {@link ProxyConfiguration.newProxyInfo}.
42
- *
43
- * **Example usage:**
44
- *
45
- * ```javascript
46
- * const proxyConfiguration = new ProxyConfiguration({
47
- * proxyUrls: ['...', '...'] // List of Proxy URLs to rotate
48
- * });
49
- *
50
- * // Getting proxyInfo object by calling class method directly
51
- * const proxyInfo = await proxyConfiguration.newProxyInfo();
52
- *
53
- * // In crawler
54
- * const crawler = new CheerioCrawler({
55
- * // ...
56
- * proxyConfiguration,
57
- * requestHandler({ proxyInfo }) {
58
- * // Getting used proxy URL
59
- * const proxyUrl = proxyInfo.url;
60
- *
61
- * // Getting ID of used Session
62
- * const sessionIdentifier = proxyInfo.sessionId;
63
- * }
64
- * })
65
- *
66
- * ```
67
- */
68
- export interface ProxyInfo {
69
- /**
70
- * The identifier of used {@link Session}, if used.
71
- */
72
- sessionId?: string;
73
- /**
74
- * The URL of the proxy.
75
- */
76
- url: string;
77
- /**
78
- * Username for the proxy.
79
- */
80
- username?: string;
81
- /**
82
- * User's password for the proxy.
83
- */
84
- password: string;
85
- /**
86
- * Hostname of your proxy.
87
- */
88
- hostname: string;
89
- /**
90
- * Proxy port.
91
- */
92
- port: number | string;
93
- /**
94
- * Proxy tier for the current proxy, if applicable (only for `tieredProxyUrls`).
95
- */
96
- proxyTier?: number;
97
- }
98
- interface TieredProxyOptions {
24
+ interface NewUrlOptions {
99
25
  request?: Request;
100
- proxyTier?: number;
101
- }
102
- /**
103
- * Internal class for tracking the proxy tier history for a specific domain.
104
- *
105
- * Predicts the best proxy tier for the next request based on the error history for different proxy tiers.
106
- */
107
- declare class ProxyTierTracker {
108
- private histogram;
109
- private currentTier;
110
- constructor(tieredProxyUrls: (string | null)[][]);
111
- /**
112
- * Processes a single step of the algorithm and updates the current tier prediction based on the error history.
113
- */
114
- private processStep;
115
- /**
116
- * Increases the error score for the given proxy tier. This raises the chance of picking a different proxy tier for the subsequent requests.
117
- *
118
- * The error score is increased by 10 for the given tier. This means that this tier will be disadvantaged for the next 10 requests (every new request prediction decreases the error score by 1).
119
- * @param tier The proxy tier to mark as problematic.
120
- */
121
- addError(tier: number): void;
122
- /**
123
- * Returns the best proxy tier for the next request based on the error history for different proxy tiers.
124
- * @returns The proxy tier prediction
125
- */
126
- predictTier(): number;
127
26
  }
128
27
  /**
129
28
  * Configures connection to a proxy server with the provided options. Proxy servers are used to prevent target websites from blocking
@@ -156,13 +55,11 @@ declare class ProxyTierTracker {
156
55
  export declare class ProxyConfiguration {
157
56
  isManInTheMiddle: boolean;
158
57
  protected nextCustomUrlIndex: number;
159
- protected proxyUrls?: string[];
160
- protected tieredProxyUrls?: (string | null)[][];
161
- protected usedProxyUrls: Map<string, string>;
58
+ protected proxyUrls?: UrlList;
59
+ protected usedProxyUrls: Map<string, string | null>;
162
60
  protected newUrlFunction?: ProxyConfigurationFunction;
163
61
  // @ts-ignore optional peer dependency or compatibility with es2022
164
- protected log: import("@apify/log").Log;
165
- protected domainTiers: Map<string, ProxyTierTracker>;
62
+ protected log: import("@crawlee/types").CrawleeLogger;
166
63
  /**
167
64
  * Creates a {@link ProxyConfiguration} instance based on the provided options. Proxy servers are used to prevent target websites from
168
65
  * blocking your crawlers based on IP address rate limits or blacklists. Setting proxy configuration in your crawlers automatically configures
@@ -190,53 +87,22 @@ export declare class ProxyConfiguration {
190
87
  * the currently used proxy via the requestHandler parameter `proxyInfo`.
191
88
  * Use it if you want to work with a rich representation of a proxy URL.
192
89
  * If you need the URL string only, use {@link ProxyConfiguration.newUrl}.
193
- * @param [sessionId]
194
- * Represents the identifier of user {@link Session} that can be managed by the {@link SessionPool} or
195
- * you can use the Apify Proxy [Session](https://docs.apify.com/proxy#sessions) identifier.
196
- * When the provided sessionId is a number, it's converted to a string. Property sessionId of
197
- * {@link ProxyInfo} is always returned as a type string.
198
90
  *
199
- * All the HTTP requests going through the proxy with the same session identifier
200
- * will use the same target proxy server (i.e. the same IP address).
201
- * The identifier must not be longer than 50 characters and include only the following: `0-9`, `a-z`, `A-Z`, `"."`, `"_"` and `"~"`.
202
91
  * @return Represents information about used proxy and its configuration.
203
92
  */
204
- newProxyInfo(sessionId?: string | number, options?: TieredProxyOptions): Promise<ProxyInfo | undefined>;
93
+ newProxyInfo(options?: NewUrlOptions): Promise<ProxyInfo | undefined>;
205
94
  /**
206
- * Given a session identifier and a request / proxy tier, this function returns a new proxy URL based on the provided configuration options.
207
- * @param _sessionId Session identifier
208
- * @param options Options for the tiered proxy rotation
209
- * @returns An object with the proxy URL and the proxy tier used.
210
- */
211
- protected _handleTieredUrl(_sessionId: string, options?: TieredProxyOptions): TieredProxy;
212
- /**
213
- * Given a `Request` object, this function returns the tier of the proxy that should be used for the request.
95
+ * Returns a new proxy URL based on provided configuration options.
214
96
  *
215
- * This returns `null` if `tieredProxyUrls` option is not set.
216
- */
217
- protected predictProxyTier(request: Request): number | null;
218
- /**
219
- * Returns a new proxy URL based on provided configuration options and the `sessionId` parameter.
220
- * @param [sessionId]
221
- * Represents the identifier of user {@link Session} that can be managed by the {@link SessionPool} or
222
- * you can use the Apify Proxy [Session](https://docs.apify.com/proxy#sessions) identifier.
223
- * When the provided sessionId is a number, it's converted to a string.
224
- *
225
- * All the HTTP requests going through the proxy with the same session identifier
226
- * will use the same target proxy server (i.e. the same IP address).
227
- * The identifier must not be longer than 50 characters and include only the following: `0-9`, `a-z`, `A-Z`, `"."`, `"_"` and `"~"`.
228
97
  * @return A string with a proxy URL, including authentication credentials and port number.
229
98
  * For example, `http://bob:password123@proxy.example.com:8000`
230
99
  */
231
- newUrl(sessionId?: string | number, options?: TieredProxyOptions): Promise<string | undefined>;
232
- /**
233
- * Handles custom url rotation with session
234
- */
235
- protected _handleCustomUrl(sessionId?: string): string;
100
+ newUrl(options?: NewUrlOptions): Promise<string | undefined>;
101
+ protected _handleProxyUrlsList(): string | null;
236
102
  /**
237
103
  * Calls the custom newUrlFunction and checks format of its return value
238
104
  */
239
- protected _callNewUrlFunction(sessionId?: string, options?: {
105
+ protected _callNewUrlFunction(options?: {
240
106
  request?: Request;
241
107
  }): Promise<string | null>;
242
108
  protected _throwCannotCombineCustomMethods(): never;
@@ -1 +1 @@
1
- {"version":3,"file":"proxy_configuration.d.ts","sourceRoot":"","sources":["../src/proxy_configuration.ts"],"names":[],"mappings":"AAMA,OAAO,KAAK,EAAE,OAAO,EAAE,MAAM,cAAc,CAAC;AAE5C,MAAM,WAAW,0BAA0B;IACvC,CAAC,SAAS,EAAE,MAAM,GAAG,MAAM,EAAE,OAAO,CAAC,EAAE;QAAE,OAAO,CAAC,EAAE,OAAO,CAAA;KAAE,GAAG,MAAM,GAAG,IAAI,GAAG,OAAO,CAAC,MAAM,GAAG,IAAI,CAAC,CAAC;CACzG;AAED,MAAM,WAAW,yBAAyB;IACtC;;;;OAIG;IACH,SAAS,CAAC,EAAE,MAAM,EAAE,CAAC;IAErB;;;;;OAKG;IACH,cAAc,CAAC,EAAE,0BAA0B,CAAC;IAE5C;;;;;;;;;;OAUG;IACH,eAAe,CAAC,EAAE,CAAC,MAAM,GAAG,IAAI,CAAC,EAAE,EAAE,CAAC;CACzC;AAED,MAAM,WAAW,WAAW;IACxB,QAAQ,EAAE,MAAM,GAAG,IAAI,CAAC;IACxB,SAAS,CAAC,EAAE,MAAM,CAAC;CACtB;AAED;;;;;;;;;;;;;;;;;;;;;;;;;;;;;GA6BG;AACH,MAAM,WAAW,SAAS;IACtB;;OAEG;IACH,SAAS,CAAC,EAAE,MAAM,CAAC;IAEnB;;OAEG;IACH,GAAG,EAAE,MAAM,CAAC;IAEZ;;OAEG;IACH,QAAQ,CAAC,EAAE,MAAM,CAAC;IAElB;;OAEG;IACH,QAAQ,EAAE,MAAM,CAAC;IAEjB;;OAEG;IACH,QAAQ,EAAE,MAAM,CAAC;IAEjB;;OAEG;IACH,IAAI,EAAE,MAAM,GAAG,MAAM,CAAC;IAEtB;;OAEG;IACH,SAAS,CAAC,EAAE,MAAM,CAAC;CACtB;AAED,UAAU,kBAAkB;IACxB,OAAO,CAAC,EAAE,OAAO,CAAC;IAClB,SAAS,CAAC,EAAE,MAAM,CAAC;CACtB;AAED;;;;GAIG;AACH,cAAM,gBAAgB;IAClB,OAAO,CAAC,SAAS,CAAW;IAC5B,OAAO,CAAC,WAAW,CAAS;gBAEhB,eAAe,EAAE,CAAC,MAAM,GAAG,IAAI,CAAC,EAAE,EAAE;IAKhD;;OAEG;IACH,OAAO,CAAC,WAAW;IAgBnB;;;;;OAKG;IACH,QAAQ,CAAC,IAAI,EAAE,MAAM;IAIrB;;;OAGG;IACH,WAAW;CAId;AAED;;;;;;;;;;;;;;;;;;;;;;;;;;;GA2BG;AACH,qBAAa,kBAAkB;IAC3B,gBAAgB,UAAS;IACzB,SAAS,CAAC,kBAAkB,SAAK;IACjC,SAAS,CAAC,SAAS,CAAC,EAAE,MAAM,EAAE,CAAC;IAC/B,SAAS,CAAC,eAAe,CAAC,EAAE,CAAC,MAAM,GAAG,IAAI,CAAC,EAAE,EAAE,CAAC;IAChD,SAAS,CAAC,aAAa,sBAA6B;IACpD,SAAS,CAAC,cAAc,CAAC,EAAE,0BAA0B,CAAC;IACtD,SAAS,CAAC,GAAG,2BAA+C;IAC5D,SAAS,CAAC,WAAW,gCAAuC;IAE5D;;;;;;;;;;;;;;;;;;;OAmBG;gBACS,OAAO,GAAE,yBAA8B;IAwBnD;;;;;;;;;;;;;;;;OAgBG;IACG,YAAY,CAAC,SAAS,CAAC,EAAE,MAAM,GAAG,MAAM,EAAE,OAAO,CAAC,EAAE,kBAAkB,GAAG,OAAO,CAAC,SAAS,GAAG,SAAS,CAAC;IA4B7G;;;;;OAKG;IACH,SAAS,CAAC,gBAAgB,CAAC,UAAU,EAAE,MAAM,EAAE,OAAO,CAAC,EAAE,kBAAkB,GAAG,WAAW;IAwBzF;;;;OAIG;IACH,SAAS,CAAC,gBAAgB,CAAC,OAAO,EAAE,OAAO,GAAG,MAAM,GAAG,IAAI;IAiC3D;;;;;;;;;;;;OAYG;IACG,MAAM,CAAC,SAAS,CAAC,EAAE,MAAM,GAAG,MAAM,EAAE,OAAO,CAAC,EAAE,kBAAkB,GAAG,OAAO,CAAC,MAAM,GAAG,SAAS,CAAC;IAcpG;;OAEG;IACH,SAAS,CAAC,gBAAgB,CAAC,SAAS,CAAC,EAAE,MAAM,GAAG,MAAM;IAiBtD;;OAEG;cACa,mBAAmB,CAAC,SAAS,CAAC,EAAE,MAAM,EAAE,OAAO,CAAC,EAAE;QAAE,OAAO,CAAC,EAAE,OAAO,CAAA;KAAE;IAcvF,SAAS,CAAC,gCAAgC,IAAI,KAAK;IAMnD,SAAS,CAAC,uBAAuB,IAAI,KAAK;CAG7C"}
1
+ {"version":3,"file":"proxy_configuration.d.ts","sourceRoot":"","sources":["../src/proxy_configuration.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAc,SAAS,EAAE,MAAM,gBAAgB,CAAC;AAG5D,OAAO,KAAK,EAAE,OAAO,EAAE,MAAM,cAAc,CAAC;AAG5C,MAAM,WAAW,0BAA0B;IACvC,CAAC,OAAO,CAAC,EAAE;QAAE,OAAO,CAAC,EAAE,OAAO,CAAA;KAAE,GAAG,MAAM,GAAG,IAAI,GAAG,OAAO,CAAC,MAAM,GAAG,IAAI,CAAC,CAAC;CAC7E;AAED,KAAK,OAAO,GAAG,CAAC,MAAM,GAAG,IAAI,CAAC,EAAE,CAAC;AAEjC,MAAM,WAAW,yBAAyB;IACtC;;;;OAIG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IAEpB;;;;;OAKG;IACH,cAAc,CAAC,EAAE,0BAA0B,CAAC;CAC/C;AAED,UAAU,aAAa;IACnB,OAAO,CAAC,EAAE,OAAO,CAAC;CACrB;AAED;;;;;;;;;;;;;;;;;;;;;;;;;;;GA2BG;AACH,qBAAa,kBAAkB;IAC3B,gBAAgB,UAAS;IACzB,SAAS,CAAC,kBAAkB,SAAK;IACjC,SAAS,CAAC,SAAS,CAAC,EAAE,OAAO,CAAC;IAC9B,SAAS,CAAC,aAAa,6BAAoC;IAC3D,SAAS,CAAC,cAAc,CAAC,EAAE,0BAA0B,CAAC;IACtD,SAAS,CAAC,GAAG,yCAAsE;IAEnF;;;;;;;;;;;;;;;;;;;OAmBG;gBACS,OAAO,GAAE,yBAA8B;IA2BnD;;;;;;;;OAQG;IACG,YAAY,CAAC,OAAO,CAAC,EAAE,aAAa,GAAG,OAAO,CAAC,SAAS,GAAG,SAAS,CAAC;IAe3E;;;;;OAKG;IACG,MAAM,CAAC,OAAO,CAAC,EAAE,aAAa,GAAG,OAAO,CAAC,MAAM,GAAG,SAAS,CAAC;IAQlE,SAAS,CAAC,oBAAoB,IAAI,MAAM,GAAG,IAAI;IAI/C;;OAEG;cACa,mBAAmB,CAAC,OAAO,CAAC,EAAE;QAAE,OAAO,CAAC,EAAE,OAAO,CAAA;KAAE;IAcnE,SAAS,CAAC,gCAAgC,IAAI,KAAK;IAMnD,SAAS,CAAC,uBAAuB,IAAI,KAAK;CAG7C"}