@slicemachine/manager 0.25.4-beta.5 → 0.25.4-beta.7
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/_node_modules/@amplitude/analytics-core/_node_modules/tslib/tslib.es6.cjs +69 -88
- package/dist/_node_modules/@amplitude/analytics-core/_node_modules/tslib/tslib.es6.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/_node_modules/tslib/tslib.es6.js +69 -88
- package/dist/_node_modules/@amplitude/analytics-core/_node_modules/tslib/tslib.es6.js.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/config.cjs +4 -4
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/config.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/config.js +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/config.js.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/core-client.cjs +3 -6
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/core-client.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/core-client.js +3 -6
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/core-client.js.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/identify.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/identify.js.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/logger.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/logger.js.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/plugins/destination.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/plugins/destination.js.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/revenue.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/revenue.js.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/timeline.cjs +6 -12
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/timeline.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/timeline.js +6 -12
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/timeline.js.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/transports/base.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/transports/base.js.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/utils/chunk.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/utils/chunk.js.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/utils/debug.cjs +4 -11
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/utils/debug.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/utils/debug.js +5 -12
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/utils/debug.js.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/utils/event-builder.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/utils/event-builder.js.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/utils/result-builder.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/utils/result-builder.js.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/utils/return-wrapper.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/utils/return-wrapper.js.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/utils/uuid.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/utils/uuid.js.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/utils/valid-properties.cjs +2 -4
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/utils/valid-properties.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/utils/valid-properties.js +2 -4
- package/dist/_node_modules/@amplitude/analytics-core/lib/esm/utils/valid-properties.js.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-node/_node_modules/tslib/tslib.es6.cjs +52 -63
- package/dist/_node_modules/@amplitude/analytics-node/_node_modules/tslib/tslib.es6.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-node/_node_modules/tslib/tslib.es6.js +52 -63
- package/dist/_node_modules/@amplitude/analytics-node/_node_modules/tslib/tslib.es6.js.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-node/lib/esm/config.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-node/lib/esm/config.js.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-node/lib/esm/index.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-node/lib/esm/node-client.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-node/lib/esm/node-client.js.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-node/lib/esm/plugins/context.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-node/lib/esm/plugins/context.js.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-node/lib/esm/transports/http.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-node/lib/esm/transports/http.js.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-types/lib/esm/event.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-types/lib/esm/event.js.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-types/lib/esm/logger.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-types/lib/esm/plugin.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-types/lib/esm/server-zone.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-types/lib/esm/status.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/analytics-types/lib/esm/transport.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-core/dist/experiment-core.esm.cjs +101 -146
- package/dist/_node_modules/@amplitude/experiment-core/dist/experiment-core.esm.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-core/dist/experiment-core.esm.js +101 -146
- package/dist/_node_modules/@amplitude/experiment-core/dist/experiment-core.esm.js.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/gen/version.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/gen/version.js.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/assignment/assignment-filter.cjs +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/assignment/assignment-filter.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/assignment/assignment-filter.js +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/assignment/assignment-filter.js.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/assignment/assignment-service.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/assignment/assignment-service.js.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/assignment/assignment.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/assignment/assignment.js.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/cookie.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/cookie.js.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/factory.cjs +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/factory.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/factory.js +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/factory.js.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/index.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/index.js.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/local/cache.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/local/cache.js.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/local/client.cjs +12 -18
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/local/client.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/local/client.js +12 -18
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/local/client.js.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/local/fetcher.cjs +2 -2
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/local/fetcher.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/local/fetcher.js +2 -2
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/local/fetcher.js.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/local/poller.cjs +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/local/poller.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/local/poller.js +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/local/poller.js.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/remote/client.cjs +3 -3
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/remote/client.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/remote/client.js +3 -3
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/remote/client.js.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/transport/http.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/transport/http.js.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/types/config.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/types/config.js.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/util/backoff.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/util/backoff.js.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/util/cache.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/util/cache.js.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/util/hash.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/util/hash.js.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/util/logger.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/util/logger.js.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/util/time.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/util/time.js.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/util/user.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/util/user.js.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/util/variant.cjs.map +1 -1
- package/dist/_node_modules/@amplitude/experiment-node-server/dist/src/util/variant.js.map +1 -1
- package/dist/_node_modules/@msgpack/msgpack/dist.es5_esm/CachedKeyDecoder.cjs +8 -9
- package/dist/_node_modules/@msgpack/msgpack/dist.es5_esm/CachedKeyDecoder.cjs.map +1 -1
- package/dist/_node_modules/@msgpack/msgpack/dist.es5_esm/CachedKeyDecoder.js +8 -9
- package/dist/_node_modules/@msgpack/msgpack/dist.es5_esm/CachedKeyDecoder.js.map +1 -1
- package/dist/_node_modules/@msgpack/msgpack/dist.es5_esm/DecodeError.cjs +2 -4
- package/dist/_node_modules/@msgpack/msgpack/dist.es5_esm/DecodeError.cjs.map +1 -1
- package/dist/_node_modules/@msgpack/msgpack/dist.es5_esm/DecodeError.js +2 -4
- package/dist/_node_modules/@msgpack/msgpack/dist.es5_esm/DecodeError.js.map +1 -1
- package/dist/_node_modules/@msgpack/msgpack/dist.es5_esm/Decoder.cjs +220 -239
- package/dist/_node_modules/@msgpack/msgpack/dist.es5_esm/Decoder.cjs.map +1 -1
- package/dist/_node_modules/@msgpack/msgpack/dist.es5_esm/Decoder.js +221 -240
- package/dist/_node_modules/@msgpack/msgpack/dist.es5_esm/Decoder.js.map +1 -1
- package/dist/_node_modules/@msgpack/msgpack/dist.es5_esm/Encoder.cjs.map +1 -1
- package/dist/_node_modules/@msgpack/msgpack/dist.es5_esm/Encoder.js.map +1 -1
- package/dist/_node_modules/@msgpack/msgpack/dist.es5_esm/ExtData.cjs +1 -1
- package/dist/_node_modules/@msgpack/msgpack/dist.es5_esm/ExtData.cjs.map +1 -1
- package/dist/_node_modules/@msgpack/msgpack/dist.es5_esm/ExtData.js +1 -1
- package/dist/_node_modules/@msgpack/msgpack/dist.es5_esm/ExtData.js.map +1 -1
- package/dist/_node_modules/@msgpack/msgpack/dist.es5_esm/ExtensionCodec.cjs.map +1 -1
- package/dist/_node_modules/@msgpack/msgpack/dist.es5_esm/ExtensionCodec.js.map +1 -1
- package/dist/_node_modules/@msgpack/msgpack/dist.es5_esm/decode.cjs.map +1 -1
- package/dist/_node_modules/@msgpack/msgpack/dist.es5_esm/decode.js.map +1 -1
- package/dist/_node_modules/@msgpack/msgpack/dist.es5_esm/encode.cjs.map +1 -1
- package/dist/_node_modules/@msgpack/msgpack/dist.es5_esm/encode.js.map +1 -1
- package/dist/_node_modules/@msgpack/msgpack/dist.es5_esm/timestamp.cjs.map +1 -1
- package/dist/_node_modules/@msgpack/msgpack/dist.es5_esm/timestamp.js +1 -1
- package/dist/_node_modules/@msgpack/msgpack/dist.es5_esm/timestamp.js.map +1 -1
- package/dist/_node_modules/@msgpack/msgpack/dist.es5_esm/utils/typedArrays.cjs.map +1 -1
- package/dist/_node_modules/@msgpack/msgpack/dist.es5_esm/utils/typedArrays.js.map +1 -1
- package/dist/_node_modules/@msgpack/msgpack/dist.es5_esm/utils/utf8.cjs.map +1 -1
- package/dist/_node_modules/@msgpack/msgpack/dist.es5_esm/utils/utf8.js.map +1 -1
- package/dist/_node_modules/common-tags/es/TemplateTag/TemplateTag.cjs +4 -7
- package/dist/_node_modules/common-tags/es/TemplateTag/TemplateTag.cjs.map +1 -1
- package/dist/_node_modules/common-tags/es/TemplateTag/TemplateTag.js +4 -7
- package/dist/_node_modules/common-tags/es/TemplateTag/TemplateTag.js.map +1 -1
- package/dist/_node_modules/common-tags/es/commaLists/commaLists.cjs.map +1 -1
- package/dist/_node_modules/common-tags/es/commaLists/commaLists.js.map +1 -1
- package/dist/_node_modules/common-tags/es/inlineArrayTransformer/inlineArrayTransformer.cjs.map +1 -1
- package/dist/_node_modules/common-tags/es/inlineArrayTransformer/inlineArrayTransformer.js.map +1 -1
- package/dist/_node_modules/common-tags/es/oneLineCommaListsAnd/oneLineCommaListsAnd.cjs.map +1 -1
- package/dist/_node_modules/common-tags/es/oneLineCommaListsAnd/oneLineCommaListsAnd.js.map +1 -1
- package/dist/_node_modules/common-tags/es/oneLineCommaListsOr/oneLineCommaListsOr.cjs.map +1 -1
- package/dist/_node_modules/common-tags/es/oneLineCommaListsOr/oneLineCommaListsOr.js.map +1 -1
- package/dist/_node_modules/common-tags/es/removeNonPrintingValuesTransformer/removeNonPrintingValuesTransformer.cjs.map +1 -1
- package/dist/_node_modules/common-tags/es/removeNonPrintingValuesTransformer/removeNonPrintingValuesTransformer.js.map +1 -1
- package/dist/_node_modules/common-tags/es/replaceResultTransformer/replaceResultTransformer.cjs.map +1 -1
- package/dist/_node_modules/common-tags/es/replaceResultTransformer/replaceResultTransformer.js.map +1 -1
- package/dist/_node_modules/common-tags/es/replaceSubstitutionTransformer/replaceSubstitutionTransformer.cjs.map +1 -1
- package/dist/_node_modules/common-tags/es/replaceSubstitutionTransformer/replaceSubstitutionTransformer.js.map +1 -1
- package/dist/_node_modules/common-tags/es/splitStringTransformer/splitStringTransformer.cjs +1 -3
- package/dist/_node_modules/common-tags/es/splitStringTransformer/splitStringTransformer.cjs.map +1 -1
- package/dist/_node_modules/common-tags/es/splitStringTransformer/splitStringTransformer.js +1 -3
- package/dist/_node_modules/common-tags/es/splitStringTransformer/splitStringTransformer.js.map +1 -1
- package/dist/_node_modules/common-tags/es/stripIndent/stripIndent.cjs +1 -2
- package/dist/_node_modules/common-tags/es/stripIndent/stripIndent.cjs.map +1 -1
- package/dist/_node_modules/common-tags/es/stripIndent/stripIndent.js +1 -2
- package/dist/_node_modules/common-tags/es/stripIndent/stripIndent.js.map +1 -1
- package/dist/_node_modules/common-tags/es/stripIndentTransformer/stripIndentTransformer.cjs.map +1 -1
- package/dist/_node_modules/common-tags/es/stripIndentTransformer/stripIndentTransformer.js.map +1 -1
- package/dist/_node_modules/common-tags/es/trimResultTransformer/trimResultTransformer.cjs.map +1 -1
- package/dist/_node_modules/common-tags/es/trimResultTransformer/trimResultTransformer.js.map +1 -1
- package/dist/_node_modules/cross-spawn/index.cjs.map +1 -1
- package/dist/_node_modules/cross-spawn/index.js.map +1 -1
- package/dist/_node_modules/cross-spawn/lib/enoent.cjs.map +1 -1
- package/dist/_node_modules/cross-spawn/lib/enoent.js.map +1 -1
- package/dist/_node_modules/cross-spawn/lib/parse.cjs.map +1 -1
- package/dist/_node_modules/cross-spawn/lib/parse.js.map +1 -1
- package/dist/_node_modules/cross-spawn/lib/util/escape.cjs.map +1 -1
- package/dist/_node_modules/cross-spawn/lib/util/escape.js.map +1 -1
- package/dist/_node_modules/cross-spawn/lib/util/readShebang.cjs.map +1 -1
- package/dist/_node_modules/cross-spawn/lib/util/readShebang.js.map +1 -1
- package/dist/_node_modules/cross-spawn/lib/util/resolveCommand.cjs.map +1 -1
- package/dist/_node_modules/cross-spawn/lib/util/resolveCommand.js.map +1 -1
- package/dist/_node_modules/data-uri-to-buffer/dist/index.cjs.map +1 -1
- package/dist/_node_modules/data-uri-to-buffer/dist/index.js.map +1 -1
- package/dist/_node_modules/execa/index.cjs +1 -1
- package/dist/_node_modules/execa/index.cjs.map +1 -1
- package/dist/_node_modules/execa/index.js +1 -1
- package/dist/_node_modules/execa/index.js.map +1 -1
- package/dist/_node_modules/execa/lib/command.cjs.map +1 -1
- package/dist/_node_modules/execa/lib/command.js.map +1 -1
- package/dist/_node_modules/execa/lib/error.cjs.map +1 -1
- package/dist/_node_modules/execa/lib/error.js.map +1 -1
- package/dist/_node_modules/execa/lib/kill.cjs.map +1 -1
- package/dist/_node_modules/execa/lib/kill.js.map +1 -1
- package/dist/_node_modules/execa/lib/pipe.cjs.map +1 -1
- package/dist/_node_modules/execa/lib/pipe.js.map +1 -1
- package/dist/_node_modules/execa/lib/promise.cjs +2 -2
- package/dist/_node_modules/execa/lib/promise.cjs.map +1 -1
- package/dist/_node_modules/execa/lib/promise.js +2 -2
- package/dist/_node_modules/execa/lib/promise.js.map +1 -1
- package/dist/_node_modules/execa/lib/stdio.cjs.map +1 -1
- package/dist/_node_modules/execa/lib/stdio.js.map +1 -1
- package/dist/_node_modules/execa/lib/stream.cjs.map +1 -1
- package/dist/_node_modules/execa/lib/stream.js.map +1 -1
- package/dist/_node_modules/execa/lib/verbose.cjs.map +1 -1
- package/dist/_node_modules/execa/lib/verbose.js.map +1 -1
- package/dist/_node_modules/fetch-blob/file.cjs +9 -21
- package/dist/_node_modules/fetch-blob/file.cjs.map +1 -1
- package/dist/_node_modules/fetch-blob/file.js +8 -20
- package/dist/_node_modules/fetch-blob/file.js.map +1 -1
- package/dist/_node_modules/fetch-blob/from.cjs +9 -0
- package/dist/_node_modules/fetch-blob/from.cjs.map +1 -0
- package/dist/_node_modules/fetch-blob/from.js +9 -0
- package/dist/_node_modules/fetch-blob/from.js.map +1 -0
- package/dist/_node_modules/fetch-blob/index.cjs +8 -21
- package/dist/_node_modules/fetch-blob/index.cjs.map +1 -1
- package/dist/_node_modules/fetch-blob/index.js +8 -21
- package/dist/_node_modules/fetch-blob/index.js.map +1 -1
- package/dist/_node_modules/fetch-blob/streams.cjs.map +1 -1
- package/dist/_node_modules/fetch-blob/streams.js.map +1 -1
- package/dist/_node_modules/file-type/core.cjs.map +1 -1
- package/dist/_node_modules/file-type/core.js.map +1 -1
- package/dist/_node_modules/file-type/util.cjs.map +1 -1
- package/dist/_node_modules/file-type/util.js.map +1 -1
- package/dist/_node_modules/formdata-polyfill/esm.min.cjs +12 -29
- package/dist/_node_modules/formdata-polyfill/esm.min.cjs.map +1 -1
- package/dist/_node_modules/formdata-polyfill/esm.min.js +12 -29
- package/dist/_node_modules/formdata-polyfill/esm.min.js.map +1 -1
- package/dist/_node_modules/fp-ts/es6/Either.cjs.map +1 -1
- package/dist/_node_modules/fp-ts/es6/Either.js.map +1 -1
- package/dist/_node_modules/fp-ts/es6/function.cjs +7 -9
- package/dist/_node_modules/fp-ts/es6/function.cjs.map +1 -1
- package/dist/_node_modules/fp-ts/es6/function.js +7 -9
- package/dist/_node_modules/fp-ts/es6/function.js.map +1 -1
- package/dist/_node_modules/fp-ts/es6/internal.cjs +7 -9
- package/dist/_node_modules/fp-ts/es6/internal.cjs.map +1 -1
- package/dist/_node_modules/fp-ts/es6/internal.js +7 -9
- package/dist/_node_modules/fp-ts/es6/internal.js.map +1 -1
- package/dist/_node_modules/get-port/index.cjs.map +1 -1
- package/dist/_node_modules/get-port/index.js.map +1 -1
- package/dist/_node_modules/get-stream/buffer-stream.cjs.map +1 -1
- package/dist/_node_modules/get-stream/buffer-stream.js.map +1 -1
- package/dist/_node_modules/get-stream/index.cjs +4 -4
- package/dist/_node_modules/get-stream/index.cjs.map +1 -1
- package/dist/_node_modules/get-stream/index.js +4 -4
- package/dist/_node_modules/get-stream/index.js.map +1 -1
- package/dist/_node_modules/human-signals/build/src/main.cjs.map +1 -1
- package/dist/_node_modules/human-signals/build/src/main.js.map +1 -1
- package/dist/_node_modules/human-signals/build/src/signals.cjs.map +1 -1
- package/dist/_node_modules/isexe/index.cjs.map +1 -1
- package/dist/_node_modules/isexe/index.js.map +1 -1
- package/dist/_node_modules/isexe/mode.cjs +1 -2
- package/dist/_node_modules/isexe/mode.cjs.map +1 -1
- package/dist/_node_modules/isexe/mode.js +1 -2
- package/dist/_node_modules/isexe/mode.js.map +1 -1
- package/dist/_node_modules/isexe/windows.cjs +1 -2
- package/dist/_node_modules/isexe/windows.cjs.map +1 -1
- package/dist/_node_modules/isexe/windows.js +1 -2
- package/dist/_node_modules/isexe/windows.js.map +1 -1
- package/dist/_node_modules/merge-stream/index.cjs.map +1 -1
- package/dist/_node_modules/merge-stream/index.js.map +1 -1
- package/dist/_node_modules/mimic-fn/index.cjs.map +1 -1
- package/dist/_node_modules/mimic-fn/index.js.map +1 -1
- package/dist/_node_modules/node-domexception/index.cjs.map +1 -1
- package/dist/_node_modules/node-domexception/index.js.map +1 -1
- package/dist/_node_modules/node-fetch/src/body.cjs +4 -7
- package/dist/_node_modules/node-fetch/src/body.cjs.map +1 -1
- package/dist/_node_modules/node-fetch/src/body.js +5 -8
- package/dist/_node_modules/node-fetch/src/body.js.map +1 -1
- package/dist/_node_modules/node-fetch/src/errors/abort-error.cjs.map +1 -1
- package/dist/_node_modules/node-fetch/src/errors/abort-error.js.map +1 -1
- package/dist/_node_modules/node-fetch/src/errors/base.cjs.map +1 -1
- package/dist/_node_modules/node-fetch/src/errors/base.js.map +1 -1
- package/dist/_node_modules/node-fetch/src/errors/fetch-error.cjs.map +1 -1
- package/dist/_node_modules/node-fetch/src/errors/fetch-error.js.map +1 -1
- package/dist/_node_modules/node-fetch/src/headers.cjs +1 -2
- package/dist/_node_modules/node-fetch/src/headers.cjs.map +1 -1
- package/dist/_node_modules/node-fetch/src/headers.js +1 -2
- package/dist/_node_modules/node-fetch/src/headers.js.map +1 -1
- package/dist/_node_modules/node-fetch/src/index.cjs +2 -4
- package/dist/_node_modules/node-fetch/src/index.cjs.map +1 -1
- package/dist/_node_modules/node-fetch/src/index.js +3 -5
- package/dist/_node_modules/node-fetch/src/index.js.map +1 -1
- package/dist/_node_modules/node-fetch/src/request.cjs.map +1 -1
- package/dist/_node_modules/node-fetch/src/request.js.map +1 -1
- package/dist/_node_modules/node-fetch/src/response.cjs.map +1 -1
- package/dist/_node_modules/node-fetch/src/response.js.map +1 -1
- package/dist/_node_modules/node-fetch/src/utils/get-search.cjs.map +1 -1
- package/dist/_node_modules/node-fetch/src/utils/get-search.js.map +1 -1
- package/dist/_node_modules/node-fetch/src/utils/multipart-parser.cjs +2 -5
- package/dist/_node_modules/node-fetch/src/utils/multipart-parser.cjs.map +1 -1
- package/dist/_node_modules/node-fetch/src/utils/multipart-parser.js +2 -5
- package/dist/_node_modules/node-fetch/src/utils/multipart-parser.js.map +1 -1
- package/dist/_node_modules/node-fetch/src/utils/referrer.cjs.map +1 -1
- package/dist/_node_modules/node-fetch/src/utils/referrer.js.map +1 -1
- package/dist/_node_modules/npm-run-path/_node_modules/path-key/index.cjs.map +1 -1
- package/dist/_node_modules/npm-run-path/_node_modules/path-key/index.js.map +1 -1
- package/dist/_node_modules/npm-run-path/index.cjs.map +1 -1
- package/dist/_node_modules/npm-run-path/index.js.map +1 -1
- package/dist/_node_modules/onetime/index.cjs.map +1 -1
- package/dist/_node_modules/onetime/index.js.map +1 -1
- package/dist/_node_modules/p-limit/_node_modules/yocto-queue/index.cjs +11 -25
- package/dist/_node_modules/p-limit/_node_modules/yocto-queue/index.cjs.map +1 -1
- package/dist/_node_modules/p-limit/_node_modules/yocto-queue/index.js +11 -25
- package/dist/_node_modules/p-limit/_node_modules/yocto-queue/index.js.map +1 -1
- package/dist/_node_modules/p-limit/index.cjs.map +1 -1
- package/dist/_node_modules/p-limit/index.js.map +1 -1
- package/dist/_node_modules/path-key/index.cjs.map +1 -1
- package/dist/_node_modules/path-key/index.js.map +1 -1
- package/dist/_node_modules/peek-readable/lib/EndOfFileStream.cjs.map +1 -1
- package/dist/_node_modules/peek-readable/lib/EndOfFileStream.js.map +1 -1
- package/dist/_node_modules/r19/dist/R19Error.cjs.map +1 -1
- package/dist/_node_modules/r19/dist/R19Error.js.map +1 -1
- package/dist/_node_modules/r19/dist/client/createRPCClient.cjs.map +1 -1
- package/dist/_node_modules/r19/dist/client/createRPCClient.js.map +1 -1
- package/dist/_node_modules/r19/dist/createRPCMiddleware.cjs.map +1 -1
- package/dist/_node_modules/r19/dist/createRPCMiddleware.js.map +1 -1
- package/dist/_node_modules/r19/dist/handleRPCRequest.cjs.map +1 -1
- package/dist/_node_modules/r19/dist/handleRPCRequest.js.map +1 -1
- package/dist/_node_modules/r19/dist/lib/replaceLeaves.cjs.map +1 -1
- package/dist/_node_modules/r19/dist/lib/replaceLeaves.js.map +1 -1
- package/dist/_node_modules/r19/dist/proceduresFromInstance.cjs.map +1 -1
- package/dist/_node_modules/r19/dist/proceduresFromInstance.js.map +1 -1
- package/dist/_node_modules/shebang-command/index.cjs.map +1 -1
- package/dist/_node_modules/shebang-command/index.js.map +1 -1
- package/dist/_node_modules/signal-exit/index.cjs.map +1 -1
- package/dist/_node_modules/signal-exit/index.js +1 -1
- package/dist/_node_modules/signal-exit/index.js.map +1 -1
- package/dist/_node_modules/signal-exit/signals.cjs +1 -2
- package/dist/_node_modules/signal-exit/signals.cjs.map +1 -1
- package/dist/_node_modules/signal-exit/signals.js +1 -2
- package/dist/_node_modules/signal-exit/signals.js.map +1 -1
- package/dist/_node_modules/strip-final-newline/index.cjs.map +1 -1
- package/dist/_node_modules/strip-final-newline/index.js.map +1 -1
- package/dist/_node_modules/strtok3/lib/AbstractTokenizer.cjs.map +1 -1
- package/dist/_node_modules/strtok3/lib/AbstractTokenizer.js.map +1 -1
- package/dist/_node_modules/strtok3/lib/BufferTokenizer.cjs.map +1 -1
- package/dist/_node_modules/strtok3/lib/BufferTokenizer.js.map +1 -1
- package/dist/_node_modules/strtok3/lib/core.cjs.map +1 -1
- package/dist/_node_modules/token-types/lib/index.cjs.map +1 -1
- package/dist/_node_modules/token-types/lib/index.js.map +1 -1
- package/dist/_node_modules/web-streams-polyfill/dist/ponyfill.es2018.cjs +5 -10
- package/dist/_node_modules/web-streams-polyfill/dist/ponyfill.es2018.cjs.map +1 -1
- package/dist/_node_modules/web-streams-polyfill/dist/ponyfill.es2018.js +5 -10
- package/dist/_node_modules/web-streams-polyfill/dist/ponyfill.es2018.js.map +1 -1
- package/dist/_node_modules/which/which.cjs.map +1 -1
- package/dist/_node_modules/which/which.js.map +1 -1
- package/dist/_node_modules/zod/lib/index.cjs +11 -20
- package/dist/_node_modules/zod/lib/index.cjs.map +1 -1
- package/dist/_node_modules/zod/lib/index.js +11 -20
- package/dist/_node_modules/zod/lib/index.js.map +1 -1
- package/dist/_virtual/_commonjsHelpers.cjs +2 -4
- package/dist/_virtual/_commonjsHelpers.cjs.map +1 -1
- package/dist/_virtual/_commonjsHelpers.js +2 -4
- package/dist/_virtual/_commonjsHelpers.js.map +1 -1
- package/dist/auth/PrismicAuthManager.cjs +1 -4
- package/dist/auth/PrismicAuthManager.cjs.map +1 -1
- package/dist/auth/PrismicAuthManager.js +1 -4
- package/dist/auth/PrismicAuthManager.js.map +1 -1
- package/dist/auth/createPrismicAuthManager.cjs.map +1 -1
- package/dist/auth/createPrismicAuthManager.js.map +1 -1
- package/dist/auth/createPrismicAuthManagerMiddleware.cjs.map +1 -1
- package/dist/auth/createPrismicAuthManagerMiddleware.js.map +1 -1
- package/dist/constants/API_ENDPOINTS.cjs.map +1 -1
- package/dist/constants/API_ENDPOINTS.js.map +1 -1
- package/dist/constants/API_TOKENS.cjs.map +1 -1
- package/dist/constants/API_TOKENS.js.map +1 -1
- package/dist/errors.cjs +1 -4
- package/dist/errors.cjs.map +1 -1
- package/dist/errors.js +1 -4
- package/dist/errors.js.map +1 -1
- package/dist/getEnvironmentInfo.cjs.map +1 -1
- package/dist/getEnvironmentInfo.js.map +1 -1
- package/dist/lib/DecodeError.cjs +1 -4
- package/dist/lib/DecodeError.cjs.map +1 -1
- package/dist/lib/DecodeError.js +1 -4
- package/dist/lib/DecodeError.js.map +1 -1
- package/dist/lib/assertPluginsInitialized.cjs.map +1 -1
- package/dist/lib/assertPluginsInitialized.js.map +1 -1
- package/dist/lib/bufferCodec.cjs.map +1 -1
- package/dist/lib/buildPrismicRepositoryAPIEndpoint.cjs.map +1 -1
- package/dist/lib/buildPrismicRepositoryAPIEndpoint.js.map +1 -1
- package/dist/lib/castArray.cjs.map +1 -1
- package/dist/lib/castArray.js.map +1 -1
- package/dist/lib/createContentDigest.cjs.map +1 -1
- package/dist/lib/createContentDigest.js.map +1 -1
- package/dist/lib/decode.cjs.map +1 -1
- package/dist/lib/decode.js.map +1 -1
- package/dist/lib/decodeHookResult.cjs.map +1 -1
- package/dist/lib/decodeHookResult.js.map +1 -1
- package/dist/lib/decodePackageJSON.cjs.map +1 -1
- package/dist/lib/decodePackageJSON.js.map +1 -1
- package/dist/lib/decodeSliceMachineConfig.cjs.map +1 -1
- package/dist/lib/decodeSliceMachineConfig.js.map +1 -1
- package/dist/lib/fetch.cjs +1 -2
- package/dist/lib/fetch.cjs.map +1 -1
- package/dist/lib/fetch.js +3 -4
- package/dist/lib/fetch.js.map +1 -1
- package/dist/lib/fetchGitHubReleaseBodyForRelease.cjs.map +1 -1
- package/dist/lib/fetchGitHubReleaseBodyForRelease.js.map +1 -1
- package/dist/lib/fetchNPMPackageVersions.cjs.map +1 -1
- package/dist/lib/fetchNPMPackageVersions.js.map +1 -1
- package/dist/lib/findEnvironment.cjs.map +1 -1
- package/dist/lib/findEnvironment.js.map +1 -1
- package/dist/lib/format.cjs +1 -4
- package/dist/lib/format.cjs.map +1 -1
- package/dist/lib/format.js +1 -4
- package/dist/lib/format.js.map +1 -1
- package/dist/lib/installDependencies.cjs.map +1 -1
- package/dist/lib/installDependencies.js.map +1 -1
- package/dist/lib/locateFileUpward.cjs.map +1 -1
- package/dist/lib/locateFileUpward.js.map +1 -1
- package/dist/lib/mockSlice.cjs.map +1 -1
- package/dist/lib/mockSlice.js.map +1 -1
- package/dist/lib/prismicrc.cjs.map +1 -1
- package/dist/lib/prismicrc.js.map +1 -1
- package/dist/lib/requireResolve.cjs.map +1 -1
- package/dist/lib/requireResolve.js.map +1 -1
- package/dist/lib/serializeCookies.cjs.map +1 -1
- package/dist/lib/serializeCookies.js.map +1 -1
- package/dist/lib/trailingSlash.cjs.map +1 -1
- package/dist/lib/trailingSlash.js.map +1 -1
- package/dist/managers/BaseManager.cjs +1 -4
- package/dist/managers/BaseManager.cjs.map +1 -1
- package/dist/managers/BaseManager.js +1 -4
- package/dist/managers/BaseManager.js.map +1 -1
- package/dist/managers/SliceMachineManager.cjs +1 -4
- package/dist/managers/SliceMachineManager.cjs.map +1 -1
- package/dist/managers/SliceMachineManager.js +1 -4
- package/dist/managers/SliceMachineManager.js.map +1 -1
- package/dist/managers/createSliceMachineManager.cjs.map +1 -1
- package/dist/managers/createSliceMachineManager.js.map +1 -1
- package/dist/managers/createSliceMachineManagerClient.cjs.map +1 -1
- package/dist/managers/createSliceMachineManagerClient.js.map +1 -1
- package/dist/managers/createSliceMachineManagerMiddleware.cjs.map +1 -1
- package/dist/managers/createSliceMachineManagerMiddleware.js.map +1 -1
- package/dist/managers/customTypes/CustomTypesManager.cjs.map +1 -1
- package/dist/managers/customTypes/CustomTypesManager.js +1 -1
- package/dist/managers/customTypes/CustomTypesManager.js.map +1 -1
- package/dist/managers/documentation/DocumentationManager.cjs.map +1 -1
- package/dist/managers/documentation/DocumentationManager.js.map +1 -1
- package/dist/managers/plugins/PluginsManager.cjs +1 -4
- package/dist/managers/plugins/PluginsManager.cjs.map +1 -1
- package/dist/managers/plugins/PluginsManager.js +1 -4
- package/dist/managers/plugins/PluginsManager.js.map +1 -1
- package/dist/managers/prismicRepository/PrismicRepositoryManager.cjs.map +1 -1
- package/dist/managers/prismicRepository/PrismicRepositoryManager.js.map +1 -1
- package/dist/managers/prismicRepository/sortEnvironments.cjs.map +1 -1
- package/dist/managers/prismicRepository/sortEnvironments.js.map +1 -1
- package/dist/managers/prismicRepository/types.cjs.map +1 -1
- package/dist/managers/prismicRepository/types.js.map +1 -1
- package/dist/managers/project/ProjectManager.cjs +1 -4
- package/dist/managers/project/ProjectManager.cjs.map +1 -1
- package/dist/managers/project/ProjectManager.js +1 -4
- package/dist/managers/project/ProjectManager.js.map +1 -1
- package/dist/managers/screenshots/ScreenshotsManager.cjs +6 -11
- package/dist/managers/screenshots/ScreenshotsManager.cjs.map +1 -1
- package/dist/managers/screenshots/ScreenshotsManager.js +3 -8
- package/dist/managers/screenshots/ScreenshotsManager.js.map +1 -1
- package/dist/managers/simulator/SimulatorManager.cjs.map +1 -1
- package/dist/managers/simulator/SimulatorManager.js.map +1 -1
- package/dist/managers/sliceTemplateLibrary/SliceTemplateLibraryManager.cjs.map +1 -1
- package/dist/managers/sliceTemplateLibrary/SliceTemplateLibraryManager.js.map +1 -1
- package/dist/managers/slices/SlicesManager.cjs.map +1 -1
- package/dist/managers/slices/SlicesManager.js.map +1 -1
- package/dist/managers/snippets/SnippetsManager.cjs.map +1 -1
- package/dist/managers/snippets/SnippetsManager.js.map +1 -1
- package/dist/managers/telemetry/TelemetryManager.cjs +1 -13
- package/dist/managers/telemetry/TelemetryManager.cjs.map +1 -1
- package/dist/managers/telemetry/TelemetryManager.js +1 -13
- package/dist/managers/telemetry/TelemetryManager.js.map +1 -1
- package/dist/managers/telemetry/types.cjs +6 -2
- package/dist/managers/telemetry/types.cjs.map +1 -1
- package/dist/managers/telemetry/types.d.ts +11 -1
- package/dist/managers/telemetry/types.js +6 -2
- package/dist/managers/telemetry/types.js.map +1 -1
- package/dist/managers/user/UserManager.cjs +1 -4
- package/dist/managers/user/UserManager.cjs.map +1 -1
- package/dist/managers/user/UserManager.js +1 -4
- package/dist/managers/user/UserManager.js.map +1 -1
- package/dist/managers/versions/VersionsManager.cjs +1 -4
- package/dist/managers/versions/VersionsManager.cjs.map +1 -1
- package/dist/managers/versions/VersionsManager.js +1 -4
- package/dist/managers/versions/VersionsManager.js.map +1 -1
- package/dist/test/createSliceMachineManagerMSWHandler.cjs.map +1 -1
- package/dist/test/createSliceMachineManagerMSWHandler.js.map +1 -1
- package/package.json +3 -3
- package/src/managers/telemetry/types.ts +23 -1
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"assertPluginsInitialized.cjs","sources":["../../../src/lib/assertPluginsInitialized.ts"],"sourcesContent":["import { SliceMachinePluginRunner } from \"@slicemachine/plugin-kit\";\n\nexport function assertPluginsInitialized(\n\tpluginRunner: SliceMachinePluginRunner | undefined,\n): asserts pluginRunner is NonNullable<typeof pluginRunner> {\n\tif (pluginRunner == undefined) {\n\t\tthrow new Error(\n\t\t\t\"Plugins have not been initialized. Run `SliceMachineManager.plugins.prototype.initPlugins()` before re-calling this method.\",\n\t\t);\n\t}\n}\n"],"names":[],"mappings":";;AAEM,SAAU,yBACf,cAAkD;AAElD,MAAI,gBAAgB,QAAW;
|
1
|
+
{"version":3,"file":"assertPluginsInitialized.cjs","sources":["../../../src/lib/assertPluginsInitialized.ts"],"sourcesContent":["import { SliceMachinePluginRunner } from \"@slicemachine/plugin-kit\";\n\nexport function assertPluginsInitialized(\n\tpluginRunner: SliceMachinePluginRunner | undefined,\n): asserts pluginRunner is NonNullable<typeof pluginRunner> {\n\tif (pluginRunner == undefined) {\n\t\tthrow new Error(\n\t\t\t\"Plugins have not been initialized. Run `SliceMachineManager.plugins.prototype.initPlugins()` before re-calling this method.\",\n\t\t);\n\t}\n}\n"],"names":[],"mappings":";;AAEM,SAAU,yBACf,cAAkD;AAElD,MAAI,gBAAgB,QAAW;AAC9B,UAAM,IAAI,MACT,6HAA6H;AAAA,EAE/H;AACD;;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"assertPluginsInitialized.js","sources":["../../../src/lib/assertPluginsInitialized.ts"],"sourcesContent":["import { SliceMachinePluginRunner } from \"@slicemachine/plugin-kit\";\n\nexport function assertPluginsInitialized(\n\tpluginRunner: SliceMachinePluginRunner | undefined,\n): asserts pluginRunner is NonNullable<typeof pluginRunner> {\n\tif (pluginRunner == undefined) {\n\t\tthrow new Error(\n\t\t\t\"Plugins have not been initialized. Run `SliceMachineManager.plugins.prototype.initPlugins()` before re-calling this method.\",\n\t\t);\n\t}\n}\n"],"names":[],"mappings":"AAEM,SAAU,yBACf,cAAkD;AAElD,MAAI,gBAAgB,QAAW;
|
1
|
+
{"version":3,"file":"assertPluginsInitialized.js","sources":["../../../src/lib/assertPluginsInitialized.ts"],"sourcesContent":["import { SliceMachinePluginRunner } from \"@slicemachine/plugin-kit\";\n\nexport function assertPluginsInitialized(\n\tpluginRunner: SliceMachinePluginRunner | undefined,\n): asserts pluginRunner is NonNullable<typeof pluginRunner> {\n\tif (pluginRunner == undefined) {\n\t\tthrow new Error(\n\t\t\t\"Plugins have not been initialized. Run `SliceMachineManager.plugins.prototype.initPlugins()` before re-calling this method.\",\n\t\t);\n\t}\n}\n"],"names":[],"mappings":"AAEM,SAAU,yBACf,cAAkD;AAElD,MAAI,gBAAgB,QAAW;AAC9B,UAAM,IAAI,MACT,6HAA6H;AAAA,EAE/H;AACD;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"bufferCodec.cjs","sources":["../../../src/lib/bufferCodec.ts"],"sourcesContent":["import { Buffer } from \"node:buffer\";\nimport * as t from \"io-ts\";\n\nexport const bufferCodec = new t.Type<Buffer, Buffer, unknown>(\n\t\"Buffer\",\n\t(input): input is Buffer => Buffer.isBuffer(input),\n\t(input, context) =>\n\t\tBuffer.isBuffer(input) ? t.success(input) : t.failure(input, context),\n\tt.identity,\n);\n"],"names":["t","Buffer"],"mappings":";;;;;;;;;;;;;;;;;;;;;AAGO,MAAM,cAAc,IAAIA,aAAE,KAChC,UACA,CAAC,UAA2BC,
|
1
|
+
{"version":3,"file":"bufferCodec.cjs","sources":["../../../src/lib/bufferCodec.ts"],"sourcesContent":["import { Buffer } from \"node:buffer\";\nimport * as t from \"io-ts\";\n\nexport const bufferCodec = new t.Type<Buffer, Buffer, unknown>(\n\t\"Buffer\",\n\t(input): input is Buffer => Buffer.isBuffer(input),\n\t(input, context) =>\n\t\tBuffer.isBuffer(input) ? t.success(input) : t.failure(input, context),\n\tt.identity,\n);\n"],"names":["t","Buffer"],"mappings":";;;;;;;;;;;;;;;;;;;;;AAGO,MAAM,cAAc,IAAIA,aAAE,KAChC,UACA,CAAC,UAA2BC,YAAAA,OAAO,SAAS,KAAK,GACjD,CAAC,OAAO,YACPA,mBAAO,SAAS,KAAK,IAAID,aAAE,QAAQ,KAAK,IAAIA,aAAE,QAAQ,OAAO,OAAO,GACrEA,aAAE,QAAQ;;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"buildPrismicRepositoryAPIEndpoint.cjs","sources":["../../../src/lib/buildPrismicRepositoryAPIEndpoint.ts"],"sourcesContent":["import { API_ENDPOINTS } from \"../constants/API_ENDPOINTS\";\n\nexport type BuildPrismicRepositoryAPIEndpointConfig = {\n\twithCDN?: boolean;\n};\n\nexport const buildPrismicRepositoryAPIEndpoint = (\n\trepositoryName: string,\n\t{ withCDN = true }: BuildPrismicRepositoryAPIEndpointConfig = {},\n): string => {\n\tconst baseEndpoint = new URL(API_ENDPOINTS.PrismicWroom);\n\n\treturn `${baseEndpoint.protocol}//${repositoryName}${withCDN ? \".cdn\" : \"\"}.${\n\t\tbaseEndpoint.host\n\t}/api/v2`;\n};\n"],"names":["API_ENDPOINTS"],"mappings":";;;
|
1
|
+
{"version":3,"file":"buildPrismicRepositoryAPIEndpoint.cjs","sources":["../../../src/lib/buildPrismicRepositoryAPIEndpoint.ts"],"sourcesContent":["import { API_ENDPOINTS } from \"../constants/API_ENDPOINTS\";\n\nexport type BuildPrismicRepositoryAPIEndpointConfig = {\n\twithCDN?: boolean;\n};\n\nexport const buildPrismicRepositoryAPIEndpoint = (\n\trepositoryName: string,\n\t{ withCDN = true }: BuildPrismicRepositoryAPIEndpointConfig = {},\n): string => {\n\tconst baseEndpoint = new URL(API_ENDPOINTS.PrismicWroom);\n\n\treturn `${baseEndpoint.protocol}//${repositoryName}${withCDN ? \".cdn\" : \"\"}.${\n\t\tbaseEndpoint.host\n\t}/api/v2`;\n};\n"],"names":["API_ENDPOINTS"],"mappings":";;;AAMO,MAAM,oCAAoC,CAChD,gBACA,EAAE,UAAU,KAAA,IAAkD,CAAA,MACnD;AACX,QAAM,eAAe,IAAI,IAAIA,cAAAA,cAAc,YAAY;AAEvD,SAAO,GAAG,aAAa,QAAQ,KAAK,cAAc,GAAG,UAAU,SAAS,EAAE,IACzE,aAAa,IACd;AACD;;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"buildPrismicRepositoryAPIEndpoint.js","sources":["../../../src/lib/buildPrismicRepositoryAPIEndpoint.ts"],"sourcesContent":["import { API_ENDPOINTS } from \"../constants/API_ENDPOINTS\";\n\nexport type BuildPrismicRepositoryAPIEndpointConfig = {\n\twithCDN?: boolean;\n};\n\nexport const buildPrismicRepositoryAPIEndpoint = (\n\trepositoryName: string,\n\t{ withCDN = true }: BuildPrismicRepositoryAPIEndpointConfig = {},\n): string => {\n\tconst baseEndpoint = new URL(API_ENDPOINTS.PrismicWroom);\n\n\treturn `${baseEndpoint.protocol}//${repositoryName}${withCDN ? \".cdn\" : \"\"}.${\n\t\tbaseEndpoint.host\n\t}/api/v2`;\n};\n"],"names":[],"mappings":";
|
1
|
+
{"version":3,"file":"buildPrismicRepositoryAPIEndpoint.js","sources":["../../../src/lib/buildPrismicRepositoryAPIEndpoint.ts"],"sourcesContent":["import { API_ENDPOINTS } from \"../constants/API_ENDPOINTS\";\n\nexport type BuildPrismicRepositoryAPIEndpointConfig = {\n\twithCDN?: boolean;\n};\n\nexport const buildPrismicRepositoryAPIEndpoint = (\n\trepositoryName: string,\n\t{ withCDN = true }: BuildPrismicRepositoryAPIEndpointConfig = {},\n): string => {\n\tconst baseEndpoint = new URL(API_ENDPOINTS.PrismicWroom);\n\n\treturn `${baseEndpoint.protocol}//${repositoryName}${withCDN ? \".cdn\" : \"\"}.${\n\t\tbaseEndpoint.host\n\t}/api/v2`;\n};\n"],"names":[],"mappings":";AAMO,MAAM,oCAAoC,CAChD,gBACA,EAAE,UAAU,KAAA,IAAkD,CAAA,MACnD;AACX,QAAM,eAAe,IAAI,IAAI,cAAc,YAAY;AAEvD,SAAO,GAAG,aAAa,QAAQ,KAAK,cAAc,GAAG,UAAU,SAAS,EAAE,IACzE,aAAa,IACd;AACD;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"castArray.cjs","sources":["../../../src/lib/castArray.ts"],"sourcesContent":["export const castArray = <T>(input: T | readonly T[]): T[] => {\n\treturn Array.isArray(input) ? input : ([input] as T[]);\n};\n"],"names":[],"mappings":";;
|
1
|
+
{"version":3,"file":"castArray.cjs","sources":["../../../src/lib/castArray.ts"],"sourcesContent":["export const castArray = <T>(input: T | readonly T[]): T[] => {\n\treturn Array.isArray(input) ? input : ([input] as T[]);\n};\n"],"names":[],"mappings":";;AAAO,MAAM,YAAY,CAAI,UAAgC;AAC5D,SAAO,MAAM,QAAQ,KAAK,IAAI,QAAS,CAAC,KAAK;AAC9C;;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"castArray.js","sources":["../../../src/lib/castArray.ts"],"sourcesContent":["export const castArray = <T>(input: T | readonly T[]): T[] => {\n\treturn Array.isArray(input) ? input : ([input] as T[]);\n};\n"],"names":[],"mappings":"
|
1
|
+
{"version":3,"file":"castArray.js","sources":["../../../src/lib/castArray.ts"],"sourcesContent":["export const castArray = <T>(input: T | readonly T[]): T[] => {\n\treturn Array.isArray(input) ? input : ([input] as T[]);\n};\n"],"names":[],"mappings":"AAAO,MAAM,YAAY,CAAI,UAAgC;AAC5D,SAAO,MAAM,QAAQ,KAAK,IAAI,QAAS,CAAC,KAAK;AAC9C;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"createContentDigest.cjs","sources":["../../../src/lib/createContentDigest.ts"],"sourcesContent":["import * as crypto from \"node:crypto\";\n\n/**\n * Creates a content digest for a given input.\n *\n * @param input - The value used to create a digest digest.\n *\n * @returns The content digest of `input`.\n */\nexport const createContentDigest = (input: crypto.BinaryLike): string => {\n\treturn crypto.createHash(\"sha1\").update(input).digest(\"hex\");\n};\n"],"names":["crypto"],"mappings":";;;;;;;;;;;;;;;;;;;;
|
1
|
+
{"version":3,"file":"createContentDigest.cjs","sources":["../../../src/lib/createContentDigest.ts"],"sourcesContent":["import * as crypto from \"node:crypto\";\n\n/**\n * Creates a content digest for a given input.\n *\n * @param input - The value used to create a digest digest.\n *\n * @returns The content digest of `input`.\n */\nexport const createContentDigest = (input: crypto.BinaryLike): string => {\n\treturn crypto.createHash(\"sha1\").update(input).digest(\"hex\");\n};\n"],"names":["crypto"],"mappings":";;;;;;;;;;;;;;;;;;;;AASO,MAAM,sBAAsB,CAAC,UAAoC;AACvE,SAAOA,kBAAO,WAAW,MAAM,EAAE,OAAO,KAAK,EAAE,OAAO,KAAK;AAC5D;;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"createContentDigest.js","sources":["../../../src/lib/createContentDigest.ts"],"sourcesContent":["import * as crypto from \"node:crypto\";\n\n/**\n * Creates a content digest for a given input.\n *\n * @param input - The value used to create a digest digest.\n *\n * @returns The content digest of `input`.\n */\nexport const createContentDigest = (input: crypto.BinaryLike): string => {\n\treturn crypto.createHash(\"sha1\").update(input).digest(\"hex\");\n};\n"],"names":[],"mappings":";
|
1
|
+
{"version":3,"file":"createContentDigest.js","sources":["../../../src/lib/createContentDigest.ts"],"sourcesContent":["import * as crypto from \"node:crypto\";\n\n/**\n * Creates a content digest for a given input.\n *\n * @param input - The value used to create a digest digest.\n *\n * @returns The content digest of `input`.\n */\nexport const createContentDigest = (input: crypto.BinaryLike): string => {\n\treturn crypto.createHash(\"sha1\").update(input).digest(\"hex\");\n};\n"],"names":[],"mappings":";AASO,MAAM,sBAAsB,CAAC,UAAoC;AACvE,SAAO,OAAO,WAAW,MAAM,EAAE,OAAO,KAAK,EAAE,OAAO,KAAK;AAC5D;"}
|
package/dist/lib/decode.cjs.map
CHANGED
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"decode.cjs","sources":["../../../src/lib/decode.ts"],"sourcesContent":["import * as t from \"io-ts\";\nimport { ZodType, ZodTypeDef } from \"zod\";\nimport * as E from \"fp-ts/Either\";\nimport { pipe } from \"fp-ts/function\";\n\nimport { DecodeError } from \"./DecodeError\";\n\nexport type DecodeReturnType<A, _O, I> =\n\t| {\n\t\t\tvalue: A;\n\t\t\terror?: never;\n\t }\n\t| {\n\t\t\tvalue?: never;\n\t\t\terror: DecodeError<I>;\n\t };\n\nfunction isZodSchema(value: unknown): value is ZodType<unknown> {\n\treturn (\n\t\ttypeof (value as ZodType<unknown>).safeParse === \"function\" &&\n\t\tvalue instanceof ZodType\n\t);\n}\n\nexport function decode<A, O, I>(\n\tcodec: ZodType<A, ZodTypeDef, unknown>,\n\tinput: I,\n): DecodeReturnType<A, O, I>;\nexport function decode<A, O, I>(\n\tcodec: t.Type<A, O, I> | ZodType<A, ZodTypeDef, unknown>,\n\tinput: I,\n): DecodeReturnType<A, O, I>;\nexport function decode<A, O, I>(\n\tcodec: t.Type<A, O, I> | ZodType<A, ZodTypeDef, unknown>,\n\tinput: I,\n): DecodeReturnType<A, O, I> {\n\tif (isZodSchema(codec)) {\n\t\tconst parsed = codec.safeParse(input);\n\n\t\tif (parsed.success) {\n\t\t\treturn { value: parsed.data };\n\t\t}\n\n\t\treturn { error: new DecodeError({ input, errors: parsed.error.errors }) };\n\t}\n\n\treturn pipe(\n\t\tcodec.decode(input),\n\t\tE.foldW(\n\t\t\t(errors) => {\n\t\t\t\treturn {\n\t\t\t\t\terror: new DecodeError({ input, errors }),\n\t\t\t\t};\n\t\t\t},\n\t\t\t(value) => {\n\t\t\t\treturn {\n\t\t\t\t\tvalue,\n\t\t\t\t};\n\t\t\t},\n\t\t),\n\t);\n}\n"],"names":["ZodType","DecodeError","pipe","E.foldW"],"mappings":";;;;;;AAiBA,SAAS,YAAY,OAAc;AAClC,SACC,OAAQ,MAA2B,cAAc,cACjD,iBAAiBA,MAAAA;AAEnB;
|
1
|
+
{"version":3,"file":"decode.cjs","sources":["../../../src/lib/decode.ts"],"sourcesContent":["import * as t from \"io-ts\";\nimport { ZodType, ZodTypeDef } from \"zod\";\nimport * as E from \"fp-ts/Either\";\nimport { pipe } from \"fp-ts/function\";\n\nimport { DecodeError } from \"./DecodeError\";\n\nexport type DecodeReturnType<A, _O, I> =\n\t| {\n\t\t\tvalue: A;\n\t\t\terror?: never;\n\t }\n\t| {\n\t\t\tvalue?: never;\n\t\t\terror: DecodeError<I>;\n\t };\n\nfunction isZodSchema(value: unknown): value is ZodType<unknown> {\n\treturn (\n\t\ttypeof (value as ZodType<unknown>).safeParse === \"function\" &&\n\t\tvalue instanceof ZodType\n\t);\n}\n\nexport function decode<A, O, I>(\n\tcodec: ZodType<A, ZodTypeDef, unknown>,\n\tinput: I,\n): DecodeReturnType<A, O, I>;\nexport function decode<A, O, I>(\n\tcodec: t.Type<A, O, I> | ZodType<A, ZodTypeDef, unknown>,\n\tinput: I,\n): DecodeReturnType<A, O, I>;\nexport function decode<A, O, I>(\n\tcodec: t.Type<A, O, I> | ZodType<A, ZodTypeDef, unknown>,\n\tinput: I,\n): DecodeReturnType<A, O, I> {\n\tif (isZodSchema(codec)) {\n\t\tconst parsed = codec.safeParse(input);\n\n\t\tif (parsed.success) {\n\t\t\treturn { value: parsed.data };\n\t\t}\n\n\t\treturn { error: new DecodeError({ input, errors: parsed.error.errors }) };\n\t}\n\n\treturn pipe(\n\t\tcodec.decode(input),\n\t\tE.foldW(\n\t\t\t(errors) => {\n\t\t\t\treturn {\n\t\t\t\t\terror: new DecodeError({ input, errors }),\n\t\t\t\t};\n\t\t\t},\n\t\t\t(value) => {\n\t\t\t\treturn {\n\t\t\t\t\tvalue,\n\t\t\t\t};\n\t\t\t},\n\t\t),\n\t);\n}\n"],"names":["ZodType","DecodeError","pipe","E.foldW"],"mappings":";;;;;;AAiBA,SAAS,YAAY,OAAc;AAClC,SACC,OAAQ,MAA2B,cAAc,cACjD,iBAAiBA,MAAAA;AAEnB;AAUM,SAAU,OACf,OACA,OAAQ;AAER,MAAI,YAAY,KAAK,GAAG;AACvB,UAAM,SAAS,MAAM,UAAU,KAAK;AAEpC,QAAI,OAAO,SAAS;AACnB,aAAO,EAAE,OAAO,OAAO,KAAA;AAAA,IACxB;AAEA,WAAO,EAAE,OAAO,IAAIC,YAAAA,YAAY,EAAE,OAAO,QAAQ,OAAO,MAAM,OAAA,CAAQ,EAAA;AAAA,EACvE;AAEA,SAAOC,UAAAA,KACN,MAAM,OAAO,KAAK,GAClBC,OAAAA,MACC,CAAC,WAAU;AACV,WAAO;AAAA,MACN,OAAO,IAAIF,YAAAA,YAAY,EAAE,OAAO,QAAQ;AAAA,IAAA;AAAA,EAE1C,GACA,CAAC,UAAS;AACT,WAAO;AAAA,MACN;AAAA,IAAA;AAAA,EAEF,CAAC,CACD;AAEH;;"}
|
package/dist/lib/decode.js.map
CHANGED
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"decode.js","sources":["../../../src/lib/decode.ts"],"sourcesContent":["import * as t from \"io-ts\";\nimport { ZodType, ZodTypeDef } from \"zod\";\nimport * as E from \"fp-ts/Either\";\nimport { pipe } from \"fp-ts/function\";\n\nimport { DecodeError } from \"./DecodeError\";\n\nexport type DecodeReturnType<A, _O, I> =\n\t| {\n\t\t\tvalue: A;\n\t\t\terror?: never;\n\t }\n\t| {\n\t\t\tvalue?: never;\n\t\t\terror: DecodeError<I>;\n\t };\n\nfunction isZodSchema(value: unknown): value is ZodType<unknown> {\n\treturn (\n\t\ttypeof (value as ZodType<unknown>).safeParse === \"function\" &&\n\t\tvalue instanceof ZodType\n\t);\n}\n\nexport function decode<A, O, I>(\n\tcodec: ZodType<A, ZodTypeDef, unknown>,\n\tinput: I,\n): DecodeReturnType<A, O, I>;\nexport function decode<A, O, I>(\n\tcodec: t.Type<A, O, I> | ZodType<A, ZodTypeDef, unknown>,\n\tinput: I,\n): DecodeReturnType<A, O, I>;\nexport function decode<A, O, I>(\n\tcodec: t.Type<A, O, I> | ZodType<A, ZodTypeDef, unknown>,\n\tinput: I,\n): DecodeReturnType<A, O, I> {\n\tif (isZodSchema(codec)) {\n\t\tconst parsed = codec.safeParse(input);\n\n\t\tif (parsed.success) {\n\t\t\treturn { value: parsed.data };\n\t\t}\n\n\t\treturn { error: new DecodeError({ input, errors: parsed.error.errors }) };\n\t}\n\n\treturn pipe(\n\t\tcodec.decode(input),\n\t\tE.foldW(\n\t\t\t(errors) => {\n\t\t\t\treturn {\n\t\t\t\t\terror: new DecodeError({ input, errors }),\n\t\t\t\t};\n\t\t\t},\n\t\t\t(value) => {\n\t\t\t\treturn {\n\t\t\t\t\tvalue,\n\t\t\t\t};\n\t\t\t},\n\t\t),\n\t);\n}\n"],"names":["E.foldW"],"mappings":";;;;AAiBA,SAAS,YAAY,OAAc;AAClC,SACC,OAAQ,MAA2B,cAAc,cACjD,iBAAiB;AAEnB;
|
1
|
+
{"version":3,"file":"decode.js","sources":["../../../src/lib/decode.ts"],"sourcesContent":["import * as t from \"io-ts\";\nimport { ZodType, ZodTypeDef } from \"zod\";\nimport * as E from \"fp-ts/Either\";\nimport { pipe } from \"fp-ts/function\";\n\nimport { DecodeError } from \"./DecodeError\";\n\nexport type DecodeReturnType<A, _O, I> =\n\t| {\n\t\t\tvalue: A;\n\t\t\terror?: never;\n\t }\n\t| {\n\t\t\tvalue?: never;\n\t\t\terror: DecodeError<I>;\n\t };\n\nfunction isZodSchema(value: unknown): value is ZodType<unknown> {\n\treturn (\n\t\ttypeof (value as ZodType<unknown>).safeParse === \"function\" &&\n\t\tvalue instanceof ZodType\n\t);\n}\n\nexport function decode<A, O, I>(\n\tcodec: ZodType<A, ZodTypeDef, unknown>,\n\tinput: I,\n): DecodeReturnType<A, O, I>;\nexport function decode<A, O, I>(\n\tcodec: t.Type<A, O, I> | ZodType<A, ZodTypeDef, unknown>,\n\tinput: I,\n): DecodeReturnType<A, O, I>;\nexport function decode<A, O, I>(\n\tcodec: t.Type<A, O, I> | ZodType<A, ZodTypeDef, unknown>,\n\tinput: I,\n): DecodeReturnType<A, O, I> {\n\tif (isZodSchema(codec)) {\n\t\tconst parsed = codec.safeParse(input);\n\n\t\tif (parsed.success) {\n\t\t\treturn { value: parsed.data };\n\t\t}\n\n\t\treturn { error: new DecodeError({ input, errors: parsed.error.errors }) };\n\t}\n\n\treturn pipe(\n\t\tcodec.decode(input),\n\t\tE.foldW(\n\t\t\t(errors) => {\n\t\t\t\treturn {\n\t\t\t\t\terror: new DecodeError({ input, errors }),\n\t\t\t\t};\n\t\t\t},\n\t\t\t(value) => {\n\t\t\t\treturn {\n\t\t\t\t\tvalue,\n\t\t\t\t};\n\t\t\t},\n\t\t),\n\t);\n}\n"],"names":["E.foldW"],"mappings":";;;;AAiBA,SAAS,YAAY,OAAc;AAClC,SACC,OAAQ,MAA2B,cAAc,cACjD,iBAAiB;AAEnB;AAUM,SAAU,OACf,OACA,OAAQ;AAER,MAAI,YAAY,KAAK,GAAG;AACvB,UAAM,SAAS,MAAM,UAAU,KAAK;AAEpC,QAAI,OAAO,SAAS;AACnB,aAAO,EAAE,OAAO,OAAO,KAAA;AAAA,IACxB;AAEA,WAAO,EAAE,OAAO,IAAI,YAAY,EAAE,OAAO,QAAQ,OAAO,MAAM,OAAA,CAAQ,EAAA;AAAA,EACvE;AAEA,SAAO,KACN,MAAM,OAAO,KAAK,GAClBA,MACC,CAAC,WAAU;AACV,WAAO;AAAA,MACN,OAAO,IAAI,YAAY,EAAE,OAAO,QAAQ;AAAA,IAAA;AAAA,EAE1C,GACA,CAAC,UAAS;AACT,WAAO;AAAA,MACN;AAAA,IAAA;AAAA,EAEF,CAAC,CACD;AAEH;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"decodeHookResult.cjs","sources":["../../../src/lib/decodeHookResult.ts"],"sourcesContent":["import * as t from \"io-ts\";\nimport { CallHookReturnType, HookError } from \"@slicemachine/plugin-kit\";\n\nimport { DecodeError } from \"./DecodeError\";\nimport { decode } from \"./decode\";\n\nexport const decodeHookResult = <\n\tA,\n\tO,\n\tI,\n\tTHookResult extends Awaited<CallHookReturnType>,\n>(\n\tcodec: t.Type<A, O, I>,\n\thookResult: THookResult,\n): {\n\tdata: A[];\n\terrors: (HookError | DecodeError<I>)[];\n} => {\n\tconst data: A[] = [];\n\tconst errors: DecodeError<I>[] = [];\n\n\tfor (const dataElement of hookResult.data) {\n\t\tconst { value, error } = decode(codec, dataElement);\n\n\t\tif (error) {\n\t\t\terrors.push(error);\n\t\t} else {\n\t\t\tdata.push(value);\n\t\t}\n\t}\n\n\treturn {\n\t\tdata,\n\t\terrors: [...errors, ...hookResult.errors],\n\t};\n};\n"],"names":["decode"],"mappings":";;;
|
1
|
+
{"version":3,"file":"decodeHookResult.cjs","sources":["../../../src/lib/decodeHookResult.ts"],"sourcesContent":["import * as t from \"io-ts\";\nimport { CallHookReturnType, HookError } from \"@slicemachine/plugin-kit\";\n\nimport { DecodeError } from \"./DecodeError\";\nimport { decode } from \"./decode\";\n\nexport const decodeHookResult = <\n\tA,\n\tO,\n\tI,\n\tTHookResult extends Awaited<CallHookReturnType>,\n>(\n\tcodec: t.Type<A, O, I>,\n\thookResult: THookResult,\n): {\n\tdata: A[];\n\terrors: (HookError | DecodeError<I>)[];\n} => {\n\tconst data: A[] = [];\n\tconst errors: DecodeError<I>[] = [];\n\n\tfor (const dataElement of hookResult.data) {\n\t\tconst { value, error } = decode(codec, dataElement);\n\n\t\tif (error) {\n\t\t\terrors.push(error);\n\t\t} else {\n\t\t\tdata.push(value);\n\t\t}\n\t}\n\n\treturn {\n\t\tdata,\n\t\terrors: [...errors, ...hookResult.errors],\n\t};\n};\n"],"names":["decode"],"mappings":";;;AAMO,MAAM,mBAAmB,CAM/B,OACA,eAIG;AACH,QAAM,OAAY,CAAA;AAClB,QAAM,SAA2B,CAAA;AAEjC,aAAW,eAAe,WAAW,MAAM;AAC1C,UAAM,EAAE,OAAO,MAAA,IAAUA,OAAAA,OAAO,OAAO,WAAW;AAElD,QAAI,OAAO;AACV,aAAO,KAAK,KAAK;AAAA,IAClB,OAAO;AACN,WAAK,KAAK,KAAK;AAAA,IAChB;AAAA,EACD;AAEA,SAAO;AAAA,IACN;AAAA,IACA,QAAQ,CAAC,GAAG,QAAQ,GAAG,WAAW,MAAM;AAAA,EAAA;AAE1C;;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"decodeHookResult.js","sources":["../../../src/lib/decodeHookResult.ts"],"sourcesContent":["import * as t from \"io-ts\";\nimport { CallHookReturnType, HookError } from \"@slicemachine/plugin-kit\";\n\nimport { DecodeError } from \"./DecodeError\";\nimport { decode } from \"./decode\";\n\nexport const decodeHookResult = <\n\tA,\n\tO,\n\tI,\n\tTHookResult extends Awaited<CallHookReturnType>,\n>(\n\tcodec: t.Type<A, O, I>,\n\thookResult: THookResult,\n): {\n\tdata: A[];\n\terrors: (HookError | DecodeError<I>)[];\n} => {\n\tconst data: A[] = [];\n\tconst errors: DecodeError<I>[] = [];\n\n\tfor (const dataElement of hookResult.data) {\n\t\tconst { value, error } = decode(codec, dataElement);\n\n\t\tif (error) {\n\t\t\terrors.push(error);\n\t\t} else {\n\t\t\tdata.push(value);\n\t\t}\n\t}\n\n\treturn {\n\t\tdata,\n\t\terrors: [...errors, ...hookResult.errors],\n\t};\n};\n"],"names":[],"mappings":";
|
1
|
+
{"version":3,"file":"decodeHookResult.js","sources":["../../../src/lib/decodeHookResult.ts"],"sourcesContent":["import * as t from \"io-ts\";\nimport { CallHookReturnType, HookError } from \"@slicemachine/plugin-kit\";\n\nimport { DecodeError } from \"./DecodeError\";\nimport { decode } from \"./decode\";\n\nexport const decodeHookResult = <\n\tA,\n\tO,\n\tI,\n\tTHookResult extends Awaited<CallHookReturnType>,\n>(\n\tcodec: t.Type<A, O, I>,\n\thookResult: THookResult,\n): {\n\tdata: A[];\n\terrors: (HookError | DecodeError<I>)[];\n} => {\n\tconst data: A[] = [];\n\tconst errors: DecodeError<I>[] = [];\n\n\tfor (const dataElement of hookResult.data) {\n\t\tconst { value, error } = decode(codec, dataElement);\n\n\t\tif (error) {\n\t\t\terrors.push(error);\n\t\t} else {\n\t\t\tdata.push(value);\n\t\t}\n\t}\n\n\treturn {\n\t\tdata,\n\t\terrors: [...errors, ...hookResult.errors],\n\t};\n};\n"],"names":[],"mappings":";AAMO,MAAM,mBAAmB,CAM/B,OACA,eAIG;AACH,QAAM,OAAY,CAAA;AAClB,QAAM,SAA2B,CAAA;AAEjC,aAAW,eAAe,WAAW,MAAM;AAC1C,UAAM,EAAE,OAAO,MAAA,IAAU,OAAO,OAAO,WAAW;AAElD,QAAI,OAAO;AACV,aAAO,KAAK,KAAK;AAAA,IAClB,OAAO;AACN,WAAK,KAAK,KAAK;AAAA,IAChB;AAAA,EACD;AAEA,SAAO;AAAA,IACN;AAAA,IACA,QAAQ,CAAC,GAAG,QAAQ,GAAG,WAAW,MAAM;AAAA,EAAA;AAE1C;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"decodePackageJSON.cjs","sources":["../../../src/lib/decodePackageJSON.ts"],"sourcesContent":["import * as t from \"io-ts\";\n\nimport { decode, DecodeReturnType } from \"./decode\";\n\n/**\n * A minimally defined codec for package.json files. Only data needed for\n * version detection is defined.\n */\nconst PackageJSONCodec = t.type({\n\tversion: t.string,\n});\ntype PackageJSON = t.TypeOf<typeof PackageJSONCodec>;\n\nexport const decodePackageJSON = (\n\tinput: unknown,\n): DecodeReturnType<PackageJSON, PackageJSON, unknown> => {\n\treturn decode(PackageJSONCodec, input);\n};\n"],"names":["t","decode"],"mappings":";;;;;;;;;;;;;;;;;;;;;AAQA,MAAM,mBAAmBA,aAAE,KAAK;AAAA,EAC/B,SAASA,aAAE;AACX,CAAA;
|
1
|
+
{"version":3,"file":"decodePackageJSON.cjs","sources":["../../../src/lib/decodePackageJSON.ts"],"sourcesContent":["import * as t from \"io-ts\";\n\nimport { decode, DecodeReturnType } from \"./decode\";\n\n/**\n * A minimally defined codec for package.json files. Only data needed for\n * version detection is defined.\n */\nconst PackageJSONCodec = t.type({\n\tversion: t.string,\n});\ntype PackageJSON = t.TypeOf<typeof PackageJSONCodec>;\n\nexport const decodePackageJSON = (\n\tinput: unknown,\n): DecodeReturnType<PackageJSON, PackageJSON, unknown> => {\n\treturn decode(PackageJSONCodec, input);\n};\n"],"names":["t","decode"],"mappings":";;;;;;;;;;;;;;;;;;;;;AAQA,MAAM,mBAAmBA,aAAE,KAAK;AAAA,EAC/B,SAASA,aAAE;AACX,CAAA;AAGM,MAAM,oBAAoB,CAChC,UACwD;AACxD,SAAOC,OAAAA,OAAO,kBAAkB,KAAK;AACtC;;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"decodePackageJSON.js","sources":["../../../src/lib/decodePackageJSON.ts"],"sourcesContent":["import * as t from \"io-ts\";\n\nimport { decode, DecodeReturnType } from \"./decode\";\n\n/**\n * A minimally defined codec for package.json files. Only data needed for\n * version detection is defined.\n */\nconst PackageJSONCodec = t.type({\n\tversion: t.string,\n});\ntype PackageJSON = t.TypeOf<typeof PackageJSONCodec>;\n\nexport const decodePackageJSON = (\n\tinput: unknown,\n): DecodeReturnType<PackageJSON, PackageJSON, unknown> => {\n\treturn decode(PackageJSONCodec, input);\n};\n"],"names":[],"mappings":";;AAQA,MAAM,mBAAmB,EAAE,KAAK;AAAA,EAC/B,SAAS,EAAE;AACX,CAAA;
|
1
|
+
{"version":3,"file":"decodePackageJSON.js","sources":["../../../src/lib/decodePackageJSON.ts"],"sourcesContent":["import * as t from \"io-ts\";\n\nimport { decode, DecodeReturnType } from \"./decode\";\n\n/**\n * A minimally defined codec for package.json files. Only data needed for\n * version detection is defined.\n */\nconst PackageJSONCodec = t.type({\n\tversion: t.string,\n});\ntype PackageJSON = t.TypeOf<typeof PackageJSONCodec>;\n\nexport const decodePackageJSON = (\n\tinput: unknown,\n): DecodeReturnType<PackageJSON, PackageJSON, unknown> => {\n\treturn decode(PackageJSONCodec, input);\n};\n"],"names":[],"mappings":";;AAQA,MAAM,mBAAmB,EAAE,KAAK;AAAA,EAC/B,SAAS,EAAE;AACX,CAAA;AAGM,MAAM,oBAAoB,CAChC,UACwD;AACxD,SAAO,OAAO,kBAAkB,KAAK;AACtC;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"decodeSliceMachineConfig.cjs","sources":["../../../src/lib/decodeSliceMachineConfig.ts"],"sourcesContent":["import * as t from \"io-ts\";\n\nimport { SliceMachineConfig } from \"../types\";\n\nimport { decode, DecodeReturnType } from \"./decode\";\n\nconst SliceMachineConfigPluginRegistrationCodec = t.union([\n\tt.string,\n\tt.intersection([\n\t\tt.type({\n\t\t\tresolve: t.string,\n\t\t}),\n\t\tt.partial({\n\t\t\toptions: t.UnknownRecord,\n\t\t}),\n\t]),\n]);\n\nconst SliceMachineConfigCodec = t.intersection([\n\tt.type({\n\t\trepositoryName: t.string,\n\t\tadapter: SliceMachineConfigPluginRegistrationCodec,\n\t}),\n\tt.partial({\n\t\tapiEndpoint: t.string,\n\t\tlibraries: t.array(t.string),\n\t\tlocalSliceSimulatorURL: t.string,\n\t\tplugins: t.array(SliceMachineConfigPluginRegistrationCodec),\n\t\tlabs: t.partial({ legacySliceUpgrader: t.boolean }),\n\t}),\n]);\n\n// TODO: Maybe rename \"decode\" to \"validate\". \"decode\" exposes the `io-ts`\n// internals.\nexport const decodeSliceMachineConfig = (\n\tinput: unknown,\n): DecodeReturnType<SliceMachineConfig, SliceMachineConfig, unknown> => {\n\treturn decode(SliceMachineConfigCodec, input);\n};\n"],"names":["t","decode"],"mappings":";;;;;;;;;;;;;;;;;;;;;AAMA,MAAM,4CAA4CA,aAAE,MAAM;AAAA,EACzDA,aAAE;AAAA,EACFA,aAAE,aAAa;AAAA,IACdA,aAAE,KAAK;AAAA,MACN,SAASA,aAAE;AAAA,IAAA,CACX;AAAA,IACDA,aAAE,QAAQ;AAAA,MACT,SAASA,aAAE;AAAA,IAAA,CACX;AAAA,EAAA,CACD;AACD,CAAA;AAED,MAAM,0BAA0BA,aAAE,aAAa;AAAA,EAC9CA,aAAE,KAAK;AAAA,IACN,gBAAgBA,aAAE;AAAA,IAClB,SAAS;AAAA,EAAA,CACT;AAAA,EACDA,aAAE,QAAQ;AAAA,IACT,aAAaA,aAAE;AAAA,IACf,WAAWA,aAAE,MAAMA,aAAE,MAAM;AAAA,IAC3B,wBAAwBA,aAAE;AAAA,IAC1B,SAASA,aAAE,MAAM,yCAAyC;AAAA,IAC1D,MAAMA,aAAE,QAAQ,EAAE,qBAAqBA,aAAE,SAAS;AAAA,EAAA,CAClD;AACD,CAAA;
|
1
|
+
{"version":3,"file":"decodeSliceMachineConfig.cjs","sources":["../../../src/lib/decodeSliceMachineConfig.ts"],"sourcesContent":["import * as t from \"io-ts\";\n\nimport { SliceMachineConfig } from \"../types\";\n\nimport { decode, DecodeReturnType } from \"./decode\";\n\nconst SliceMachineConfigPluginRegistrationCodec = t.union([\n\tt.string,\n\tt.intersection([\n\t\tt.type({\n\t\t\tresolve: t.string,\n\t\t}),\n\t\tt.partial({\n\t\t\toptions: t.UnknownRecord,\n\t\t}),\n\t]),\n]);\n\nconst SliceMachineConfigCodec = t.intersection([\n\tt.type({\n\t\trepositoryName: t.string,\n\t\tadapter: SliceMachineConfigPluginRegistrationCodec,\n\t}),\n\tt.partial({\n\t\tapiEndpoint: t.string,\n\t\tlibraries: t.array(t.string),\n\t\tlocalSliceSimulatorURL: t.string,\n\t\tplugins: t.array(SliceMachineConfigPluginRegistrationCodec),\n\t\tlabs: t.partial({ legacySliceUpgrader: t.boolean }),\n\t}),\n]);\n\n// TODO: Maybe rename \"decode\" to \"validate\". \"decode\" exposes the `io-ts`\n// internals.\nexport const decodeSliceMachineConfig = (\n\tinput: unknown,\n): DecodeReturnType<SliceMachineConfig, SliceMachineConfig, unknown> => {\n\treturn decode(SliceMachineConfigCodec, input);\n};\n"],"names":["t","decode"],"mappings":";;;;;;;;;;;;;;;;;;;;;AAMA,MAAM,4CAA4CA,aAAE,MAAM;AAAA,EACzDA,aAAE;AAAA,EACFA,aAAE,aAAa;AAAA,IACdA,aAAE,KAAK;AAAA,MACN,SAASA,aAAE;AAAA,IAAA,CACX;AAAA,IACDA,aAAE,QAAQ;AAAA,MACT,SAASA,aAAE;AAAA,IAAA,CACX;AAAA,EAAA,CACD;AACD,CAAA;AAED,MAAM,0BAA0BA,aAAE,aAAa;AAAA,EAC9CA,aAAE,KAAK;AAAA,IACN,gBAAgBA,aAAE;AAAA,IAClB,SAAS;AAAA,EAAA,CACT;AAAA,EACDA,aAAE,QAAQ;AAAA,IACT,aAAaA,aAAE;AAAA,IACf,WAAWA,aAAE,MAAMA,aAAE,MAAM;AAAA,IAC3B,wBAAwBA,aAAE;AAAA,IAC1B,SAASA,aAAE,MAAM,yCAAyC;AAAA,IAC1D,MAAMA,aAAE,QAAQ,EAAE,qBAAqBA,aAAE,SAAS;AAAA,EAAA,CAClD;AACD,CAAA;AAIM,MAAM,2BAA2B,CACvC,UACsE;AACtE,SAAOC,OAAAA,OAAO,yBAAyB,KAAK;AAC7C;;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"decodeSliceMachineConfig.js","sources":["../../../src/lib/decodeSliceMachineConfig.ts"],"sourcesContent":["import * as t from \"io-ts\";\n\nimport { SliceMachineConfig } from \"../types\";\n\nimport { decode, DecodeReturnType } from \"./decode\";\n\nconst SliceMachineConfigPluginRegistrationCodec = t.union([\n\tt.string,\n\tt.intersection([\n\t\tt.type({\n\t\t\tresolve: t.string,\n\t\t}),\n\t\tt.partial({\n\t\t\toptions: t.UnknownRecord,\n\t\t}),\n\t]),\n]);\n\nconst SliceMachineConfigCodec = t.intersection([\n\tt.type({\n\t\trepositoryName: t.string,\n\t\tadapter: SliceMachineConfigPluginRegistrationCodec,\n\t}),\n\tt.partial({\n\t\tapiEndpoint: t.string,\n\t\tlibraries: t.array(t.string),\n\t\tlocalSliceSimulatorURL: t.string,\n\t\tplugins: t.array(SliceMachineConfigPluginRegistrationCodec),\n\t\tlabs: t.partial({ legacySliceUpgrader: t.boolean }),\n\t}),\n]);\n\n// TODO: Maybe rename \"decode\" to \"validate\". \"decode\" exposes the `io-ts`\n// internals.\nexport const decodeSliceMachineConfig = (\n\tinput: unknown,\n): DecodeReturnType<SliceMachineConfig, SliceMachineConfig, unknown> => {\n\treturn decode(SliceMachineConfigCodec, input);\n};\n"],"names":[],"mappings":";;AAMA,MAAM,4CAA4C,EAAE,MAAM;AAAA,EACzD,EAAE;AAAA,EACF,EAAE,aAAa;AAAA,IACd,EAAE,KAAK;AAAA,MACN,SAAS,EAAE;AAAA,IAAA,CACX;AAAA,IACD,EAAE,QAAQ;AAAA,MACT,SAAS,EAAE;AAAA,IAAA,CACX;AAAA,EAAA,CACD;AACD,CAAA;AAED,MAAM,0BAA0B,EAAE,aAAa;AAAA,EAC9C,EAAE,KAAK;AAAA,IACN,gBAAgB,EAAE;AAAA,IAClB,SAAS;AAAA,EAAA,CACT;AAAA,EACD,EAAE,QAAQ;AAAA,IACT,aAAa,EAAE;AAAA,IACf,WAAW,EAAE,MAAM,EAAE,MAAM;AAAA,IAC3B,wBAAwB,EAAE;AAAA,IAC1B,SAAS,EAAE,MAAM,yCAAyC;AAAA,IAC1D,MAAM,EAAE,QAAQ,EAAE,qBAAqB,EAAE,SAAS;AAAA,EAAA,CAClD;AACD,CAAA;
|
1
|
+
{"version":3,"file":"decodeSliceMachineConfig.js","sources":["../../../src/lib/decodeSliceMachineConfig.ts"],"sourcesContent":["import * as t from \"io-ts\";\n\nimport { SliceMachineConfig } from \"../types\";\n\nimport { decode, DecodeReturnType } from \"./decode\";\n\nconst SliceMachineConfigPluginRegistrationCodec = t.union([\n\tt.string,\n\tt.intersection([\n\t\tt.type({\n\t\t\tresolve: t.string,\n\t\t}),\n\t\tt.partial({\n\t\t\toptions: t.UnknownRecord,\n\t\t}),\n\t]),\n]);\n\nconst SliceMachineConfigCodec = t.intersection([\n\tt.type({\n\t\trepositoryName: t.string,\n\t\tadapter: SliceMachineConfigPluginRegistrationCodec,\n\t}),\n\tt.partial({\n\t\tapiEndpoint: t.string,\n\t\tlibraries: t.array(t.string),\n\t\tlocalSliceSimulatorURL: t.string,\n\t\tplugins: t.array(SliceMachineConfigPluginRegistrationCodec),\n\t\tlabs: t.partial({ legacySliceUpgrader: t.boolean }),\n\t}),\n]);\n\n// TODO: Maybe rename \"decode\" to \"validate\". \"decode\" exposes the `io-ts`\n// internals.\nexport const decodeSliceMachineConfig = (\n\tinput: unknown,\n): DecodeReturnType<SliceMachineConfig, SliceMachineConfig, unknown> => {\n\treturn decode(SliceMachineConfigCodec, input);\n};\n"],"names":[],"mappings":";;AAMA,MAAM,4CAA4C,EAAE,MAAM;AAAA,EACzD,EAAE;AAAA,EACF,EAAE,aAAa;AAAA,IACd,EAAE,KAAK;AAAA,MACN,SAAS,EAAE;AAAA,IAAA,CACX;AAAA,IACD,EAAE,QAAQ;AAAA,MACT,SAAS,EAAE;AAAA,IAAA,CACX;AAAA,EAAA,CACD;AACD,CAAA;AAED,MAAM,0BAA0B,EAAE,aAAa;AAAA,EAC9C,EAAE,KAAK;AAAA,IACN,gBAAgB,EAAE;AAAA,IAClB,SAAS;AAAA,EAAA,CACT;AAAA,EACD,EAAE,QAAQ;AAAA,IACT,aAAa,EAAE;AAAA,IACf,WAAW,EAAE,MAAM,EAAE,MAAM;AAAA,IAC3B,wBAAwB,EAAE;AAAA,IAC1B,SAAS,EAAE,MAAM,yCAAyC;AAAA,IAC1D,MAAM,EAAE,QAAQ,EAAE,qBAAqB,EAAE,SAAS;AAAA,EAAA,CAClD;AACD,CAAA;AAIM,MAAM,2BAA2B,CACvC,UACsE;AACtE,SAAO,OAAO,yBAAyB,KAAK;AAC7C;"}
|
package/dist/lib/fetch.cjs
CHANGED
package/dist/lib/fetch.cjs.map
CHANGED
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"fetch.cjs","sources":["../../../src/lib/fetch.ts"],"sourcesContent":["// This temporary wrapper around `node-fetch` fixes an issue where quick\n// consecutive network requests cause failed requests.\n//\n// See https://github.com/node-fetch/node-fetch/issues/1735 for more details.\n//\n// TODO: Remove this wrapper and replace all imports with `node-fetch` if https://github.com/node-fetch/node-fetch/pull/1736 is merged.\n\nimport * as http from \"node:http\";\nimport * as https from \"node:https\";\nimport baseFetch from \"node-fetch\";\n\nexport * from \"node-fetch\";\n\n/**\n * The default HTTP Agent with `keepAlive: true` used in `fetch()` requests.\n */\nconst DEFAULT_HTTP_AGENT = new http.Agent({ keepAlive: true });\n\n/**\n * The default HTTPS Agent with `keepAlive: true` used in `fetch()` requests.\n */\nconst DEFAULT_HTTPS_AGENT = new https.Agent({ keepAlive: true });\n\n/**\n * Patched `fetch()` from `node-fetch` that fixes a bug where quick consecutive\n * network requests cause failed requests.\n *\n * Use this `fetch()` in place of `node-fetch`'s `fetch()`.\n *\n * @remarks\n * `fetch()` is patched by setting an HTTP/HTTPS Agent with `keepAlive: true`.\n * If you need to assign an Agent, be sure to retain the `keepAlive: true`\n * option.\n */\nconst fetch: typeof baseFetch = (url, init) => {\n\treturn baseFetch(url, {\n\t\tagent: (parsedURL) => {\n\t\t\treturn parsedURL.protocol === \"http:\"\n\t\t\t\t? DEFAULT_HTTP_AGENT\n\t\t\t\t: DEFAULT_HTTPS_AGENT;\n\t\t},\n\t\t...init,\n\t});\n};\n\nexport default fetch;\n"],"names":["http","https","baseFetch"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;AAgBA,MAAM,qBAAqB,IAAIA,gBAAK,MAAM,EAAE,WAAW,MAAM;AAK7D,MAAM,sBAAsB,IAAIC,iBAAM,MAAM,EAAE,WAAW,MAAM;AAa/D,MAAM,QAA0B,CAAC,KAAK,SAAQ;AAC7C,SAAOC,QAAAA,QAAU,KAAK;AAAA,IACrB,OAAO,CAAC,cAAa;
|
1
|
+
{"version":3,"file":"fetch.cjs","sources":["../../../src/lib/fetch.ts"],"sourcesContent":["// This temporary wrapper around `node-fetch` fixes an issue where quick\n// consecutive network requests cause failed requests.\n//\n// See https://github.com/node-fetch/node-fetch/issues/1735 for more details.\n//\n// TODO: Remove this wrapper and replace all imports with `node-fetch` if https://github.com/node-fetch/node-fetch/pull/1736 is merged.\n\nimport * as http from \"node:http\";\nimport * as https from \"node:https\";\nimport baseFetch from \"node-fetch\";\n\nexport * from \"node-fetch\";\n\n/**\n * The default HTTP Agent with `keepAlive: true` used in `fetch()` requests.\n */\nconst DEFAULT_HTTP_AGENT = new http.Agent({ keepAlive: true });\n\n/**\n * The default HTTPS Agent with `keepAlive: true` used in `fetch()` requests.\n */\nconst DEFAULT_HTTPS_AGENT = new https.Agent({ keepAlive: true });\n\n/**\n * Patched `fetch()` from `node-fetch` that fixes a bug where quick consecutive\n * network requests cause failed requests.\n *\n * Use this `fetch()` in place of `node-fetch`'s `fetch()`.\n *\n * @remarks\n * `fetch()` is patched by setting an HTTP/HTTPS Agent with `keepAlive: true`.\n * If you need to assign an Agent, be sure to retain the `keepAlive: true`\n * option.\n */\nconst fetch: typeof baseFetch = (url, init) => {\n\treturn baseFetch(url, {\n\t\tagent: (parsedURL) => {\n\t\t\treturn parsedURL.protocol === \"http:\"\n\t\t\t\t? DEFAULT_HTTP_AGENT\n\t\t\t\t: DEFAULT_HTTPS_AGENT;\n\t\t},\n\t\t...init,\n\t});\n};\n\nexport default fetch;\n"],"names":["http","https","baseFetch"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;AAgBA,MAAM,qBAAqB,IAAIA,gBAAK,MAAM,EAAE,WAAW,MAAM;AAK7D,MAAM,sBAAsB,IAAIC,iBAAM,MAAM,EAAE,WAAW,MAAM;AAa/D,MAAM,QAA0B,CAAC,KAAK,SAAQ;AAC7C,SAAOC,QAAAA,QAAU,KAAK;AAAA,IACrB,OAAO,CAAC,cAAa;AACpB,aAAO,UAAU,aAAa,UAC3B,qBACA;AAAA,IACJ;AAAA,IACA,GAAG;AAAA,EAAA,CACH;AACF;;"}
|
package/dist/lib/fetch.js
CHANGED
@@ -1,18 +1,17 @@
|
|
1
1
|
import * as http from "node:http";
|
2
2
|
import * as https from "node:https";
|
3
|
-
import fetch$
|
3
|
+
import fetch$1 from './../_node_modules/node-fetch/src/index.js';
|
4
4
|
const DEFAULT_HTTP_AGENT = new http.Agent({ keepAlive: true });
|
5
5
|
const DEFAULT_HTTPS_AGENT = new https.Agent({ keepAlive: true });
|
6
6
|
const fetch = (url, init) => {
|
7
|
-
return fetch$
|
7
|
+
return fetch$1(url, {
|
8
8
|
agent: (parsedURL) => {
|
9
9
|
return parsedURL.protocol === "http:" ? DEFAULT_HTTP_AGENT : DEFAULT_HTTPS_AGENT;
|
10
10
|
},
|
11
11
|
...init
|
12
12
|
});
|
13
13
|
};
|
14
|
-
const fetch$1 = fetch;
|
15
14
|
export {
|
16
|
-
fetch
|
15
|
+
fetch as default
|
17
16
|
};
|
18
17
|
//# sourceMappingURL=fetch.js.map
|
package/dist/lib/fetch.js.map
CHANGED
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"fetch.js","sources":["../../../src/lib/fetch.ts"],"sourcesContent":["// This temporary wrapper around `node-fetch` fixes an issue where quick\n// consecutive network requests cause failed requests.\n//\n// See https://github.com/node-fetch/node-fetch/issues/1735 for more details.\n//\n// TODO: Remove this wrapper and replace all imports with `node-fetch` if https://github.com/node-fetch/node-fetch/pull/1736 is merged.\n\nimport * as http from \"node:http\";\nimport * as https from \"node:https\";\nimport baseFetch from \"node-fetch\";\n\nexport * from \"node-fetch\";\n\n/**\n * The default HTTP Agent with `keepAlive: true` used in `fetch()` requests.\n */\nconst DEFAULT_HTTP_AGENT = new http.Agent({ keepAlive: true });\n\n/**\n * The default HTTPS Agent with `keepAlive: true` used in `fetch()` requests.\n */\nconst DEFAULT_HTTPS_AGENT = new https.Agent({ keepAlive: true });\n\n/**\n * Patched `fetch()` from `node-fetch` that fixes a bug where quick consecutive\n * network requests cause failed requests.\n *\n * Use this `fetch()` in place of `node-fetch`'s `fetch()`.\n *\n * @remarks\n * `fetch()` is patched by setting an HTTP/HTTPS Agent with `keepAlive: true`.\n * If you need to assign an Agent, be sure to retain the `keepAlive: true`\n * option.\n */\nconst fetch: typeof baseFetch = (url, init) => {\n\treturn baseFetch(url, {\n\t\tagent: (parsedURL) => {\n\t\t\treturn parsedURL.protocol === \"http:\"\n\t\t\t\t? DEFAULT_HTTP_AGENT\n\t\t\t\t: DEFAULT_HTTPS_AGENT;\n\t\t},\n\t\t...init,\n\t});\n};\n\nexport default fetch;\n"],"names":["baseFetch"],"mappings":";;;AAgBA,MAAM,qBAAqB,IAAI,KAAK,MAAM,EAAE,WAAW,MAAM;AAK7D,MAAM,sBAAsB,IAAI,MAAM,MAAM,EAAE,WAAW,MAAM;AAa/D,MAAM,QAA0B,CAAC,KAAK,SAAQ;AAC7C,SAAOA,QAAU,KAAK;AAAA,IACrB,OAAO,CAAC,cAAa;
|
1
|
+
{"version":3,"file":"fetch.js","sources":["../../../src/lib/fetch.ts"],"sourcesContent":["// This temporary wrapper around `node-fetch` fixes an issue where quick\n// consecutive network requests cause failed requests.\n//\n// See https://github.com/node-fetch/node-fetch/issues/1735 for more details.\n//\n// TODO: Remove this wrapper and replace all imports with `node-fetch` if https://github.com/node-fetch/node-fetch/pull/1736 is merged.\n\nimport * as http from \"node:http\";\nimport * as https from \"node:https\";\nimport baseFetch from \"node-fetch\";\n\nexport * from \"node-fetch\";\n\n/**\n * The default HTTP Agent with `keepAlive: true` used in `fetch()` requests.\n */\nconst DEFAULT_HTTP_AGENT = new http.Agent({ keepAlive: true });\n\n/**\n * The default HTTPS Agent with `keepAlive: true` used in `fetch()` requests.\n */\nconst DEFAULT_HTTPS_AGENT = new https.Agent({ keepAlive: true });\n\n/**\n * Patched `fetch()` from `node-fetch` that fixes a bug where quick consecutive\n * network requests cause failed requests.\n *\n * Use this `fetch()` in place of `node-fetch`'s `fetch()`.\n *\n * @remarks\n * `fetch()` is patched by setting an HTTP/HTTPS Agent with `keepAlive: true`.\n * If you need to assign an Agent, be sure to retain the `keepAlive: true`\n * option.\n */\nconst fetch: typeof baseFetch = (url, init) => {\n\treturn baseFetch(url, {\n\t\tagent: (parsedURL) => {\n\t\t\treturn parsedURL.protocol === \"http:\"\n\t\t\t\t? DEFAULT_HTTP_AGENT\n\t\t\t\t: DEFAULT_HTTPS_AGENT;\n\t\t},\n\t\t...init,\n\t});\n};\n\nexport default fetch;\n"],"names":["baseFetch"],"mappings":";;;AAgBA,MAAM,qBAAqB,IAAI,KAAK,MAAM,EAAE,WAAW,MAAM;AAK7D,MAAM,sBAAsB,IAAI,MAAM,MAAM,EAAE,WAAW,MAAM;AAa/D,MAAM,QAA0B,CAAC,KAAK,SAAQ;AAC7C,SAAOA,QAAU,KAAK;AAAA,IACrB,OAAO,CAAC,cAAa;AACpB,aAAO,UAAU,aAAa,UAC3B,qBACA;AAAA,IACJ;AAAA,IACA,GAAG;AAAA,EAAA,CACH;AACF;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"fetchGitHubReleaseBodyForRelease.cjs","sources":["../../../src/lib/fetchGitHubReleaseBodyForRelease.ts"],"sourcesContent":["import * as t from \"io-ts\";\nimport fetch from \"./fetch\";\nimport pLimit from \"p-limit\";\n\nimport { decode } from \"./decode\";\n\nconst GITHUB_JSON_ACCEPT_HEADER = \"application/vnd.github+json\";\n\n/**\n * A minimally defined codec for GitHub release metadata.\n *\n * @see https://docs.github.com/en/rest/releases/releases#get-a-release-by-tag-name\n */\nconst GitHubReleaseMetadata = t.type({\n\tname: t.string,\n\tbody: t.union([t.null, t.string]),\n});\nexport type GitHubReleaseMetadata = t.TypeOf<typeof GitHubReleaseMetadata>;\n\ntype FetchAllGitHubReleasesArgs = {\n\trepositoryOwner: string;\n\trepositoryName: string;\n};\n\nconst fetchAllGitHubReleases = async (\n\targs: FetchAllGitHubReleasesArgs,\n): Promise<GitHubReleaseMetadata[]> => {\n\tconst res = await fetch(\n\t\t`https://api.github.com/repos/${args.repositoryOwner}/${args.repositoryName}/releases`,\n\t\t{\n\t\t\theaders: {\n\t\t\t\tAccept: GITHUB_JSON_ACCEPT_HEADER,\n\t\t\t},\n\t\t},\n\t);\n\n\tif (res.ok) {\n\t\tconst json = await res.json();\n\n\t\tconst { value, error } = decode(t.array(GitHubReleaseMetadata), json);\n\n\t\tif (error) {\n\t\t\tthrow new Error(`Invalid GitHub Release response.`, { cause: error });\n\t\t}\n\n\t\treturn value;\n\t} else {\n\t\tconst text = await res.text();\n\t\tthrow new Error(`Invalid GitHub Release response.`, {\n\t\t\tcause: text,\n\t\t});\n\t}\n};\n\ntype FetchGitHubReleaseByVersionArgs = {\n\trepositoryOwner: string;\n\trepositoryName: string;\n\tpackageName?: string;\n\tversion: string;\n};\n\nconst fetchGitHubReleaseByVersion = async (\n\targs: FetchGitHubReleaseByVersionArgs,\n): Promise<GitHubReleaseMetadata | undefined> => {\n\tlet url: string;\n\n\tif (args.packageName) {\n\t\turl = `https://api.github.com/repos/${args.repositoryOwner}/${args.repositoryName}/releases/tags/${args.packageName}@${args.version}`;\n\t} else {\n\t\turl = `https://api.github.com/repos/${args.repositoryOwner}/${args.repositoryName}/releases/tags/${args.version}`;\n\t}\n\n\tconst res = await fetch(url, {\n\t\theaders: {\n\t\t\tAccept: GITHUB_JSON_ACCEPT_HEADER,\n\t\t},\n\t});\n\n\tif (res.ok) {\n\t\tconst json = await res.json();\n\n\t\tconst { value, error } = decode(GitHubReleaseMetadata, json);\n\n\t\tif (error) {\n\t\t\tthrow new Error(`Invalid GitHub Release response.`, { cause: error });\n\t\t}\n\n\t\treturn value;\n\t}\n};\n\ntype FetchGitHubReleaseBodyForReleaseArgs = {\n\trepositoryOwner: string;\n\trepositoryName: string;\n\tpackageName?: string;\n\tversion: string;\n\tcache?: Record<string, GitHubReleaseMetadata | undefined>;\n};\n\nconst _fetchGitHubReleaseBodyForRelease = async (\n\targs: FetchGitHubReleaseBodyForReleaseArgs,\n): Promise<string | undefined> => {\n\tconst cache = args.cache || {};\n\n\tif (Object.keys(cache).length < 1) {\n\t\ttry {\n\t\t\tconst releases = await fetchAllGitHubReleases({\n\t\t\t\trepositoryOwner: args.repositoryOwner,\n\t\t\t\trepositoryName: args.repositoryName,\n\t\t\t});\n\n\t\t\tfor (const release of releases) {\n\t\t\t\tcache[release.name] = release;\n\t\t\t}\n\t\t} catch (error) {\n\t\t\t// noop - Fetch all releases failed, no need to track this error in Sentry.\n\t\t\treturn undefined;\n\t\t}\n\t}\n\n\tif (args.version in cache) {\n\t\tconst release = cache[args.version];\n\n\t\treturn release?.body ?? undefined;\n\t} else {\n\t\ttry {\n\t\t\tconst version = await fetchGitHubReleaseByVersion({\n\t\t\t\trepositoryOwner: args.repositoryOwner,\n\t\t\t\trepositoryName: args.repositoryName,\n\t\t\t\tpackageName: args.packageName,\n\t\t\t\tversion: args.version,\n\t\t\t});\n\n\t\t\tcache[args.version] = version;\n\n\t\t\treturn version?.body ?? undefined;\n\t\t} catch {\n\t\t\tcache[args.version] = undefined;\n\n\t\t\treturn undefined;\n\t\t}\n\t}\n};\n\nconst limit = pLimit(1);\n\nexport const fetchGitHubReleaseBodyForRelease = async (\n\t...args: Parameters<typeof _fetchGitHubReleaseBodyForRelease>\n): ReturnType<typeof _fetchGitHubReleaseBodyForRelease> => {\n\treturn await limit(() => _fetchGitHubReleaseBodyForRelease(...args));\n};\n"],"names":["t","fetch","decode","pLimit"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;AAMA,MAAM,4BAA4B;AAOlC,MAAM,wBAAwBA,aAAE,KAAK;AAAA,EACpC,MAAMA,aAAE;AAAA,EACR,MAAMA,aAAE,MAAM,CAACA,aAAE,MAAMA,aAAE,MAAM,CAAC;AAChC,CAAA;AAQD,MAAM,yBAAyB,OAC9B,SACqC;
|
1
|
+
{"version":3,"file":"fetchGitHubReleaseBodyForRelease.cjs","sources":["../../../src/lib/fetchGitHubReleaseBodyForRelease.ts"],"sourcesContent":["import * as t from \"io-ts\";\nimport fetch from \"./fetch\";\nimport pLimit from \"p-limit\";\n\nimport { decode } from \"./decode\";\n\nconst GITHUB_JSON_ACCEPT_HEADER = \"application/vnd.github+json\";\n\n/**\n * A minimally defined codec for GitHub release metadata.\n *\n * @see https://docs.github.com/en/rest/releases/releases#get-a-release-by-tag-name\n */\nconst GitHubReleaseMetadata = t.type({\n\tname: t.string,\n\tbody: t.union([t.null, t.string]),\n});\nexport type GitHubReleaseMetadata = t.TypeOf<typeof GitHubReleaseMetadata>;\n\ntype FetchAllGitHubReleasesArgs = {\n\trepositoryOwner: string;\n\trepositoryName: string;\n};\n\nconst fetchAllGitHubReleases = async (\n\targs: FetchAllGitHubReleasesArgs,\n): Promise<GitHubReleaseMetadata[]> => {\n\tconst res = await fetch(\n\t\t`https://api.github.com/repos/${args.repositoryOwner}/${args.repositoryName}/releases`,\n\t\t{\n\t\t\theaders: {\n\t\t\t\tAccept: GITHUB_JSON_ACCEPT_HEADER,\n\t\t\t},\n\t\t},\n\t);\n\n\tif (res.ok) {\n\t\tconst json = await res.json();\n\n\t\tconst { value, error } = decode(t.array(GitHubReleaseMetadata), json);\n\n\t\tif (error) {\n\t\t\tthrow new Error(`Invalid GitHub Release response.`, { cause: error });\n\t\t}\n\n\t\treturn value;\n\t} else {\n\t\tconst text = await res.text();\n\t\tthrow new Error(`Invalid GitHub Release response.`, {\n\t\t\tcause: text,\n\t\t});\n\t}\n};\n\ntype FetchGitHubReleaseByVersionArgs = {\n\trepositoryOwner: string;\n\trepositoryName: string;\n\tpackageName?: string;\n\tversion: string;\n};\n\nconst fetchGitHubReleaseByVersion = async (\n\targs: FetchGitHubReleaseByVersionArgs,\n): Promise<GitHubReleaseMetadata | undefined> => {\n\tlet url: string;\n\n\tif (args.packageName) {\n\t\turl = `https://api.github.com/repos/${args.repositoryOwner}/${args.repositoryName}/releases/tags/${args.packageName}@${args.version}`;\n\t} else {\n\t\turl = `https://api.github.com/repos/${args.repositoryOwner}/${args.repositoryName}/releases/tags/${args.version}`;\n\t}\n\n\tconst res = await fetch(url, {\n\t\theaders: {\n\t\t\tAccept: GITHUB_JSON_ACCEPT_HEADER,\n\t\t},\n\t});\n\n\tif (res.ok) {\n\t\tconst json = await res.json();\n\n\t\tconst { value, error } = decode(GitHubReleaseMetadata, json);\n\n\t\tif (error) {\n\t\t\tthrow new Error(`Invalid GitHub Release response.`, { cause: error });\n\t\t}\n\n\t\treturn value;\n\t}\n};\n\ntype FetchGitHubReleaseBodyForReleaseArgs = {\n\trepositoryOwner: string;\n\trepositoryName: string;\n\tpackageName?: string;\n\tversion: string;\n\tcache?: Record<string, GitHubReleaseMetadata | undefined>;\n};\n\nconst _fetchGitHubReleaseBodyForRelease = async (\n\targs: FetchGitHubReleaseBodyForReleaseArgs,\n): Promise<string | undefined> => {\n\tconst cache = args.cache || {};\n\n\tif (Object.keys(cache).length < 1) {\n\t\ttry {\n\t\t\tconst releases = await fetchAllGitHubReleases({\n\t\t\t\trepositoryOwner: args.repositoryOwner,\n\t\t\t\trepositoryName: args.repositoryName,\n\t\t\t});\n\n\t\t\tfor (const release of releases) {\n\t\t\t\tcache[release.name] = release;\n\t\t\t}\n\t\t} catch (error) {\n\t\t\t// noop - Fetch all releases failed, no need to track this error in Sentry.\n\t\t\treturn undefined;\n\t\t}\n\t}\n\n\tif (args.version in cache) {\n\t\tconst release = cache[args.version];\n\n\t\treturn release?.body ?? undefined;\n\t} else {\n\t\ttry {\n\t\t\tconst version = await fetchGitHubReleaseByVersion({\n\t\t\t\trepositoryOwner: args.repositoryOwner,\n\t\t\t\trepositoryName: args.repositoryName,\n\t\t\t\tpackageName: args.packageName,\n\t\t\t\tversion: args.version,\n\t\t\t});\n\n\t\t\tcache[args.version] = version;\n\n\t\t\treturn version?.body ?? undefined;\n\t\t} catch {\n\t\t\tcache[args.version] = undefined;\n\n\t\t\treturn undefined;\n\t\t}\n\t}\n};\n\nconst limit = pLimit(1);\n\nexport const fetchGitHubReleaseBodyForRelease = async (\n\t...args: Parameters<typeof _fetchGitHubReleaseBodyForRelease>\n): ReturnType<typeof _fetchGitHubReleaseBodyForRelease> => {\n\treturn await limit(() => _fetchGitHubReleaseBodyForRelease(...args));\n};\n"],"names":["t","fetch","decode","pLimit"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;AAMA,MAAM,4BAA4B;AAOlC,MAAM,wBAAwBA,aAAE,KAAK;AAAA,EACpC,MAAMA,aAAE;AAAA,EACR,MAAMA,aAAE,MAAM,CAACA,aAAE,MAAMA,aAAE,MAAM,CAAC;AAChC,CAAA;AAQD,MAAM,yBAAyB,OAC9B,SACqC;AACrC,QAAM,MAAM,MAAMC,cACjB,gCAAgC,KAAK,eAAe,IAAI,KAAK,cAAc,aAC3E;AAAA,IACC,SAAS;AAAA,MACR,QAAQ;AAAA,IAAA;AAAA,EACR,CACD;AAGF,MAAI,IAAI,IAAI;AACX,UAAM,OAAO,MAAM,IAAI,KAAA;AAEvB,UAAM,EAAE,OAAO,UAAUC,OAAAA,OAAOF,aAAE,MAAM,qBAAqB,GAAG,IAAI;AAEpE,QAAI,OAAO;AACV,YAAM,IAAI,MAAM,oCAAoC,EAAE,OAAO,OAAO;AAAA,IACrE;AAEA,WAAO;AAAA,EACR,OAAO;AACN,UAAM,OAAO,MAAM,IAAI,KAAA;AACvB,UAAM,IAAI,MAAM,oCAAoC;AAAA,MACnD,OAAO;AAAA,IAAA,CACP;AAAA,EACF;AACD;AASA,MAAM,8BAA8B,OACnC,SAC+C;AAC/C,MAAI;AAEJ,MAAI,KAAK,aAAa;AACrB,UAAM,gCAAgC,KAAK,eAAe,IAAI,KAAK,cAAc,kBAAkB,KAAK,WAAW,IAAI,KAAK,OAAO;AAAA,EACpI,OAAO;AACN,UAAM,gCAAgC,KAAK,eAAe,IAAI,KAAK,cAAc,kBAAkB,KAAK,OAAO;AAAA,EAChH;AAEA,QAAM,MAAM,MAAMC,MAAAA,QAAM,KAAK;AAAA,IAC5B,SAAS;AAAA,MACR,QAAQ;AAAA,IAAA;AAAA,EACR,CACD;AAED,MAAI,IAAI,IAAI;AACX,UAAM,OAAO,MAAM,IAAI,KAAA;AAEvB,UAAM,EAAE,OAAO,MAAA,IAAUC,OAAAA,OAAO,uBAAuB,IAAI;AAE3D,QAAI,OAAO;AACV,YAAM,IAAI,MAAM,oCAAoC,EAAE,OAAO,OAAO;AAAA,IACrE;AAEA,WAAO;AAAA,EACR;AACD;AAUA,MAAM,oCAAoC,OACzC,SACgC;AAChC,QAAM,QAAQ,KAAK,SAAS,CAAA;AAE5B,MAAI,OAAO,KAAK,KAAK,EAAE,SAAS,GAAG;AAClC,QAAI;AACH,YAAM,WAAW,MAAM,uBAAuB;AAAA,QAC7C,iBAAiB,KAAK;AAAA,QACtB,gBAAgB,KAAK;AAAA,MAAA,CACrB;AAED,iBAAW,WAAW,UAAU;AAC/B,cAAM,QAAQ,IAAI,IAAI;AAAA,MACvB;AAAA,IACD,SAAS,OAAO;AAEf,aAAO;AAAA,IACR;AAAA,EACD;AAEA,MAAI,KAAK,WAAW,OAAO;AAC1B,UAAM,UAAU,MAAM,KAAK,OAAO;AAElC,YAAO,mCAAS,SAAQ;AAAA,EACzB,OAAO;AACN,QAAI;AACH,YAAM,UAAU,MAAM,4BAA4B;AAAA,QACjD,iBAAiB,KAAK;AAAA,QACtB,gBAAgB,KAAK;AAAA,QACrB,aAAa,KAAK;AAAA,QAClB,SAAS,KAAK;AAAA,MAAA,CACd;AAED,YAAM,KAAK,OAAO,IAAI;AAEtB,cAAO,mCAAS,SAAQ;AAAA,IACzB,QAAQ;AACP,YAAM,KAAK,OAAO,IAAI;AAEtB,aAAO;AAAA,IACR;AAAA,EACD;AACD;AAEA,MAAM,QAAQC,MAAO,CAAC;AAEf,MAAM,mCAAmC,UAC5C,SACsD;AACzD,SAAO,MAAM,MAAM,MAAM,kCAAkC,GAAG,IAAI,CAAC;AACpE;;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"fetchGitHubReleaseBodyForRelease.js","sources":["../../../src/lib/fetchGitHubReleaseBodyForRelease.ts"],"sourcesContent":["import * as t from \"io-ts\";\nimport fetch from \"./fetch\";\nimport pLimit from \"p-limit\";\n\nimport { decode } from \"./decode\";\n\nconst GITHUB_JSON_ACCEPT_HEADER = \"application/vnd.github+json\";\n\n/**\n * A minimally defined codec for GitHub release metadata.\n *\n * @see https://docs.github.com/en/rest/releases/releases#get-a-release-by-tag-name\n */\nconst GitHubReleaseMetadata = t.type({\n\tname: t.string,\n\tbody: t.union([t.null, t.string]),\n});\nexport type GitHubReleaseMetadata = t.TypeOf<typeof GitHubReleaseMetadata>;\n\ntype FetchAllGitHubReleasesArgs = {\n\trepositoryOwner: string;\n\trepositoryName: string;\n};\n\nconst fetchAllGitHubReleases = async (\n\targs: FetchAllGitHubReleasesArgs,\n): Promise<GitHubReleaseMetadata[]> => {\n\tconst res = await fetch(\n\t\t`https://api.github.com/repos/${args.repositoryOwner}/${args.repositoryName}/releases`,\n\t\t{\n\t\t\theaders: {\n\t\t\t\tAccept: GITHUB_JSON_ACCEPT_HEADER,\n\t\t\t},\n\t\t},\n\t);\n\n\tif (res.ok) {\n\t\tconst json = await res.json();\n\n\t\tconst { value, error } = decode(t.array(GitHubReleaseMetadata), json);\n\n\t\tif (error) {\n\t\t\tthrow new Error(`Invalid GitHub Release response.`, { cause: error });\n\t\t}\n\n\t\treturn value;\n\t} else {\n\t\tconst text = await res.text();\n\t\tthrow new Error(`Invalid GitHub Release response.`, {\n\t\t\tcause: text,\n\t\t});\n\t}\n};\n\ntype FetchGitHubReleaseByVersionArgs = {\n\trepositoryOwner: string;\n\trepositoryName: string;\n\tpackageName?: string;\n\tversion: string;\n};\n\nconst fetchGitHubReleaseByVersion = async (\n\targs: FetchGitHubReleaseByVersionArgs,\n): Promise<GitHubReleaseMetadata | undefined> => {\n\tlet url: string;\n\n\tif (args.packageName) {\n\t\turl = `https://api.github.com/repos/${args.repositoryOwner}/${args.repositoryName}/releases/tags/${args.packageName}@${args.version}`;\n\t} else {\n\t\turl = `https://api.github.com/repos/${args.repositoryOwner}/${args.repositoryName}/releases/tags/${args.version}`;\n\t}\n\n\tconst res = await fetch(url, {\n\t\theaders: {\n\t\t\tAccept: GITHUB_JSON_ACCEPT_HEADER,\n\t\t},\n\t});\n\n\tif (res.ok) {\n\t\tconst json = await res.json();\n\n\t\tconst { value, error } = decode(GitHubReleaseMetadata, json);\n\n\t\tif (error) {\n\t\t\tthrow new Error(`Invalid GitHub Release response.`, { cause: error });\n\t\t}\n\n\t\treturn value;\n\t}\n};\n\ntype FetchGitHubReleaseBodyForReleaseArgs = {\n\trepositoryOwner: string;\n\trepositoryName: string;\n\tpackageName?: string;\n\tversion: string;\n\tcache?: Record<string, GitHubReleaseMetadata | undefined>;\n};\n\nconst _fetchGitHubReleaseBodyForRelease = async (\n\targs: FetchGitHubReleaseBodyForReleaseArgs,\n): Promise<string | undefined> => {\n\tconst cache = args.cache || {};\n\n\tif (Object.keys(cache).length < 1) {\n\t\ttry {\n\t\t\tconst releases = await fetchAllGitHubReleases({\n\t\t\t\trepositoryOwner: args.repositoryOwner,\n\t\t\t\trepositoryName: args.repositoryName,\n\t\t\t});\n\n\t\t\tfor (const release of releases) {\n\t\t\t\tcache[release.name] = release;\n\t\t\t}\n\t\t} catch (error) {\n\t\t\t// noop - Fetch all releases failed, no need to track this error in Sentry.\n\t\t\treturn undefined;\n\t\t}\n\t}\n\n\tif (args.version in cache) {\n\t\tconst release = cache[args.version];\n\n\t\treturn release?.body ?? undefined;\n\t} else {\n\t\ttry {\n\t\t\tconst version = await fetchGitHubReleaseByVersion({\n\t\t\t\trepositoryOwner: args.repositoryOwner,\n\t\t\t\trepositoryName: args.repositoryName,\n\t\t\t\tpackageName: args.packageName,\n\t\t\t\tversion: args.version,\n\t\t\t});\n\n\t\t\tcache[args.version] = version;\n\n\t\t\treturn version?.body ?? undefined;\n\t\t} catch {\n\t\t\tcache[args.version] = undefined;\n\n\t\t\treturn undefined;\n\t\t}\n\t}\n};\n\nconst limit = pLimit(1);\n\nexport const fetchGitHubReleaseBodyForRelease = async (\n\t...args: Parameters<typeof _fetchGitHubReleaseBodyForRelease>\n): ReturnType<typeof _fetchGitHubReleaseBodyForRelease> => {\n\treturn await limit(() => _fetchGitHubReleaseBodyForRelease(...args));\n};\n"],"names":[],"mappings":";;;;AAMA,MAAM,4BAA4B;AAOlC,MAAM,wBAAwB,EAAE,KAAK;AAAA,EACpC,MAAM,EAAE;AAAA,EACR,MAAM,EAAE,MAAM,CAAC,EAAE,MAAM,EAAE,MAAM,CAAC;AAChC,CAAA;AAQD,MAAM,yBAAyB,OAC9B,SACqC;
|
1
|
+
{"version":3,"file":"fetchGitHubReleaseBodyForRelease.js","sources":["../../../src/lib/fetchGitHubReleaseBodyForRelease.ts"],"sourcesContent":["import * as t from \"io-ts\";\nimport fetch from \"./fetch\";\nimport pLimit from \"p-limit\";\n\nimport { decode } from \"./decode\";\n\nconst GITHUB_JSON_ACCEPT_HEADER = \"application/vnd.github+json\";\n\n/**\n * A minimally defined codec for GitHub release metadata.\n *\n * @see https://docs.github.com/en/rest/releases/releases#get-a-release-by-tag-name\n */\nconst GitHubReleaseMetadata = t.type({\n\tname: t.string,\n\tbody: t.union([t.null, t.string]),\n});\nexport type GitHubReleaseMetadata = t.TypeOf<typeof GitHubReleaseMetadata>;\n\ntype FetchAllGitHubReleasesArgs = {\n\trepositoryOwner: string;\n\trepositoryName: string;\n};\n\nconst fetchAllGitHubReleases = async (\n\targs: FetchAllGitHubReleasesArgs,\n): Promise<GitHubReleaseMetadata[]> => {\n\tconst res = await fetch(\n\t\t`https://api.github.com/repos/${args.repositoryOwner}/${args.repositoryName}/releases`,\n\t\t{\n\t\t\theaders: {\n\t\t\t\tAccept: GITHUB_JSON_ACCEPT_HEADER,\n\t\t\t},\n\t\t},\n\t);\n\n\tif (res.ok) {\n\t\tconst json = await res.json();\n\n\t\tconst { value, error } = decode(t.array(GitHubReleaseMetadata), json);\n\n\t\tif (error) {\n\t\t\tthrow new Error(`Invalid GitHub Release response.`, { cause: error });\n\t\t}\n\n\t\treturn value;\n\t} else {\n\t\tconst text = await res.text();\n\t\tthrow new Error(`Invalid GitHub Release response.`, {\n\t\t\tcause: text,\n\t\t});\n\t}\n};\n\ntype FetchGitHubReleaseByVersionArgs = {\n\trepositoryOwner: string;\n\trepositoryName: string;\n\tpackageName?: string;\n\tversion: string;\n};\n\nconst fetchGitHubReleaseByVersion = async (\n\targs: FetchGitHubReleaseByVersionArgs,\n): Promise<GitHubReleaseMetadata | undefined> => {\n\tlet url: string;\n\n\tif (args.packageName) {\n\t\turl = `https://api.github.com/repos/${args.repositoryOwner}/${args.repositoryName}/releases/tags/${args.packageName}@${args.version}`;\n\t} else {\n\t\turl = `https://api.github.com/repos/${args.repositoryOwner}/${args.repositoryName}/releases/tags/${args.version}`;\n\t}\n\n\tconst res = await fetch(url, {\n\t\theaders: {\n\t\t\tAccept: GITHUB_JSON_ACCEPT_HEADER,\n\t\t},\n\t});\n\n\tif (res.ok) {\n\t\tconst json = await res.json();\n\n\t\tconst { value, error } = decode(GitHubReleaseMetadata, json);\n\n\t\tif (error) {\n\t\t\tthrow new Error(`Invalid GitHub Release response.`, { cause: error });\n\t\t}\n\n\t\treturn value;\n\t}\n};\n\ntype FetchGitHubReleaseBodyForReleaseArgs = {\n\trepositoryOwner: string;\n\trepositoryName: string;\n\tpackageName?: string;\n\tversion: string;\n\tcache?: Record<string, GitHubReleaseMetadata | undefined>;\n};\n\nconst _fetchGitHubReleaseBodyForRelease = async (\n\targs: FetchGitHubReleaseBodyForReleaseArgs,\n): Promise<string | undefined> => {\n\tconst cache = args.cache || {};\n\n\tif (Object.keys(cache).length < 1) {\n\t\ttry {\n\t\t\tconst releases = await fetchAllGitHubReleases({\n\t\t\t\trepositoryOwner: args.repositoryOwner,\n\t\t\t\trepositoryName: args.repositoryName,\n\t\t\t});\n\n\t\t\tfor (const release of releases) {\n\t\t\t\tcache[release.name] = release;\n\t\t\t}\n\t\t} catch (error) {\n\t\t\t// noop - Fetch all releases failed, no need to track this error in Sentry.\n\t\t\treturn undefined;\n\t\t}\n\t}\n\n\tif (args.version in cache) {\n\t\tconst release = cache[args.version];\n\n\t\treturn release?.body ?? undefined;\n\t} else {\n\t\ttry {\n\t\t\tconst version = await fetchGitHubReleaseByVersion({\n\t\t\t\trepositoryOwner: args.repositoryOwner,\n\t\t\t\trepositoryName: args.repositoryName,\n\t\t\t\tpackageName: args.packageName,\n\t\t\t\tversion: args.version,\n\t\t\t});\n\n\t\t\tcache[args.version] = version;\n\n\t\t\treturn version?.body ?? undefined;\n\t\t} catch {\n\t\t\tcache[args.version] = undefined;\n\n\t\t\treturn undefined;\n\t\t}\n\t}\n};\n\nconst limit = pLimit(1);\n\nexport const fetchGitHubReleaseBodyForRelease = async (\n\t...args: Parameters<typeof _fetchGitHubReleaseBodyForRelease>\n): ReturnType<typeof _fetchGitHubReleaseBodyForRelease> => {\n\treturn await limit(() => _fetchGitHubReleaseBodyForRelease(...args));\n};\n"],"names":[],"mappings":";;;;AAMA,MAAM,4BAA4B;AAOlC,MAAM,wBAAwB,EAAE,KAAK;AAAA,EACpC,MAAM,EAAE;AAAA,EACR,MAAM,EAAE,MAAM,CAAC,EAAE,MAAM,EAAE,MAAM,CAAC;AAChC,CAAA;AAQD,MAAM,yBAAyB,OAC9B,SACqC;AACrC,QAAM,MAAM,MAAM,MACjB,gCAAgC,KAAK,eAAe,IAAI,KAAK,cAAc,aAC3E;AAAA,IACC,SAAS;AAAA,MACR,QAAQ;AAAA,IAAA;AAAA,EACR,CACD;AAGF,MAAI,IAAI,IAAI;AACX,UAAM,OAAO,MAAM,IAAI,KAAA;AAEvB,UAAM,EAAE,OAAO,UAAU,OAAO,EAAE,MAAM,qBAAqB,GAAG,IAAI;AAEpE,QAAI,OAAO;AACV,YAAM,IAAI,MAAM,oCAAoC,EAAE,OAAO,OAAO;AAAA,IACrE;AAEA,WAAO;AAAA,EACR,OAAO;AACN,UAAM,OAAO,MAAM,IAAI,KAAA;AACvB,UAAM,IAAI,MAAM,oCAAoC;AAAA,MACnD,OAAO;AAAA,IAAA,CACP;AAAA,EACF;AACD;AASA,MAAM,8BAA8B,OACnC,SAC+C;AAC/C,MAAI;AAEJ,MAAI,KAAK,aAAa;AACrB,UAAM,gCAAgC,KAAK,eAAe,IAAI,KAAK,cAAc,kBAAkB,KAAK,WAAW,IAAI,KAAK,OAAO;AAAA,EACpI,OAAO;AACN,UAAM,gCAAgC,KAAK,eAAe,IAAI,KAAK,cAAc,kBAAkB,KAAK,OAAO;AAAA,EAChH;AAEA,QAAM,MAAM,MAAM,MAAM,KAAK;AAAA,IAC5B,SAAS;AAAA,MACR,QAAQ;AAAA,IAAA;AAAA,EACR,CACD;AAED,MAAI,IAAI,IAAI;AACX,UAAM,OAAO,MAAM,IAAI,KAAA;AAEvB,UAAM,EAAE,OAAO,MAAA,IAAU,OAAO,uBAAuB,IAAI;AAE3D,QAAI,OAAO;AACV,YAAM,IAAI,MAAM,oCAAoC,EAAE,OAAO,OAAO;AAAA,IACrE;AAEA,WAAO;AAAA,EACR;AACD;AAUA,MAAM,oCAAoC,OACzC,SACgC;AAChC,QAAM,QAAQ,KAAK,SAAS,CAAA;AAE5B,MAAI,OAAO,KAAK,KAAK,EAAE,SAAS,GAAG;AAClC,QAAI;AACH,YAAM,WAAW,MAAM,uBAAuB;AAAA,QAC7C,iBAAiB,KAAK;AAAA,QACtB,gBAAgB,KAAK;AAAA,MAAA,CACrB;AAED,iBAAW,WAAW,UAAU;AAC/B,cAAM,QAAQ,IAAI,IAAI;AAAA,MACvB;AAAA,IACD,SAAS,OAAO;AAEf,aAAO;AAAA,IACR;AAAA,EACD;AAEA,MAAI,KAAK,WAAW,OAAO;AAC1B,UAAM,UAAU,MAAM,KAAK,OAAO;AAElC,YAAO,mCAAS,SAAQ;AAAA,EACzB,OAAO;AACN,QAAI;AACH,YAAM,UAAU,MAAM,4BAA4B;AAAA,QACjD,iBAAiB,KAAK;AAAA,QACtB,gBAAgB,KAAK;AAAA,QACrB,aAAa,KAAK;AAAA,QAClB,SAAS,KAAK;AAAA,MAAA,CACd;AAED,YAAM,KAAK,OAAO,IAAI;AAEtB,cAAO,mCAAS,SAAQ;AAAA,IACzB,QAAQ;AACP,YAAM,KAAK,OAAO,IAAI;AAEtB,aAAO;AAAA,IACR;AAAA,EACD;AACD;AAEA,MAAM,QAAQ,OAAO,CAAC;AAEf,MAAM,mCAAmC,UAC5C,SACsD;AACzD,SAAO,MAAM,MAAM,MAAM,kCAAkC,GAAG,IAAI,CAAC;AACpE;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"fetchNPMPackageVersions.cjs","sources":["../../../src/lib/fetchNPMPackageVersions.ts"],"sourcesContent":["import * as t from \"io-ts\";\nimport fetch from \"./fetch\";\n\nimport { decode } from \"./decode\";\n\n/**\n * The Accept header value used when only metadata required for package\n * installation is needed. The response includes all versions and some of their\n * metadata.\n *\n * @see https://github.com/npm/registry/blob/cfe04736f34db9274a780184d1cdb2fb3e4ead2a/docs/responses/package-metadata.md#package-metadata\n */\nconst NPM_REGISTRY_ABBREVIATED_METADATA_ACCEPT_HEADER =\n\t\"application/vnd.npm.install-v1+json; q=1.0, application/json; q=0.8, */*\";\n\n/**\n * A minimally defined codec for NPM registry package metadata. Only data needed\n * for version detection is defined.\n */\nconst NPMRegistryPackageMetadataCodec = t.type({\n\tversions: t.UnknownRecord,\n});\n\ntype FetchNPMPackageVersionsArgs = {\n\tpackageName: string;\n};\n\nexport const fetchNPMPackageVersions = async (\n\targs: FetchNPMPackageVersionsArgs,\n): Promise<string[]> => {\n\tconst res = await fetch(`https://registry.npmjs.org/${args.packageName}`, {\n\t\theaders: {\n\t\t\tAccept: NPM_REGISTRY_ABBREVIATED_METADATA_ACCEPT_HEADER,\n\t\t},\n\t});\n\n\tconst json = await res.json();\n\n\tconst { value, error } = decode(NPMRegistryPackageMetadataCodec, json);\n\n\tif (error) {\n\t\tthrow new Error(`Invalid NPM registry response.`, { cause: error });\n\t}\n\n\treturn Object.keys(value.versions);\n};\n"],"names":["t","fetch","decode"],"mappings":";;;;;;;;;;;;;;;;;;;;;;AAYA,MAAM,kDACL;AAMD,MAAM,kCAAkCA,aAAE,KAAK;AAAA,EAC9C,UAAUA,aAAE;AACZ,CAAA;
|
1
|
+
{"version":3,"file":"fetchNPMPackageVersions.cjs","sources":["../../../src/lib/fetchNPMPackageVersions.ts"],"sourcesContent":["import * as t from \"io-ts\";\nimport fetch from \"./fetch\";\n\nimport { decode } from \"./decode\";\n\n/**\n * The Accept header value used when only metadata required for package\n * installation is needed. The response includes all versions and some of their\n * metadata.\n *\n * @see https://github.com/npm/registry/blob/cfe04736f34db9274a780184d1cdb2fb3e4ead2a/docs/responses/package-metadata.md#package-metadata\n */\nconst NPM_REGISTRY_ABBREVIATED_METADATA_ACCEPT_HEADER =\n\t\"application/vnd.npm.install-v1+json; q=1.0, application/json; q=0.8, */*\";\n\n/**\n * A minimally defined codec for NPM registry package metadata. Only data needed\n * for version detection is defined.\n */\nconst NPMRegistryPackageMetadataCodec = t.type({\n\tversions: t.UnknownRecord,\n});\n\ntype FetchNPMPackageVersionsArgs = {\n\tpackageName: string;\n};\n\nexport const fetchNPMPackageVersions = async (\n\targs: FetchNPMPackageVersionsArgs,\n): Promise<string[]> => {\n\tconst res = await fetch(`https://registry.npmjs.org/${args.packageName}`, {\n\t\theaders: {\n\t\t\tAccept: NPM_REGISTRY_ABBREVIATED_METADATA_ACCEPT_HEADER,\n\t\t},\n\t});\n\n\tconst json = await res.json();\n\n\tconst { value, error } = decode(NPMRegistryPackageMetadataCodec, json);\n\n\tif (error) {\n\t\tthrow new Error(`Invalid NPM registry response.`, { cause: error });\n\t}\n\n\treturn Object.keys(value.versions);\n};\n"],"names":["t","fetch","decode"],"mappings":";;;;;;;;;;;;;;;;;;;;;;AAYA,MAAM,kDACL;AAMD,MAAM,kCAAkCA,aAAE,KAAK;AAAA,EAC9C,UAAUA,aAAE;AACZ,CAAA;AAMM,MAAM,0BAA0B,OACtC,SACsB;AACtB,QAAM,MAAM,MAAMC,cAAM,8BAA8B,KAAK,WAAW,IAAI;AAAA,IACzE,SAAS;AAAA,MACR,QAAQ;AAAA,IAAA;AAAA,EACR,CACD;AAED,QAAM,OAAO,MAAM,IAAI,KAAA;AAEvB,QAAM,EAAE,OAAO,MAAA,IAAUC,OAAAA,OAAO,iCAAiC,IAAI;AAErE,MAAI,OAAO;AACV,UAAM,IAAI,MAAM,kCAAkC,EAAE,OAAO,OAAO;AAAA,EACnE;AAEA,SAAO,OAAO,KAAK,MAAM,QAAQ;AAClC;;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"fetchNPMPackageVersions.js","sources":["../../../src/lib/fetchNPMPackageVersions.ts"],"sourcesContent":["import * as t from \"io-ts\";\nimport fetch from \"./fetch\";\n\nimport { decode } from \"./decode\";\n\n/**\n * The Accept header value used when only metadata required for package\n * installation is needed. The response includes all versions and some of their\n * metadata.\n *\n * @see https://github.com/npm/registry/blob/cfe04736f34db9274a780184d1cdb2fb3e4ead2a/docs/responses/package-metadata.md#package-metadata\n */\nconst NPM_REGISTRY_ABBREVIATED_METADATA_ACCEPT_HEADER =\n\t\"application/vnd.npm.install-v1+json; q=1.0, application/json; q=0.8, */*\";\n\n/**\n * A minimally defined codec for NPM registry package metadata. Only data needed\n * for version detection is defined.\n */\nconst NPMRegistryPackageMetadataCodec = t.type({\n\tversions: t.UnknownRecord,\n});\n\ntype FetchNPMPackageVersionsArgs = {\n\tpackageName: string;\n};\n\nexport const fetchNPMPackageVersions = async (\n\targs: FetchNPMPackageVersionsArgs,\n): Promise<string[]> => {\n\tconst res = await fetch(`https://registry.npmjs.org/${args.packageName}`, {\n\t\theaders: {\n\t\t\tAccept: NPM_REGISTRY_ABBREVIATED_METADATA_ACCEPT_HEADER,\n\t\t},\n\t});\n\n\tconst json = await res.json();\n\n\tconst { value, error } = decode(NPMRegistryPackageMetadataCodec, json);\n\n\tif (error) {\n\t\tthrow new Error(`Invalid NPM registry response.`, { cause: error });\n\t}\n\n\treturn Object.keys(value.versions);\n};\n"],"names":[],"mappings":";;;AAYA,MAAM,kDACL;AAMD,MAAM,kCAAkC,EAAE,KAAK;AAAA,EAC9C,UAAU,EAAE;AACZ,CAAA;
|
1
|
+
{"version":3,"file":"fetchNPMPackageVersions.js","sources":["../../../src/lib/fetchNPMPackageVersions.ts"],"sourcesContent":["import * as t from \"io-ts\";\nimport fetch from \"./fetch\";\n\nimport { decode } from \"./decode\";\n\n/**\n * The Accept header value used when only metadata required for package\n * installation is needed. The response includes all versions and some of their\n * metadata.\n *\n * @see https://github.com/npm/registry/blob/cfe04736f34db9274a780184d1cdb2fb3e4ead2a/docs/responses/package-metadata.md#package-metadata\n */\nconst NPM_REGISTRY_ABBREVIATED_METADATA_ACCEPT_HEADER =\n\t\"application/vnd.npm.install-v1+json; q=1.0, application/json; q=0.8, */*\";\n\n/**\n * A minimally defined codec for NPM registry package metadata. Only data needed\n * for version detection is defined.\n */\nconst NPMRegistryPackageMetadataCodec = t.type({\n\tversions: t.UnknownRecord,\n});\n\ntype FetchNPMPackageVersionsArgs = {\n\tpackageName: string;\n};\n\nexport const fetchNPMPackageVersions = async (\n\targs: FetchNPMPackageVersionsArgs,\n): Promise<string[]> => {\n\tconst res = await fetch(`https://registry.npmjs.org/${args.packageName}`, {\n\t\theaders: {\n\t\t\tAccept: NPM_REGISTRY_ABBREVIATED_METADATA_ACCEPT_HEADER,\n\t\t},\n\t});\n\n\tconst json = await res.json();\n\n\tconst { value, error } = decode(NPMRegistryPackageMetadataCodec, json);\n\n\tif (error) {\n\t\tthrow new Error(`Invalid NPM registry response.`, { cause: error });\n\t}\n\n\treturn Object.keys(value.versions);\n};\n"],"names":[],"mappings":";;;AAYA,MAAM,kDACL;AAMD,MAAM,kCAAkC,EAAE,KAAK;AAAA,EAC9C,UAAU,EAAE;AACZ,CAAA;AAMM,MAAM,0BAA0B,OACtC,SACsB;AACtB,QAAM,MAAM,MAAM,MAAM,8BAA8B,KAAK,WAAW,IAAI;AAAA,IACzE,SAAS;AAAA,MACR,QAAQ;AAAA,IAAA;AAAA,EACR,CACD;AAED,QAAM,OAAO,MAAM,IAAI,KAAA;AAEvB,QAAM,EAAE,OAAO,MAAA,IAAU,OAAO,iCAAiC,IAAI;AAErE,MAAI,OAAO;AACV,UAAM,IAAI,MAAM,kCAAkC,EAAE,OAAO,OAAO;AAAA,EACnE;AAEA,SAAO,OAAO,KAAK,MAAM,QAAQ;AAClC;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"findEnvironment.cjs","sources":["../../../src/lib/findEnvironment.ts"],"sourcesContent":["import { Environment } from \"../managers/prismicRepository/types\";\n\nexport const findEnvironment = (\n\tenvironmentDomain: string | undefined,\n\tenvironments: Environment[],\n): Environment | undefined => {\n\treturn environments?.find((environment) => {\n\t\tif (environmentDomain === undefined) {\n\t\t\treturn environment.kind === \"prod\";\n\t\t}\n\n\t\treturn environment.domain === environmentDomain;\n\t});\n};\n"],"names":[],"mappings":";;
|
1
|
+
{"version":3,"file":"findEnvironment.cjs","sources":["../../../src/lib/findEnvironment.ts"],"sourcesContent":["import { Environment } from \"../managers/prismicRepository/types\";\n\nexport const findEnvironment = (\n\tenvironmentDomain: string | undefined,\n\tenvironments: Environment[],\n): Environment | undefined => {\n\treturn environments?.find((environment) => {\n\t\tif (environmentDomain === undefined) {\n\t\t\treturn environment.kind === \"prod\";\n\t\t}\n\n\t\treturn environment.domain === environmentDomain;\n\t});\n};\n"],"names":[],"mappings":";;AAEO,MAAM,kBAAkB,CAC9B,mBACA,iBAC4B;AAC5B,SAAO,6CAAc,KAAK,CAAC,gBAAe;AACzC,QAAI,sBAAsB,QAAW;AACpC,aAAO,YAAY,SAAS;AAAA,IAC7B;AAEA,WAAO,YAAY,WAAW;AAAA,EAC/B;AACD;;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"findEnvironment.js","sources":["../../../src/lib/findEnvironment.ts"],"sourcesContent":["import { Environment } from \"../managers/prismicRepository/types\";\n\nexport const findEnvironment = (\n\tenvironmentDomain: string | undefined,\n\tenvironments: Environment[],\n): Environment | undefined => {\n\treturn environments?.find((environment) => {\n\t\tif (environmentDomain === undefined) {\n\t\t\treturn environment.kind === \"prod\";\n\t\t}\n\n\t\treturn environment.domain === environmentDomain;\n\t});\n};\n"],"names":[],"mappings":"
|
1
|
+
{"version":3,"file":"findEnvironment.js","sources":["../../../src/lib/findEnvironment.ts"],"sourcesContent":["import { Environment } from \"../managers/prismicRepository/types\";\n\nexport const findEnvironment = (\n\tenvironmentDomain: string | undefined,\n\tenvironments: Environment[],\n): Environment | undefined => {\n\treturn environments?.find((environment) => {\n\t\tif (environmentDomain === undefined) {\n\t\t\treturn environment.kind === \"prod\";\n\t\t}\n\n\t\treturn environment.domain === environmentDomain;\n\t});\n};\n"],"names":[],"mappings":"AAEO,MAAM,kBAAkB,CAC9B,mBACA,iBAC4B;AAC5B,SAAO,6CAAc,KAAK,CAAC,gBAAe;AACzC,QAAI,sBAAsB,QAAW;AACpC,aAAO,YAAY,SAAS;AAAA,IAC7B;AAEA,WAAO,YAAY,WAAW;AAAA,EAC/B;AACD;"}
|
package/dist/lib/format.cjs
CHANGED
@@ -22,11 +22,8 @@ const format = async (source, filePath, options) => {
|
|
22
22
|
formatted = await prettier.format(formatted, {
|
23
23
|
...prettierOptions,
|
24
24
|
filepath: filePath,
|
25
|
-
...
|
25
|
+
...{}
|
26
26
|
});
|
27
|
-
if ((options == null ? void 0 : options.includeNewlineAtEnd) === false) {
|
28
|
-
formatted.replace(/[\r\n]+$/, "");
|
29
|
-
}
|
30
27
|
return formatted;
|
31
28
|
};
|
32
29
|
exports.format = format;
|
package/dist/lib/format.cjs.map
CHANGED
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"format.cjs","sources":["../../../src/lib/format.ts"],"sourcesContent":["import prettier from \"prettier\";\nimport { stripIndent } from \"common-tags\";\n\ntype FormatOptions = {\n\tprettier?: prettier.Options;\n\t/**\n\t * Determines if a newline is included at the end of the formatted result.\n\t *\n\t * @defaultValue `true`\n\t */\n\tincludeNewlineAtEnd?: boolean;\n};\n\nexport const format = async (\n\tsource: string,\n\tfilePath: string,\n\toptions?: FormatOptions,\n): Promise<string> => {\n\tlet formatted = stripIndent(source);\n\n\tconst prettierOptions = await prettier.resolveConfig(filePath);\n\n\tformatted = await prettier.format(formatted, {\n\t\t...prettierOptions,\n\t\tfilepath: filePath,\n\t\t...(options?.prettier ?? {}),\n\t});\n\n\tif (options?.includeNewlineAtEnd === false) {\n\t\tformatted.replace(/[\\r\\n]+$/, \"\");\n\t}\n\n\treturn formatted;\n};\n"],"names":[],"mappings":";;;;;;;;;;;;;;;;;;AAaO,MAAM,SAAS,OACrB,QACA,UACA,YACoB;
|
1
|
+
{"version":3,"file":"format.cjs","sources":["../../../src/lib/format.ts"],"sourcesContent":["import prettier from \"prettier\";\nimport { stripIndent } from \"common-tags\";\n\ntype FormatOptions = {\n\tprettier?: prettier.Options;\n\t/**\n\t * Determines if a newline is included at the end of the formatted result.\n\t *\n\t * @defaultValue `true`\n\t */\n\tincludeNewlineAtEnd?: boolean;\n};\n\nexport const format = async (\n\tsource: string,\n\tfilePath: string,\n\toptions?: FormatOptions,\n): Promise<string> => {\n\tlet formatted = stripIndent(source);\n\n\tconst prettierOptions = await prettier.resolveConfig(filePath);\n\n\tformatted = await prettier.format(formatted, {\n\t\t...prettierOptions,\n\t\tfilepath: filePath,\n\t\t...(options?.prettier ?? {}),\n\t});\n\n\tif (options?.includeNewlineAtEnd === false) {\n\t\tformatted.replace(/[\\r\\n]+$/, \"\");\n\t}\n\n\treturn formatted;\n};\n"],"names":[],"mappings":";;;;;;;;;;;;;;;;;;AAaO,MAAM,SAAS,OACrB,QACA,UACA,YACoB;AACpB,MAAI,YAAY,YAAY,MAAM;AAElC,QAAM,kBAAkB,MAAM,SAAS,cAAc,QAAQ;AAE7D,cAAY,MAAM,SAAS,OAAO,WAAW;AAAA,IAC5C,GAAG;AAAA,IACH,UAAU;AAAA,IACV,GAAyB,CAAA;AAAA,GACzB;AAMD,SAAO;AACR;;"}
|
package/dist/lib/format.js
CHANGED
@@ -20,11 +20,8 @@ const format = async (source, filePath, options) => {
|
|
20
20
|
formatted = await prettier.format(formatted, {
|
21
21
|
...prettierOptions,
|
22
22
|
filepath: filePath,
|
23
|
-
...
|
23
|
+
...{}
|
24
24
|
});
|
25
|
-
if ((options == null ? void 0 : options.includeNewlineAtEnd) === false) {
|
26
|
-
formatted.replace(/[\r\n]+$/, "");
|
27
|
-
}
|
28
25
|
return formatted;
|
29
26
|
};
|
30
27
|
export {
|
package/dist/lib/format.js.map
CHANGED
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"format.js","sources":["../../../src/lib/format.ts"],"sourcesContent":["import prettier from \"prettier\";\nimport { stripIndent } from \"common-tags\";\n\ntype FormatOptions = {\n\tprettier?: prettier.Options;\n\t/**\n\t * Determines if a newline is included at the end of the formatted result.\n\t *\n\t * @defaultValue `true`\n\t */\n\tincludeNewlineAtEnd?: boolean;\n};\n\nexport const format = async (\n\tsource: string,\n\tfilePath: string,\n\toptions?: FormatOptions,\n): Promise<string> => {\n\tlet formatted = stripIndent(source);\n\n\tconst prettierOptions = await prettier.resolveConfig(filePath);\n\n\tformatted = await prettier.format(formatted, {\n\t\t...prettierOptions,\n\t\tfilepath: filePath,\n\t\t...(options?.prettier ?? {}),\n\t});\n\n\tif (options?.includeNewlineAtEnd === false) {\n\t\tformatted.replace(/[\\r\\n]+$/, \"\");\n\t}\n\n\treturn formatted;\n};\n"],"names":[],"mappings":";;;;;;;;;;;;;;;;AAaO,MAAM,SAAS,OACrB,QACA,UACA,YACoB;
|
1
|
+
{"version":3,"file":"format.js","sources":["../../../src/lib/format.ts"],"sourcesContent":["import prettier from \"prettier\";\nimport { stripIndent } from \"common-tags\";\n\ntype FormatOptions = {\n\tprettier?: prettier.Options;\n\t/**\n\t * Determines if a newline is included at the end of the formatted result.\n\t *\n\t * @defaultValue `true`\n\t */\n\tincludeNewlineAtEnd?: boolean;\n};\n\nexport const format = async (\n\tsource: string,\n\tfilePath: string,\n\toptions?: FormatOptions,\n): Promise<string> => {\n\tlet formatted = stripIndent(source);\n\n\tconst prettierOptions = await prettier.resolveConfig(filePath);\n\n\tformatted = await prettier.format(formatted, {\n\t\t...prettierOptions,\n\t\tfilepath: filePath,\n\t\t...(options?.prettier ?? {}),\n\t});\n\n\tif (options?.includeNewlineAtEnd === false) {\n\t\tformatted.replace(/[\\r\\n]+$/, \"\");\n\t}\n\n\treturn formatted;\n};\n"],"names":[],"mappings":";;;;;;;;;;;;;;;;AAaO,MAAM,SAAS,OACrB,QACA,UACA,YACoB;AACpB,MAAI,YAAY,YAAY,MAAM;AAElC,QAAM,kBAAkB,MAAM,SAAS,cAAc,QAAQ;AAE7D,cAAY,MAAM,SAAS,OAAO,WAAW;AAAA,IAC5C,GAAG;AAAA,IACH,UAAU;AAAA,IACV,GAAyB,CAAA;AAAA,GACzB;AAMD,SAAO;AACR;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"installDependencies.cjs","sources":["../../../src/lib/installDependencies.ts"],"sourcesContent":["import { parseNi } from \"@antfu/ni\";\nimport {\n\tExecaChildProcess,\n\texecaCommand,\n\tOptions as ExacaOptions,\n} from \"execa\";\n\nimport { PackageManager } from \"../types\";\n\nconst EXTRA_INSTALL_FLAGS: Record<PackageManager, string[]> = {\n\tnpm: [\"--color=always\", \"--loglevel=info\"], // Default: `--color=true --loglevel=notice`.\n\tpnpm: [], // Default: `--color=auto --loglevel=info`.\n\tyarn: [], // TODO: Add the correct options.\n\t\"yarn@berry\": [], // TODO: Add the correct options.\n\t\"pnpm@6\": [], // // TODO: Add the correct options.\n\tbun: [], // TODO: Add the correct options.\n};\n\ntype InstallDependenciesArgs = {\n\tpackageManager: PackageManager;\n\tdependencies: Record<string, string>;\n\tdev?: boolean;\n\texeca?: ExacaOptions;\n};\n\ntype InstallDependenciesReturnType = {\n\texecaProcess: ExecaChildProcess;\n};\n\nexport const installDependencies = async (\n\targs: InstallDependenciesArgs,\n): Promise<InstallDependenciesReturnType> => {\n\tconst commandArgs = Object.entries(args.dependencies).map(\n\t\t([pkg, range]) => `${pkg}@${range}`,\n\t);\n\n\tif (commandArgs.length && args.dev) {\n\t\tcommandArgs.unshift(\"-D\");\n\t}\n\n\tcommandArgs.push(...EXTRA_INSTALL_FLAGS[args.packageManager]);\n\n\tconst command = await parseNi(args.packageManager, commandArgs);\n\n\tif (!command) {\n\t\tthrow new Error(\n\t\t\t\"Failed to begin dependency installation (could not parse command)\",\n\t\t\t{\n\t\t\t\tcause: {\n\t\t\t\t\tpackageManager: args.packageManager,\n\t\t\t\t\tdependencies: args.dependencies,\n\t\t\t\t},\n\t\t\t},\n\t\t);\n\t}\n\n\tconst execaProcess = execaCommand(command, {\n\t\tencoding: \"utf-8\",\n\t\t...args.execa,\n\t});\n\n\treturn {\n\t\texecaProcess,\n\t};\n};\n"],"names":["parseNi","execaCommand"],"mappings":";;;;AASA,MAAM,sBAAwD;AAAA,EAC7D,KAAK,CAAC,kBAAkB,iBAAiB;AAAA;AAAA,EACzC,MAAM,
|
1
|
+
{"version":3,"file":"installDependencies.cjs","sources":["../../../src/lib/installDependencies.ts"],"sourcesContent":["import { parseNi } from \"@antfu/ni\";\nimport {\n\tExecaChildProcess,\n\texecaCommand,\n\tOptions as ExacaOptions,\n} from \"execa\";\n\nimport { PackageManager } from \"../types\";\n\nconst EXTRA_INSTALL_FLAGS: Record<PackageManager, string[]> = {\n\tnpm: [\"--color=always\", \"--loglevel=info\"], // Default: `--color=true --loglevel=notice`.\n\tpnpm: [], // Default: `--color=auto --loglevel=info`.\n\tyarn: [], // TODO: Add the correct options.\n\t\"yarn@berry\": [], // TODO: Add the correct options.\n\t\"pnpm@6\": [], // // TODO: Add the correct options.\n\tbun: [], // TODO: Add the correct options.\n};\n\ntype InstallDependenciesArgs = {\n\tpackageManager: PackageManager;\n\tdependencies: Record<string, string>;\n\tdev?: boolean;\n\texeca?: ExacaOptions;\n};\n\ntype InstallDependenciesReturnType = {\n\texecaProcess: ExecaChildProcess;\n};\n\nexport const installDependencies = async (\n\targs: InstallDependenciesArgs,\n): Promise<InstallDependenciesReturnType> => {\n\tconst commandArgs = Object.entries(args.dependencies).map(\n\t\t([pkg, range]) => `${pkg}@${range}`,\n\t);\n\n\tif (commandArgs.length && args.dev) {\n\t\tcommandArgs.unshift(\"-D\");\n\t}\n\n\tcommandArgs.push(...EXTRA_INSTALL_FLAGS[args.packageManager]);\n\n\tconst command = await parseNi(args.packageManager, commandArgs);\n\n\tif (!command) {\n\t\tthrow new Error(\n\t\t\t\"Failed to begin dependency installation (could not parse command)\",\n\t\t\t{\n\t\t\t\tcause: {\n\t\t\t\t\tpackageManager: args.packageManager,\n\t\t\t\t\tdependencies: args.dependencies,\n\t\t\t\t},\n\t\t\t},\n\t\t);\n\t}\n\n\tconst execaProcess = execaCommand(command, {\n\t\tencoding: \"utf-8\",\n\t\t...args.execa,\n\t});\n\n\treturn {\n\t\texecaProcess,\n\t};\n};\n"],"names":["parseNi","execaCommand"],"mappings":";;;;AASA,MAAM,sBAAwD;AAAA,EAC7D,KAAK,CAAC,kBAAkB,iBAAiB;AAAA;AAAA,EACzC,MAAM,CAAA;AAAA;AAAA,EACN,MAAM,CAAA;AAAA;AAAA,EACN,cAAc,CAAA;AAAA;AAAA,EACd,UAAU,CAAA;AAAA;AAAA,EACV,KAAK,CAAA;AAAA;;AAcC,MAAM,sBAAsB,OAClC,SAC2C;AAC3C,QAAM,cAAc,OAAO,QAAQ,KAAK,YAAY,EAAE,IACrD,CAAC,CAAC,KAAK,KAAK,MAAM,GAAG,GAAG,IAAI,KAAK,EAAE;AAGpC,MAAI,YAAY,UAAU,KAAK,KAAK;AACnC,gBAAY,QAAQ,IAAI;AAAA,EACzB;AAEA,cAAY,KAAK,GAAG,oBAAoB,KAAK,cAAc,CAAC;AAE5D,QAAM,UAAU,MAAMA,GAAAA,QAAQ,KAAK,gBAAgB,WAAW;AAE9D,MAAI,CAAC,SAAS;AACb,UAAM,IAAI,MACT,qEACA;AAAA,MACC,OAAO;AAAA,QACN,gBAAgB,KAAK;AAAA,QACrB,cAAc,KAAK;AAAA,MAAA;AAAA,IACnB,CACD;AAAA,EAEH;AAEA,QAAM,eAAeC,MAAAA,aAAa,SAAS;AAAA,IAC1C,UAAU;AAAA,IACV,GAAG,KAAK;AAAA,EAAA,CACR;AAED,SAAO;AAAA,IACN;AAAA,EAAA;AAEF;;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"installDependencies.js","sources":["../../../src/lib/installDependencies.ts"],"sourcesContent":["import { parseNi } from \"@antfu/ni\";\nimport {\n\tExecaChildProcess,\n\texecaCommand,\n\tOptions as ExacaOptions,\n} from \"execa\";\n\nimport { PackageManager } from \"../types\";\n\nconst EXTRA_INSTALL_FLAGS: Record<PackageManager, string[]> = {\n\tnpm: [\"--color=always\", \"--loglevel=info\"], // Default: `--color=true --loglevel=notice`.\n\tpnpm: [], // Default: `--color=auto --loglevel=info`.\n\tyarn: [], // TODO: Add the correct options.\n\t\"yarn@berry\": [], // TODO: Add the correct options.\n\t\"pnpm@6\": [], // // TODO: Add the correct options.\n\tbun: [], // TODO: Add the correct options.\n};\n\ntype InstallDependenciesArgs = {\n\tpackageManager: PackageManager;\n\tdependencies: Record<string, string>;\n\tdev?: boolean;\n\texeca?: ExacaOptions;\n};\n\ntype InstallDependenciesReturnType = {\n\texecaProcess: ExecaChildProcess;\n};\n\nexport const installDependencies = async (\n\targs: InstallDependenciesArgs,\n): Promise<InstallDependenciesReturnType> => {\n\tconst commandArgs = Object.entries(args.dependencies).map(\n\t\t([pkg, range]) => `${pkg}@${range}`,\n\t);\n\n\tif (commandArgs.length && args.dev) {\n\t\tcommandArgs.unshift(\"-D\");\n\t}\n\n\tcommandArgs.push(...EXTRA_INSTALL_FLAGS[args.packageManager]);\n\n\tconst command = await parseNi(args.packageManager, commandArgs);\n\n\tif (!command) {\n\t\tthrow new Error(\n\t\t\t\"Failed to begin dependency installation (could not parse command)\",\n\t\t\t{\n\t\t\t\tcause: {\n\t\t\t\t\tpackageManager: args.packageManager,\n\t\t\t\t\tdependencies: args.dependencies,\n\t\t\t\t},\n\t\t\t},\n\t\t);\n\t}\n\n\tconst execaProcess = execaCommand(command, {\n\t\tencoding: \"utf-8\",\n\t\t...args.execa,\n\t});\n\n\treturn {\n\t\texecaProcess,\n\t};\n};\n"],"names":[],"mappings":";;AASA,MAAM,sBAAwD;AAAA,EAC7D,KAAK,CAAC,kBAAkB,iBAAiB;AAAA;AAAA,EACzC,MAAM,
|
1
|
+
{"version":3,"file":"installDependencies.js","sources":["../../../src/lib/installDependencies.ts"],"sourcesContent":["import { parseNi } from \"@antfu/ni\";\nimport {\n\tExecaChildProcess,\n\texecaCommand,\n\tOptions as ExacaOptions,\n} from \"execa\";\n\nimport { PackageManager } from \"../types\";\n\nconst EXTRA_INSTALL_FLAGS: Record<PackageManager, string[]> = {\n\tnpm: [\"--color=always\", \"--loglevel=info\"], // Default: `--color=true --loglevel=notice`.\n\tpnpm: [], // Default: `--color=auto --loglevel=info`.\n\tyarn: [], // TODO: Add the correct options.\n\t\"yarn@berry\": [], // TODO: Add the correct options.\n\t\"pnpm@6\": [], // // TODO: Add the correct options.\n\tbun: [], // TODO: Add the correct options.\n};\n\ntype InstallDependenciesArgs = {\n\tpackageManager: PackageManager;\n\tdependencies: Record<string, string>;\n\tdev?: boolean;\n\texeca?: ExacaOptions;\n};\n\ntype InstallDependenciesReturnType = {\n\texecaProcess: ExecaChildProcess;\n};\n\nexport const installDependencies = async (\n\targs: InstallDependenciesArgs,\n): Promise<InstallDependenciesReturnType> => {\n\tconst commandArgs = Object.entries(args.dependencies).map(\n\t\t([pkg, range]) => `${pkg}@${range}`,\n\t);\n\n\tif (commandArgs.length && args.dev) {\n\t\tcommandArgs.unshift(\"-D\");\n\t}\n\n\tcommandArgs.push(...EXTRA_INSTALL_FLAGS[args.packageManager]);\n\n\tconst command = await parseNi(args.packageManager, commandArgs);\n\n\tif (!command) {\n\t\tthrow new Error(\n\t\t\t\"Failed to begin dependency installation (could not parse command)\",\n\t\t\t{\n\t\t\t\tcause: {\n\t\t\t\t\tpackageManager: args.packageManager,\n\t\t\t\t\tdependencies: args.dependencies,\n\t\t\t\t},\n\t\t\t},\n\t\t);\n\t}\n\n\tconst execaProcess = execaCommand(command, {\n\t\tencoding: \"utf-8\",\n\t\t...args.execa,\n\t});\n\n\treturn {\n\t\texecaProcess,\n\t};\n};\n"],"names":[],"mappings":";;AASA,MAAM,sBAAwD;AAAA,EAC7D,KAAK,CAAC,kBAAkB,iBAAiB;AAAA;AAAA,EACzC,MAAM,CAAA;AAAA;AAAA,EACN,MAAM,CAAA;AAAA;AAAA,EACN,cAAc,CAAA;AAAA;AAAA,EACd,UAAU,CAAA;AAAA;AAAA,EACV,KAAK,CAAA;AAAA;;AAcC,MAAM,sBAAsB,OAClC,SAC2C;AAC3C,QAAM,cAAc,OAAO,QAAQ,KAAK,YAAY,EAAE,IACrD,CAAC,CAAC,KAAK,KAAK,MAAM,GAAG,GAAG,IAAI,KAAK,EAAE;AAGpC,MAAI,YAAY,UAAU,KAAK,KAAK;AACnC,gBAAY,QAAQ,IAAI;AAAA,EACzB;AAEA,cAAY,KAAK,GAAG,oBAAoB,KAAK,cAAc,CAAC;AAE5D,QAAM,UAAU,MAAM,QAAQ,KAAK,gBAAgB,WAAW;AAE9D,MAAI,CAAC,SAAS;AACb,UAAM,IAAI,MACT,qEACA;AAAA,MACC,OAAO;AAAA,QACN,gBAAgB,KAAK;AAAA,QACrB,cAAc,KAAK;AAAA,MAAA;AAAA,IACnB,CACD;AAAA,EAEH;AAEA,QAAM,eAAe,aAAa,SAAS;AAAA,IAC1C,UAAU;AAAA,IACV,GAAG,KAAK;AAAA,EAAA,CACR;AAED,SAAO;AAAA,IACN;AAAA,EAAA;AAEF;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"locateFileUpward.cjs","sources":["../../../src/lib/locateFileUpward.ts"],"sourcesContent":["import * as path from \"node:path\";\nimport * as fs from \"node:fs/promises\";\n\nimport { castArray } from \"./castArray\";\n\ntype LocateFileUpwardConfig = {\n\tstartDir?: string;\n\tstopDir?: string;\n\t/**\n\t * @internal\n\t */\n\t_originalStartDir?: string;\n};\n\nexport const locateFileUpward = async (\n\tfilePathOrPaths: string | readonly string[],\n\t{\n\t\tstartDir = process.cwd(),\n\t\tstopDir = path.resolve(startDir, \"/\"),\n\t\t_originalStartDir,\n\t}: LocateFileUpwardConfig = {},\n): Promise<string> => {\n\tconst originalStartDir = _originalStartDir ?? startDir;\n\n\tconst filePaths = castArray(filePathOrPaths);\n\n\tfor (const filePath of filePaths) {\n\t\tconst resolvedFilePath = path.resolve(startDir, filePath);\n\t\ttry {\n\t\t\tawait fs.access(resolvedFilePath);\n\n\t\t\treturn resolvedFilePath;\n\t\t} catch {\n\t\t\tcontinue;\n\t\t}\n\t}\n\n\tif (startDir === stopDir) {\n\t\tconst formattedFilePaths = filePaths\n\t\t\t.map((filePath) => \"`\" + filePath + \"`\")\n\t\t\t.join(\" or \");\n\n\t\tthrow new Error(\n\t\t\t`Could not locate ${formattedFilePaths} between \\`${originalStartDir}\\` and \\`${stopDir}\\`.`,\n\t\t);\n\t}\n\n\treturn locateFileUpward(filePathOrPaths, {\n\t\tstartDir: path.resolve(startDir, \"..\"),\n\t\tstopDir,\n\t\t_originalStartDir: originalStartDir,\n\t});\n};\n"],"names":["path","castArray","fs"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;AAcO,MAAM,mBAAmB,OAC/B,iBACA,EACC,WAAW,QAAQ,
|
1
|
+
{"version":3,"file":"locateFileUpward.cjs","sources":["../../../src/lib/locateFileUpward.ts"],"sourcesContent":["import * as path from \"node:path\";\nimport * as fs from \"node:fs/promises\";\n\nimport { castArray } from \"./castArray\";\n\ntype LocateFileUpwardConfig = {\n\tstartDir?: string;\n\tstopDir?: string;\n\t/**\n\t * @internal\n\t */\n\t_originalStartDir?: string;\n};\n\nexport const locateFileUpward = async (\n\tfilePathOrPaths: string | readonly string[],\n\t{\n\t\tstartDir = process.cwd(),\n\t\tstopDir = path.resolve(startDir, \"/\"),\n\t\t_originalStartDir,\n\t}: LocateFileUpwardConfig = {},\n): Promise<string> => {\n\tconst originalStartDir = _originalStartDir ?? startDir;\n\n\tconst filePaths = castArray(filePathOrPaths);\n\n\tfor (const filePath of filePaths) {\n\t\tconst resolvedFilePath = path.resolve(startDir, filePath);\n\t\ttry {\n\t\t\tawait fs.access(resolvedFilePath);\n\n\t\t\treturn resolvedFilePath;\n\t\t} catch {\n\t\t\tcontinue;\n\t\t}\n\t}\n\n\tif (startDir === stopDir) {\n\t\tconst formattedFilePaths = filePaths\n\t\t\t.map((filePath) => \"`\" + filePath + \"`\")\n\t\t\t.join(\" or \");\n\n\t\tthrow new Error(\n\t\t\t`Could not locate ${formattedFilePaths} between \\`${originalStartDir}\\` and \\`${stopDir}\\`.`,\n\t\t);\n\t}\n\n\treturn locateFileUpward(filePathOrPaths, {\n\t\tstartDir: path.resolve(startDir, \"..\"),\n\t\tstopDir,\n\t\t_originalStartDir: originalStartDir,\n\t});\n};\n"],"names":["path","castArray","fs"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;AAcO,MAAM,mBAAmB,OAC/B,iBACA,EACC,WAAW,QAAQ,IAAA,GACnB,UAAUA,gBAAK,QAAQ,UAAU,GAAG,GACpC,kBAAA,IAC2B,CAAA,MACR;AACpB,QAAM,mBAAmB,qBAAqB;AAE9C,QAAM,YAAYC,UAAAA,UAAU,eAAe;AAE3C,aAAW,YAAY,WAAW;AACjC,UAAM,mBAAmBD,gBAAK,QAAQ,UAAU,QAAQ;AACxD,QAAI;AACH,YAAME,cAAG,OAAO,gBAAgB;AAEhC,aAAO;AAAA,IACR,QAAQ;AACP;AAAA,IACD;AAAA,EACD;AAEA,MAAI,aAAa,SAAS;AACzB,UAAM,qBAAqB,UACzB,IAAI,CAAC,aAAa,MAAM,WAAW,GAAG,EACtC,KAAK,MAAM;AAEb,UAAM,IAAI,MACT,oBAAoB,kBAAkB,cAAc,gBAAgB,YAAY,OAAO,KAAK;AAAA,EAE9F;AAEA,SAAO,iBAAiB,iBAAiB;AAAA,IACxC,UAAUF,gBAAK,QAAQ,UAAU,IAAI;AAAA,IACrC;AAAA,IACA,mBAAmB;AAAA,EAAA,CACnB;AACF;;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"locateFileUpward.js","sources":["../../../src/lib/locateFileUpward.ts"],"sourcesContent":["import * as path from \"node:path\";\nimport * as fs from \"node:fs/promises\";\n\nimport { castArray } from \"./castArray\";\n\ntype LocateFileUpwardConfig = {\n\tstartDir?: string;\n\tstopDir?: string;\n\t/**\n\t * @internal\n\t */\n\t_originalStartDir?: string;\n};\n\nexport const locateFileUpward = async (\n\tfilePathOrPaths: string | readonly string[],\n\t{\n\t\tstartDir = process.cwd(),\n\t\tstopDir = path.resolve(startDir, \"/\"),\n\t\t_originalStartDir,\n\t}: LocateFileUpwardConfig = {},\n): Promise<string> => {\n\tconst originalStartDir = _originalStartDir ?? startDir;\n\n\tconst filePaths = castArray(filePathOrPaths);\n\n\tfor (const filePath of filePaths) {\n\t\tconst resolvedFilePath = path.resolve(startDir, filePath);\n\t\ttry {\n\t\t\tawait fs.access(resolvedFilePath);\n\n\t\t\treturn resolvedFilePath;\n\t\t} catch {\n\t\t\tcontinue;\n\t\t}\n\t}\n\n\tif (startDir === stopDir) {\n\t\tconst formattedFilePaths = filePaths\n\t\t\t.map((filePath) => \"`\" + filePath + \"`\")\n\t\t\t.join(\" or \");\n\n\t\tthrow new Error(\n\t\t\t`Could not locate ${formattedFilePaths} between \\`${originalStartDir}\\` and \\`${stopDir}\\`.`,\n\t\t);\n\t}\n\n\treturn locateFileUpward(filePathOrPaths, {\n\t\tstartDir: path.resolve(startDir, \"..\"),\n\t\tstopDir,\n\t\t_originalStartDir: originalStartDir,\n\t});\n};\n"],"names":[],"mappings":";;;AAcO,MAAM,mBAAmB,OAC/B,iBACA,EACC,WAAW,QAAQ,
|
1
|
+
{"version":3,"file":"locateFileUpward.js","sources":["../../../src/lib/locateFileUpward.ts"],"sourcesContent":["import * as path from \"node:path\";\nimport * as fs from \"node:fs/promises\";\n\nimport { castArray } from \"./castArray\";\n\ntype LocateFileUpwardConfig = {\n\tstartDir?: string;\n\tstopDir?: string;\n\t/**\n\t * @internal\n\t */\n\t_originalStartDir?: string;\n};\n\nexport const locateFileUpward = async (\n\tfilePathOrPaths: string | readonly string[],\n\t{\n\t\tstartDir = process.cwd(),\n\t\tstopDir = path.resolve(startDir, \"/\"),\n\t\t_originalStartDir,\n\t}: LocateFileUpwardConfig = {},\n): Promise<string> => {\n\tconst originalStartDir = _originalStartDir ?? startDir;\n\n\tconst filePaths = castArray(filePathOrPaths);\n\n\tfor (const filePath of filePaths) {\n\t\tconst resolvedFilePath = path.resolve(startDir, filePath);\n\t\ttry {\n\t\t\tawait fs.access(resolvedFilePath);\n\n\t\t\treturn resolvedFilePath;\n\t\t} catch {\n\t\t\tcontinue;\n\t\t}\n\t}\n\n\tif (startDir === stopDir) {\n\t\tconst formattedFilePaths = filePaths\n\t\t\t.map((filePath) => \"`\" + filePath + \"`\")\n\t\t\t.join(\" or \");\n\n\t\tthrow new Error(\n\t\t\t`Could not locate ${formattedFilePaths} between \\`${originalStartDir}\\` and \\`${stopDir}\\`.`,\n\t\t);\n\t}\n\n\treturn locateFileUpward(filePathOrPaths, {\n\t\tstartDir: path.resolve(startDir, \"..\"),\n\t\tstopDir,\n\t\t_originalStartDir: originalStartDir,\n\t});\n};\n"],"names":[],"mappings":";;;AAcO,MAAM,mBAAmB,OAC/B,iBACA,EACC,WAAW,QAAQ,IAAA,GACnB,UAAU,KAAK,QAAQ,UAAU,GAAG,GACpC,kBAAA,IAC2B,CAAA,MACR;AACpB,QAAM,mBAAmB,qBAAqB;AAE9C,QAAM,YAAY,UAAU,eAAe;AAE3C,aAAW,YAAY,WAAW;AACjC,UAAM,mBAAmB,KAAK,QAAQ,UAAU,QAAQ;AACxD,QAAI;AACH,YAAM,GAAG,OAAO,gBAAgB;AAEhC,aAAO;AAAA,IACR,QAAQ;AACP;AAAA,IACD;AAAA,EACD;AAEA,MAAI,aAAa,SAAS;AACzB,UAAM,qBAAqB,UACzB,IAAI,CAAC,aAAa,MAAM,WAAW,GAAG,EACtC,KAAK,MAAM;AAEb,UAAM,IAAI,MACT,oBAAoB,kBAAkB,cAAc,gBAAgB,YAAY,OAAO,KAAK;AAAA,EAE9F;AAEA,SAAO,iBAAiB,iBAAiB;AAAA,IACxC,UAAU,KAAK,QAAQ,UAAU,IAAI;AAAA,IACrC;AAAA,IACA,mBAAmB;AAAA,EAAA,CACnB;AACF;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"mockSlice.cjs","sources":["../../../src/lib/mockSlice.ts"],"sourcesContent":["import { SharedSliceMock } from \"@prismicio/mocks\";\n\nimport { SharedSliceContent } from \"@prismicio/types-internal/lib/content\";\nimport {\n\tSharedSlice,\n\tSliceDiff,\n} from \"@prismicio/types-internal/lib/customtypes\";\n\ntype mockSliceArgs = {\n\tmodel: SharedSlice;\n\tmocks?: SharedSliceContent[];\n\tdiff?: SliceDiff;\n};\n\nexport const mockSlice = (args: mockSliceArgs): SharedSliceContent[] => {\n\tconst { model, mocks, diff } = args;\n\n\treturn model.variations.map((variation) => {\n\t\tconst variationMock = mocks?.find((m) => m.variation === variation.id);\n\n\t\tconst mockConfig = {\n\t\t\ttype: \"SharedSlice\",\n\t\t\tvariation: variation.id,\n\t\t} as const;\n\n\t\tif (!variationMock) {\n\t\t\treturn SharedSliceMock.generate(model, mockConfig);\n\t\t}\n\t\tif (!diff) {\n\t\t\treturn variationMock;\n\t\t}\n\n\t\t// Enforce at least one item to be created if none are present and variation has item fields\n\t\tif (\n\t\t\tObject.keys(variation.items || {}).length &&\n\t\t\tvariationMock.items.length === 0\n\t\t) {\n\t\t\tvariationMock.items.push({ __TYPE__: \"GroupItemContent\", value: [] });\n\t\t}\n\t\tconst patched = SharedSliceMock.patch(diff, variationMock, mockConfig);\n\t\tif (!patched.ok) {\n\t\t\treturn variationMock;\n\t\t}\n\t\tif (!patched.result) {\n\t\t\treturn SharedSliceMock.generate(model, mockConfig);\n\t\t}\n\n\t\treturn patched.result;\n\t});\n};\n"],"names":["mocks","SharedSliceMock"],"mappings":";;;
|
1
|
+
{"version":3,"file":"mockSlice.cjs","sources":["../../../src/lib/mockSlice.ts"],"sourcesContent":["import { SharedSliceMock } from \"@prismicio/mocks\";\n\nimport { SharedSliceContent } from \"@prismicio/types-internal/lib/content\";\nimport {\n\tSharedSlice,\n\tSliceDiff,\n} from \"@prismicio/types-internal/lib/customtypes\";\n\ntype mockSliceArgs = {\n\tmodel: SharedSlice;\n\tmocks?: SharedSliceContent[];\n\tdiff?: SliceDiff;\n};\n\nexport const mockSlice = (args: mockSliceArgs): SharedSliceContent[] => {\n\tconst { model, mocks, diff } = args;\n\n\treturn model.variations.map((variation) => {\n\t\tconst variationMock = mocks?.find((m) => m.variation === variation.id);\n\n\t\tconst mockConfig = {\n\t\t\ttype: \"SharedSlice\",\n\t\t\tvariation: variation.id,\n\t\t} as const;\n\n\t\tif (!variationMock) {\n\t\t\treturn SharedSliceMock.generate(model, mockConfig);\n\t\t}\n\t\tif (!diff) {\n\t\t\treturn variationMock;\n\t\t}\n\n\t\t// Enforce at least one item to be created if none are present and variation has item fields\n\t\tif (\n\t\t\tObject.keys(variation.items || {}).length &&\n\t\t\tvariationMock.items.length === 0\n\t\t) {\n\t\t\tvariationMock.items.push({ __TYPE__: \"GroupItemContent\", value: [] });\n\t\t}\n\t\tconst patched = SharedSliceMock.patch(diff, variationMock, mockConfig);\n\t\tif (!patched.ok) {\n\t\t\treturn variationMock;\n\t\t}\n\t\tif (!patched.result) {\n\t\t\treturn SharedSliceMock.generate(model, mockConfig);\n\t\t}\n\n\t\treturn patched.result;\n\t});\n};\n"],"names":["mocks","SharedSliceMock"],"mappings":";;;AAcO,MAAM,YAAY,CAAC,SAA6C;AACtE,QAAM,EAAE,OAAA,OAAOA,SAAO,KAAA,IAAS;AAE/B,SAAO,MAAM,WAAW,IAAI,CAAC,cAAa;AACzC,UAAM,gBAAgBA,mCAAO,KAAK,CAAC,MAAM,EAAE,cAAc,UAAU;AAEnE,UAAM,aAAa;AAAA,MAClB,MAAM;AAAA,MACN,WAAW,UAAU;AAAA,IAAA;AAGtB,QAAI,CAAC,eAAe;AACnB,aAAOC,sBAAgB,SAAS,OAAO,UAAU;AAAA,IAClD;AACA,QAAI,CAAC,MAAM;AACV,aAAO;AAAA,IACR;AAGA,QACC,OAAO,KAAK,UAAU,SAAS,CAAA,CAAE,EAAE,UACnC,cAAc,MAAM,WAAW,GAC9B;AACD,oBAAc,MAAM,KAAK,EAAE,UAAU,oBAAoB,OAAO,CAAA,GAAI;AAAA,IACrE;AACA,UAAM,UAAUA,MAAAA,gBAAgB,MAAM,MAAM,eAAe,UAAU;AACrE,QAAI,CAAC,QAAQ,IAAI;AAChB,aAAO;AAAA,IACR;AACA,QAAI,CAAC,QAAQ,QAAQ;AACpB,aAAOA,sBAAgB,SAAS,OAAO,UAAU;AAAA,IAClD;AAEA,WAAO,QAAQ;AAAA,EAChB,CAAC;AACF;;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"mockSlice.js","sources":["../../../src/lib/mockSlice.ts"],"sourcesContent":["import { SharedSliceMock } from \"@prismicio/mocks\";\n\nimport { SharedSliceContent } from \"@prismicio/types-internal/lib/content\";\nimport {\n\tSharedSlice,\n\tSliceDiff,\n} from \"@prismicio/types-internal/lib/customtypes\";\n\ntype mockSliceArgs = {\n\tmodel: SharedSlice;\n\tmocks?: SharedSliceContent[];\n\tdiff?: SliceDiff;\n};\n\nexport const mockSlice = (args: mockSliceArgs): SharedSliceContent[] => {\n\tconst { model, mocks, diff } = args;\n\n\treturn model.variations.map((variation) => {\n\t\tconst variationMock = mocks?.find((m) => m.variation === variation.id);\n\n\t\tconst mockConfig = {\n\t\t\ttype: \"SharedSlice\",\n\t\t\tvariation: variation.id,\n\t\t} as const;\n\n\t\tif (!variationMock) {\n\t\t\treturn SharedSliceMock.generate(model, mockConfig);\n\t\t}\n\t\tif (!diff) {\n\t\t\treturn variationMock;\n\t\t}\n\n\t\t// Enforce at least one item to be created if none are present and variation has item fields\n\t\tif (\n\t\t\tObject.keys(variation.items || {}).length &&\n\t\t\tvariationMock.items.length === 0\n\t\t) {\n\t\t\tvariationMock.items.push({ __TYPE__: \"GroupItemContent\", value: [] });\n\t\t}\n\t\tconst patched = SharedSliceMock.patch(diff, variationMock, mockConfig);\n\t\tif (!patched.ok) {\n\t\t\treturn variationMock;\n\t\t}\n\t\tif (!patched.result) {\n\t\t\treturn SharedSliceMock.generate(model, mockConfig);\n\t\t}\n\n\t\treturn patched.result;\n\t});\n};\n"],"names":[],"mappings":";
|
1
|
+
{"version":3,"file":"mockSlice.js","sources":["../../../src/lib/mockSlice.ts"],"sourcesContent":["import { SharedSliceMock } from \"@prismicio/mocks\";\n\nimport { SharedSliceContent } from \"@prismicio/types-internal/lib/content\";\nimport {\n\tSharedSlice,\n\tSliceDiff,\n} from \"@prismicio/types-internal/lib/customtypes\";\n\ntype mockSliceArgs = {\n\tmodel: SharedSlice;\n\tmocks?: SharedSliceContent[];\n\tdiff?: SliceDiff;\n};\n\nexport const mockSlice = (args: mockSliceArgs): SharedSliceContent[] => {\n\tconst { model, mocks, diff } = args;\n\n\treturn model.variations.map((variation) => {\n\t\tconst variationMock = mocks?.find((m) => m.variation === variation.id);\n\n\t\tconst mockConfig = {\n\t\t\ttype: \"SharedSlice\",\n\t\t\tvariation: variation.id,\n\t\t} as const;\n\n\t\tif (!variationMock) {\n\t\t\treturn SharedSliceMock.generate(model, mockConfig);\n\t\t}\n\t\tif (!diff) {\n\t\t\treturn variationMock;\n\t\t}\n\n\t\t// Enforce at least one item to be created if none are present and variation has item fields\n\t\tif (\n\t\t\tObject.keys(variation.items || {}).length &&\n\t\t\tvariationMock.items.length === 0\n\t\t) {\n\t\t\tvariationMock.items.push({ __TYPE__: \"GroupItemContent\", value: [] });\n\t\t}\n\t\tconst patched = SharedSliceMock.patch(diff, variationMock, mockConfig);\n\t\tif (!patched.ok) {\n\t\t\treturn variationMock;\n\t\t}\n\t\tif (!patched.result) {\n\t\t\treturn SharedSliceMock.generate(model, mockConfig);\n\t\t}\n\n\t\treturn patched.result;\n\t});\n};\n"],"names":[],"mappings":";AAcO,MAAM,YAAY,CAAC,SAA6C;AACtE,QAAM,EAAE,OAAO,OAAO,KAAA,IAAS;AAE/B,SAAO,MAAM,WAAW,IAAI,CAAC,cAAa;AACzC,UAAM,gBAAgB,+BAAO,KAAK,CAAC,MAAM,EAAE,cAAc,UAAU;AAEnE,UAAM,aAAa;AAAA,MAClB,MAAM;AAAA,MACN,WAAW,UAAU;AAAA,IAAA;AAGtB,QAAI,CAAC,eAAe;AACnB,aAAO,gBAAgB,SAAS,OAAO,UAAU;AAAA,IAClD;AACA,QAAI,CAAC,MAAM;AACV,aAAO;AAAA,IACR;AAGA,QACC,OAAO,KAAK,UAAU,SAAS,CAAA,CAAE,EAAE,UACnC,cAAc,MAAM,WAAW,GAC9B;AACD,oBAAc,MAAM,KAAK,EAAE,UAAU,oBAAoB,OAAO,CAAA,GAAI;AAAA,IACrE;AACA,UAAM,UAAU,gBAAgB,MAAM,MAAM,eAAe,UAAU;AACrE,QAAI,CAAC,QAAQ,IAAI;AAChB,aAAO;AAAA,IACR;AACA,QAAI,CAAC,QAAQ,QAAQ;AACpB,aAAO,gBAAgB,SAAS,OAAO,UAAU;AAAA,IAClD;AAEA,WAAO,QAAQ;AAAA,EAChB,CAAC;AACF;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"prismicrc.cjs","sources":["../../../src/lib/prismicrc.ts"],"sourcesContent":["import * as path from \"node:path\";\nimport * as t from \"io-ts\";\nimport * as rc9 from \"rc9\";\n\nimport { decode } from \"./decode\";\n\nconst PRISMICRC = \".prismicrc\";\n\nconst Prismicrc = t.partial({\n\ttelemetry: t.boolean,\n});\ntype Prismicrc = t.TypeOf<typeof Prismicrc>;\n\nexport const readRawPrismicrc = (dir?: string): Prismicrc => {\n\tconst rawPrismicrc = dir\n\t\t? rc9.read({ dir, name: PRISMICRC })\n\t\t: rc9.readUser(PRISMICRC);\n\n\tconst { value: prismicrc, error } = decode(Prismicrc, rawPrismicrc);\n\n\tif (error) {\n\t\tthrow new Error(\n\t\t\t`Failed to parse ${\n\t\t\t\tdir ? path.resolve(dir, PRISMICRC) : `~/${PRISMICRC}`\n\t\t\t}: ${error.errors.join(\", \")}`,\n\t\t);\n\t}\n\n\treturn prismicrc;\n};\n\nexport const readPrismicrc = (dir: string): Prismicrc => {\n\tconst userPrismicrc = readRawPrismicrc();\n\tconst projectPrismicrc = readRawPrismicrc(dir);\n\n\treturn {\n\t\t...userPrismicrc,\n\t\t...projectPrismicrc,\n\t};\n};\n\nexport const writePrismicrc = (config: Prismicrc, dir?: string): void => {\n\tconst { value: validatedConfig, error } = decode(Prismicrc, config);\n\n\tif (error) {\n\t\tthrow new Error(`Failed to validate config: ${error.errors.join(\", \")}`);\n\t}\n\n\tdir\n\t\t? rc9.write(validatedConfig, { dir, name: PRISMICRC })\n\t\t: rc9.writeUser(validatedConfig, PRISMICRC);\n};\n\nexport const updatePrismicrc = (config: Prismicrc, dir?: string): Prismicrc => {\n\tconst { value: validatedConfig, error } = decode(Prismicrc, config);\n\n\tif (error) {\n\t\tthrow new Error(`Failed to validate config: ${error.errors.join(\", \")}`);\n\t}\n\n\treturn dir\n\t\t? rc9.update(validatedConfig, { dir, name: PRISMICRC })\n\t\t: rc9.updateUser(validatedConfig, PRISMICRC);\n};\n"],"names":["t","rc9","decode","path"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;AAMA,MAAM,YAAY;AAElB,MAAM,YAAYA,aAAE,QAAQ;AAAA,EAC3B,WAAWA,aAAE;AACb,CAAA;
|
1
|
+
{"version":3,"file":"prismicrc.cjs","sources":["../../../src/lib/prismicrc.ts"],"sourcesContent":["import * as path from \"node:path\";\nimport * as t from \"io-ts\";\nimport * as rc9 from \"rc9\";\n\nimport { decode } from \"./decode\";\n\nconst PRISMICRC = \".prismicrc\";\n\nconst Prismicrc = t.partial({\n\ttelemetry: t.boolean,\n});\ntype Prismicrc = t.TypeOf<typeof Prismicrc>;\n\nexport const readRawPrismicrc = (dir?: string): Prismicrc => {\n\tconst rawPrismicrc = dir\n\t\t? rc9.read({ dir, name: PRISMICRC })\n\t\t: rc9.readUser(PRISMICRC);\n\n\tconst { value: prismicrc, error } = decode(Prismicrc, rawPrismicrc);\n\n\tif (error) {\n\t\tthrow new Error(\n\t\t\t`Failed to parse ${\n\t\t\t\tdir ? path.resolve(dir, PRISMICRC) : `~/${PRISMICRC}`\n\t\t\t}: ${error.errors.join(\", \")}`,\n\t\t);\n\t}\n\n\treturn prismicrc;\n};\n\nexport const readPrismicrc = (dir: string): Prismicrc => {\n\tconst userPrismicrc = readRawPrismicrc();\n\tconst projectPrismicrc = readRawPrismicrc(dir);\n\n\treturn {\n\t\t...userPrismicrc,\n\t\t...projectPrismicrc,\n\t};\n};\n\nexport const writePrismicrc = (config: Prismicrc, dir?: string): void => {\n\tconst { value: validatedConfig, error } = decode(Prismicrc, config);\n\n\tif (error) {\n\t\tthrow new Error(`Failed to validate config: ${error.errors.join(\", \")}`);\n\t}\n\n\tdir\n\t\t? rc9.write(validatedConfig, { dir, name: PRISMICRC })\n\t\t: rc9.writeUser(validatedConfig, PRISMICRC);\n};\n\nexport const updatePrismicrc = (config: Prismicrc, dir?: string): Prismicrc => {\n\tconst { value: validatedConfig, error } = decode(Prismicrc, config);\n\n\tif (error) {\n\t\tthrow new Error(`Failed to validate config: ${error.errors.join(\", \")}`);\n\t}\n\n\treturn dir\n\t\t? rc9.update(validatedConfig, { dir, name: PRISMICRC })\n\t\t: rc9.updateUser(validatedConfig, PRISMICRC);\n};\n"],"names":["t","rc9","decode","path"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;AAMA,MAAM,YAAY;AAElB,MAAM,YAAYA,aAAE,QAAQ;AAAA,EAC3B,WAAWA,aAAE;AACb,CAAA;AAGM,MAAM,mBAAmB,CAAC,QAA2B;AAC3D,QAAM,eAAe,MAClBC,eAAI,KAAK,EAAE,KAAK,MAAM,UAAA,CAAW,IACjCA,eAAI,SAAS,SAAS;AAEzB,QAAM,EAAE,OAAO,WAAW,UAAUC,OAAAA,OAAO,WAAW,YAAY;AAElE,MAAI,OAAO;AACV,UAAM,IAAI,MACT,mBACC,MAAMC,gBAAK,QAAQ,KAAK,SAAS,IAAI,KAAK,SAAS,EACpD,KAAK,MAAM,OAAO,KAAK,IAAI,CAAC,EAAE;AAAA,EAEhC;AAEA,SAAO;AACR;AAEO,MAAM,gBAAgB,CAAC,QAA0B;AACvD,QAAM,gBAAgB,iBAAA;AACtB,QAAM,mBAAmB,iBAAiB,GAAG;AAE7C,SAAO;AAAA,IACN,GAAG;AAAA,IACH,GAAG;AAAA,EAAA;AAEL;;;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"prismicrc.js","sources":["../../../src/lib/prismicrc.ts"],"sourcesContent":["import * as path from \"node:path\";\nimport * as t from \"io-ts\";\nimport * as rc9 from \"rc9\";\n\nimport { decode } from \"./decode\";\n\nconst PRISMICRC = \".prismicrc\";\n\nconst Prismicrc = t.partial({\n\ttelemetry: t.boolean,\n});\ntype Prismicrc = t.TypeOf<typeof Prismicrc>;\n\nexport const readRawPrismicrc = (dir?: string): Prismicrc => {\n\tconst rawPrismicrc = dir\n\t\t? rc9.read({ dir, name: PRISMICRC })\n\t\t: rc9.readUser(PRISMICRC);\n\n\tconst { value: prismicrc, error } = decode(Prismicrc, rawPrismicrc);\n\n\tif (error) {\n\t\tthrow new Error(\n\t\t\t`Failed to parse ${\n\t\t\t\tdir ? path.resolve(dir, PRISMICRC) : `~/${PRISMICRC}`\n\t\t\t}: ${error.errors.join(\", \")}`,\n\t\t);\n\t}\n\n\treturn prismicrc;\n};\n\nexport const readPrismicrc = (dir: string): Prismicrc => {\n\tconst userPrismicrc = readRawPrismicrc();\n\tconst projectPrismicrc = readRawPrismicrc(dir);\n\n\treturn {\n\t\t...userPrismicrc,\n\t\t...projectPrismicrc,\n\t};\n};\n\nexport const writePrismicrc = (config: Prismicrc, dir?: string): void => {\n\tconst { value: validatedConfig, error } = decode(Prismicrc, config);\n\n\tif (error) {\n\t\tthrow new Error(`Failed to validate config: ${error.errors.join(\", \")}`);\n\t}\n\n\tdir\n\t\t? rc9.write(validatedConfig, { dir, name: PRISMICRC })\n\t\t: rc9.writeUser(validatedConfig, PRISMICRC);\n};\n\nexport const updatePrismicrc = (config: Prismicrc, dir?: string): Prismicrc => {\n\tconst { value: validatedConfig, error } = decode(Prismicrc, config);\n\n\tif (error) {\n\t\tthrow new Error(`Failed to validate config: ${error.errors.join(\", \")}`);\n\t}\n\n\treturn dir\n\t\t? rc9.update(validatedConfig, { dir, name: PRISMICRC })\n\t\t: rc9.updateUser(validatedConfig, PRISMICRC);\n};\n"],"names":[],"mappings":";;;;AAMA,MAAM,YAAY;AAElB,MAAM,YAAY,EAAE,QAAQ;AAAA,EAC3B,WAAW,EAAE;AACb,CAAA;
|
1
|
+
{"version":3,"file":"prismicrc.js","sources":["../../../src/lib/prismicrc.ts"],"sourcesContent":["import * as path from \"node:path\";\nimport * as t from \"io-ts\";\nimport * as rc9 from \"rc9\";\n\nimport { decode } from \"./decode\";\n\nconst PRISMICRC = \".prismicrc\";\n\nconst Prismicrc = t.partial({\n\ttelemetry: t.boolean,\n});\ntype Prismicrc = t.TypeOf<typeof Prismicrc>;\n\nexport const readRawPrismicrc = (dir?: string): Prismicrc => {\n\tconst rawPrismicrc = dir\n\t\t? rc9.read({ dir, name: PRISMICRC })\n\t\t: rc9.readUser(PRISMICRC);\n\n\tconst { value: prismicrc, error } = decode(Prismicrc, rawPrismicrc);\n\n\tif (error) {\n\t\tthrow new Error(\n\t\t\t`Failed to parse ${\n\t\t\t\tdir ? path.resolve(dir, PRISMICRC) : `~/${PRISMICRC}`\n\t\t\t}: ${error.errors.join(\", \")}`,\n\t\t);\n\t}\n\n\treturn prismicrc;\n};\n\nexport const readPrismicrc = (dir: string): Prismicrc => {\n\tconst userPrismicrc = readRawPrismicrc();\n\tconst projectPrismicrc = readRawPrismicrc(dir);\n\n\treturn {\n\t\t...userPrismicrc,\n\t\t...projectPrismicrc,\n\t};\n};\n\nexport const writePrismicrc = (config: Prismicrc, dir?: string): void => {\n\tconst { value: validatedConfig, error } = decode(Prismicrc, config);\n\n\tif (error) {\n\t\tthrow new Error(`Failed to validate config: ${error.errors.join(\", \")}`);\n\t}\n\n\tdir\n\t\t? rc9.write(validatedConfig, { dir, name: PRISMICRC })\n\t\t: rc9.writeUser(validatedConfig, PRISMICRC);\n};\n\nexport const updatePrismicrc = (config: Prismicrc, dir?: string): Prismicrc => {\n\tconst { value: validatedConfig, error } = decode(Prismicrc, config);\n\n\tif (error) {\n\t\tthrow new Error(`Failed to validate config: ${error.errors.join(\", \")}`);\n\t}\n\n\treturn dir\n\t\t? rc9.update(validatedConfig, { dir, name: PRISMICRC })\n\t\t: rc9.updateUser(validatedConfig, PRISMICRC);\n};\n"],"names":[],"mappings":";;;;AAMA,MAAM,YAAY;AAElB,MAAM,YAAY,EAAE,QAAQ;AAAA,EAC3B,WAAW,EAAE;AACb,CAAA;AAGM,MAAM,mBAAmB,CAAC,QAA2B;AAC3D,QAAM,eAAe,MAClB,IAAI,KAAK,EAAE,KAAK,MAAM,UAAA,CAAW,IACjC,IAAI,SAAS,SAAS;AAEzB,QAAM,EAAE,OAAO,WAAW,UAAU,OAAO,WAAW,YAAY;AAElE,MAAI,OAAO;AACV,UAAM,IAAI,MACT,mBACC,MAAM,KAAK,QAAQ,KAAK,SAAS,IAAI,KAAK,SAAS,EACpD,KAAK,MAAM,OAAO,KAAK,IAAI,CAAC,EAAE;AAAA,EAEhC;AAEA,SAAO;AACR;AAEO,MAAM,gBAAgB,CAAC,QAA0B;AACvD,QAAM,gBAAgB,iBAAA;AACtB,QAAM,mBAAmB,iBAAiB,GAAG;AAE7C,SAAO;AAAA,IACN,GAAG;AAAA,IACH,GAAG;AAAA,EAAA;AAEL;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"requireResolve.cjs","sources":["../../../src/lib/requireResolve.ts"],"sourcesContent":["import _module, { createRequire } from \"node:module\";\n\n/**\n * Resolves a module path with `createRequire().resolve()` with Yarn PnP\n * support.\n *\n * @param id - Module to resolve.\n * @param from - Location to resolve the module from.\n *\n * @returns - Resolved module path.\n */\nexport const requireResolve = (id: string, from: string): string => {\n\tlet resolvedID = id;\n\n\t// Support Yarn PnP\n\tif (\n\t\tprocess.versions.pnp &&\n\t\t\"findPnpApi\" in _module &&\n\t\ttypeof _module.findPnpApi === \"function\"\n\t) {\n\t\tconst pnpApi = _module.findPnpApi(from);\n\t\tif (pnpApi) {\n\t\t\tresolvedID = pnpApi.resolveRequest(id, from);\n\t\t}\n\t}\n\n\tconst require = createRequire(from);\n\n\treturn require.resolve(resolvedID);\n};\n"],"names":["require","createRequire"],"mappings":";;;
|
1
|
+
{"version":3,"file":"requireResolve.cjs","sources":["../../../src/lib/requireResolve.ts"],"sourcesContent":["import _module, { createRequire } from \"node:module\";\n\n/**\n * Resolves a module path with `createRequire().resolve()` with Yarn PnP\n * support.\n *\n * @param id - Module to resolve.\n * @param from - Location to resolve the module from.\n *\n * @returns - Resolved module path.\n */\nexport const requireResolve = (id: string, from: string): string => {\n\tlet resolvedID = id;\n\n\t// Support Yarn PnP\n\tif (\n\t\tprocess.versions.pnp &&\n\t\t\"findPnpApi\" in _module &&\n\t\ttypeof _module.findPnpApi === \"function\"\n\t) {\n\t\tconst pnpApi = _module.findPnpApi(from);\n\t\tif (pnpApi) {\n\t\t\tresolvedID = pnpApi.resolveRequest(id, from);\n\t\t}\n\t}\n\n\tconst require = createRequire(from);\n\n\treturn require.resolve(resolvedID);\n};\n"],"names":["require","createRequire"],"mappings":";;;AAWO,MAAM,iBAAiB,CAAC,IAAY,SAAwB;AAClE,MAAI,aAAa;AAGjB,MACC,QAAQ,SAAS,OACjB,gBAAgB,WAChB,OAAO,QAAQ,eAAe,YAC7B;AACD,UAAM,SAAS,QAAQ,WAAW,IAAI;AACtC,QAAI,QAAQ;AACX,mBAAa,OAAO,eAAe,IAAI,IAAI;AAAA,IAC5C;AAAA,EACD;AAEA,QAAMA,WAAUC,QAAAA,cAAc,IAAI;AAElC,SAAOD,SAAQ,QAAQ,UAAU;AAClC;;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"requireResolve.js","sources":["../../../src/lib/requireResolve.ts"],"sourcesContent":["import _module, { createRequire } from \"node:module\";\n\n/**\n * Resolves a module path with `createRequire().resolve()` with Yarn PnP\n * support.\n *\n * @param id - Module to resolve.\n * @param from - Location to resolve the module from.\n *\n * @returns - Resolved module path.\n */\nexport const requireResolve = (id: string, from: string): string => {\n\tlet resolvedID = id;\n\n\t// Support Yarn PnP\n\tif (\n\t\tprocess.versions.pnp &&\n\t\t\"findPnpApi\" in _module &&\n\t\ttypeof _module.findPnpApi === \"function\"\n\t) {\n\t\tconst pnpApi = _module.findPnpApi(from);\n\t\tif (pnpApi) {\n\t\t\tresolvedID = pnpApi.resolveRequest(id, from);\n\t\t}\n\t}\n\n\tconst require = createRequire(from);\n\n\treturn require.resolve(resolvedID);\n};\n"],"names":["require"],"mappings":";
|
1
|
+
{"version":3,"file":"requireResolve.js","sources":["../../../src/lib/requireResolve.ts"],"sourcesContent":["import _module, { createRequire } from \"node:module\";\n\n/**\n * Resolves a module path with `createRequire().resolve()` with Yarn PnP\n * support.\n *\n * @param id - Module to resolve.\n * @param from - Location to resolve the module from.\n *\n * @returns - Resolved module path.\n */\nexport const requireResolve = (id: string, from: string): string => {\n\tlet resolvedID = id;\n\n\t// Support Yarn PnP\n\tif (\n\t\tprocess.versions.pnp &&\n\t\t\"findPnpApi\" in _module &&\n\t\ttypeof _module.findPnpApi === \"function\"\n\t) {\n\t\tconst pnpApi = _module.findPnpApi(from);\n\t\tif (pnpApi) {\n\t\t\tresolvedID = pnpApi.resolveRequest(id, from);\n\t\t}\n\t}\n\n\tconst require = createRequire(from);\n\n\treturn require.resolve(resolvedID);\n};\n"],"names":["require"],"mappings":";AAWO,MAAM,iBAAiB,CAAC,IAAY,SAAwB;AAClE,MAAI,aAAa;AAGjB,MACC,QAAQ,SAAS,OACjB,gBAAgB,WAChB,OAAO,QAAQ,eAAe,YAC7B;AACD,UAAM,SAAS,QAAQ,WAAW,IAAI;AACtC,QAAI,QAAQ;AACX,mBAAa,OAAO,eAAe,IAAI,IAAI;AAAA,IAC5C;AAAA,EACD;AAEA,QAAMA,WAAU,cAAc,IAAI;AAElC,SAAOA,SAAQ,QAAQ,UAAU;AAClC;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"serializeCookies.cjs","sources":["../../../src/lib/serializeCookies.ts"],"sourcesContent":["import cookie from \"cookie\";\n\nconst COOKIE_SEPARATOR = \"; \";\n\ntype Cookies = string | string[] | Record<string, string>;\n\nconst castParsedCookies = (\n\tcookies: Cookies,\n): Record<string, string | undefined> => {\n\tif (Array.isArray(cookies)) {\n\t\treturn cookie.parse(cookies.join(COOKIE_SEPARATOR));\n\t} else if (typeof cookies === \"string\") {\n\t\treturn cookie.parse(cookies);\n\t} else {\n\t\treturn cookies;\n\t}\n};\n\ntype SerializeCookiesArgs = {\n\tcookieJar?: Cookies;\n};\n\n// TODO: If the `cookieJar` and multiple input types are not used anywhere,\n// simplify this function to only serialize a given object of cookies into a\n// string.\nexport const serializeCookies = (\n\tcookies: Cookies,\n\targs: SerializeCookiesArgs = {},\n): string => {\n\tconst cookiesToSerialize = {\n\t\t...castParsedCookies(args.cookieJar || {}),\n\t\t...castParsedCookies(cookies),\n\t};\n\n\tconst items: string[] = [];\n\n\tfor (const name in cookiesToSerialize) {\n\t\tconst cookieValue = cookiesToSerialize[name];\n\t\tif (cookieValue) {\n\t\t\titems.push(\n\t\t\t\tcookie.serialize(name, cookieValue, {\n\t\t\t\t\t// Cookies need be stored raw (not encoded or escaped), so that consumers can format them the way they want them to be formatted.\n\t\t\t\t\tencode: (cookie) => cookie,\n\t\t\t\t}),\n\t\t\t);\n\t\t}\n\t}\n\n\treturn items.join(COOKIE_SEPARATOR);\n};\n"],"names":["cookie"],"mappings":";;;AAEA,MAAM,mBAAmB;AAIzB,MAAM,oBAAoB,CACzB,YACuC;
|
1
|
+
{"version":3,"file":"serializeCookies.cjs","sources":["../../../src/lib/serializeCookies.ts"],"sourcesContent":["import cookie from \"cookie\";\n\nconst COOKIE_SEPARATOR = \"; \";\n\ntype Cookies = string | string[] | Record<string, string>;\n\nconst castParsedCookies = (\n\tcookies: Cookies,\n): Record<string, string | undefined> => {\n\tif (Array.isArray(cookies)) {\n\t\treturn cookie.parse(cookies.join(COOKIE_SEPARATOR));\n\t} else if (typeof cookies === \"string\") {\n\t\treturn cookie.parse(cookies);\n\t} else {\n\t\treturn cookies;\n\t}\n};\n\ntype SerializeCookiesArgs = {\n\tcookieJar?: Cookies;\n};\n\n// TODO: If the `cookieJar` and multiple input types are not used anywhere,\n// simplify this function to only serialize a given object of cookies into a\n// string.\nexport const serializeCookies = (\n\tcookies: Cookies,\n\targs: SerializeCookiesArgs = {},\n): string => {\n\tconst cookiesToSerialize = {\n\t\t...castParsedCookies(args.cookieJar || {}),\n\t\t...castParsedCookies(cookies),\n\t};\n\n\tconst items: string[] = [];\n\n\tfor (const name in cookiesToSerialize) {\n\t\tconst cookieValue = cookiesToSerialize[name];\n\t\tif (cookieValue) {\n\t\t\titems.push(\n\t\t\t\tcookie.serialize(name, cookieValue, {\n\t\t\t\t\t// Cookies need be stored raw (not encoded or escaped), so that consumers can format them the way they want them to be formatted.\n\t\t\t\t\tencode: (cookie) => cookie,\n\t\t\t\t}),\n\t\t\t);\n\t\t}\n\t}\n\n\treturn items.join(COOKIE_SEPARATOR);\n};\n"],"names":["cookie"],"mappings":";;;AAEA,MAAM,mBAAmB;AAIzB,MAAM,oBAAoB,CACzB,YACuC;AACvC,MAAI,MAAM,QAAQ,OAAO,GAAG;AAC3B,WAAO,OAAO,MAAM,QAAQ,KAAK,gBAAgB,CAAC;AAAA,EACnD,WAAW,OAAO,YAAY,UAAU;AACvC,WAAO,OAAO,MAAM,OAAO;AAAA,EAC5B,OAAO;AACN,WAAO;AAAA,EACR;AACD;AASO,MAAM,mBAAmB,CAC/B,SACA,OAA6B,OAClB;AACX,QAAM,qBAAqB;AAAA,IAC1B,GAAG,kBAAkB,KAAK,aAAa,EAAE;AAAA,IACzC,GAAG,kBAAkB,OAAO;AAAA,EAAA;AAG7B,QAAM,QAAkB,CAAA;AAExB,aAAW,QAAQ,oBAAoB;AACtC,UAAM,cAAc,mBAAmB,IAAI;AAC3C,QAAI,aAAa;AAChB,YAAM,KACL,OAAO,UAAU,MAAM,aAAa;AAAA;AAAA,QAEnC,QAAQ,CAACA,YAAWA;AAAAA,MAAA,CACpB,CAAC;AAAA,IAEJ;AAAA,EACD;AAEA,SAAO,MAAM,KAAK,gBAAgB;AACnC;;"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"file":"serializeCookies.js","sources":["../../../src/lib/serializeCookies.ts"],"sourcesContent":["import cookie from \"cookie\";\n\nconst COOKIE_SEPARATOR = \"; \";\n\ntype Cookies = string | string[] | Record<string, string>;\n\nconst castParsedCookies = (\n\tcookies: Cookies,\n): Record<string, string | undefined> => {\n\tif (Array.isArray(cookies)) {\n\t\treturn cookie.parse(cookies.join(COOKIE_SEPARATOR));\n\t} else if (typeof cookies === \"string\") {\n\t\treturn cookie.parse(cookies);\n\t} else {\n\t\treturn cookies;\n\t}\n};\n\ntype SerializeCookiesArgs = {\n\tcookieJar?: Cookies;\n};\n\n// TODO: If the `cookieJar` and multiple input types are not used anywhere,\n// simplify this function to only serialize a given object of cookies into a\n// string.\nexport const serializeCookies = (\n\tcookies: Cookies,\n\targs: SerializeCookiesArgs = {},\n): string => {\n\tconst cookiesToSerialize = {\n\t\t...castParsedCookies(args.cookieJar || {}),\n\t\t...castParsedCookies(cookies),\n\t};\n\n\tconst items: string[] = [];\n\n\tfor (const name in cookiesToSerialize) {\n\t\tconst cookieValue = cookiesToSerialize[name];\n\t\tif (cookieValue) {\n\t\t\titems.push(\n\t\t\t\tcookie.serialize(name, cookieValue, {\n\t\t\t\t\t// Cookies need be stored raw (not encoded or escaped), so that consumers can format them the way they want them to be formatted.\n\t\t\t\t\tencode: (cookie) => cookie,\n\t\t\t\t}),\n\t\t\t);\n\t\t}\n\t}\n\n\treturn items.join(COOKIE_SEPARATOR);\n};\n"],"names":["cookie"],"mappings":";AAEA,MAAM,mBAAmB;AAIzB,MAAM,oBAAoB,CACzB,YACuC;
|
1
|
+
{"version":3,"file":"serializeCookies.js","sources":["../../../src/lib/serializeCookies.ts"],"sourcesContent":["import cookie from \"cookie\";\n\nconst COOKIE_SEPARATOR = \"; \";\n\ntype Cookies = string | string[] | Record<string, string>;\n\nconst castParsedCookies = (\n\tcookies: Cookies,\n): Record<string, string | undefined> => {\n\tif (Array.isArray(cookies)) {\n\t\treturn cookie.parse(cookies.join(COOKIE_SEPARATOR));\n\t} else if (typeof cookies === \"string\") {\n\t\treturn cookie.parse(cookies);\n\t} else {\n\t\treturn cookies;\n\t}\n};\n\ntype SerializeCookiesArgs = {\n\tcookieJar?: Cookies;\n};\n\n// TODO: If the `cookieJar` and multiple input types are not used anywhere,\n// simplify this function to only serialize a given object of cookies into a\n// string.\nexport const serializeCookies = (\n\tcookies: Cookies,\n\targs: SerializeCookiesArgs = {},\n): string => {\n\tconst cookiesToSerialize = {\n\t\t...castParsedCookies(args.cookieJar || {}),\n\t\t...castParsedCookies(cookies),\n\t};\n\n\tconst items: string[] = [];\n\n\tfor (const name in cookiesToSerialize) {\n\t\tconst cookieValue = cookiesToSerialize[name];\n\t\tif (cookieValue) {\n\t\t\titems.push(\n\t\t\t\tcookie.serialize(name, cookieValue, {\n\t\t\t\t\t// Cookies need be stored raw (not encoded or escaped), so that consumers can format them the way they want them to be formatted.\n\t\t\t\t\tencode: (cookie) => cookie,\n\t\t\t\t}),\n\t\t\t);\n\t\t}\n\t}\n\n\treturn items.join(COOKIE_SEPARATOR);\n};\n"],"names":["cookie"],"mappings":";AAEA,MAAM,mBAAmB;AAIzB,MAAM,oBAAoB,CACzB,YACuC;AACvC,MAAI,MAAM,QAAQ,OAAO,GAAG;AAC3B,WAAO,OAAO,MAAM,QAAQ,KAAK,gBAAgB,CAAC;AAAA,EACnD,WAAW,OAAO,YAAY,UAAU;AACvC,WAAO,OAAO,MAAM,OAAO;AAAA,EAC5B,OAAO;AACN,WAAO;AAAA,EACR;AACD;AASO,MAAM,mBAAmB,CAC/B,SACA,OAA6B,OAClB;AACX,QAAM,qBAAqB;AAAA,IAC1B,GAAG,kBAAkB,KAAK,aAAa,EAAE;AAAA,IACzC,GAAG,kBAAkB,OAAO;AAAA,EAAA;AAG7B,QAAM,QAAkB,CAAA;AAExB,aAAW,QAAQ,oBAAoB;AACtC,UAAM,cAAc,mBAAmB,IAAI;AAC3C,QAAI,aAAa;AAChB,YAAM,KACL,OAAO,UAAU,MAAM,aAAa;AAAA;AAAA,QAEnC,QAAQ,CAACA,YAAWA;AAAAA,MAAA,CACpB,CAAC;AAAA,IAEJ;AAAA,EACD;AAEA,SAAO,MAAM,KAAK,gBAAgB;AACnC;"}
|