@huggingface/tasks 0.13.0-test2 → 0.13.0-test3
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/dataset-libraries.d.cts +87 -0
- package/dist/dataset-libraries.d.cts.map +1 -0
- package/dist/dataset-libraries.d.ts +87 -0
- package/dist/dataset-libraries.d.ts.map +1 -0
- package/dist/default-widget-inputs.d.cts +6 -0
- package/dist/default-widget-inputs.d.cts.map +1 -0
- package/dist/default-widget-inputs.d.ts +6 -0
- package/dist/default-widget-inputs.d.ts.map +1 -0
- package/dist/gguf.d.cts +35 -0
- package/dist/gguf.d.cts.map +1 -0
- package/dist/gguf.d.ts +35 -0
- package/dist/gguf.d.ts.map +1 -0
- package/dist/hardware.d.cts +478 -0
- package/dist/hardware.d.cts.map +1 -0
- package/dist/hardware.d.ts +478 -0
- package/dist/hardware.d.ts.map +1 -0
- package/dist/index.d.cts +21 -4778
- package/dist/index.d.cts.map +1 -0
- package/dist/index.d.ts +21 -4778
- package/dist/index.d.ts.map +1 -0
- package/dist/library-to-tasks.d.cts +11 -0
- package/dist/library-to-tasks.d.cts.map +1 -0
- package/dist/library-to-tasks.d.ts +11 -0
- package/dist/library-to-tasks.d.ts.map +1 -0
- package/dist/local-apps.d.cts +195 -0
- package/dist/local-apps.d.cts.map +1 -0
- package/dist/local-apps.d.ts +195 -0
- package/dist/local-apps.d.ts.map +1 -0
- package/dist/model-data.d.cts +146 -0
- package/dist/model-data.d.cts.map +1 -0
- package/dist/model-data.d.ts +146 -0
- package/dist/model-data.d.ts.map +1 -0
- package/dist/model-libraries-downloads.d.cts +18 -0
- package/dist/model-libraries-downloads.d.cts.map +1 -0
- package/dist/model-libraries-downloads.d.ts +18 -0
- package/dist/model-libraries-downloads.d.ts.map +1 -0
- package/dist/model-libraries-snippets.d.cts +72 -0
- package/dist/model-libraries-snippets.d.cts.map +1 -0
- package/dist/model-libraries-snippets.d.ts +72 -0
- package/dist/model-libraries-snippets.d.ts.map +1 -0
- package/dist/model-libraries.d.cts +804 -0
- package/dist/model-libraries.d.cts.map +1 -0
- package/dist/model-libraries.d.ts +804 -0
- package/dist/model-libraries.d.ts.map +1 -0
- package/dist/pipelines.d.cts +425 -0
- package/dist/pipelines.d.cts.map +1 -0
- package/dist/pipelines.d.ts +425 -0
- package/dist/pipelines.d.ts.map +1 -0
- package/dist/snippets/common.d.cts +14 -0
- package/dist/snippets/common.d.cts.map +1 -0
- package/dist/snippets/common.d.ts +14 -0
- package/dist/snippets/common.d.ts.map +1 -0
- package/dist/snippets/curl.d.cts +17 -0
- package/dist/snippets/curl.d.cts.map +1 -0
- package/dist/snippets/curl.d.ts +17 -0
- package/dist/snippets/curl.d.ts.map +1 -0
- package/dist/snippets/curl.spec.d.cts +2 -0
- package/dist/snippets/curl.spec.d.cts.map +1 -0
- package/dist/snippets/curl.spec.d.ts +2 -0
- package/dist/snippets/curl.spec.d.ts.map +1 -0
- package/dist/snippets/index.d.cts +6 -0
- package/dist/snippets/index.d.cts.map +1 -0
- package/dist/snippets/index.d.ts +6 -0
- package/dist/snippets/index.d.ts.map +1 -0
- package/dist/snippets/inputs.d.cts +4 -0
- package/dist/snippets/inputs.d.cts.map +1 -0
- package/dist/snippets/inputs.d.ts +4 -0
- package/dist/snippets/inputs.d.ts.map +1 -0
- package/dist/snippets/js.d.cts +19 -0
- package/dist/snippets/js.d.cts.map +1 -0
- package/dist/snippets/js.d.ts +19 -0
- package/dist/snippets/js.d.ts.map +1 -0
- package/dist/snippets/js.spec.d.cts +2 -0
- package/dist/snippets/js.spec.d.cts.map +1 -0
- package/dist/snippets/js.spec.d.ts +2 -0
- package/dist/snippets/js.spec.d.ts.map +1 -0
- package/dist/snippets/python.d.cts +22 -0
- package/dist/snippets/python.d.cts.map +1 -0
- package/dist/snippets/python.d.ts +22 -0
- package/dist/snippets/python.d.ts.map +1 -0
- package/dist/snippets/python.spec.d.cts +2 -0
- package/dist/snippets/python.spec.d.cts.map +1 -0
- package/dist/snippets/python.spec.d.ts +2 -0
- package/dist/snippets/python.spec.d.ts.map +1 -0
- package/dist/snippets/types.d.cts +12 -0
- package/dist/snippets/types.d.cts.map +1 -0
- package/dist/snippets/types.d.ts +12 -0
- package/dist/snippets/types.d.ts.map +1 -0
- package/dist/tasks/audio-classification/data.d.cts +4 -0
- package/dist/tasks/audio-classification/data.d.cts.map +1 -0
- package/dist/tasks/audio-classification/data.d.ts +4 -0
- package/dist/tasks/audio-classification/data.d.ts.map +1 -0
- package/dist/tasks/audio-classification/inference.d.cts +53 -0
- package/dist/tasks/audio-classification/inference.d.cts.map +1 -0
- package/dist/tasks/audio-classification/inference.d.ts +53 -0
- package/dist/tasks/audio-classification/inference.d.ts.map +1 -0
- package/dist/tasks/audio-to-audio/data.d.cts +4 -0
- package/dist/tasks/audio-to-audio/data.d.cts.map +1 -0
- package/dist/tasks/audio-to-audio/data.d.ts +4 -0
- package/dist/tasks/audio-to-audio/data.d.ts.map +1 -0
- package/dist/tasks/automatic-speech-recognition/data.d.cts +4 -0
- package/dist/tasks/automatic-speech-recognition/data.d.cts.map +1 -0
- package/dist/tasks/automatic-speech-recognition/data.d.ts +4 -0
- package/dist/tasks/automatic-speech-recognition/data.d.ts.map +1 -0
- package/dist/tasks/automatic-speech-recognition/inference.d.cts +155 -0
- package/dist/tasks/automatic-speech-recognition/inference.d.cts.map +1 -0
- package/dist/tasks/automatic-speech-recognition/inference.d.ts +155 -0
- package/dist/tasks/automatic-speech-recognition/inference.d.ts.map +1 -0
- package/dist/tasks/chat-completion/inference.d.cts +291 -0
- package/dist/tasks/chat-completion/inference.d.cts.map +1 -0
- package/dist/tasks/chat-completion/inference.d.ts +291 -0
- package/dist/tasks/chat-completion/inference.d.ts.map +1 -0
- package/dist/tasks/depth-estimation/data.d.cts +4 -0
- package/dist/tasks/depth-estimation/data.d.cts.map +1 -0
- package/dist/tasks/depth-estimation/data.d.ts +4 -0
- package/dist/tasks/depth-estimation/data.d.ts.map +1 -0
- package/dist/tasks/depth-estimation/inference.d.cts +36 -0
- package/dist/tasks/depth-estimation/inference.d.cts.map +1 -0
- package/dist/tasks/depth-estimation/inference.d.ts +36 -0
- package/dist/tasks/depth-estimation/inference.d.ts.map +1 -0
- package/dist/tasks/document-question-answering/data.d.cts +4 -0
- package/dist/tasks/document-question-answering/data.d.cts.map +1 -0
- package/dist/tasks/document-question-answering/data.d.ts +4 -0
- package/dist/tasks/document-question-answering/data.d.ts.map +1 -0
- package/dist/tasks/document-question-answering/inference.d.cts +111 -0
- package/dist/tasks/document-question-answering/inference.d.cts.map +1 -0
- package/dist/tasks/document-question-answering/inference.d.ts +111 -0
- package/dist/tasks/document-question-answering/inference.d.ts.map +1 -0
- package/dist/tasks/feature-extraction/data.d.cts +4 -0
- package/dist/tasks/feature-extraction/data.d.cts.map +1 -0
- package/dist/tasks/feature-extraction/data.d.ts +4 -0
- package/dist/tasks/feature-extraction/data.d.ts.map +1 -0
- package/dist/tasks/feature-extraction/inference.d.cts +38 -0
- package/dist/tasks/feature-extraction/inference.d.cts.map +1 -0
- package/dist/tasks/feature-extraction/inference.d.ts +38 -0
- package/dist/tasks/feature-extraction/inference.d.ts.map +1 -0
- package/dist/tasks/fill-mask/data.d.cts +4 -0
- package/dist/tasks/fill-mask/data.d.cts.map +1 -0
- package/dist/tasks/fill-mask/data.d.ts +4 -0
- package/dist/tasks/fill-mask/data.d.ts.map +1 -0
- package/dist/tasks/fill-mask/inference.d.cts +63 -0
- package/dist/tasks/fill-mask/inference.d.cts.map +1 -0
- package/dist/tasks/fill-mask/inference.d.ts +63 -0
- package/dist/tasks/fill-mask/inference.d.ts.map +1 -0
- package/dist/tasks/image-classification/data.d.cts +4 -0
- package/dist/tasks/image-classification/data.d.cts.map +1 -0
- package/dist/tasks/image-classification/data.d.ts +4 -0
- package/dist/tasks/image-classification/data.d.ts.map +1 -0
- package/dist/tasks/image-classification/inference.d.cts +53 -0
- package/dist/tasks/image-classification/inference.d.cts.map +1 -0
- package/dist/tasks/image-classification/inference.d.ts +53 -0
- package/dist/tasks/image-classification/inference.d.ts.map +1 -0
- package/dist/tasks/image-feature-extraction/data.d.cts +4 -0
- package/dist/tasks/image-feature-extraction/data.d.cts.map +1 -0
- package/dist/tasks/image-feature-extraction/data.d.ts +4 -0
- package/dist/tasks/image-feature-extraction/data.d.ts.map +1 -0
- package/dist/tasks/image-segmentation/data.d.cts +4 -0
- package/dist/tasks/image-segmentation/data.d.cts.map +1 -0
- package/dist/tasks/image-segmentation/data.d.ts +4 -0
- package/dist/tasks/image-segmentation/data.d.ts.map +1 -0
- package/dist/tasks/image-segmentation/inference.d.cts +70 -0
- package/dist/tasks/image-segmentation/inference.d.cts.map +1 -0
- package/dist/tasks/image-segmentation/inference.d.ts +70 -0
- package/dist/tasks/image-segmentation/inference.d.ts.map +1 -0
- package/dist/tasks/image-text-to-text/data.d.cts +4 -0
- package/dist/tasks/image-text-to-text/data.d.cts.map +1 -0
- package/dist/tasks/image-text-to-text/data.d.ts +4 -0
- package/dist/tasks/image-text-to-text/data.d.ts.map +1 -0
- package/dist/tasks/image-to-3d/data.d.cts +4 -0
- package/dist/tasks/image-to-3d/data.d.cts.map +1 -0
- package/dist/tasks/image-to-3d/data.d.ts +4 -0
- package/dist/tasks/image-to-3d/data.d.ts.map +1 -0
- package/dist/tasks/image-to-image/data.d.cts +4 -0
- package/dist/tasks/image-to-image/data.d.cts.map +1 -0
- package/dist/tasks/image-to-image/data.d.ts +4 -0
- package/dist/tasks/image-to-image/data.d.ts.map +1 -0
- package/dist/tasks/image-to-image/inference.d.cts +65 -0
- package/dist/tasks/image-to-image/inference.d.cts.map +1 -0
- package/dist/tasks/image-to-image/inference.d.ts +65 -0
- package/dist/tasks/image-to-image/inference.d.ts.map +1 -0
- package/dist/tasks/image-to-text/data.d.cts +4 -0
- package/dist/tasks/image-to-text/data.d.cts.map +1 -0
- package/dist/tasks/image-to-text/data.d.ts +4 -0
- package/dist/tasks/image-to-text/data.d.ts.map +1 -0
- package/dist/tasks/image-to-text/inference.d.cts +139 -0
- package/dist/tasks/image-to-text/inference.d.cts.map +1 -0
- package/dist/tasks/image-to-text/inference.d.ts +139 -0
- package/dist/tasks/image-to-text/inference.d.ts.map +1 -0
- package/dist/tasks/index.d.cts +87 -0
- package/dist/tasks/index.d.cts.map +1 -0
- package/dist/tasks/index.d.ts +87 -0
- package/dist/tasks/index.d.ts.map +1 -0
- package/dist/tasks/keypoint-detection/data.d.cts +4 -0
- package/dist/tasks/keypoint-detection/data.d.cts.map +1 -0
- package/dist/tasks/keypoint-detection/data.d.ts +4 -0
- package/dist/tasks/keypoint-detection/data.d.ts.map +1 -0
- package/dist/tasks/mask-generation/data.d.cts +4 -0
- package/dist/tasks/mask-generation/data.d.cts.map +1 -0
- package/dist/tasks/mask-generation/data.d.ts +4 -0
- package/dist/tasks/mask-generation/data.d.ts.map +1 -0
- package/dist/tasks/object-detection/data.d.cts +4 -0
- package/dist/tasks/object-detection/data.d.cts.map +1 -0
- package/dist/tasks/object-detection/data.d.ts +4 -0
- package/dist/tasks/object-detection/data.d.ts.map +1 -0
- package/dist/tasks/object-detection/inference.d.cts +76 -0
- package/dist/tasks/object-detection/inference.d.cts.map +1 -0
- package/dist/tasks/object-detection/inference.d.ts +76 -0
- package/dist/tasks/object-detection/inference.d.ts.map +1 -0
- package/dist/tasks/placeholder/data.d.cts +4 -0
- package/dist/tasks/placeholder/data.d.cts.map +1 -0
- package/dist/tasks/placeholder/data.d.ts +4 -0
- package/dist/tasks/placeholder/data.d.ts.map +1 -0
- package/dist/tasks/question-answering/data.d.cts +4 -0
- package/dist/tasks/question-answering/data.d.cts.map +1 -0
- package/dist/tasks/question-answering/data.d.ts +4 -0
- package/dist/tasks/question-answering/data.d.ts.map +1 -0
- package/dist/tasks/question-answering/inference.d.cts +100 -0
- package/dist/tasks/question-answering/inference.d.cts.map +1 -0
- package/dist/tasks/question-answering/inference.d.ts +100 -0
- package/dist/tasks/question-answering/inference.d.ts.map +1 -0
- package/dist/tasks/reinforcement-learning/data.d.cts +4 -0
- package/dist/tasks/reinforcement-learning/data.d.cts.map +1 -0
- package/dist/tasks/reinforcement-learning/data.d.ts +4 -0
- package/dist/tasks/reinforcement-learning/data.d.ts.map +1 -0
- package/dist/tasks/sentence-similarity/data.d.cts +4 -0
- package/dist/tasks/sentence-similarity/data.d.cts.map +1 -0
- package/dist/tasks/sentence-similarity/data.d.ts +4 -0
- package/dist/tasks/sentence-similarity/data.d.ts.map +1 -0
- package/dist/tasks/sentence-similarity/inference.d.cts +32 -0
- package/dist/tasks/sentence-similarity/inference.d.cts.map +1 -0
- package/dist/tasks/sentence-similarity/inference.d.ts +32 -0
- package/dist/tasks/sentence-similarity/inference.d.ts.map +1 -0
- package/dist/tasks/summarization/data.d.cts +4 -0
- package/dist/tasks/summarization/data.d.cts.map +1 -0
- package/dist/tasks/summarization/data.d.ts +4 -0
- package/dist/tasks/summarization/data.d.ts.map +1 -0
- package/dist/tasks/summarization/inference.d.cts +56 -0
- package/dist/tasks/summarization/inference.d.cts.map +1 -0
- package/dist/tasks/summarization/inference.d.ts +56 -0
- package/dist/tasks/summarization/inference.d.ts.map +1 -0
- package/dist/tasks/table-question-answering/data.d.cts +4 -0
- package/dist/tasks/table-question-answering/data.d.cts.map +1 -0
- package/dist/tasks/table-question-answering/data.d.ts +4 -0
- package/dist/tasks/table-question-answering/data.d.ts.map +1 -0
- package/dist/tasks/table-question-answering/inference.d.cts +62 -0
- package/dist/tasks/table-question-answering/inference.d.cts.map +1 -0
- package/dist/tasks/table-question-answering/inference.d.ts +62 -0
- package/dist/tasks/table-question-answering/inference.d.ts.map +1 -0
- package/dist/tasks/tabular-classification/data.d.cts +4 -0
- package/dist/tasks/tabular-classification/data.d.cts.map +1 -0
- package/dist/tasks/tabular-classification/data.d.ts +4 -0
- package/dist/tasks/tabular-classification/data.d.ts.map +1 -0
- package/dist/tasks/tabular-regression/data.d.cts +4 -0
- package/dist/tasks/tabular-regression/data.d.cts.map +1 -0
- package/dist/tasks/tabular-regression/data.d.ts +4 -0
- package/dist/tasks/tabular-regression/data.d.ts.map +1 -0
- package/dist/tasks/text-classification/data.d.cts +4 -0
- package/dist/tasks/text-classification/data.d.cts.map +1 -0
- package/dist/tasks/text-classification/data.d.ts +4 -0
- package/dist/tasks/text-classification/data.d.ts.map +1 -0
- package/dist/tasks/text-classification/inference.d.cts +52 -0
- package/dist/tasks/text-classification/inference.d.cts.map +1 -0
- package/dist/tasks/text-classification/inference.d.ts +52 -0
- package/dist/tasks/text-classification/inference.d.ts.map +1 -0
- package/dist/tasks/text-generation/data.d.cts +4 -0
- package/dist/tasks/text-generation/data.d.cts.map +1 -0
- package/dist/tasks/text-generation/data.d.ts +4 -0
- package/dist/tasks/text-generation/data.d.ts.map +1 -0
- package/dist/tasks/text-generation/inference.d.cts +188 -0
- package/dist/tasks/text-generation/inference.d.cts.map +1 -0
- package/dist/tasks/text-generation/inference.d.ts +188 -0
- package/dist/tasks/text-generation/inference.d.ts.map +1 -0
- package/dist/tasks/text-to-3d/data.d.cts +4 -0
- package/dist/tasks/text-to-3d/data.d.cts.map +1 -0
- package/dist/tasks/text-to-3d/data.d.ts +4 -0
- package/dist/tasks/text-to-3d/data.d.ts.map +1 -0
- package/dist/tasks/text-to-audio/inference.d.cts +139 -0
- package/dist/tasks/text-to-audio/inference.d.cts.map +1 -0
- package/dist/tasks/text-to-audio/inference.d.ts +139 -0
- package/dist/tasks/text-to-audio/inference.d.ts.map +1 -0
- package/dist/tasks/text-to-image/data.d.cts +4 -0
- package/dist/tasks/text-to-image/data.d.cts.map +1 -0
- package/dist/tasks/text-to-image/data.d.ts +4 -0
- package/dist/tasks/text-to-image/data.d.ts.map +1 -0
- package/dist/tasks/text-to-image/inference.d.cts +72 -0
- package/dist/tasks/text-to-image/inference.d.cts.map +1 -0
- package/dist/tasks/text-to-image/inference.d.ts +72 -0
- package/dist/tasks/text-to-image/inference.d.ts.map +1 -0
- package/dist/tasks/text-to-speech/data.d.cts +4 -0
- package/dist/tasks/text-to-speech/data.d.cts.map +1 -0
- package/dist/tasks/text-to-speech/data.d.ts +4 -0
- package/dist/tasks/text-to-speech/data.d.ts.map +1 -0
- package/dist/tasks/text-to-speech/inference.d.cts +141 -0
- package/dist/tasks/text-to-speech/inference.d.cts.map +1 -0
- package/dist/tasks/text-to-speech/inference.d.ts +141 -0
- package/dist/tasks/text-to-speech/inference.d.ts.map +1 -0
- package/dist/tasks/text-to-video/data.d.cts +4 -0
- package/dist/tasks/text-to-video/data.d.cts.map +1 -0
- package/dist/tasks/text-to-video/data.d.ts +4 -0
- package/dist/tasks/text-to-video/data.d.ts.map +1 -0
- package/dist/tasks/text2text-generation/inference.d.cts +54 -0
- package/dist/tasks/text2text-generation/inference.d.cts.map +1 -0
- package/dist/tasks/text2text-generation/inference.d.ts +54 -0
- package/dist/tasks/text2text-generation/inference.d.ts.map +1 -0
- package/dist/tasks/token-classification/data.d.cts +4 -0
- package/dist/tasks/token-classification/data.d.cts.map +1 -0
- package/dist/tasks/token-classification/data.d.ts +4 -0
- package/dist/tasks/token-classification/data.d.ts.map +1 -0
- package/dist/tasks/token-classification/inference.d.cts +86 -0
- package/dist/tasks/token-classification/inference.d.cts.map +1 -0
- package/dist/tasks/token-classification/inference.d.ts +86 -0
- package/dist/tasks/token-classification/inference.d.ts.map +1 -0
- package/dist/tasks/translation/data.d.cts +4 -0
- package/dist/tasks/translation/data.d.cts.map +1 -0
- package/dist/tasks/translation/data.d.ts +4 -0
- package/dist/tasks/translation/data.d.ts.map +1 -0
- package/dist/tasks/translation/inference.d.cts +66 -0
- package/dist/tasks/translation/inference.d.cts.map +1 -0
- package/dist/tasks/translation/inference.d.ts +66 -0
- package/dist/tasks/translation/inference.d.ts.map +1 -0
- package/dist/tasks/unconditional-image-generation/data.d.cts +4 -0
- package/dist/tasks/unconditional-image-generation/data.d.cts.map +1 -0
- package/dist/tasks/unconditional-image-generation/data.d.ts +4 -0
- package/dist/tasks/unconditional-image-generation/data.d.ts.map +1 -0
- package/dist/tasks/video-classification/data.d.cts +4 -0
- package/dist/tasks/video-classification/data.d.cts.map +1 -0
- package/dist/tasks/video-classification/data.d.ts +4 -0
- package/dist/tasks/video-classification/data.d.ts.map +1 -0
- package/dist/tasks/video-classification/inference.d.cts +60 -0
- package/dist/tasks/video-classification/inference.d.cts.map +1 -0
- package/dist/tasks/video-classification/inference.d.ts +60 -0
- package/dist/tasks/video-classification/inference.d.ts.map +1 -0
- package/dist/tasks/video-text-to-text/data.d.cts +4 -0
- package/dist/tasks/video-text-to-text/data.d.cts.map +1 -0
- package/dist/tasks/video-text-to-text/data.d.ts +4 -0
- package/dist/tasks/video-text-to-text/data.d.ts.map +1 -0
- package/dist/tasks/visual-question-answering/data.d.cts +4 -0
- package/dist/tasks/visual-question-answering/data.d.cts.map +1 -0
- package/dist/tasks/visual-question-answering/data.d.ts +4 -0
- package/dist/tasks/visual-question-answering/data.d.ts.map +1 -0
- package/dist/tasks/visual-question-answering/inference.d.cts +63 -0
- package/dist/tasks/visual-question-answering/inference.d.cts.map +1 -0
- package/dist/tasks/visual-question-answering/inference.d.ts +63 -0
- package/dist/tasks/visual-question-answering/inference.d.ts.map +1 -0
- package/dist/tasks/zero-shot-classification/data.d.cts +4 -0
- package/dist/tasks/zero-shot-classification/data.d.cts.map +1 -0
- package/dist/tasks/zero-shot-classification/data.d.ts +4 -0
- package/dist/tasks/zero-shot-classification/data.d.ts.map +1 -0
- package/dist/tasks/zero-shot-classification/inference.d.cts +68 -0
- package/dist/tasks/zero-shot-classification/inference.d.cts.map +1 -0
- package/dist/tasks/zero-shot-classification/inference.d.ts +68 -0
- package/dist/tasks/zero-shot-classification/inference.d.ts.map +1 -0
- package/dist/tasks/zero-shot-image-classification/data.d.cts +4 -0
- package/dist/tasks/zero-shot-image-classification/data.d.cts.map +1 -0
- package/dist/tasks/zero-shot-image-classification/data.d.ts +4 -0
- package/dist/tasks/zero-shot-image-classification/data.d.ts.map +1 -0
- package/dist/tasks/zero-shot-image-classification/inference.d.cts +62 -0
- package/dist/tasks/zero-shot-image-classification/inference.d.cts.map +1 -0
- package/dist/tasks/zero-shot-image-classification/inference.d.ts +62 -0
- package/dist/tasks/zero-shot-image-classification/inference.d.ts.map +1 -0
- package/dist/tasks/zero-shot-object-detection/data.d.cts +4 -0
- package/dist/tasks/zero-shot-object-detection/data.d.cts.map +1 -0
- package/dist/tasks/zero-shot-object-detection/data.d.ts +4 -0
- package/dist/tasks/zero-shot-object-detection/data.d.ts.map +1 -0
- package/dist/tasks/zero-shot-object-detection/inference.d.cts +67 -0
- package/dist/tasks/zero-shot-object-detection/inference.d.cts.map +1 -0
- package/dist/tasks/zero-shot-object-detection/inference.d.ts +67 -0
- package/dist/tasks/zero-shot-object-detection/inference.d.ts.map +1 -0
- package/dist/tokenizer-data.d.cts +26 -0
- package/dist/tokenizer-data.d.cts.map +1 -0
- package/dist/tokenizer-data.d.ts +26 -0
- package/dist/tokenizer-data.d.ts.map +1 -0
- package/dist/widget-example.d.cts +83 -0
- package/dist/widget-example.d.cts.map +1 -0
- package/dist/widget-example.d.ts +83 -0
- package/dist/widget-example.d.ts.map +1 -0
- package/package.json +5 -2
- package/tsconfig.json +1 -1
|
@@ -0,0 +1,22 @@
|
|
|
1
|
+
import type { PipelineType } from "../pipelines.js";
|
|
2
|
+
import type { ChatCompletionInputMessage, GenerationParameters } from "../tasks/index.js";
|
|
3
|
+
import type { InferenceSnippet, ModelDataMinimal } from "./types.js";
|
|
4
|
+
export declare const snippetConversational: (model: ModelDataMinimal, accessToken: string, opts?: {
|
|
5
|
+
streaming?: boolean;
|
|
6
|
+
messages?: ChatCompletionInputMessage[];
|
|
7
|
+
temperature?: GenerationParameters["temperature"];
|
|
8
|
+
max_tokens?: GenerationParameters["max_tokens"];
|
|
9
|
+
top_p?: GenerationParameters["top_p"];
|
|
10
|
+
}) => InferenceSnippet[];
|
|
11
|
+
export declare const snippetZeroShotClassification: (model: ModelDataMinimal) => InferenceSnippet;
|
|
12
|
+
export declare const snippetZeroShotImageClassification: (model: ModelDataMinimal) => InferenceSnippet;
|
|
13
|
+
export declare const snippetBasic: (model: ModelDataMinimal) => InferenceSnippet;
|
|
14
|
+
export declare const snippetFile: (model: ModelDataMinimal) => InferenceSnippet;
|
|
15
|
+
export declare const snippetTextToImage: (model: ModelDataMinimal) => InferenceSnippet;
|
|
16
|
+
export declare const snippetTabular: (model: ModelDataMinimal) => InferenceSnippet;
|
|
17
|
+
export declare const snippetTextToAudio: (model: ModelDataMinimal) => InferenceSnippet;
|
|
18
|
+
export declare const snippetDocumentQuestionAnswering: (model: ModelDataMinimal) => InferenceSnippet;
|
|
19
|
+
export declare const pythonSnippets: Partial<Record<PipelineType, (model: ModelDataMinimal, accessToken: string, opts?: Record<string, unknown>) => InferenceSnippet | InferenceSnippet[]>>;
|
|
20
|
+
export declare function getPythonInferenceSnippet(model: ModelDataMinimal, accessToken: string, opts?: Record<string, unknown>): InferenceSnippet | InferenceSnippet[];
|
|
21
|
+
export declare function hasPythonInferenceSnippet(model: ModelDataMinimal): boolean;
|
|
22
|
+
//# sourceMappingURL=python.d.cts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"python.d.cts","sourceRoot":"","sources":["../../src/snippets/python.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,YAAY,EAAE,MAAM,iBAAiB,CAAC;AACpD,OAAO,KAAK,EAAE,0BAA0B,EAAE,oBAAoB,EAAE,MAAM,mBAAmB,CAAC;AAG1F,OAAO,KAAK,EAAE,gBAAgB,EAAE,gBAAgB,EAAE,MAAM,YAAY,CAAC;AAErE,eAAO,MAAM,qBAAqB,UAC1B,gBAAgB,eACV,MAAM,SACZ;IACN,SAAS,CAAC,EAAE,OAAO,CAAC;IACpB,QAAQ,CAAC,EAAE,0BAA0B,EAAE,CAAC;IACxC,WAAW,CAAC,EAAE,oBAAoB,CAAC,aAAa,CAAC,CAAC;IAClD,UAAU,CAAC,EAAE,oBAAoB,CAAC,YAAY,CAAC,CAAC;IAChD,KAAK,CAAC,EAAE,oBAAoB,CAAC,OAAO,CAAC,CAAC;CACtC,KACC,gBAAgB,EAiGlB,CAAC;AAEF,eAAO,MAAM,6BAA6B,UAAW,gBAAgB,KAAG,gBAStE,CAAC;AAEH,eAAO,MAAM,kCAAkC,UAAW,gBAAgB,KAAG,gBAe3E,CAAC;AAEH,eAAO,MAAM,YAAY,UAAW,gBAAgB,KAAG,gBAQrD,CAAC;AAEH,eAAO,MAAM,WAAW,UAAW,gBAAgB,KAAG,gBAQpD,CAAC;AAEH,eAAO,MAAM,kBAAkB,UAAW,gBAAgB,KAAG,gBAW3D,CAAC;AAEH,eAAO,MAAM,cAAc,UAAW,gBAAgB,KAAG,gBAOvD,CAAC;AAEH,eAAO,MAAM,kBAAkB,UAAW,gBAAgB,KAAG,gBA+B5D,CAAC;AAEF,eAAO,MAAM,gCAAgC,UAAW,gBAAgB,KAAG,gBAWzE,CAAC;AAEH,eAAO,MAAM,cAAc,EAAE,OAAO,CACnC,MAAM,CACL,YAAY,EACZ,CACC,KAAK,EAAE,gBAAgB,EACvB,WAAW,EAAE,MAAM,EACnB,IAAI,CAAC,EAAE,MAAM,CAAC,MAAM,EAAE,OAAO,CAAC,KAC1B,gBAAgB,GAAG,gBAAgB,EAAE,CAC1C,CA8BD,CAAC;AAEF,wBAAgB,yBAAyB,CACxC,KAAK,EAAE,gBAAgB,EACvB,WAAW,EAAE,MAAM,EACnB,IAAI,CAAC,EAAE,MAAM,CAAC,MAAM,EAAE,OAAO,CAAC,GAC5B,gBAAgB,GAAG,gBAAgB,EAAE,CAwBvC;AAED,wBAAgB,yBAAyB,CAAC,KAAK,EAAE,gBAAgB,GAAG,OAAO,CAE1E"}
|
|
@@ -0,0 +1,22 @@
|
|
|
1
|
+
import type { PipelineType } from "../pipelines.js";
|
|
2
|
+
import type { ChatCompletionInputMessage, GenerationParameters } from "../tasks/index.js";
|
|
3
|
+
import type { InferenceSnippet, ModelDataMinimal } from "./types.js";
|
|
4
|
+
export declare const snippetConversational: (model: ModelDataMinimal, accessToken: string, opts?: {
|
|
5
|
+
streaming?: boolean;
|
|
6
|
+
messages?: ChatCompletionInputMessage[];
|
|
7
|
+
temperature?: GenerationParameters["temperature"];
|
|
8
|
+
max_tokens?: GenerationParameters["max_tokens"];
|
|
9
|
+
top_p?: GenerationParameters["top_p"];
|
|
10
|
+
}) => InferenceSnippet[];
|
|
11
|
+
export declare const snippetZeroShotClassification: (model: ModelDataMinimal) => InferenceSnippet;
|
|
12
|
+
export declare const snippetZeroShotImageClassification: (model: ModelDataMinimal) => InferenceSnippet;
|
|
13
|
+
export declare const snippetBasic: (model: ModelDataMinimal) => InferenceSnippet;
|
|
14
|
+
export declare const snippetFile: (model: ModelDataMinimal) => InferenceSnippet;
|
|
15
|
+
export declare const snippetTextToImage: (model: ModelDataMinimal) => InferenceSnippet;
|
|
16
|
+
export declare const snippetTabular: (model: ModelDataMinimal) => InferenceSnippet;
|
|
17
|
+
export declare const snippetTextToAudio: (model: ModelDataMinimal) => InferenceSnippet;
|
|
18
|
+
export declare const snippetDocumentQuestionAnswering: (model: ModelDataMinimal) => InferenceSnippet;
|
|
19
|
+
export declare const pythonSnippets: Partial<Record<PipelineType, (model: ModelDataMinimal, accessToken: string, opts?: Record<string, unknown>) => InferenceSnippet | InferenceSnippet[]>>;
|
|
20
|
+
export declare function getPythonInferenceSnippet(model: ModelDataMinimal, accessToken: string, opts?: Record<string, unknown>): InferenceSnippet | InferenceSnippet[];
|
|
21
|
+
export declare function hasPythonInferenceSnippet(model: ModelDataMinimal): boolean;
|
|
22
|
+
//# sourceMappingURL=python.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"python.d.ts","sourceRoot":"","sources":["../../src/snippets/python.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,YAAY,EAAE,MAAM,iBAAiB,CAAC;AACpD,OAAO,KAAK,EAAE,0BAA0B,EAAE,oBAAoB,EAAE,MAAM,mBAAmB,CAAC;AAG1F,OAAO,KAAK,EAAE,gBAAgB,EAAE,gBAAgB,EAAE,MAAM,YAAY,CAAC;AAErE,eAAO,MAAM,qBAAqB,UAC1B,gBAAgB,eACV,MAAM,SACZ;IACN,SAAS,CAAC,EAAE,OAAO,CAAC;IACpB,QAAQ,CAAC,EAAE,0BAA0B,EAAE,CAAC;IACxC,WAAW,CAAC,EAAE,oBAAoB,CAAC,aAAa,CAAC,CAAC;IAClD,UAAU,CAAC,EAAE,oBAAoB,CAAC,YAAY,CAAC,CAAC;IAChD,KAAK,CAAC,EAAE,oBAAoB,CAAC,OAAO,CAAC,CAAC;CACtC,KACC,gBAAgB,EAiGlB,CAAC;AAEF,eAAO,MAAM,6BAA6B,UAAW,gBAAgB,KAAG,gBAStE,CAAC;AAEH,eAAO,MAAM,kCAAkC,UAAW,gBAAgB,KAAG,gBAe3E,CAAC;AAEH,eAAO,MAAM,YAAY,UAAW,gBAAgB,KAAG,gBAQrD,CAAC;AAEH,eAAO,MAAM,WAAW,UAAW,gBAAgB,KAAG,gBAQpD,CAAC;AAEH,eAAO,MAAM,kBAAkB,UAAW,gBAAgB,KAAG,gBAW3D,CAAC;AAEH,eAAO,MAAM,cAAc,UAAW,gBAAgB,KAAG,gBAOvD,CAAC;AAEH,eAAO,MAAM,kBAAkB,UAAW,gBAAgB,KAAG,gBA+B5D,CAAC;AAEF,eAAO,MAAM,gCAAgC,UAAW,gBAAgB,KAAG,gBAWzE,CAAC;AAEH,eAAO,MAAM,cAAc,EAAE,OAAO,CACnC,MAAM,CACL,YAAY,EACZ,CACC,KAAK,EAAE,gBAAgB,EACvB,WAAW,EAAE,MAAM,EACnB,IAAI,CAAC,EAAE,MAAM,CAAC,MAAM,EAAE,OAAO,CAAC,KAC1B,gBAAgB,GAAG,gBAAgB,EAAE,CAC1C,CA8BD,CAAC;AAEF,wBAAgB,yBAAyB,CACxC,KAAK,EAAE,gBAAgB,EACvB,WAAW,EAAE,MAAM,EACnB,IAAI,CAAC,EAAE,MAAM,CAAC,MAAM,EAAE,OAAO,CAAC,GAC5B,gBAAgB,GAAG,gBAAgB,EAAE,CAwBvC;AAED,wBAAgB,yBAAyB,CAAC,KAAK,EAAE,gBAAgB,GAAG,OAAO,CAE1E"}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"python.spec.d.cts","sourceRoot":"","sources":["../../src/snippets/python.spec.ts"],"names":[],"mappings":""}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"python.spec.d.ts","sourceRoot":"","sources":["../../src/snippets/python.spec.ts"],"names":[],"mappings":""}
|
|
@@ -0,0 +1,12 @@
|
|
|
1
|
+
import type { ModelData } from "../model-data";
|
|
2
|
+
/**
|
|
3
|
+
* Minimal model data required for snippets.
|
|
4
|
+
*
|
|
5
|
+
* Add more fields as needed.
|
|
6
|
+
*/
|
|
7
|
+
export type ModelDataMinimal = Pick<ModelData, "id" | "pipeline_tag" | "mask_token" | "library_name" | "config" | "tags" | "inference">;
|
|
8
|
+
export interface InferenceSnippet {
|
|
9
|
+
content: string;
|
|
10
|
+
client?: string;
|
|
11
|
+
}
|
|
12
|
+
//# sourceMappingURL=types.d.cts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"types.d.cts","sourceRoot":"","sources":["../../src/snippets/types.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,eAAe,CAAC;AAE/C;;;;GAIG;AACH,MAAM,MAAM,gBAAgB,GAAG,IAAI,CAClC,SAAS,EACT,IAAI,GAAG,cAAc,GAAG,YAAY,GAAG,cAAc,GAAG,QAAQ,GAAG,MAAM,GAAG,WAAW,CACvF,CAAC;AAEF,MAAM,WAAW,gBAAgB;IAChC,OAAO,EAAE,MAAM,CAAC;IAChB,MAAM,CAAC,EAAE,MAAM,CAAC;CAChB"}
|
|
@@ -0,0 +1,12 @@
|
|
|
1
|
+
import type { ModelData } from "../model-data";
|
|
2
|
+
/**
|
|
3
|
+
* Minimal model data required for snippets.
|
|
4
|
+
*
|
|
5
|
+
* Add more fields as needed.
|
|
6
|
+
*/
|
|
7
|
+
export type ModelDataMinimal = Pick<ModelData, "id" | "pipeline_tag" | "mask_token" | "library_name" | "config" | "tags" | "inference">;
|
|
8
|
+
export interface InferenceSnippet {
|
|
9
|
+
content: string;
|
|
10
|
+
client?: string;
|
|
11
|
+
}
|
|
12
|
+
//# sourceMappingURL=types.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"types.d.ts","sourceRoot":"","sources":["../../src/snippets/types.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,eAAe,CAAC;AAE/C;;;;GAIG;AACH,MAAM,MAAM,gBAAgB,GAAG,IAAI,CAClC,SAAS,EACT,IAAI,GAAG,cAAc,GAAG,YAAY,GAAG,cAAc,GAAG,QAAQ,GAAG,MAAM,GAAG,WAAW,CACvF,CAAC;AAEF,MAAM,WAAW,gBAAgB;IAChC,OAAO,EAAE,MAAM,CAAC;IAChB,MAAM,CAAC,EAAE,MAAM,CAAC;CAChB"}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"data.d.cts","sourceRoot":"","sources":["../../../src/tasks/audio-classification/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,IAAI,CAAC;AAEzC,QAAA,MAAM,QAAQ,EAAE,cA4Ef,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../src/tasks/audio-classification/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,IAAI,CAAC;AAEzC,QAAA,MAAM,QAAQ,EAAE,cA4Ef,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -0,0 +1,53 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Inference code generated from the JSON schema spec in ./spec
|
|
3
|
+
*
|
|
4
|
+
* Using src/scripts/inference-codegen
|
|
5
|
+
*/
|
|
6
|
+
/**
|
|
7
|
+
* Inputs for Audio Classification inference
|
|
8
|
+
*/
|
|
9
|
+
export interface AudioClassificationInput {
|
|
10
|
+
/**
|
|
11
|
+
* The input audio data as a base64-encoded string. If no `parameters` are provided, you can
|
|
12
|
+
* also provide the audio data as a raw bytes payload.
|
|
13
|
+
*/
|
|
14
|
+
inputs: string;
|
|
15
|
+
/**
|
|
16
|
+
* Additional inference parameters
|
|
17
|
+
*/
|
|
18
|
+
parameters?: AudioClassificationParameters;
|
|
19
|
+
[property: string]: unknown;
|
|
20
|
+
}
|
|
21
|
+
/**
|
|
22
|
+
* Additional inference parameters
|
|
23
|
+
*
|
|
24
|
+
* Additional inference parameters for Audio Classification
|
|
25
|
+
*/
|
|
26
|
+
export interface AudioClassificationParameters {
|
|
27
|
+
function_to_apply?: ClassificationOutputTransform;
|
|
28
|
+
/**
|
|
29
|
+
* When specified, limits the output to the top K most probable classes.
|
|
30
|
+
*/
|
|
31
|
+
top_k?: number;
|
|
32
|
+
[property: string]: unknown;
|
|
33
|
+
}
|
|
34
|
+
/**
|
|
35
|
+
* The function to apply to the model outputs in order to retrieve the scores.
|
|
36
|
+
*/
|
|
37
|
+
export type ClassificationOutputTransform = "sigmoid" | "softmax" | "none";
|
|
38
|
+
export type AudioClassificationOutput = AudioClassificationOutputElement[];
|
|
39
|
+
/**
|
|
40
|
+
* Outputs for Audio Classification inference
|
|
41
|
+
*/
|
|
42
|
+
export interface AudioClassificationOutputElement {
|
|
43
|
+
/**
|
|
44
|
+
* The predicted class label.
|
|
45
|
+
*/
|
|
46
|
+
label: string;
|
|
47
|
+
/**
|
|
48
|
+
* The corresponding probability.
|
|
49
|
+
*/
|
|
50
|
+
score: number;
|
|
51
|
+
[property: string]: unknown;
|
|
52
|
+
}
|
|
53
|
+
//# sourceMappingURL=inference.d.cts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"inference.d.cts","sourceRoot":"","sources":["../../../src/tasks/audio-classification/inference.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AACH;;GAEG;AACH,MAAM,WAAW,wBAAwB;IACxC;;;OAGG;IACH,MAAM,EAAE,MAAM,CAAC;IACf;;OAEG;IACH,UAAU,CAAC,EAAE,6BAA6B,CAAC;IAC3C,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AACD;;;;GAIG;AACH,MAAM,WAAW,6BAA6B;IAC7C,iBAAiB,CAAC,EAAE,6BAA6B,CAAC;IAClD;;OAEG;IACH,KAAK,CAAC,EAAE,MAAM,CAAC;IACf,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AACD;;GAEG;AACH,MAAM,MAAM,6BAA6B,GAAG,SAAS,GAAG,SAAS,GAAG,MAAM,CAAC;AAC3E,MAAM,MAAM,yBAAyB,GAAG,gCAAgC,EAAE,CAAC;AAC3E;;GAEG;AACH,MAAM,WAAW,gCAAgC;IAChD;;OAEG;IACH,KAAK,EAAE,MAAM,CAAC;IACd;;OAEG;IACH,KAAK,EAAE,MAAM,CAAC;IACd,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B"}
|
|
@@ -0,0 +1,53 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Inference code generated from the JSON schema spec in ./spec
|
|
3
|
+
*
|
|
4
|
+
* Using src/scripts/inference-codegen
|
|
5
|
+
*/
|
|
6
|
+
/**
|
|
7
|
+
* Inputs for Audio Classification inference
|
|
8
|
+
*/
|
|
9
|
+
export interface AudioClassificationInput {
|
|
10
|
+
/**
|
|
11
|
+
* The input audio data as a base64-encoded string. If no `parameters` are provided, you can
|
|
12
|
+
* also provide the audio data as a raw bytes payload.
|
|
13
|
+
*/
|
|
14
|
+
inputs: string;
|
|
15
|
+
/**
|
|
16
|
+
* Additional inference parameters
|
|
17
|
+
*/
|
|
18
|
+
parameters?: AudioClassificationParameters;
|
|
19
|
+
[property: string]: unknown;
|
|
20
|
+
}
|
|
21
|
+
/**
|
|
22
|
+
* Additional inference parameters
|
|
23
|
+
*
|
|
24
|
+
* Additional inference parameters for Audio Classification
|
|
25
|
+
*/
|
|
26
|
+
export interface AudioClassificationParameters {
|
|
27
|
+
function_to_apply?: ClassificationOutputTransform;
|
|
28
|
+
/**
|
|
29
|
+
* When specified, limits the output to the top K most probable classes.
|
|
30
|
+
*/
|
|
31
|
+
top_k?: number;
|
|
32
|
+
[property: string]: unknown;
|
|
33
|
+
}
|
|
34
|
+
/**
|
|
35
|
+
* The function to apply to the model outputs in order to retrieve the scores.
|
|
36
|
+
*/
|
|
37
|
+
export type ClassificationOutputTransform = "sigmoid" | "softmax" | "none";
|
|
38
|
+
export type AudioClassificationOutput = AudioClassificationOutputElement[];
|
|
39
|
+
/**
|
|
40
|
+
* Outputs for Audio Classification inference
|
|
41
|
+
*/
|
|
42
|
+
export interface AudioClassificationOutputElement {
|
|
43
|
+
/**
|
|
44
|
+
* The predicted class label.
|
|
45
|
+
*/
|
|
46
|
+
label: string;
|
|
47
|
+
/**
|
|
48
|
+
* The corresponding probability.
|
|
49
|
+
*/
|
|
50
|
+
score: number;
|
|
51
|
+
[property: string]: unknown;
|
|
52
|
+
}
|
|
53
|
+
//# sourceMappingURL=inference.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"inference.d.ts","sourceRoot":"","sources":["../../../src/tasks/audio-classification/inference.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AACH;;GAEG;AACH,MAAM,WAAW,wBAAwB;IACxC;;;OAGG;IACH,MAAM,EAAE,MAAM,CAAC;IACf;;OAEG;IACH,UAAU,CAAC,EAAE,6BAA6B,CAAC;IAC3C,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AACD;;;;GAIG;AACH,MAAM,WAAW,6BAA6B;IAC7C,iBAAiB,CAAC,EAAE,6BAA6B,CAAC;IAClD;;OAEG;IACH,KAAK,CAAC,EAAE,MAAM,CAAC;IACf,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AACD;;GAEG;AACH,MAAM,MAAM,6BAA6B,GAAG,SAAS,GAAG,SAAS,GAAG,MAAM,CAAC;AAC3E,MAAM,MAAM,yBAAyB,GAAG,gCAAgC,EAAE,CAAC;AAC3E;;GAEG;AACH,MAAM,WAAW,gCAAgC;IAChD;;OAEG;IACH,KAAK,EAAE,MAAM,CAAC;IACd;;OAEG;IACH,KAAK,EAAE,MAAM,CAAC;IACd,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B"}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"data.d.cts","sourceRoot":"","sources":["../../../src/tasks/audio-to-audio/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,IAAI,CAAC;AAEzC,QAAA,MAAM,QAAQ,EAAE,cAiEf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../src/tasks/audio-to-audio/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,IAAI,CAAC;AAEzC,QAAA,MAAM,QAAQ,EAAE,cAiEf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"data.d.cts","sourceRoot":"","sources":["../../../src/tasks/automatic-speech-recognition/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,IAAI,CAAC;AAEzC,QAAA,MAAM,QAAQ,EAAE,cA6Ef,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../src/tasks/automatic-speech-recognition/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,IAAI,CAAC;AAEzC,QAAA,MAAM,QAAQ,EAAE,cA6Ef,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -0,0 +1,155 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Inference code generated from the JSON schema spec in ./spec
|
|
3
|
+
*
|
|
4
|
+
* Using src/scripts/inference-codegen
|
|
5
|
+
*/
|
|
6
|
+
/**
|
|
7
|
+
* Inputs for Automatic Speech Recognition inference
|
|
8
|
+
*/
|
|
9
|
+
export interface AutomaticSpeechRecognitionInput {
|
|
10
|
+
/**
|
|
11
|
+
* The input audio data as a base64-encoded string. If no `parameters` are provided, you can
|
|
12
|
+
* also provide the audio data as a raw bytes payload.
|
|
13
|
+
*/
|
|
14
|
+
inputs: string;
|
|
15
|
+
/**
|
|
16
|
+
* Additional inference parameters
|
|
17
|
+
*/
|
|
18
|
+
parameters?: AutomaticSpeechRecognitionParameters;
|
|
19
|
+
[property: string]: unknown;
|
|
20
|
+
}
|
|
21
|
+
/**
|
|
22
|
+
* Additional inference parameters
|
|
23
|
+
*
|
|
24
|
+
* Additional inference parameters for Automatic Speech Recognition
|
|
25
|
+
*/
|
|
26
|
+
export interface AutomaticSpeechRecognitionParameters {
|
|
27
|
+
/**
|
|
28
|
+
* Parametrization of the text generation process
|
|
29
|
+
*/
|
|
30
|
+
generation_parameters?: GenerationParameters;
|
|
31
|
+
/**
|
|
32
|
+
* Whether to output corresponding timestamps with the generated text
|
|
33
|
+
*/
|
|
34
|
+
return_timestamps?: boolean;
|
|
35
|
+
[property: string]: unknown;
|
|
36
|
+
}
|
|
37
|
+
/**
|
|
38
|
+
* Parametrization of the text generation process
|
|
39
|
+
*
|
|
40
|
+
* Ad-hoc parametrization of the text generation process
|
|
41
|
+
*/
|
|
42
|
+
export interface GenerationParameters {
|
|
43
|
+
/**
|
|
44
|
+
* Whether to use sampling instead of greedy decoding when generating new tokens.
|
|
45
|
+
*/
|
|
46
|
+
do_sample?: boolean;
|
|
47
|
+
/**
|
|
48
|
+
* Controls the stopping condition for beam-based methods.
|
|
49
|
+
*/
|
|
50
|
+
early_stopping?: EarlyStoppingUnion;
|
|
51
|
+
/**
|
|
52
|
+
* If set to float strictly between 0 and 1, only tokens with a conditional probability
|
|
53
|
+
* greater than epsilon_cutoff will be sampled. In the paper, suggested values range from
|
|
54
|
+
* 3e-4 to 9e-4, depending on the size of the model. See [Truncation Sampling as Language
|
|
55
|
+
* Model Desmoothing](https://hf.co/papers/2210.15191) for more details.
|
|
56
|
+
*/
|
|
57
|
+
epsilon_cutoff?: number;
|
|
58
|
+
/**
|
|
59
|
+
* Eta sampling is a hybrid of locally typical sampling and epsilon sampling. If set to
|
|
60
|
+
* float strictly between 0 and 1, a token is only considered if it is greater than either
|
|
61
|
+
* eta_cutoff or sqrt(eta_cutoff) * exp(-entropy(softmax(next_token_logits))). The latter
|
|
62
|
+
* term is intuitively the expected next token probability, scaled by sqrt(eta_cutoff). In
|
|
63
|
+
* the paper, suggested values range from 3e-4 to 2e-3, depending on the size of the model.
|
|
64
|
+
* See [Truncation Sampling as Language Model Desmoothing](https://hf.co/papers/2210.15191)
|
|
65
|
+
* for more details.
|
|
66
|
+
*/
|
|
67
|
+
eta_cutoff?: number;
|
|
68
|
+
/**
|
|
69
|
+
* The maximum length (in tokens) of the generated text, including the input.
|
|
70
|
+
*/
|
|
71
|
+
max_length?: number;
|
|
72
|
+
/**
|
|
73
|
+
* The maximum number of tokens to generate. Takes precedence over max_length.
|
|
74
|
+
*/
|
|
75
|
+
max_new_tokens?: number;
|
|
76
|
+
/**
|
|
77
|
+
* The minimum length (in tokens) of the generated text, including the input.
|
|
78
|
+
*/
|
|
79
|
+
min_length?: number;
|
|
80
|
+
/**
|
|
81
|
+
* The minimum number of tokens to generate. Takes precedence over min_length.
|
|
82
|
+
*/
|
|
83
|
+
min_new_tokens?: number;
|
|
84
|
+
/**
|
|
85
|
+
* Number of groups to divide num_beams into in order to ensure diversity among different
|
|
86
|
+
* groups of beams. See [this paper](https://hf.co/papers/1610.02424) for more details.
|
|
87
|
+
*/
|
|
88
|
+
num_beam_groups?: number;
|
|
89
|
+
/**
|
|
90
|
+
* Number of beams to use for beam search.
|
|
91
|
+
*/
|
|
92
|
+
num_beams?: number;
|
|
93
|
+
/**
|
|
94
|
+
* The value balances the model confidence and the degeneration penalty in contrastive
|
|
95
|
+
* search decoding.
|
|
96
|
+
*/
|
|
97
|
+
penalty_alpha?: number;
|
|
98
|
+
/**
|
|
99
|
+
* The value used to modulate the next token probabilities.
|
|
100
|
+
*/
|
|
101
|
+
temperature?: number;
|
|
102
|
+
/**
|
|
103
|
+
* The number of highest probability vocabulary tokens to keep for top-k-filtering.
|
|
104
|
+
*/
|
|
105
|
+
top_k?: number;
|
|
106
|
+
/**
|
|
107
|
+
* If set to float < 1, only the smallest set of most probable tokens with probabilities
|
|
108
|
+
* that add up to top_p or higher are kept for generation.
|
|
109
|
+
*/
|
|
110
|
+
top_p?: number;
|
|
111
|
+
/**
|
|
112
|
+
* Local typicality measures how similar the conditional probability of predicting a target
|
|
113
|
+
* token next is to the expected conditional probability of predicting a random token next,
|
|
114
|
+
* given the partial text already generated. If set to float < 1, the smallest set of the
|
|
115
|
+
* most locally typical tokens with probabilities that add up to typical_p or higher are
|
|
116
|
+
* kept for generation. See [this paper](https://hf.co/papers/2202.00666) for more details.
|
|
117
|
+
*/
|
|
118
|
+
typical_p?: number;
|
|
119
|
+
/**
|
|
120
|
+
* Whether the model should use the past last key/values attentions to speed up decoding
|
|
121
|
+
*/
|
|
122
|
+
use_cache?: boolean;
|
|
123
|
+
[property: string]: unknown;
|
|
124
|
+
}
|
|
125
|
+
/**
|
|
126
|
+
* Controls the stopping condition for beam-based methods.
|
|
127
|
+
*/
|
|
128
|
+
export type EarlyStoppingUnion = boolean | "never";
|
|
129
|
+
/**
|
|
130
|
+
* Outputs of inference for the Automatic Speech Recognition task
|
|
131
|
+
*/
|
|
132
|
+
export interface AutomaticSpeechRecognitionOutput {
|
|
133
|
+
/**
|
|
134
|
+
* When returnTimestamps is enabled, chunks contains a list of audio chunks identified by
|
|
135
|
+
* the model.
|
|
136
|
+
*/
|
|
137
|
+
chunks?: AutomaticSpeechRecognitionOutputChunk[];
|
|
138
|
+
/**
|
|
139
|
+
* The recognized text.
|
|
140
|
+
*/
|
|
141
|
+
text: string;
|
|
142
|
+
[property: string]: unknown;
|
|
143
|
+
}
|
|
144
|
+
export interface AutomaticSpeechRecognitionOutputChunk {
|
|
145
|
+
/**
|
|
146
|
+
* A chunk of text identified by the model
|
|
147
|
+
*/
|
|
148
|
+
text: string;
|
|
149
|
+
/**
|
|
150
|
+
* The start and end timestamps corresponding with the text
|
|
151
|
+
*/
|
|
152
|
+
timestamps: number[];
|
|
153
|
+
[property: string]: unknown;
|
|
154
|
+
}
|
|
155
|
+
//# sourceMappingURL=inference.d.cts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"inference.d.cts","sourceRoot":"","sources":["../../../src/tasks/automatic-speech-recognition/inference.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAEH;;GAEG;AACH,MAAM,WAAW,+BAA+B;IAC/C;;;OAGG;IACH,MAAM,EAAE,MAAM,CAAC;IACf;;OAEG;IACH,UAAU,CAAC,EAAE,oCAAoC,CAAC;IAClD,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED;;;;GAIG;AACH,MAAM,WAAW,oCAAoC;IACpD;;OAEG;IACH,qBAAqB,CAAC,EAAE,oBAAoB,CAAC;IAC7C;;OAEG;IACH,iBAAiB,CAAC,EAAE,OAAO,CAAC;IAC5B,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED;;;;GAIG;AACH,MAAM,WAAW,oBAAoB;IACpC;;OAEG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IACpB;;OAEG;IACH,cAAc,CAAC,EAAE,kBAAkB,CAAC;IACpC;;;;;OAKG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;;;;;;;OAQG;IACH,UAAU,CAAC,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,UAAU,CAAC,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;OAEG;IACH,UAAU,CAAC,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;;OAGG;IACH,eAAe,CAAC,EAAE,MAAM,CAAC;IACzB;;OAEG;IACH,SAAS,CAAC,EAAE,MAAM,CAAC;IACnB;;;OAGG;IACH,aAAa,CAAC,EAAE,MAAM,CAAC;IACvB;;OAEG;IACH,WAAW,CAAC,EAAE,MAAM,CAAC;IACrB;;OAEG;IACH,KAAK,CAAC,EAAE,MAAM,CAAC;IACf;;;OAGG;IACH,KAAK,CAAC,EAAE,MAAM,CAAC;IACf;;;;;;OAMG;IACH,SAAS,CAAC,EAAE,MAAM,CAAC;IACnB;;OAEG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IACpB,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED;;GAEG;AACH,MAAM,MAAM,kBAAkB,GAAG,OAAO,GAAG,OAAO,CAAC;AAEnD;;GAEG;AACH,MAAM,WAAW,gCAAgC;IAChD;;;OAGG;IACH,MAAM,CAAC,EAAE,qCAAqC,EAAE,CAAC;IACjD;;OAEG;IACH,IAAI,EAAE,MAAM,CAAC;IACb,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,WAAW,qCAAqC;IACrD;;OAEG;IACH,IAAI,EAAE,MAAM,CAAC;IACb;;OAEG;IACH,UAAU,EAAE,MAAM,EAAE,CAAC;IACrB,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B"}
|
|
@@ -0,0 +1,155 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Inference code generated from the JSON schema spec in ./spec
|
|
3
|
+
*
|
|
4
|
+
* Using src/scripts/inference-codegen
|
|
5
|
+
*/
|
|
6
|
+
/**
|
|
7
|
+
* Inputs for Automatic Speech Recognition inference
|
|
8
|
+
*/
|
|
9
|
+
export interface AutomaticSpeechRecognitionInput {
|
|
10
|
+
/**
|
|
11
|
+
* The input audio data as a base64-encoded string. If no `parameters` are provided, you can
|
|
12
|
+
* also provide the audio data as a raw bytes payload.
|
|
13
|
+
*/
|
|
14
|
+
inputs: string;
|
|
15
|
+
/**
|
|
16
|
+
* Additional inference parameters
|
|
17
|
+
*/
|
|
18
|
+
parameters?: AutomaticSpeechRecognitionParameters;
|
|
19
|
+
[property: string]: unknown;
|
|
20
|
+
}
|
|
21
|
+
/**
|
|
22
|
+
* Additional inference parameters
|
|
23
|
+
*
|
|
24
|
+
* Additional inference parameters for Automatic Speech Recognition
|
|
25
|
+
*/
|
|
26
|
+
export interface AutomaticSpeechRecognitionParameters {
|
|
27
|
+
/**
|
|
28
|
+
* Parametrization of the text generation process
|
|
29
|
+
*/
|
|
30
|
+
generation_parameters?: GenerationParameters;
|
|
31
|
+
/**
|
|
32
|
+
* Whether to output corresponding timestamps with the generated text
|
|
33
|
+
*/
|
|
34
|
+
return_timestamps?: boolean;
|
|
35
|
+
[property: string]: unknown;
|
|
36
|
+
}
|
|
37
|
+
/**
|
|
38
|
+
* Parametrization of the text generation process
|
|
39
|
+
*
|
|
40
|
+
* Ad-hoc parametrization of the text generation process
|
|
41
|
+
*/
|
|
42
|
+
export interface GenerationParameters {
|
|
43
|
+
/**
|
|
44
|
+
* Whether to use sampling instead of greedy decoding when generating new tokens.
|
|
45
|
+
*/
|
|
46
|
+
do_sample?: boolean;
|
|
47
|
+
/**
|
|
48
|
+
* Controls the stopping condition for beam-based methods.
|
|
49
|
+
*/
|
|
50
|
+
early_stopping?: EarlyStoppingUnion;
|
|
51
|
+
/**
|
|
52
|
+
* If set to float strictly between 0 and 1, only tokens with a conditional probability
|
|
53
|
+
* greater than epsilon_cutoff will be sampled. In the paper, suggested values range from
|
|
54
|
+
* 3e-4 to 9e-4, depending on the size of the model. See [Truncation Sampling as Language
|
|
55
|
+
* Model Desmoothing](https://hf.co/papers/2210.15191) for more details.
|
|
56
|
+
*/
|
|
57
|
+
epsilon_cutoff?: number;
|
|
58
|
+
/**
|
|
59
|
+
* Eta sampling is a hybrid of locally typical sampling and epsilon sampling. If set to
|
|
60
|
+
* float strictly between 0 and 1, a token is only considered if it is greater than either
|
|
61
|
+
* eta_cutoff or sqrt(eta_cutoff) * exp(-entropy(softmax(next_token_logits))). The latter
|
|
62
|
+
* term is intuitively the expected next token probability, scaled by sqrt(eta_cutoff). In
|
|
63
|
+
* the paper, suggested values range from 3e-4 to 2e-3, depending on the size of the model.
|
|
64
|
+
* See [Truncation Sampling as Language Model Desmoothing](https://hf.co/papers/2210.15191)
|
|
65
|
+
* for more details.
|
|
66
|
+
*/
|
|
67
|
+
eta_cutoff?: number;
|
|
68
|
+
/**
|
|
69
|
+
* The maximum length (in tokens) of the generated text, including the input.
|
|
70
|
+
*/
|
|
71
|
+
max_length?: number;
|
|
72
|
+
/**
|
|
73
|
+
* The maximum number of tokens to generate. Takes precedence over max_length.
|
|
74
|
+
*/
|
|
75
|
+
max_new_tokens?: number;
|
|
76
|
+
/**
|
|
77
|
+
* The minimum length (in tokens) of the generated text, including the input.
|
|
78
|
+
*/
|
|
79
|
+
min_length?: number;
|
|
80
|
+
/**
|
|
81
|
+
* The minimum number of tokens to generate. Takes precedence over min_length.
|
|
82
|
+
*/
|
|
83
|
+
min_new_tokens?: number;
|
|
84
|
+
/**
|
|
85
|
+
* Number of groups to divide num_beams into in order to ensure diversity among different
|
|
86
|
+
* groups of beams. See [this paper](https://hf.co/papers/1610.02424) for more details.
|
|
87
|
+
*/
|
|
88
|
+
num_beam_groups?: number;
|
|
89
|
+
/**
|
|
90
|
+
* Number of beams to use for beam search.
|
|
91
|
+
*/
|
|
92
|
+
num_beams?: number;
|
|
93
|
+
/**
|
|
94
|
+
* The value balances the model confidence and the degeneration penalty in contrastive
|
|
95
|
+
* search decoding.
|
|
96
|
+
*/
|
|
97
|
+
penalty_alpha?: number;
|
|
98
|
+
/**
|
|
99
|
+
* The value used to modulate the next token probabilities.
|
|
100
|
+
*/
|
|
101
|
+
temperature?: number;
|
|
102
|
+
/**
|
|
103
|
+
* The number of highest probability vocabulary tokens to keep for top-k-filtering.
|
|
104
|
+
*/
|
|
105
|
+
top_k?: number;
|
|
106
|
+
/**
|
|
107
|
+
* If set to float < 1, only the smallest set of most probable tokens with probabilities
|
|
108
|
+
* that add up to top_p or higher are kept for generation.
|
|
109
|
+
*/
|
|
110
|
+
top_p?: number;
|
|
111
|
+
/**
|
|
112
|
+
* Local typicality measures how similar the conditional probability of predicting a target
|
|
113
|
+
* token next is to the expected conditional probability of predicting a random token next,
|
|
114
|
+
* given the partial text already generated. If set to float < 1, the smallest set of the
|
|
115
|
+
* most locally typical tokens with probabilities that add up to typical_p or higher are
|
|
116
|
+
* kept for generation. See [this paper](https://hf.co/papers/2202.00666) for more details.
|
|
117
|
+
*/
|
|
118
|
+
typical_p?: number;
|
|
119
|
+
/**
|
|
120
|
+
* Whether the model should use the past last key/values attentions to speed up decoding
|
|
121
|
+
*/
|
|
122
|
+
use_cache?: boolean;
|
|
123
|
+
[property: string]: unknown;
|
|
124
|
+
}
|
|
125
|
+
/**
|
|
126
|
+
* Controls the stopping condition for beam-based methods.
|
|
127
|
+
*/
|
|
128
|
+
export type EarlyStoppingUnion = boolean | "never";
|
|
129
|
+
/**
|
|
130
|
+
* Outputs of inference for the Automatic Speech Recognition task
|
|
131
|
+
*/
|
|
132
|
+
export interface AutomaticSpeechRecognitionOutput {
|
|
133
|
+
/**
|
|
134
|
+
* When returnTimestamps is enabled, chunks contains a list of audio chunks identified by
|
|
135
|
+
* the model.
|
|
136
|
+
*/
|
|
137
|
+
chunks?: AutomaticSpeechRecognitionOutputChunk[];
|
|
138
|
+
/**
|
|
139
|
+
* The recognized text.
|
|
140
|
+
*/
|
|
141
|
+
text: string;
|
|
142
|
+
[property: string]: unknown;
|
|
143
|
+
}
|
|
144
|
+
export interface AutomaticSpeechRecognitionOutputChunk {
|
|
145
|
+
/**
|
|
146
|
+
* A chunk of text identified by the model
|
|
147
|
+
*/
|
|
148
|
+
text: string;
|
|
149
|
+
/**
|
|
150
|
+
* The start and end timestamps corresponding with the text
|
|
151
|
+
*/
|
|
152
|
+
timestamps: number[];
|
|
153
|
+
[property: string]: unknown;
|
|
154
|
+
}
|
|
155
|
+
//# sourceMappingURL=inference.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"inference.d.ts","sourceRoot":"","sources":["../../../src/tasks/automatic-speech-recognition/inference.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAEH;;GAEG;AACH,MAAM,WAAW,+BAA+B;IAC/C;;;OAGG;IACH,MAAM,EAAE,MAAM,CAAC;IACf;;OAEG;IACH,UAAU,CAAC,EAAE,oCAAoC,CAAC;IAClD,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED;;;;GAIG;AACH,MAAM,WAAW,oCAAoC;IACpD;;OAEG;IACH,qBAAqB,CAAC,EAAE,oBAAoB,CAAC;IAC7C;;OAEG;IACH,iBAAiB,CAAC,EAAE,OAAO,CAAC;IAC5B,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED;;;;GAIG;AACH,MAAM,WAAW,oBAAoB;IACpC;;OAEG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IACpB;;OAEG;IACH,cAAc,CAAC,EAAE,kBAAkB,CAAC;IACpC;;;;;OAKG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;;;;;;;OAQG;IACH,UAAU,CAAC,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,UAAU,CAAC,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;OAEG;IACH,UAAU,CAAC,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;;OAGG;IACH,eAAe,CAAC,EAAE,MAAM,CAAC;IACzB;;OAEG;IACH,SAAS,CAAC,EAAE,MAAM,CAAC;IACnB;;;OAGG;IACH,aAAa,CAAC,EAAE,MAAM,CAAC;IACvB;;OAEG;IACH,WAAW,CAAC,EAAE,MAAM,CAAC;IACrB;;OAEG;IACH,KAAK,CAAC,EAAE,MAAM,CAAC;IACf;;;OAGG;IACH,KAAK,CAAC,EAAE,MAAM,CAAC;IACf;;;;;;OAMG;IACH,SAAS,CAAC,EAAE,MAAM,CAAC;IACnB;;OAEG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IACpB,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED;;GAEG;AACH,MAAM,MAAM,kBAAkB,GAAG,OAAO,GAAG,OAAO,CAAC;AAEnD;;GAEG;AACH,MAAM,WAAW,gCAAgC;IAChD;;;OAGG;IACH,MAAM,CAAC,EAAE,qCAAqC,EAAE,CAAC;IACjD;;OAEG;IACH,IAAI,EAAE,MAAM,CAAC;IACb,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,WAAW,qCAAqC;IACrD;;OAEG;IACH,IAAI,EAAE,MAAM,CAAC;IACb;;OAEG;IACH,UAAU,EAAE,MAAM,EAAE,CAAC;IACrB,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B"}
|