@huggingface/tasks 0.11.3 → 0.11.4
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.cjs +22 -5
- package/dist/index.js +22 -5
- package/dist/src/dataset-libraries.d.ts +6 -0
- package/dist/src/dataset-libraries.d.ts.map +1 -1
- package/dist/src/tasks/document-question-answering/data.d.ts.map +1 -1
- package/dist/src/tasks/image-feature-extraction/data.d.ts.map +1 -1
- package/dist/src/tasks/image-text-to-text/data.d.ts.map +1 -1
- package/package.json +1 -1
- package/src/dataset-libraries.ts +6 -0
- package/src/tasks/document-question-answering/data.ts +6 -3
- package/src/tasks/image-feature-extraction/data.ts +4 -0
- package/src/tasks/image-text-to-text/data.ts +6 -2
package/dist/index.cjs
CHANGED
|
@@ -1617,7 +1617,10 @@ var data_default3 = taskData3;
|
|
|
1617
1617
|
var taskData4 = {
|
|
1618
1618
|
datasets: [
|
|
1619
1619
|
{
|
|
1620
|
-
|
|
1620
|
+
description: "Largest document understanding dataset.",
|
|
1621
|
+
id: "HuggingFaceM4/Docmatix"
|
|
1622
|
+
},
|
|
1623
|
+
{
|
|
1621
1624
|
description: "Dataset from the 2020 DocVQA challenge. The documents are taken from the UCSF Industry Documents Library.",
|
|
1622
1625
|
id: "eliolio/docvqa"
|
|
1623
1626
|
}
|
|
@@ -1658,8 +1661,8 @@ var taskData4 = {
|
|
|
1658
1661
|
id: "impira/layoutlm-document-qa"
|
|
1659
1662
|
},
|
|
1660
1663
|
{
|
|
1661
|
-
description: "A special model for OCR-free Document QA task.
|
|
1662
|
-
id: "
|
|
1664
|
+
description: "A special model for OCR-free Document QA task.",
|
|
1665
|
+
id: "microsoft/udop-large"
|
|
1663
1666
|
},
|
|
1664
1667
|
{
|
|
1665
1668
|
description: "A powerful model for document question answering.",
|
|
@@ -1939,6 +1942,10 @@ var taskData8 = {
|
|
|
1939
1942
|
{
|
|
1940
1943
|
description: "A robust image feature extraction models.",
|
|
1941
1944
|
id: "facebook/dino-vitb16"
|
|
1945
|
+
},
|
|
1946
|
+
{
|
|
1947
|
+
description: "Strong image-text-to-text model made for information retrieval from documents.",
|
|
1948
|
+
id: "vidore/colpali"
|
|
1942
1949
|
}
|
|
1943
1950
|
],
|
|
1944
1951
|
spaces: [],
|
|
@@ -2165,8 +2172,8 @@ var taskData11 = {
|
|
|
2165
2172
|
metrics: [],
|
|
2166
2173
|
models: [
|
|
2167
2174
|
{
|
|
2168
|
-
description: "Cutting-edge vision language model that can
|
|
2169
|
-
id: "
|
|
2175
|
+
description: "Cutting-edge vision language model that can take multiple image inputs.",
|
|
2176
|
+
id: "facebook/chameleon-7b"
|
|
2170
2177
|
},
|
|
2171
2178
|
{
|
|
2172
2179
|
description: "Cutting-edge conversational vision language model that can take multiple image inputs.",
|
|
@@ -2179,6 +2186,10 @@ var taskData11 = {
|
|
|
2179
2186
|
{
|
|
2180
2187
|
description: "Strong image-text-to-text model made to understand documents.",
|
|
2181
2188
|
id: "mPLUG/DocOwl1.5"
|
|
2189
|
+
},
|
|
2190
|
+
{
|
|
2191
|
+
description: "Strong image-text-to-text model.",
|
|
2192
|
+
id: "llava-hf/llava-v1.6-mistral-7b-hf"
|
|
2182
2193
|
}
|
|
2183
2194
|
],
|
|
2184
2195
|
spaces: [
|
|
@@ -6626,6 +6637,12 @@ var DATASET_LIBRARIES_UI_ELEMENTS = {
|
|
|
6626
6637
|
repoName: "argilla",
|
|
6627
6638
|
repoUrl: "https://github.com/argilla-io/argilla",
|
|
6628
6639
|
docsUrl: "https://argilla-io.github.io/argilla"
|
|
6640
|
+
},
|
|
6641
|
+
polars: {
|
|
6642
|
+
prettyLabel: "Polars",
|
|
6643
|
+
repoName: "polars",
|
|
6644
|
+
repoUrl: "https://github.com/pola-rs/polars",
|
|
6645
|
+
docsUrl: "https://docs.pola.rs/"
|
|
6629
6646
|
}
|
|
6630
6647
|
};
|
|
6631
6648
|
// Annotate the CommonJS export names for ESM import in node:
|
package/dist/index.js
CHANGED
|
@@ -1579,7 +1579,10 @@ var data_default3 = taskData3;
|
|
|
1579
1579
|
var taskData4 = {
|
|
1580
1580
|
datasets: [
|
|
1581
1581
|
{
|
|
1582
|
-
|
|
1582
|
+
description: "Largest document understanding dataset.",
|
|
1583
|
+
id: "HuggingFaceM4/Docmatix"
|
|
1584
|
+
},
|
|
1585
|
+
{
|
|
1583
1586
|
description: "Dataset from the 2020 DocVQA challenge. The documents are taken from the UCSF Industry Documents Library.",
|
|
1584
1587
|
id: "eliolio/docvqa"
|
|
1585
1588
|
}
|
|
@@ -1620,8 +1623,8 @@ var taskData4 = {
|
|
|
1620
1623
|
id: "impira/layoutlm-document-qa"
|
|
1621
1624
|
},
|
|
1622
1625
|
{
|
|
1623
|
-
description: "A special model for OCR-free Document QA task.
|
|
1624
|
-
id: "
|
|
1626
|
+
description: "A special model for OCR-free Document QA task.",
|
|
1627
|
+
id: "microsoft/udop-large"
|
|
1625
1628
|
},
|
|
1626
1629
|
{
|
|
1627
1630
|
description: "A powerful model for document question answering.",
|
|
@@ -1901,6 +1904,10 @@ var taskData8 = {
|
|
|
1901
1904
|
{
|
|
1902
1905
|
description: "A robust image feature extraction models.",
|
|
1903
1906
|
id: "facebook/dino-vitb16"
|
|
1907
|
+
},
|
|
1908
|
+
{
|
|
1909
|
+
description: "Strong image-text-to-text model made for information retrieval from documents.",
|
|
1910
|
+
id: "vidore/colpali"
|
|
1904
1911
|
}
|
|
1905
1912
|
],
|
|
1906
1913
|
spaces: [],
|
|
@@ -2127,8 +2134,8 @@ var taskData11 = {
|
|
|
2127
2134
|
metrics: [],
|
|
2128
2135
|
models: [
|
|
2129
2136
|
{
|
|
2130
|
-
description: "Cutting-edge vision language model that can
|
|
2131
|
-
id: "
|
|
2137
|
+
description: "Cutting-edge vision language model that can take multiple image inputs.",
|
|
2138
|
+
id: "facebook/chameleon-7b"
|
|
2132
2139
|
},
|
|
2133
2140
|
{
|
|
2134
2141
|
description: "Cutting-edge conversational vision language model that can take multiple image inputs.",
|
|
@@ -2141,6 +2148,10 @@ var taskData11 = {
|
|
|
2141
2148
|
{
|
|
2142
2149
|
description: "Strong image-text-to-text model made to understand documents.",
|
|
2143
2150
|
id: "mPLUG/DocOwl1.5"
|
|
2151
|
+
},
|
|
2152
|
+
{
|
|
2153
|
+
description: "Strong image-text-to-text model.",
|
|
2154
|
+
id: "llava-hf/llava-v1.6-mistral-7b-hf"
|
|
2144
2155
|
}
|
|
2145
2156
|
],
|
|
2146
2157
|
spaces: [
|
|
@@ -6588,6 +6599,12 @@ var DATASET_LIBRARIES_UI_ELEMENTS = {
|
|
|
6588
6599
|
repoName: "argilla",
|
|
6589
6600
|
repoUrl: "https://github.com/argilla-io/argilla",
|
|
6590
6601
|
docsUrl: "https://argilla-io.github.io/argilla"
|
|
6602
|
+
},
|
|
6603
|
+
polars: {
|
|
6604
|
+
prettyLabel: "Polars",
|
|
6605
|
+
repoName: "polars",
|
|
6606
|
+
repoUrl: "https://github.com/pola-rs/polars",
|
|
6607
|
+
docsUrl: "https://docs.pola.rs/"
|
|
6591
6608
|
}
|
|
6592
6609
|
};
|
|
6593
6610
|
export {
|
|
@@ -70,6 +70,12 @@ export declare const DATASET_LIBRARIES_UI_ELEMENTS: {
|
|
|
70
70
|
repoUrl: string;
|
|
71
71
|
docsUrl: string;
|
|
72
72
|
};
|
|
73
|
+
polars: {
|
|
74
|
+
prettyLabel: string;
|
|
75
|
+
repoName: string;
|
|
76
|
+
repoUrl: string;
|
|
77
|
+
docsUrl: string;
|
|
78
|
+
};
|
|
73
79
|
};
|
|
74
80
|
export type DatasetLibraryKey = keyof typeof DATASET_LIBRARIES_UI_ELEMENTS;
|
|
75
81
|
//# sourceMappingURL=dataset-libraries.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"dataset-libraries.d.ts","sourceRoot":"","sources":["../../src/dataset-libraries.ts"],"names":[],"mappings":"AAAA;;GAEG;AACH,MAAM,WAAW,uBAAuB;IACvC;;;;OAIG;IACH,WAAW,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,QAAQ,EAAE,MAAM,CAAC;IACjB;;OAEG;IACH,OAAO,EAAE,MAAM,CAAC;IAChB;;OAEG;IACH,OAAO,CAAC,EAAE,MAAM,CAAC;CACjB;AAED,eAAO,MAAM,6BAA6B
|
|
1
|
+
{"version":3,"file":"dataset-libraries.d.ts","sourceRoot":"","sources":["../../src/dataset-libraries.ts"],"names":[],"mappings":"AAAA;;GAEG;AACH,MAAM,WAAW,uBAAuB;IACvC;;;;OAIG;IACH,WAAW,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,QAAQ,EAAE,MAAM,CAAC;IACjB;;OAEG;IACH,OAAO,EAAE,MAAM,CAAC;IAChB;;OAEG;IACH,OAAO,CAAC,EAAE,MAAM,CAAC;CACjB;AAED,eAAO,MAAM,6BAA6B;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CAuDS,CAAC;AAGpD,MAAM,MAAM,iBAAiB,GAAG,MAAM,OAAO,6BAA6B,CAAC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/document-question-answering/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,IAAI,CAAC;AAEzC,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/document-question-answering/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,IAAI,CAAC;AAEzC,QAAA,MAAM,QAAQ,EAAE,cA4Ef,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/image-feature-extraction/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,IAAI,CAAC;AAEzC,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/image-feature-extraction/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,IAAI,CAAC;AAEzC,QAAA,MAAM,QAAQ,EAAE,cAkDf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/image-text-to-text/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,IAAI,CAAC;AAEzC,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/image-text-to-text/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,IAAI,CAAC;AAEzC,QAAA,MAAM,QAAQ,EAAE,cAyFf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
package/package.json
CHANGED
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@huggingface/tasks",
|
|
3
3
|
"packageManager": "pnpm@8.10.5",
|
|
4
|
-
"version": "0.11.
|
|
4
|
+
"version": "0.11.4",
|
|
5
5
|
"description": "List of ML tasks for huggingface.co/tasks",
|
|
6
6
|
"repository": "https://github.com/huggingface/huggingface.js.git",
|
|
7
7
|
"publishConfig": {
|
package/src/dataset-libraries.ts
CHANGED
|
@@ -71,6 +71,12 @@ export const DATASET_LIBRARIES_UI_ELEMENTS = {
|
|
|
71
71
|
repoUrl: "https://github.com/argilla-io/argilla",
|
|
72
72
|
docsUrl: "https://argilla-io.github.io/argilla",
|
|
73
73
|
},
|
|
74
|
+
polars: {
|
|
75
|
+
prettyLabel: "Polars",
|
|
76
|
+
repoName: "polars",
|
|
77
|
+
repoUrl: "https://github.com/pola-rs/polars",
|
|
78
|
+
docsUrl: "https://docs.pola.rs/",
|
|
79
|
+
},
|
|
74
80
|
} satisfies Record<string, DatasetLibraryUiElement>;
|
|
75
81
|
|
|
76
82
|
/// List of the dataset libraries supported by the Hub
|
|
@@ -3,7 +3,10 @@ import type { TaskDataCustom } from "..";
|
|
|
3
3
|
const taskData: TaskDataCustom = {
|
|
4
4
|
datasets: [
|
|
5
5
|
{
|
|
6
|
-
|
|
6
|
+
description: "Largest document understanding dataset.",
|
|
7
|
+
id: "HuggingFaceM4/Docmatix",
|
|
8
|
+
},
|
|
9
|
+
{
|
|
7
10
|
description:
|
|
8
11
|
"Dataset from the 2020 DocVQA challenge. The documents are taken from the UCSF Industry Documents Library.",
|
|
9
12
|
id: "eliolio/docvqa",
|
|
@@ -47,8 +50,8 @@ const taskData: TaskDataCustom = {
|
|
|
47
50
|
id: "impira/layoutlm-document-qa",
|
|
48
51
|
},
|
|
49
52
|
{
|
|
50
|
-
description: "A special model for OCR-free Document QA task.
|
|
51
|
-
id: "
|
|
53
|
+
description: "A special model for OCR-free Document QA task.",
|
|
54
|
+
id: "microsoft/udop-large",
|
|
52
55
|
},
|
|
53
56
|
{
|
|
54
57
|
description: "A powerful model for document question answering.",
|
|
@@ -42,6 +42,10 @@ const taskData: TaskDataCustom = {
|
|
|
42
42
|
description: "A robust image feature extraction models.",
|
|
43
43
|
id: "facebook/dino-vitb16",
|
|
44
44
|
},
|
|
45
|
+
{
|
|
46
|
+
description: "Strong image-text-to-text model made for information retrieval from documents.",
|
|
47
|
+
id: "vidore/colpali",
|
|
48
|
+
},
|
|
45
49
|
],
|
|
46
50
|
spaces: [],
|
|
47
51
|
summary: "Image feature extraction is the task of extracting features learnt in a computer vision model.",
|
|
@@ -43,8 +43,8 @@ const taskData: TaskDataCustom = {
|
|
|
43
43
|
metrics: [],
|
|
44
44
|
models: [
|
|
45
45
|
{
|
|
46
|
-
description: "Cutting-edge vision language model that can
|
|
47
|
-
id: "
|
|
46
|
+
description: "Cutting-edge vision language model that can take multiple image inputs.",
|
|
47
|
+
id: "facebook/chameleon-7b",
|
|
48
48
|
},
|
|
49
49
|
{
|
|
50
50
|
description: "Cutting-edge conversational vision language model that can take multiple image inputs.",
|
|
@@ -58,6 +58,10 @@ const taskData: TaskDataCustom = {
|
|
|
58
58
|
description: "Strong image-text-to-text model made to understand documents.",
|
|
59
59
|
id: "mPLUG/DocOwl1.5",
|
|
60
60
|
},
|
|
61
|
+
{
|
|
62
|
+
description: "Strong image-text-to-text model.",
|
|
63
|
+
id: "llava-hf/llava-v1.6-mistral-7b-hf",
|
|
64
|
+
},
|
|
61
65
|
],
|
|
62
66
|
spaces: [
|
|
63
67
|
{
|