@huggingface/tasks 0.11.3 → 0.11.4

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.cjs CHANGED
@@ -1617,7 +1617,10 @@ var data_default3 = taskData3;
1617
1617
  var taskData4 = {
1618
1618
  datasets: [
1619
1619
  {
1620
- // TODO write proper description
1620
+ description: "Largest document understanding dataset.",
1621
+ id: "HuggingFaceM4/Docmatix"
1622
+ },
1623
+ {
1621
1624
  description: "Dataset from the 2020 DocVQA challenge. The documents are taken from the UCSF Industry Documents Library.",
1622
1625
  id: "eliolio/docvqa"
1623
1626
  }
@@ -1658,8 +1661,8 @@ var taskData4 = {
1658
1661
  id: "impira/layoutlm-document-qa"
1659
1662
  },
1660
1663
  {
1661
- description: "A special model for OCR-free Document QA task. Donut model fine-tuned on DocVQA.",
1662
- id: "naver-clova-ix/donut-base-finetuned-docvqa"
1664
+ description: "A special model for OCR-free Document QA task.",
1665
+ id: "microsoft/udop-large"
1663
1666
  },
1664
1667
  {
1665
1668
  description: "A powerful model for document question answering.",
@@ -1939,6 +1942,10 @@ var taskData8 = {
1939
1942
  {
1940
1943
  description: "A robust image feature extraction models.",
1941
1944
  id: "facebook/dino-vitb16"
1945
+ },
1946
+ {
1947
+ description: "Strong image-text-to-text model made for information retrieval from documents.",
1948
+ id: "vidore/colpali"
1942
1949
  }
1943
1950
  ],
1944
1951
  spaces: [],
@@ -2165,8 +2172,8 @@ var taskData11 = {
2165
2172
  metrics: [],
2166
2173
  models: [
2167
2174
  {
2168
- description: "Cutting-edge vision language model that can also localize texts in images.",
2169
- id: "liuhaotian/llava-v1.6-34b"
2175
+ description: "Cutting-edge vision language model that can take multiple image inputs.",
2176
+ id: "facebook/chameleon-7b"
2170
2177
  },
2171
2178
  {
2172
2179
  description: "Cutting-edge conversational vision language model that can take multiple image inputs.",
@@ -2179,6 +2186,10 @@ var taskData11 = {
2179
2186
  {
2180
2187
  description: "Strong image-text-to-text model made to understand documents.",
2181
2188
  id: "mPLUG/DocOwl1.5"
2189
+ },
2190
+ {
2191
+ description: "Strong image-text-to-text model.",
2192
+ id: "llava-hf/llava-v1.6-mistral-7b-hf"
2182
2193
  }
2183
2194
  ],
2184
2195
  spaces: [
@@ -6626,6 +6637,12 @@ var DATASET_LIBRARIES_UI_ELEMENTS = {
6626
6637
  repoName: "argilla",
6627
6638
  repoUrl: "https://github.com/argilla-io/argilla",
6628
6639
  docsUrl: "https://argilla-io.github.io/argilla"
6640
+ },
6641
+ polars: {
6642
+ prettyLabel: "Polars",
6643
+ repoName: "polars",
6644
+ repoUrl: "https://github.com/pola-rs/polars",
6645
+ docsUrl: "https://docs.pola.rs/"
6629
6646
  }
6630
6647
  };
6631
6648
  // Annotate the CommonJS export names for ESM import in node:
package/dist/index.js CHANGED
@@ -1579,7 +1579,10 @@ var data_default3 = taskData3;
1579
1579
  var taskData4 = {
1580
1580
  datasets: [
1581
1581
  {
1582
- // TODO write proper description
1582
+ description: "Largest document understanding dataset.",
1583
+ id: "HuggingFaceM4/Docmatix"
1584
+ },
1585
+ {
1583
1586
  description: "Dataset from the 2020 DocVQA challenge. The documents are taken from the UCSF Industry Documents Library.",
1584
1587
  id: "eliolio/docvqa"
1585
1588
  }
@@ -1620,8 +1623,8 @@ var taskData4 = {
1620
1623
  id: "impira/layoutlm-document-qa"
1621
1624
  },
1622
1625
  {
1623
- description: "A special model for OCR-free Document QA task. Donut model fine-tuned on DocVQA.",
1624
- id: "naver-clova-ix/donut-base-finetuned-docvqa"
1626
+ description: "A special model for OCR-free Document QA task.",
1627
+ id: "microsoft/udop-large"
1625
1628
  },
1626
1629
  {
1627
1630
  description: "A powerful model for document question answering.",
@@ -1901,6 +1904,10 @@ var taskData8 = {
1901
1904
  {
1902
1905
  description: "A robust image feature extraction models.",
1903
1906
  id: "facebook/dino-vitb16"
1907
+ },
1908
+ {
1909
+ description: "Strong image-text-to-text model made for information retrieval from documents.",
1910
+ id: "vidore/colpali"
1904
1911
  }
1905
1912
  ],
1906
1913
  spaces: [],
@@ -2127,8 +2134,8 @@ var taskData11 = {
2127
2134
  metrics: [],
2128
2135
  models: [
2129
2136
  {
2130
- description: "Cutting-edge vision language model that can also localize texts in images.",
2131
- id: "liuhaotian/llava-v1.6-34b"
2137
+ description: "Cutting-edge vision language model that can take multiple image inputs.",
2138
+ id: "facebook/chameleon-7b"
2132
2139
  },
2133
2140
  {
2134
2141
  description: "Cutting-edge conversational vision language model that can take multiple image inputs.",
@@ -2141,6 +2148,10 @@ var taskData11 = {
2141
2148
  {
2142
2149
  description: "Strong image-text-to-text model made to understand documents.",
2143
2150
  id: "mPLUG/DocOwl1.5"
2151
+ },
2152
+ {
2153
+ description: "Strong image-text-to-text model.",
2154
+ id: "llava-hf/llava-v1.6-mistral-7b-hf"
2144
2155
  }
2145
2156
  ],
2146
2157
  spaces: [
@@ -6588,6 +6599,12 @@ var DATASET_LIBRARIES_UI_ELEMENTS = {
6588
6599
  repoName: "argilla",
6589
6600
  repoUrl: "https://github.com/argilla-io/argilla",
6590
6601
  docsUrl: "https://argilla-io.github.io/argilla"
6602
+ },
6603
+ polars: {
6604
+ prettyLabel: "Polars",
6605
+ repoName: "polars",
6606
+ repoUrl: "https://github.com/pola-rs/polars",
6607
+ docsUrl: "https://docs.pola.rs/"
6591
6608
  }
6592
6609
  };
6593
6610
  export {
@@ -70,6 +70,12 @@ export declare const DATASET_LIBRARIES_UI_ELEMENTS: {
70
70
  repoUrl: string;
71
71
  docsUrl: string;
72
72
  };
73
+ polars: {
74
+ prettyLabel: string;
75
+ repoName: string;
76
+ repoUrl: string;
77
+ docsUrl: string;
78
+ };
73
79
  };
74
80
  export type DatasetLibraryKey = keyof typeof DATASET_LIBRARIES_UI_ELEMENTS;
75
81
  //# sourceMappingURL=dataset-libraries.d.ts.map
@@ -1 +1 @@
1
- {"version":3,"file":"dataset-libraries.d.ts","sourceRoot":"","sources":["../../src/dataset-libraries.ts"],"names":[],"mappings":"AAAA;;GAEG;AACH,MAAM,WAAW,uBAAuB;IACvC;;;;OAIG;IACH,WAAW,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,QAAQ,EAAE,MAAM,CAAC;IACjB;;OAEG;IACH,OAAO,EAAE,MAAM,CAAC;IAChB;;OAEG;IACH,OAAO,CAAC,EAAE,MAAM,CAAC;CACjB;AAED,eAAO,MAAM,6BAA6B;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CAiDS,CAAC;AAGpD,MAAM,MAAM,iBAAiB,GAAG,MAAM,OAAO,6BAA6B,CAAC"}
1
+ {"version":3,"file":"dataset-libraries.d.ts","sourceRoot":"","sources":["../../src/dataset-libraries.ts"],"names":[],"mappings":"AAAA;;GAEG;AACH,MAAM,WAAW,uBAAuB;IACvC;;;;OAIG;IACH,WAAW,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,QAAQ,EAAE,MAAM,CAAC;IACjB;;OAEG;IACH,OAAO,EAAE,MAAM,CAAC;IAChB;;OAEG;IACH,OAAO,CAAC,EAAE,MAAM,CAAC;CACjB;AAED,eAAO,MAAM,6BAA6B;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CAuDS,CAAC;AAGpD,MAAM,MAAM,iBAAiB,GAAG,MAAM,OAAO,6BAA6B,CAAC"}
@@ -1 +1 @@
1
- {"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/document-question-answering/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,IAAI,CAAC;AAEzC,QAAA,MAAM,QAAQ,EAAE,cAyEf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
1
+ {"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/document-question-answering/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,IAAI,CAAC;AAEzC,QAAA,MAAM,QAAQ,EAAE,cA4Ef,CAAC;AAEF,eAAe,QAAQ,CAAC"}
@@ -1 +1 @@
1
- {"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/image-feature-extraction/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,IAAI,CAAC;AAEzC,QAAA,MAAM,QAAQ,EAAE,cA8Cf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
1
+ {"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/image-feature-extraction/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,IAAI,CAAC;AAEzC,QAAA,MAAM,QAAQ,EAAE,cAkDf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
@@ -1 +1 @@
1
- {"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/image-text-to-text/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,IAAI,CAAC;AAEzC,QAAA,MAAM,QAAQ,EAAE,cAqFf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
1
+ {"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/image-text-to-text/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,IAAI,CAAC;AAEzC,QAAA,MAAM,QAAQ,EAAE,cAyFf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
package/package.json CHANGED
@@ -1,7 +1,7 @@
1
1
  {
2
2
  "name": "@huggingface/tasks",
3
3
  "packageManager": "pnpm@8.10.5",
4
- "version": "0.11.3",
4
+ "version": "0.11.4",
5
5
  "description": "List of ML tasks for huggingface.co/tasks",
6
6
  "repository": "https://github.com/huggingface/huggingface.js.git",
7
7
  "publishConfig": {
@@ -71,6 +71,12 @@ export const DATASET_LIBRARIES_UI_ELEMENTS = {
71
71
  repoUrl: "https://github.com/argilla-io/argilla",
72
72
  docsUrl: "https://argilla-io.github.io/argilla",
73
73
  },
74
+ polars: {
75
+ prettyLabel: "Polars",
76
+ repoName: "polars",
77
+ repoUrl: "https://github.com/pola-rs/polars",
78
+ docsUrl: "https://docs.pola.rs/",
79
+ },
74
80
  } satisfies Record<string, DatasetLibraryUiElement>;
75
81
 
76
82
  /// List of the dataset libraries supported by the Hub
@@ -3,7 +3,10 @@ import type { TaskDataCustom } from "..";
3
3
  const taskData: TaskDataCustom = {
4
4
  datasets: [
5
5
  {
6
- // TODO write proper description
6
+ description: "Largest document understanding dataset.",
7
+ id: "HuggingFaceM4/Docmatix",
8
+ },
9
+ {
7
10
  description:
8
11
  "Dataset from the 2020 DocVQA challenge. The documents are taken from the UCSF Industry Documents Library.",
9
12
  id: "eliolio/docvqa",
@@ -47,8 +50,8 @@ const taskData: TaskDataCustom = {
47
50
  id: "impira/layoutlm-document-qa",
48
51
  },
49
52
  {
50
- description: "A special model for OCR-free Document QA task. Donut model fine-tuned on DocVQA.",
51
- id: "naver-clova-ix/donut-base-finetuned-docvqa",
53
+ description: "A special model for OCR-free Document QA task.",
54
+ id: "microsoft/udop-large",
52
55
  },
53
56
  {
54
57
  description: "A powerful model for document question answering.",
@@ -42,6 +42,10 @@ const taskData: TaskDataCustom = {
42
42
  description: "A robust image feature extraction models.",
43
43
  id: "facebook/dino-vitb16",
44
44
  },
45
+ {
46
+ description: "Strong image-text-to-text model made for information retrieval from documents.",
47
+ id: "vidore/colpali",
48
+ },
45
49
  ],
46
50
  spaces: [],
47
51
  summary: "Image feature extraction is the task of extracting features learnt in a computer vision model.",
@@ -43,8 +43,8 @@ const taskData: TaskDataCustom = {
43
43
  metrics: [],
44
44
  models: [
45
45
  {
46
- description: "Cutting-edge vision language model that can also localize texts in images.",
47
- id: "liuhaotian/llava-v1.6-34b",
46
+ description: "Cutting-edge vision language model that can take multiple image inputs.",
47
+ id: "facebook/chameleon-7b",
48
48
  },
49
49
  {
50
50
  description: "Cutting-edge conversational vision language model that can take multiple image inputs.",
@@ -58,6 +58,10 @@ const taskData: TaskDataCustom = {
58
58
  description: "Strong image-text-to-text model made to understand documents.",
59
59
  id: "mPLUG/DocOwl1.5",
60
60
  },
61
+ {
62
+ description: "Strong image-text-to-text model.",
63
+ id: "llava-hf/llava-v1.6-mistral-7b-hf",
64
+ },
61
65
  ],
62
66
  spaces: [
63
67
  {