@huggingface/tasks 0.10.5 → 0.10.7
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.cjs +38 -8
- package/dist/index.js +38 -8
- package/dist/src/hardware.d.ts +21 -1
- package/dist/src/hardware.d.ts.map +1 -1
- package/dist/src/library-to-tasks.d.ts.map +1 -1
- package/dist/src/local-apps.d.ts +2 -2
- package/dist/src/local-apps.d.ts.map +1 -1
- package/dist/src/model-libraries-snippets.d.ts.map +1 -1
- package/dist/src/model-libraries.d.ts +2 -2
- package/package.json +1 -1
- package/src/hardware.ts +24 -4
- package/src/library-to-tasks.ts +1 -0
- package/src/local-apps.ts +13 -4
- package/src/model-libraries-snippets.ts +2 -1
package/dist/index.cjs
CHANGED
|
@@ -104,7 +104,8 @@ var LIBRARY_TASK_MAPPING = {
|
|
|
104
104
|
"video-classification",
|
|
105
105
|
"visual-question-answering",
|
|
106
106
|
"zero-shot-classification",
|
|
107
|
-
"zero-shot-image-classification"
|
|
107
|
+
"zero-shot-image-classification",
|
|
108
|
+
"zero-shot-object-detection"
|
|
108
109
|
],
|
|
109
110
|
mindspore: ["image-classification"]
|
|
110
111
|
};
|
|
@@ -4424,7 +4425,7 @@ var transformers = (model) => {
|
|
|
4424
4425
|
`model = ${info.auto_model}.from_pretrained("${model.id}"` + remote_code_snippet + ")"
|
|
4425
4426
|
].join("\n");
|
|
4426
4427
|
}
|
|
4427
|
-
if (model.pipeline_tag) {
|
|
4428
|
+
if (model.pipeline_tag && LIBRARY_TASK_MAPPING.transformers?.includes(model.pipeline_tag)) {
|
|
4428
4429
|
const pipelineSnippet = [
|
|
4429
4430
|
"# Use a pipeline as a high-level helper",
|
|
4430
4431
|
"from transformers import pipeline",
|
|
@@ -5459,7 +5460,7 @@ var TFLOPS_THRESHOLD_WHITE_HOUSE_MODEL_TRAINING_TOTAL = 10 ** 14;
|
|
|
5459
5460
|
var TFLOPS_THRESHOLD_WHITE_HOUSE_MODEL_TRAINING_TOTAL_BIOLOGY = 10 ** 11;
|
|
5460
5461
|
var TFLOPS_THRESHOLD_WHITE_HOUSE_CLUSTER = 10 ** 8;
|
|
5461
5462
|
var TFLOPS_THRESHOLD_EU_AI_ACT_MODEL_TRAINING_TOTAL = 10 ** 13;
|
|
5462
|
-
var DEFAULT_MEMORY_OPTIONS = [8, 16, 24, 32, 40, 48, 64, 80, 96, 128];
|
|
5463
|
+
var DEFAULT_MEMORY_OPTIONS = [8, 16, 24, 32, 40, 48, 64, 80, 96, 128, 256, 512];
|
|
5463
5464
|
var SKUS = {
|
|
5464
5465
|
GPU: {
|
|
5465
5466
|
NVIDIA: {
|
|
@@ -5527,13 +5528,33 @@ var SKUS = {
|
|
|
5527
5528
|
tflops: 40,
|
|
5528
5529
|
memory: [24]
|
|
5529
5530
|
},
|
|
5531
|
+
"RTX 3080": {
|
|
5532
|
+
tflops: 30.6,
|
|
5533
|
+
memory: [12, 10]
|
|
5534
|
+
},
|
|
5530
5535
|
"RTX 3080 Ti": {
|
|
5531
5536
|
tflops: 34.1,
|
|
5532
5537
|
memory: [12]
|
|
5533
5538
|
},
|
|
5534
|
-
"RTX
|
|
5535
|
-
tflops:
|
|
5536
|
-
memory: [
|
|
5539
|
+
"RTX 3070": {
|
|
5540
|
+
tflops: 20.31,
|
|
5541
|
+
memory: [8]
|
|
5542
|
+
},
|
|
5543
|
+
"RTX 3070 Ti": {
|
|
5544
|
+
tflops: 21.75,
|
|
5545
|
+
memory: [8]
|
|
5546
|
+
},
|
|
5547
|
+
"RTX 3070 Ti Laptop": {
|
|
5548
|
+
tflops: 16.6,
|
|
5549
|
+
memory: [8]
|
|
5550
|
+
},
|
|
5551
|
+
"RTX 2060 Mobile": {
|
|
5552
|
+
tflops: 9.22,
|
|
5553
|
+
memory: [6]
|
|
5554
|
+
},
|
|
5555
|
+
"RTX Titan": {
|
|
5556
|
+
tflops: 32.62,
|
|
5557
|
+
memory: [24]
|
|
5537
5558
|
}
|
|
5538
5559
|
},
|
|
5539
5560
|
AMD: {
|
|
@@ -5738,11 +5759,20 @@ function isGgufModel(model) {
|
|
|
5738
5759
|
return model.tags.includes("gguf");
|
|
5739
5760
|
}
|
|
5740
5761
|
var snippetLlamacpp = (model) => {
|
|
5741
|
-
return
|
|
5762
|
+
return [
|
|
5763
|
+
`
|
|
5764
|
+
## Install and build llama.cpp with curl support
|
|
5765
|
+
git clone https://github.com/ggerganov/llama.cpp.git
|
|
5766
|
+
cd llama.cpp
|
|
5767
|
+
LLAMA_CURL=1 make
|
|
5768
|
+
`,
|
|
5769
|
+
`## Load and run the model
|
|
5770
|
+
./main \\
|
|
5742
5771
|
--hf-repo "${model.id}" \\
|
|
5743
5772
|
-m file.gguf \\
|
|
5744
5773
|
-p "I believe the meaning of life is" \\
|
|
5745
|
-
-n 128
|
|
5774
|
+
-n 128`
|
|
5775
|
+
];
|
|
5746
5776
|
};
|
|
5747
5777
|
var LOCAL_APPS = {
|
|
5748
5778
|
"llama.cpp": {
|
package/dist/index.js
CHANGED
|
@@ -66,7 +66,8 @@ var LIBRARY_TASK_MAPPING = {
|
|
|
66
66
|
"video-classification",
|
|
67
67
|
"visual-question-answering",
|
|
68
68
|
"zero-shot-classification",
|
|
69
|
-
"zero-shot-image-classification"
|
|
69
|
+
"zero-shot-image-classification",
|
|
70
|
+
"zero-shot-object-detection"
|
|
70
71
|
],
|
|
71
72
|
mindspore: ["image-classification"]
|
|
72
73
|
};
|
|
@@ -4386,7 +4387,7 @@ var transformers = (model) => {
|
|
|
4386
4387
|
`model = ${info.auto_model}.from_pretrained("${model.id}"` + remote_code_snippet + ")"
|
|
4387
4388
|
].join("\n");
|
|
4388
4389
|
}
|
|
4389
|
-
if (model.pipeline_tag) {
|
|
4390
|
+
if (model.pipeline_tag && LIBRARY_TASK_MAPPING.transformers?.includes(model.pipeline_tag)) {
|
|
4390
4391
|
const pipelineSnippet = [
|
|
4391
4392
|
"# Use a pipeline as a high-level helper",
|
|
4392
4393
|
"from transformers import pipeline",
|
|
@@ -5421,7 +5422,7 @@ var TFLOPS_THRESHOLD_WHITE_HOUSE_MODEL_TRAINING_TOTAL = 10 ** 14;
|
|
|
5421
5422
|
var TFLOPS_THRESHOLD_WHITE_HOUSE_MODEL_TRAINING_TOTAL_BIOLOGY = 10 ** 11;
|
|
5422
5423
|
var TFLOPS_THRESHOLD_WHITE_HOUSE_CLUSTER = 10 ** 8;
|
|
5423
5424
|
var TFLOPS_THRESHOLD_EU_AI_ACT_MODEL_TRAINING_TOTAL = 10 ** 13;
|
|
5424
|
-
var DEFAULT_MEMORY_OPTIONS = [8, 16, 24, 32, 40, 48, 64, 80, 96, 128];
|
|
5425
|
+
var DEFAULT_MEMORY_OPTIONS = [8, 16, 24, 32, 40, 48, 64, 80, 96, 128, 256, 512];
|
|
5425
5426
|
var SKUS = {
|
|
5426
5427
|
GPU: {
|
|
5427
5428
|
NVIDIA: {
|
|
@@ -5489,13 +5490,33 @@ var SKUS = {
|
|
|
5489
5490
|
tflops: 40,
|
|
5490
5491
|
memory: [24]
|
|
5491
5492
|
},
|
|
5493
|
+
"RTX 3080": {
|
|
5494
|
+
tflops: 30.6,
|
|
5495
|
+
memory: [12, 10]
|
|
5496
|
+
},
|
|
5492
5497
|
"RTX 3080 Ti": {
|
|
5493
5498
|
tflops: 34.1,
|
|
5494
5499
|
memory: [12]
|
|
5495
5500
|
},
|
|
5496
|
-
"RTX
|
|
5497
|
-
tflops:
|
|
5498
|
-
memory: [
|
|
5501
|
+
"RTX 3070": {
|
|
5502
|
+
tflops: 20.31,
|
|
5503
|
+
memory: [8]
|
|
5504
|
+
},
|
|
5505
|
+
"RTX 3070 Ti": {
|
|
5506
|
+
tflops: 21.75,
|
|
5507
|
+
memory: [8]
|
|
5508
|
+
},
|
|
5509
|
+
"RTX 3070 Ti Laptop": {
|
|
5510
|
+
tflops: 16.6,
|
|
5511
|
+
memory: [8]
|
|
5512
|
+
},
|
|
5513
|
+
"RTX 2060 Mobile": {
|
|
5514
|
+
tflops: 9.22,
|
|
5515
|
+
memory: [6]
|
|
5516
|
+
},
|
|
5517
|
+
"RTX Titan": {
|
|
5518
|
+
tflops: 32.62,
|
|
5519
|
+
memory: [24]
|
|
5499
5520
|
}
|
|
5500
5521
|
},
|
|
5501
5522
|
AMD: {
|
|
@@ -5700,11 +5721,20 @@ function isGgufModel(model) {
|
|
|
5700
5721
|
return model.tags.includes("gguf");
|
|
5701
5722
|
}
|
|
5702
5723
|
var snippetLlamacpp = (model) => {
|
|
5703
|
-
return
|
|
5724
|
+
return [
|
|
5725
|
+
`
|
|
5726
|
+
## Install and build llama.cpp with curl support
|
|
5727
|
+
git clone https://github.com/ggerganov/llama.cpp.git
|
|
5728
|
+
cd llama.cpp
|
|
5729
|
+
LLAMA_CURL=1 make
|
|
5730
|
+
`,
|
|
5731
|
+
`## Load and run the model
|
|
5732
|
+
./main \\
|
|
5704
5733
|
--hf-repo "${model.id}" \\
|
|
5705
5734
|
-m file.gguf \\
|
|
5706
5735
|
-p "I believe the meaning of life is" \\
|
|
5707
|
-
-n 128
|
|
5736
|
+
-n 128`
|
|
5737
|
+
];
|
|
5708
5738
|
};
|
|
5709
5739
|
var LOCAL_APPS = {
|
|
5710
5740
|
"llama.cpp": {
|
package/dist/src/hardware.d.ts
CHANGED
|
@@ -96,11 +96,31 @@ export declare const SKUS: {
|
|
|
96
96
|
tflops: number;
|
|
97
97
|
memory: number[];
|
|
98
98
|
};
|
|
99
|
+
"RTX 3080": {
|
|
100
|
+
tflops: number;
|
|
101
|
+
memory: number[];
|
|
102
|
+
};
|
|
99
103
|
"RTX 3080 Ti": {
|
|
100
104
|
tflops: number;
|
|
101
105
|
memory: number[];
|
|
102
106
|
};
|
|
103
|
-
"RTX
|
|
107
|
+
"RTX 3070": {
|
|
108
|
+
tflops: number;
|
|
109
|
+
memory: number[];
|
|
110
|
+
};
|
|
111
|
+
"RTX 3070 Ti": {
|
|
112
|
+
tflops: number;
|
|
113
|
+
memory: number[];
|
|
114
|
+
};
|
|
115
|
+
"RTX 3070 Ti Laptop": {
|
|
116
|
+
tflops: number;
|
|
117
|
+
memory: number[];
|
|
118
|
+
};
|
|
119
|
+
"RTX 2060 Mobile": {
|
|
120
|
+
tflops: number;
|
|
121
|
+
memory: number[];
|
|
122
|
+
};
|
|
123
|
+
"RTX Titan": {
|
|
104
124
|
tflops: number;
|
|
105
125
|
memory: number[];
|
|
106
126
|
};
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"hardware.d.ts","sourceRoot":"","sources":["../../src/hardware.ts"],"names":[],"mappings":"AAAA;;;GAGG;AACH,eAAO,MAAM,iDAAiD,QAAW,CAAC;AAC1E,eAAO,MAAM,yDAAyD,QAAW,CAAC;AAClF,eAAO,MAAM,oCAAoC,QAAU,CAAC;AAE5D;;;GAGG;AACH,eAAO,MAAM,+CAA+C,QAAW,CAAC;AAExE,MAAM,WAAW,YAAY;IAC5B;;;;;;;;;OASG;IACH,MAAM,EAAE,MAAM,CAAC;IACf;;;OAGG;IACH,MAAM,CAAC,EAAE,MAAM,EAAE,CAAC;CAClB;AAED,eAAO,MAAM,sBAAsB,
|
|
1
|
+
{"version":3,"file":"hardware.d.ts","sourceRoot":"","sources":["../../src/hardware.ts"],"names":[],"mappings":"AAAA;;;GAGG;AACH,eAAO,MAAM,iDAAiD,QAAW,CAAC;AAC1E,eAAO,MAAM,yDAAyD,QAAW,CAAC;AAClF,eAAO,MAAM,oCAAoC,QAAU,CAAC;AAE5D;;;GAGG;AACH,eAAO,MAAM,+CAA+C,QAAW,CAAC;AAExE,MAAM,WAAW,YAAY;IAC5B;;;;;;;;;OASG;IACH,MAAM,EAAE,MAAM,CAAC;IACf;;;OAGG;IACH,MAAM,CAAC,EAAE,MAAM,EAAE,CAAC;CAClB;AAED,eAAO,MAAM,sBAAsB,UAAqD,CAAC;AAEzF,eAAO,MAAM,IAAI;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CAmSuD,CAAC;AAEzE,MAAM,MAAM,OAAO,GAAG,MAAM,OAAO,IAAI,CAAC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"library-to-tasks.d.ts","sourceRoot":"","sources":["../../src/library-to-tasks.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,eAAe,EAAE,MAAM,mBAAmB,CAAC;AACzD,OAAO,KAAK,EAAE,YAAY,EAAE,MAAM,aAAa,CAAC;AAEhD;;;;;;GAMG;AACH,eAAO,MAAM,oBAAoB,EAAE,OAAO,CAAC,MAAM,CAAC,eAAe,EAAE,YAAY,EAAE,CAAC,
|
|
1
|
+
{"version":3,"file":"library-to-tasks.d.ts","sourceRoot":"","sources":["../../src/library-to-tasks.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,eAAe,EAAE,MAAM,mBAAmB,CAAC;AACzD,OAAO,KAAK,EAAE,YAAY,EAAE,MAAM,aAAa,CAAC;AAEhD;;;;;;GAMG;AACH,eAAO,MAAM,oBAAoB,EAAE,OAAO,CAAC,MAAM,CAAC,eAAe,EAAE,YAAY,EAAE,CAAC,CAiEjF,CAAC"}
|
package/dist/src/local-apps.d.ts
CHANGED
|
@@ -34,7 +34,7 @@ export type LocalApp = {
|
|
|
34
34
|
/**
|
|
35
35
|
* And if not (mostly llama.cpp), snippet to copy/paste in your terminal
|
|
36
36
|
*/
|
|
37
|
-
snippet: (model: ModelData) => string;
|
|
37
|
+
snippet: (model: ModelData) => string | string[];
|
|
38
38
|
});
|
|
39
39
|
declare function isGgufModel(model: ModelData): boolean;
|
|
40
40
|
/**
|
|
@@ -54,7 +54,7 @@ export declare const LOCAL_APPS: {
|
|
|
54
54
|
docsUrl: string;
|
|
55
55
|
mainTask: "text-generation";
|
|
56
56
|
displayOnModelPage: typeof isGgufModel;
|
|
57
|
-
snippet: (model: ModelData) => string;
|
|
57
|
+
snippet: (model: ModelData) => string[];
|
|
58
58
|
};
|
|
59
59
|
lmstudio: {
|
|
60
60
|
prettyLabel: string;
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"local-apps.d.ts","sourceRoot":"","sources":["../../src/local-apps.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,cAAc,CAAC;AAC9C,OAAO,KAAK,EAAE,YAAY,EAAE,MAAM,aAAa,CAAC;AAEhD;;GAEG;AACH,MAAM,MAAM,QAAQ,GAAG;IACtB;;OAEG;IACH,WAAW,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,OAAO,EAAE,MAAM,CAAC;IAChB;;OAEG;IACH,QAAQ,EAAE,YAAY,CAAC;IACvB;;OAEG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IAEpB,UAAU,CAAC,EAAE,OAAO,CAAC;IACrB;;OAEG;IACH,kBAAkB,EAAE,CAAC,KAAK,EAAE,SAAS,KAAK,OAAO,CAAC;CAClD,GAAG,CACD;IACA;;OAEG;IACH,QAAQ,EAAE,CAAC,KAAK,EAAE,SAAS,KAAK,GAAG,CAAC;CACnC,GACD;IACA;;OAEG;IACH,OAAO,EAAE,CAAC,KAAK,EAAE,SAAS,KAAK,MAAM,CAAC;
|
|
1
|
+
{"version":3,"file":"local-apps.d.ts","sourceRoot":"","sources":["../../src/local-apps.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,cAAc,CAAC;AAC9C,OAAO,KAAK,EAAE,YAAY,EAAE,MAAM,aAAa,CAAC;AAEhD;;GAEG;AACH,MAAM,MAAM,QAAQ,GAAG;IACtB;;OAEG;IACH,WAAW,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,OAAO,EAAE,MAAM,CAAC;IAChB;;OAEG;IACH,QAAQ,EAAE,YAAY,CAAC;IACvB;;OAEG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IAEpB,UAAU,CAAC,EAAE,OAAO,CAAC;IACrB;;OAEG;IACH,kBAAkB,EAAE,CAAC,KAAK,EAAE,SAAS,KAAK,OAAO,CAAC;CAClD,GAAG,CACD;IACA;;OAEG;IACH,QAAQ,EAAE,CAAC,KAAK,EAAE,SAAS,KAAK,GAAG,CAAC;CACnC,GACD;IACA;;OAEG;IACH,OAAO,EAAE,CAAC,KAAK,EAAE,SAAS,KAAK,MAAM,GAAG,MAAM,EAAE,CAAC;CAChD,CACH,CAAC;AAEF,iBAAS,WAAW,CAAC,KAAK,EAAE,SAAS,WAEpC;AAmBD;;;;;;;;;;GAUG;AACH,eAAO,MAAM,UAAU;;;;;;yBA5BS,SAAS,KAAG,MAAM,EAAE;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CAiFhB,CAAC;AAErC,MAAM,MAAM,WAAW,GAAG,MAAM,OAAO,UAAU,CAAC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"model-libraries-snippets.d.ts","sourceRoot":"","sources":["../../src/model-libraries-snippets.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,cAAc,CAAC;
|
|
1
|
+
{"version":3,"file":"model-libraries-snippets.d.ts","sourceRoot":"","sources":["../../src/model-libraries-snippets.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,cAAc,CAAC;AAY9C,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAKjD,CAAC;AAkBF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAKjD,CAAC;AAEF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAIjD,CAAC;AAMF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAIjD,CAAC;AA+BF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAUlD,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAMlD,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EASlD,CAAC;AAIF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAO/C,CAAC;AAEF,eAAO,MAAM,OAAO,UAAW,SAAS,KAAG,MAAM,EAMhD,CAAC;AAEF,eAAO,MAAM,KAAK,UAAW,SAAS,KAAG,MAAM,EAI9C,CAAC;AAEF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAI/C,CAAC;AAEF,eAAO,MAAM,KAAK,UAAW,SAAS,KAAG,MAAM,EAK9C,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EASlD,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAKlD,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAsBlD,CAAC;AAEF,eAAO,MAAM,uBAAuB,UAAW,SAAS,KAAG,MAAM,EAehE,CAAC;AAiBF,eAAO,MAAM,cAAc,UAAW,SAAS,KAAG,MAAM,EAKvD,CAAC;AAyBF,eAAO,MAAM,aAAa,UAAW,SAAS,KAAG,MAAM,EAOtD,CAAC;AAEF,eAAO,MAAM,IAAI,UAAW,SAAS,KAAG,MAAM,EAI7C,CAAC;AAsCF,eAAO,MAAM,OAAO,UAAW,SAAS,KAAG,MAAM,EAehD,CAAC;AAEF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAI/C,CAAC;AAEF,eAAO,MAAM,aAAa,UAAW,SAAS,KAAG,MAAM,EAEtD,CAAC;AAEF,eAAO,MAAM,oBAAoB,UAAW,SAAS,KAAG,MAAM,EAI7D,CAAC;AAEF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAI/C,CAAC;AAEF,eAAO,MAAM,KAAK,UAAW,SAAS,KAAG,MAAM,EAU9C,CAAC;AAEF,eAAO,MAAM,WAAW,UAAW,SAAS,KAAG,MAAM,EAIpD,CAAC;AAEF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAK/C,CAAC;AAkBF,eAAO,MAAM,WAAW,UAAW,SAAS,KAAG,MAAM,EAkBpD,CAAC;AAEF,eAAO,MAAM,YAAY,UAAW,SAAS,KAAG,MAAM,EAwCrD,CAAC;AAEF,eAAO,MAAM,cAAc,UAAW,SAAS,KAAG,MAAM,EAcvD,CAAC;AAiBF,eAAO,MAAM,IAAI,UAAW,SAAS,KAAG,MAAM,EAkB7C,CAAC;AAEF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAKjD,CAAC;AAEF,eAAO,MAAM,gBAAgB,UAAW,SAAS,KAAG,MAAM,EAMzD,CAAC;AAgBF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAEjD,CAAC;AAEF,eAAO,MAAM,MAAM,QAA6B,MAAM,EAMrD,CAAC;AAEF,eAAO,MAAM,UAAU,UAAW,SAAS,KAAG,MAAM,EAInD,CAAC;AAEF,eAAO,MAAM,GAAG,UAAW,SAAS,KAAG,MAAM,EAK5C,CAAC;AAEF,eAAO,MAAM,KAAK,UAAW,SAAS,KAAG,MAAM,EAI9C,CAAC;AAEF,eAAO,MAAM,IAAI,UAAW,SAAS,KAAG,MAAM,EAQ7C,CAAC;AAEF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAI/C,CAAC;AA6BF,eAAO,MAAM,UAAU,UAAW,SAAS,KAAG,MAAM,EAUnD,CAAC"}
|
|
@@ -507,6 +507,6 @@ export declare const MODEL_LIBRARIES_UI_ELEMENTS: {
|
|
|
507
507
|
};
|
|
508
508
|
};
|
|
509
509
|
export type ModelLibraryKey = keyof typeof MODEL_LIBRARIES_UI_ELEMENTS;
|
|
510
|
-
export declare const ALL_MODEL_LIBRARY_KEYS: ("
|
|
511
|
-
export declare const ALL_DISPLAY_MODEL_LIBRARY_KEYS: ("
|
|
510
|
+
export declare const ALL_MODEL_LIBRARY_KEYS: ("adapter-transformers" | "allennlp" | "asteroid" | "audiocraft" | "bertopic" | "big_vision" | "diffusers" | "doctr" | "elm" | "espnet" | "fairseq" | "fastai" | "fasttext" | "flair" | "gliner" | "grok" | "keras" | "keras-nlp" | "k2" | "mindspore" | "ml-agents" | "mlx" | "mlx-image" | "nemo" | "open_clip" | "paddlenlp" | "peft" | "pyannote-audio" | "pythae" | "sample-factory" | "sentence-transformers" | "setfit" | "sklearn" | "spacy" | "span-marker" | "speechbrain" | "stable-baselines3" | "stanza" | "tensorflowtts" | "timesfm" | "timm" | "transformers" | "transformers.js" | "unity-sentis" | "voicecraft" | "whisperkit")[];
|
|
511
|
+
export declare const ALL_DISPLAY_MODEL_LIBRARY_KEYS: ("adapter-transformers" | "allennlp" | "asteroid" | "audiocraft" | "bertopic" | "big_vision" | "diffusers" | "doctr" | "elm" | "espnet" | "fairseq" | "fastai" | "fasttext" | "flair" | "gliner" | "grok" | "keras" | "keras-nlp" | "k2" | "mindspore" | "ml-agents" | "mlx" | "mlx-image" | "nemo" | "open_clip" | "paddlenlp" | "peft" | "pyannote-audio" | "pythae" | "sample-factory" | "sentence-transformers" | "setfit" | "sklearn" | "spacy" | "span-marker" | "speechbrain" | "stable-baselines3" | "stanza" | "tensorflowtts" | "timesfm" | "timm" | "transformers" | "transformers.js" | "unity-sentis" | "voicecraft" | "whisperkit")[];
|
|
512
512
|
//# sourceMappingURL=model-libraries.d.ts.map
|
package/package.json
CHANGED
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@huggingface/tasks",
|
|
3
3
|
"packageManager": "pnpm@8.10.5",
|
|
4
|
-
"version": "0.10.
|
|
4
|
+
"version": "0.10.7",
|
|
5
5
|
"description": "List of ML tasks for huggingface.co/tasks",
|
|
6
6
|
"repository": "https://github.com/huggingface/huggingface.js.git",
|
|
7
7
|
"publishConfig": {
|
package/src/hardware.ts
CHANGED
|
@@ -31,7 +31,7 @@ export interface HardwareSpec {
|
|
|
31
31
|
memory?: number[];
|
|
32
32
|
}
|
|
33
33
|
|
|
34
|
-
export const DEFAULT_MEMORY_OPTIONS = [8, 16, 24, 32, 40, 48, 64, 80, 96, 128];
|
|
34
|
+
export const DEFAULT_MEMORY_OPTIONS = [8, 16, 24, 32, 40, 48, 64, 80, 96, 128, 256, 512];
|
|
35
35
|
|
|
36
36
|
export const SKUS = {
|
|
37
37
|
GPU: {
|
|
@@ -100,13 +100,33 @@ export const SKUS = {
|
|
|
100
100
|
tflops: 40,
|
|
101
101
|
memory: [24],
|
|
102
102
|
},
|
|
103
|
+
"RTX 3080": {
|
|
104
|
+
tflops: 30.6,
|
|
105
|
+
memory: [12, 10],
|
|
106
|
+
},
|
|
103
107
|
"RTX 3080 Ti": {
|
|
104
108
|
tflops: 34.1,
|
|
105
109
|
memory: [12],
|
|
106
110
|
},
|
|
107
|
-
"RTX
|
|
108
|
-
tflops:
|
|
109
|
-
memory: [
|
|
111
|
+
"RTX 3070": {
|
|
112
|
+
tflops: 20.31,
|
|
113
|
+
memory: [8],
|
|
114
|
+
},
|
|
115
|
+
"RTX 3070 Ti": {
|
|
116
|
+
tflops: 21.75,
|
|
117
|
+
memory: [8],
|
|
118
|
+
},
|
|
119
|
+
"RTX 3070 Ti Laptop": {
|
|
120
|
+
tflops: 16.6,
|
|
121
|
+
memory: [8],
|
|
122
|
+
},
|
|
123
|
+
"RTX 2060 Mobile": {
|
|
124
|
+
tflops: 9.22,
|
|
125
|
+
memory: [6],
|
|
126
|
+
},
|
|
127
|
+
"RTX Titan": {
|
|
128
|
+
tflops: 32.62,
|
|
129
|
+
memory: [24],
|
|
110
130
|
},
|
|
111
131
|
},
|
|
112
132
|
AMD: {
|
package/src/library-to-tasks.ts
CHANGED
|
@@ -70,6 +70,7 @@ export const LIBRARY_TASK_MAPPING: Partial<Record<ModelLibraryKey, PipelineType[
|
|
|
70
70
|
"visual-question-answering",
|
|
71
71
|
"zero-shot-classification",
|
|
72
72
|
"zero-shot-image-classification",
|
|
73
|
+
"zero-shot-object-detection",
|
|
73
74
|
],
|
|
74
75
|
mindspore: ["image-classification"],
|
|
75
76
|
};
|
package/src/local-apps.ts
CHANGED
|
@@ -38,7 +38,7 @@ export type LocalApp = {
|
|
|
38
38
|
/**
|
|
39
39
|
* And if not (mostly llama.cpp), snippet to copy/paste in your terminal
|
|
40
40
|
*/
|
|
41
|
-
snippet: (model: ModelData) => string;
|
|
41
|
+
snippet: (model: ModelData) => string | string[];
|
|
42
42
|
}
|
|
43
43
|
);
|
|
44
44
|
|
|
@@ -46,12 +46,21 @@ function isGgufModel(model: ModelData) {
|
|
|
46
46
|
return model.tags.includes("gguf");
|
|
47
47
|
}
|
|
48
48
|
|
|
49
|
-
const snippetLlamacpp = (model: ModelData): string => {
|
|
50
|
-
return
|
|
49
|
+
const snippetLlamacpp = (model: ModelData): string[] => {
|
|
50
|
+
return [
|
|
51
|
+
`
|
|
52
|
+
## Install and build llama.cpp with curl support
|
|
53
|
+
git clone https://github.com/ggerganov/llama.cpp.git
|
|
54
|
+
cd llama.cpp
|
|
55
|
+
LLAMA_CURL=1 make
|
|
56
|
+
`,
|
|
57
|
+
`## Load and run the model
|
|
58
|
+
./main \\
|
|
51
59
|
--hf-repo "${model.id}" \\
|
|
52
60
|
-m file.gguf \\
|
|
53
61
|
-p "I believe the meaning of life is" \\
|
|
54
|
-
-n 128
|
|
62
|
+
-n 128`,
|
|
63
|
+
];
|
|
55
64
|
};
|
|
56
65
|
|
|
57
66
|
/**
|
|
@@ -1,4 +1,5 @@
|
|
|
1
1
|
import type { ModelData } from "./model-data";
|
|
2
|
+
import { LIBRARY_TASK_MAPPING } from "./library-to-tasks";
|
|
2
3
|
|
|
3
4
|
const TAG_CUSTOM_CODE = "custom_code";
|
|
4
5
|
|
|
@@ -438,7 +439,7 @@ export const transformers = (model: ModelData): string[] => {
|
|
|
438
439
|
].join("\n");
|
|
439
440
|
}
|
|
440
441
|
|
|
441
|
-
if (model.pipeline_tag) {
|
|
442
|
+
if (model.pipeline_tag && LIBRARY_TASK_MAPPING.transformers?.includes(model.pipeline_tag)) {
|
|
442
443
|
const pipelineSnippet = [
|
|
443
444
|
"# Use a pipeline as a high-level helper",
|
|
444
445
|
"from transformers import pipeline",
|