@huggingface/tasks 0.10.12 → 0.10.14
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.cjs +2 -2
- package/dist/index.js +2 -2
- package/dist/src/local-apps.d.ts +1 -0
- package/dist/src/local-apps.d.ts.map +1 -1
- package/package.json +1 -1
- package/src/local-apps.ts +3 -2
package/dist/index.cjs
CHANGED
|
@@ -5937,7 +5937,7 @@ brew install llama.cpp
|
|
|
5937
5937
|
# Load and run the model
|
|
5938
5938
|
llama \\
|
|
5939
5939
|
--hf-repo "${model.id}" \\
|
|
5940
|
-
--hf-file
|
|
5940
|
+
--hf-file {{GGUF_FILE}} \\
|
|
5941
5941
|
-p "I believe the meaning of life is" \\
|
|
5942
5942
|
-n 128`,
|
|
5943
5943
|
`# Option 2: build llama.cpp from source with curl support
|
|
@@ -5948,7 +5948,7 @@ LLAMA_CURL=1 make
|
|
|
5948
5948
|
# Load and run the model
|
|
5949
5949
|
./main \\
|
|
5950
5950
|
--hf-repo "${model.id}" \\
|
|
5951
|
-
-m
|
|
5951
|
+
-m {{GGUF_FILE}} \\
|
|
5952
5952
|
-p "I believe the meaning of life is" \\
|
|
5953
5953
|
-n 128`
|
|
5954
5954
|
];
|
package/dist/index.js
CHANGED
|
@@ -5898,7 +5898,7 @@ brew install llama.cpp
|
|
|
5898
5898
|
# Load and run the model
|
|
5899
5899
|
llama \\
|
|
5900
5900
|
--hf-repo "${model.id}" \\
|
|
5901
|
-
--hf-file
|
|
5901
|
+
--hf-file {{GGUF_FILE}} \\
|
|
5902
5902
|
-p "I believe the meaning of life is" \\
|
|
5903
5903
|
-n 128`,
|
|
5904
5904
|
`# Option 2: build llama.cpp from source with curl support
|
|
@@ -5909,7 +5909,7 @@ LLAMA_CURL=1 make
|
|
|
5909
5909
|
# Load and run the model
|
|
5910
5910
|
./main \\
|
|
5911
5911
|
--hf-repo "${model.id}" \\
|
|
5912
|
-
-m
|
|
5912
|
+
-m {{GGUF_FILE}} \\
|
|
5913
5913
|
-p "I believe the meaning of life is" \\
|
|
5914
5914
|
-n 128`
|
|
5915
5915
|
];
|
package/dist/src/local-apps.d.ts
CHANGED
|
@@ -33,6 +33,7 @@ export type LocalApp = {
|
|
|
33
33
|
} | {
|
|
34
34
|
/**
|
|
35
35
|
* And if not (mostly llama.cpp), snippet to copy/paste in your terminal
|
|
36
|
+
* Support the placeholder {{GGUF_FILE}} that will be replaced by the gguf file path or the list of available files.
|
|
36
37
|
*/
|
|
37
38
|
snippet: (model: ModelData) => string | string[];
|
|
38
39
|
});
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"local-apps.d.ts","sourceRoot":"","sources":["../../src/local-apps.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,cAAc,CAAC;AAC9C,OAAO,KAAK,EAAE,YAAY,EAAE,MAAM,aAAa,CAAC;AAEhD;;GAEG;AACH,MAAM,MAAM,QAAQ,GAAG;IACtB;;OAEG;IACH,WAAW,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,OAAO,EAAE,MAAM,CAAC;IAChB;;OAEG;IACH,QAAQ,EAAE,YAAY,CAAC;IACvB;;OAEG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IAEpB,UAAU,CAAC,EAAE,OAAO,CAAC;IACrB;;OAEG;IACH,kBAAkB,EAAE,CAAC,KAAK,EAAE,SAAS,KAAK,OAAO,CAAC;CAClD,GAAG,CACD;IACA;;OAEG;IACH,QAAQ,EAAE,CAAC,KAAK,EAAE,SAAS,KAAK,GAAG,CAAC;CACnC,GACD;IACA
|
|
1
|
+
{"version":3,"file":"local-apps.d.ts","sourceRoot":"","sources":["../../src/local-apps.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,cAAc,CAAC;AAC9C,OAAO,KAAK,EAAE,YAAY,EAAE,MAAM,aAAa,CAAC;AAEhD;;GAEG;AACH,MAAM,MAAM,QAAQ,GAAG;IACtB;;OAEG;IACH,WAAW,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,OAAO,EAAE,MAAM,CAAC;IAChB;;OAEG;IACH,QAAQ,EAAE,YAAY,CAAC;IACvB;;OAEG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IAEpB,UAAU,CAAC,EAAE,OAAO,CAAC;IACrB;;OAEG;IACH,kBAAkB,EAAE,CAAC,KAAK,EAAE,SAAS,KAAK,OAAO,CAAC;CAClD,GAAG,CACD;IACA;;OAEG;IACH,QAAQ,EAAE,CAAC,KAAK,EAAE,SAAS,KAAK,GAAG,CAAC;CACnC,GACD;IACA;;;OAGG;IACH,OAAO,EAAE,CAAC,KAAK,EAAE,SAAS,KAAK,MAAM,GAAG,MAAM,EAAE,CAAC;CAChD,CACH,CAAC;AAEF,iBAAS,WAAW,CAAC,KAAK,EAAE,SAAS,WAEpC;AA2BD;;;;;;;;;;GAUG;AACH,eAAO,MAAM,UAAU;;;;;;yBApCS,SAAS,KAAG,MAAM,EAAE;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CAkIhB,CAAC;AAErC,MAAM,MAAM,WAAW,GAAG,MAAM,OAAO,UAAU,CAAC"}
|
package/package.json
CHANGED
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@huggingface/tasks",
|
|
3
3
|
"packageManager": "pnpm@8.10.5",
|
|
4
|
-
"version": "0.10.
|
|
4
|
+
"version": "0.10.14",
|
|
5
5
|
"description": "List of ML tasks for huggingface.co/tasks",
|
|
6
6
|
"repository": "https://github.com/huggingface/huggingface.js.git",
|
|
7
7
|
"publishConfig": {
|
package/src/local-apps.ts
CHANGED
|
@@ -37,6 +37,7 @@ export type LocalApp = {
|
|
|
37
37
|
| {
|
|
38
38
|
/**
|
|
39
39
|
* And if not (mostly llama.cpp), snippet to copy/paste in your terminal
|
|
40
|
+
* Support the placeholder {{GGUF_FILE}} that will be replaced by the gguf file path or the list of available files.
|
|
40
41
|
*/
|
|
41
42
|
snippet: (model: ModelData) => string | string[];
|
|
42
43
|
}
|
|
@@ -54,7 +55,7 @@ brew install llama.cpp
|
|
|
54
55
|
# Load and run the model
|
|
55
56
|
llama \\
|
|
56
57
|
--hf-repo "${model.id}" \\
|
|
57
|
-
--hf-file
|
|
58
|
+
--hf-file {{GGUF_FILE}} \\
|
|
58
59
|
-p "I believe the meaning of life is" \\
|
|
59
60
|
-n 128`,
|
|
60
61
|
`# Option 2: build llama.cpp from source with curl support
|
|
@@ -65,7 +66,7 @@ LLAMA_CURL=1 make
|
|
|
65
66
|
# Load and run the model
|
|
66
67
|
./main \\
|
|
67
68
|
--hf-repo "${model.id}" \\
|
|
68
|
-
-m
|
|
69
|
+
-m {{GGUF_FILE}} \\
|
|
69
70
|
-p "I believe the meaning of life is" \\
|
|
70
71
|
-n 128`,
|
|
71
72
|
];
|