@huggingface/inference 4.13.13 → 4.13.14

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,3 +1,3 @@
1
- export declare const PACKAGE_VERSION = "4.13.13";
1
+ export declare const PACKAGE_VERSION = "4.13.14";
2
2
  export declare const PACKAGE_NAME = "@huggingface/inference";
3
3
  //# sourceMappingURL=package.d.ts.map
@@ -2,5 +2,5 @@
2
2
  Object.defineProperty(exports, "__esModule", { value: true });
3
3
  exports.PACKAGE_NAME = exports.PACKAGE_VERSION = void 0;
4
4
  // Generated file from package.json. Issues importing JSON directly when publishing on commonjs/ESM - see https://github.com/microsoft/TypeScript/issues/51783
5
- exports.PACKAGE_VERSION = "4.13.13";
5
+ exports.PACKAGE_VERSION = "4.13.14";
6
6
  exports.PACKAGE_NAME = "@huggingface/inference";
@@ -1 +1 @@
1
- {"version":3,"file":"chatCompletion.d.ts","sourceRoot":"","sources":["../../../../src/tasks/nlp/chatCompletion.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,mBAAmB,EAAE,oBAAoB,EAAE,MAAM,oBAAoB,CAAC;AAGpF,OAAO,KAAK,EAAE,QAAQ,EAAE,OAAO,EAAE,MAAM,gBAAgB,CAAC;AAKxD;;GAEG;AACH,wBAAsB,cAAc,CACnC,IAAI,EAAE,QAAQ,GAAG,mBAAmB,EACpC,OAAO,CAAC,EAAE,OAAO,GACf,OAAO,CAAC,oBAAoB,CAAC,CAc/B"}
1
+ {"version":3,"file":"chatCompletion.d.ts","sourceRoot":"","sources":["../../../../src/tasks/nlp/chatCompletion.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,mBAAmB,EAAE,oBAAoB,EAAE,MAAM,oBAAoB,CAAC;AAGpF,OAAO,KAAK,EAAE,QAAQ,EAAE,OAAO,EAAE,MAAM,gBAAgB,CAAC;AAKxD;;GAEG;AACH,wBAAsB,cAAc,CACnC,IAAI,EAAE,QAAQ,GAAG,mBAAmB,EACpC,OAAO,CAAC,EAAE,OAAO,GACf,OAAO,CAAC,oBAAoB,CAAC,CAiB/B"}
@@ -10,7 +10,11 @@ const providerHelper_js_1 = require("../../providers/providerHelper.js");
10
10
  */
11
11
  async function chatCompletion(args, options) {
12
12
  let providerHelper;
13
- if (!args.provider || args.provider === "auto") {
13
+ if (args.endpointUrl) {
14
+ const provider = await (0, getInferenceProviderMapping_js_1.resolveProvider)(args.provider, args.model, args.endpointUrl);
15
+ providerHelper = (0, getProviderHelper_js_1.getProviderHelper)(provider, "conversational");
16
+ }
17
+ else if (!args.provider || args.provider === "auto") {
14
18
  // Special case: we have a dedicated auto-router for conversational models. No need to fetch provider mapping.
15
19
  providerHelper = new providerHelper_js_1.AutoRouterConversationalTask();
16
20
  }
@@ -1 +1 @@
1
- {"version":3,"file":"chatCompletionStream.d.ts","sourceRoot":"","sources":["../../../../src/tasks/nlp/chatCompletionStream.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,mBAAmB,EAAE,0BAA0B,EAAE,MAAM,oBAAoB,CAAC;AAG1F,OAAO,KAAK,EAAE,QAAQ,EAAE,OAAO,EAAE,MAAM,gBAAgB,CAAC;AAKxD;;GAEG;AACH,wBAAuB,oBAAoB,CAC1C,IAAI,EAAE,QAAQ,GAAG,mBAAmB,EACpC,OAAO,CAAC,EAAE,OAAO,GACf,cAAc,CAAC,0BAA0B,CAAC,CAa5C"}
1
+ {"version":3,"file":"chatCompletionStream.d.ts","sourceRoot":"","sources":["../../../../src/tasks/nlp/chatCompletionStream.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,mBAAmB,EAAE,0BAA0B,EAAE,MAAM,oBAAoB,CAAC;AAG1F,OAAO,KAAK,EAAE,QAAQ,EAAE,OAAO,EAAE,MAAM,gBAAgB,CAAC;AAKxD;;GAEG;AACH,wBAAuB,oBAAoB,CAC1C,IAAI,EAAE,QAAQ,GAAG,mBAAmB,EACpC,OAAO,CAAC,EAAE,OAAO,GACf,cAAc,CAAC,0BAA0B,CAAC,CAgB5C"}
@@ -10,7 +10,11 @@ const providerHelper_js_1 = require("../../providers/providerHelper.js");
10
10
  */
11
11
  async function* chatCompletionStream(args, options) {
12
12
  let providerHelper;
13
- if (!args.provider || args.provider === "auto") {
13
+ if (args.endpointUrl) {
14
+ const provider = await (0, getInferenceProviderMapping_js_1.resolveProvider)(args.provider, args.model, args.endpointUrl);
15
+ providerHelper = (0, getProviderHelper_js_1.getProviderHelper)(provider, "conversational");
16
+ }
17
+ else if (!args.provider || args.provider === "auto") {
14
18
  // Special case: we have a dedicated auto-router for conversational models. No need to fetch provider mapping.
15
19
  providerHelper = new providerHelper_js_1.AutoRouterConversationalTask();
16
20
  }
@@ -1,3 +1,3 @@
1
- export declare const PACKAGE_VERSION = "4.13.13";
1
+ export declare const PACKAGE_VERSION = "4.13.14";
2
2
  export declare const PACKAGE_NAME = "@huggingface/inference";
3
3
  //# sourceMappingURL=package.d.ts.map
@@ -1,3 +1,3 @@
1
1
  // Generated file from package.json. Issues importing JSON directly when publishing on commonjs/ESM - see https://github.com/microsoft/TypeScript/issues/51783
2
- export const PACKAGE_VERSION = "4.13.13";
2
+ export const PACKAGE_VERSION = "4.13.14";
3
3
  export const PACKAGE_NAME = "@huggingface/inference";
@@ -1 +1 @@
1
- {"version":3,"file":"chatCompletion.d.ts","sourceRoot":"","sources":["../../../../src/tasks/nlp/chatCompletion.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,mBAAmB,EAAE,oBAAoB,EAAE,MAAM,oBAAoB,CAAC;AAGpF,OAAO,KAAK,EAAE,QAAQ,EAAE,OAAO,EAAE,MAAM,gBAAgB,CAAC;AAKxD;;GAEG;AACH,wBAAsB,cAAc,CACnC,IAAI,EAAE,QAAQ,GAAG,mBAAmB,EACpC,OAAO,CAAC,EAAE,OAAO,GACf,OAAO,CAAC,oBAAoB,CAAC,CAc/B"}
1
+ {"version":3,"file":"chatCompletion.d.ts","sourceRoot":"","sources":["../../../../src/tasks/nlp/chatCompletion.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,mBAAmB,EAAE,oBAAoB,EAAE,MAAM,oBAAoB,CAAC;AAGpF,OAAO,KAAK,EAAE,QAAQ,EAAE,OAAO,EAAE,MAAM,gBAAgB,CAAC;AAKxD;;GAEG;AACH,wBAAsB,cAAc,CACnC,IAAI,EAAE,QAAQ,GAAG,mBAAmB,EACpC,OAAO,CAAC,EAAE,OAAO,GACf,OAAO,CAAC,oBAAoB,CAAC,CAiB/B"}
@@ -7,7 +7,11 @@ import { AutoRouterConversationalTask } from "../../providers/providerHelper.js"
7
7
  */
8
8
  export async function chatCompletion(args, options) {
9
9
  let providerHelper;
10
- if (!args.provider || args.provider === "auto") {
10
+ if (args.endpointUrl) {
11
+ const provider = await resolveProvider(args.provider, args.model, args.endpointUrl);
12
+ providerHelper = getProviderHelper(provider, "conversational");
13
+ }
14
+ else if (!args.provider || args.provider === "auto") {
11
15
  // Special case: we have a dedicated auto-router for conversational models. No need to fetch provider mapping.
12
16
  providerHelper = new AutoRouterConversationalTask();
13
17
  }
@@ -1 +1 @@
1
- {"version":3,"file":"chatCompletionStream.d.ts","sourceRoot":"","sources":["../../../../src/tasks/nlp/chatCompletionStream.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,mBAAmB,EAAE,0BAA0B,EAAE,MAAM,oBAAoB,CAAC;AAG1F,OAAO,KAAK,EAAE,QAAQ,EAAE,OAAO,EAAE,MAAM,gBAAgB,CAAC;AAKxD;;GAEG;AACH,wBAAuB,oBAAoB,CAC1C,IAAI,EAAE,QAAQ,GAAG,mBAAmB,EACpC,OAAO,CAAC,EAAE,OAAO,GACf,cAAc,CAAC,0BAA0B,CAAC,CAa5C"}
1
+ {"version":3,"file":"chatCompletionStream.d.ts","sourceRoot":"","sources":["../../../../src/tasks/nlp/chatCompletionStream.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,mBAAmB,EAAE,0BAA0B,EAAE,MAAM,oBAAoB,CAAC;AAG1F,OAAO,KAAK,EAAE,QAAQ,EAAE,OAAO,EAAE,MAAM,gBAAgB,CAAC;AAKxD;;GAEG;AACH,wBAAuB,oBAAoB,CAC1C,IAAI,EAAE,QAAQ,GAAG,mBAAmB,EACpC,OAAO,CAAC,EAAE,OAAO,GACf,cAAc,CAAC,0BAA0B,CAAC,CAgB5C"}
@@ -7,7 +7,11 @@ import { AutoRouterConversationalTask } from "../../providers/providerHelper.js"
7
7
  */
8
8
  export async function* chatCompletionStream(args, options) {
9
9
  let providerHelper;
10
- if (!args.provider || args.provider === "auto") {
10
+ if (args.endpointUrl) {
11
+ const provider = await resolveProvider(args.provider, args.model, args.endpointUrl);
12
+ providerHelper = getProviderHelper(provider, "conversational");
13
+ }
14
+ else if (!args.provider || args.provider === "auto") {
11
15
  // Special case: we have a dedicated auto-router for conversational models. No need to fetch provider mapping.
12
16
  providerHelper = new AutoRouterConversationalTask();
13
17
  }
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@huggingface/inference",
3
- "version": "4.13.13",
3
+ "version": "4.13.14",
4
4
  "description": "Typescript client for the Hugging Face Inference Providers and Inference Endpoints",
5
5
  "keywords": [
6
6
  "ai",
@@ -45,7 +45,7 @@
45
45
  },
46
46
  "dependencies": {
47
47
  "@huggingface/jinja": "^0.5.5",
48
- "@huggingface/tasks": "^0.19.85"
48
+ "@huggingface/tasks": "^0.19.88"
49
49
  },
50
50
  "devDependencies": {
51
51
  "@types/node": "18.13.0"
package/src/package.ts CHANGED
@@ -1,3 +1,3 @@
1
1
  // Generated file from package.json. Issues importing JSON directly when publishing on commonjs/ESM - see https://github.com/microsoft/TypeScript/issues/51783
2
- export const PACKAGE_VERSION = "4.13.13";
2
+ export const PACKAGE_VERSION = "4.13.14";
3
3
  export const PACKAGE_NAME = "@huggingface/inference";
@@ -14,7 +14,10 @@ export async function chatCompletion(
14
14
  options?: Options,
15
15
  ): Promise<ChatCompletionOutput> {
16
16
  let providerHelper: ConversationalTaskHelper & TaskProviderHelper;
17
- if (!args.provider || args.provider === "auto") {
17
+ if (args.endpointUrl) {
18
+ const provider = await resolveProvider(args.provider, args.model, args.endpointUrl);
19
+ providerHelper = getProviderHelper(provider, "conversational");
20
+ } else if (!args.provider || args.provider === "auto") {
18
21
  // Special case: we have a dedicated auto-router for conversational models. No need to fetch provider mapping.
19
22
  providerHelper = new AutoRouterConversationalTask();
20
23
  } else {
@@ -14,7 +14,10 @@ export async function* chatCompletionStream(
14
14
  options?: Options,
15
15
  ): AsyncGenerator<ChatCompletionStreamOutput> {
16
16
  let providerHelper: ConversationalTaskHelper & TaskProviderHelper;
17
- if (!args.provider || args.provider === "auto") {
17
+ if (args.endpointUrl) {
18
+ const provider = await resolveProvider(args.provider, args.model, args.endpointUrl);
19
+ providerHelper = getProviderHelper(provider, "conversational");
20
+ } else if (!args.provider || args.provider === "auto") {
18
21
  // Special case: we have a dedicated auto-router for conversational models. No need to fetch provider mapping.
19
22
  providerHelper = new AutoRouterConversationalTask();
20
23
  } else {