oci-generativeaiinference 2.115.0 → 2.116.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (46) hide show
  1. package/lib/model/assistant-message.d.ts +4 -0
  2. package/lib/model/assistant-message.js.map +1 -1
  3. package/lib/model/chat-choice.d.ts +2 -1
  4. package/lib/model/chat-choice.js +4 -2
  5. package/lib/model/chat-choice.js.map +1 -1
  6. package/lib/model/cohere-chat-request.d.ts +2 -1
  7. package/lib/model/cohere-chat-request.js.map +1 -1
  8. package/lib/model/developer-message.d.ts +35 -0
  9. package/lib/model/developer-message.js +57 -0
  10. package/lib/model/developer-message.js.map +1 -0
  11. package/lib/model/function-call.d.ts +1 -1
  12. package/lib/model/generic-chat-request.d.ts +41 -2
  13. package/lib/model/generic-chat-request.js +29 -0
  14. package/lib/model/generic-chat-request.js.map +1 -1
  15. package/lib/model/generic-chat-response.d.ts +1 -0
  16. package/lib/model/generic-chat-response.js +4 -2
  17. package/lib/model/generic-chat-response.js.map +1 -1
  18. package/lib/model/index.d.ts +16 -0
  19. package/lib/model/index.js +18 -2
  20. package/lib/model/index.js.map +1 -1
  21. package/lib/model/json-object-response-format.d.ts +31 -0
  22. package/lib/model/json-object-response-format.js +61 -0
  23. package/lib/model/json-object-response-format.js.map +1 -0
  24. package/lib/model/json-schema-response-format.d.ts +32 -0
  25. package/lib/model/json-schema-response-format.js +69 -0
  26. package/lib/model/json-schema-response-format.js.map +1 -0
  27. package/lib/model/message.js +4 -0
  28. package/lib/model/message.js.map +1 -1
  29. package/lib/model/prediction.d.ts +30 -0
  30. package/lib/model/prediction.js +75 -0
  31. package/lib/model/prediction.js.map +1 -0
  32. package/lib/model/response-format.d.ts +29 -0
  33. package/lib/model/response-format.js +83 -0
  34. package/lib/model/response-format.js.map +1 -0
  35. package/lib/model/response-json-schema.d.ts +45 -0
  36. package/lib/model/response-json-schema.js +36 -0
  37. package/lib/model/response-json-schema.js.map +1 -0
  38. package/lib/model/static-content.d.ts +36 -0
  39. package/lib/model/static-content.js +69 -0
  40. package/lib/model/static-content.js.map +1 -0
  41. package/lib/model/text-response-format.d.ts +31 -0
  42. package/lib/model/text-response-format.js +59 -0
  43. package/lib/model/text-response-format.js.map +1 -0
  44. package/lib/model/tool-choice-function.d.ts +1 -1
  45. package/lib/model/tool-message.d.ts +1 -1
  46. package/package.json +3 -3
@@ -0,0 +1,69 @@
1
+ "use strict";
2
+ /**
3
+ * Generative AI Service Inference API
4
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
5
+
6
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
7
+
8
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
9
+
10
+ To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
11
+
12
+ * OpenAPI spec version: 20231130
13
+ *
14
+ *
15
+ * NOTE: This class is auto generated by OracleSDKGenerator.
16
+ * Do not edit the class manually.
17
+ *
18
+ * Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
19
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
20
+ */
21
+ var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
22
+ if (k2 === undefined) k2 = k;
23
+ Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
24
+ }) : (function(o, m, k, k2) {
25
+ if (k2 === undefined) k2 = k;
26
+ o[k2] = m[k];
27
+ }));
28
+ var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
29
+ Object.defineProperty(o, "default", { enumerable: true, value: v });
30
+ }) : function(o, v) {
31
+ o["default"] = v;
32
+ });
33
+ var __importStar = (this && this.__importStar) || function (mod) {
34
+ if (mod && mod.__esModule) return mod;
35
+ var result = {};
36
+ if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
37
+ __setModuleDefault(result, mod);
38
+ return result;
39
+ };
40
+ Object.defineProperty(exports, "__esModule", { value: true });
41
+ exports.StaticContent = void 0;
42
+ const model = __importStar(require("../model"));
43
+ var StaticContent;
44
+ (function (StaticContent) {
45
+ function getJsonObj(obj, isParentJsonObj) {
46
+ const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj ? obj : model.Prediction.getJsonObj(obj))), {
47
+ "content": obj.content
48
+ ? obj.content.map(item => {
49
+ return model.TextContent.getJsonObj(item);
50
+ })
51
+ : undefined
52
+ });
53
+ return jsonObj;
54
+ }
55
+ StaticContent.getJsonObj = getJsonObj;
56
+ StaticContent.type = "CONTENT";
57
+ function getDeserializedJsonObj(obj, isParentJsonObj) {
58
+ const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj ? obj : model.Prediction.getDeserializedJsonObj(obj))), {
59
+ "content": obj.content
60
+ ? obj.content.map(item => {
61
+ return model.TextContent.getDeserializedJsonObj(item);
62
+ })
63
+ : undefined
64
+ });
65
+ return jsonObj;
66
+ }
67
+ StaticContent.getDeserializedJsonObj = getDeserializedJsonObj;
68
+ })(StaticContent = exports.StaticContent || (exports.StaticContent = {}));
69
+ //# sourceMappingURL=static-content.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"static-content.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/static-content.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAgBlC,IAAiB,aAAa,CA8B7B;AA9BD,WAAiB,aAAa;IAC5B,SAAgB,UAAU,CAAC,GAAkB,EAAE,eAAyB;QACtE,MAAM,OAAO,mCACR,CAAC,eAAe,CAAC,CAAC,CAAC,GAAG,CAAC,CAAC,CAAE,KAAK,CAAC,UAAU,CAAC,UAAU,CAAC,GAAG,CAAmB,CAAC,GAC7E;YACD,SAAS,EAAE,GAAG,CAAC,OAAO;gBACpB,CAAC,CAAC,GAAG,CAAC,OAAO,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACrB,OAAO,KAAK,CAAC,WAAW,CAAC,UAAU,CAAC,IAAI,CAAC,CAAC;gBAC5C,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAbe,wBAAU,aAazB,CAAA;IACY,kBAAI,GAAG,SAAS,CAAC;IAC9B,SAAgB,sBAAsB,CAAC,GAAkB,EAAE,eAAyB;QAClF,MAAM,OAAO,mCACR,CAAC,eAAe,CAAC,CAAC,CAAC,GAAG,CAAC,CAAC,CAAE,KAAK,CAAC,UAAU,CAAC,sBAAsB,CAAC,GAAG,CAAmB,CAAC,GACzF;YACD,SAAS,EAAE,GAAG,CAAC,OAAO;gBACpB,CAAC,CAAC,GAAG,CAAC,OAAO,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACrB,OAAO,KAAK,CAAC,WAAW,CAAC,sBAAsB,CAAC,IAAI,CAAC,CAAC;gBACxD,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAbe,oCAAsB,yBAarC,CAAA;AACH,CAAC,EA9BgB,aAAa,GAAb,qBAAa,KAAb,qBAAa,QA8B7B"}
@@ -0,0 +1,31 @@
1
+ /**
2
+ * Generative AI Service Inference API
3
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
4
+
5
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
6
+
7
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
8
+
9
+ To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
10
+
11
+ * OpenAPI spec version: 20231130
12
+ *
13
+ *
14
+ * NOTE: This class is auto generated by OracleSDKGenerator.
15
+ * Do not edit the class manually.
16
+ *
17
+ * Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
18
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
19
+ */
20
+ import * as model from "../model";
21
+ /**
22
+ * Enables TEXT mode.
23
+ */
24
+ export interface TextResponseFormat extends model.ResponseFormat {
25
+ "type": string;
26
+ }
27
+ export declare namespace TextResponseFormat {
28
+ function getJsonObj(obj: TextResponseFormat, isParentJsonObj?: boolean): object;
29
+ const type = "TEXT";
30
+ function getDeserializedJsonObj(obj: TextResponseFormat, isParentJsonObj?: boolean): object;
31
+ }
@@ -0,0 +1,59 @@
1
+ "use strict";
2
+ /**
3
+ * Generative AI Service Inference API
4
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
5
+
6
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
7
+
8
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
9
+
10
+ To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
11
+
12
+ * OpenAPI spec version: 20231130
13
+ *
14
+ *
15
+ * NOTE: This class is auto generated by OracleSDKGenerator.
16
+ * Do not edit the class manually.
17
+ *
18
+ * Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
19
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
20
+ */
21
+ var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
22
+ if (k2 === undefined) k2 = k;
23
+ Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
24
+ }) : (function(o, m, k, k2) {
25
+ if (k2 === undefined) k2 = k;
26
+ o[k2] = m[k];
27
+ }));
28
+ var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
29
+ Object.defineProperty(o, "default", { enumerable: true, value: v });
30
+ }) : function(o, v) {
31
+ o["default"] = v;
32
+ });
33
+ var __importStar = (this && this.__importStar) || function (mod) {
34
+ if (mod && mod.__esModule) return mod;
35
+ var result = {};
36
+ if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
37
+ __setModuleDefault(result, mod);
38
+ return result;
39
+ };
40
+ Object.defineProperty(exports, "__esModule", { value: true });
41
+ exports.TextResponseFormat = void 0;
42
+ const model = __importStar(require("../model"));
43
+ var TextResponseFormat;
44
+ (function (TextResponseFormat) {
45
+ function getJsonObj(obj, isParentJsonObj) {
46
+ const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj ? obj : model.ResponseFormat.getJsonObj(obj))), {});
47
+ return jsonObj;
48
+ }
49
+ TextResponseFormat.getJsonObj = getJsonObj;
50
+ TextResponseFormat.type = "TEXT";
51
+ function getDeserializedJsonObj(obj, isParentJsonObj) {
52
+ const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj
53
+ ? obj
54
+ : model.ResponseFormat.getDeserializedJsonObj(obj))), {});
55
+ return jsonObj;
56
+ }
57
+ TextResponseFormat.getDeserializedJsonObj = getDeserializedJsonObj;
58
+ })(TextResponseFormat = exports.TextResponseFormat || (exports.TextResponseFormat = {}));
59
+ //# sourceMappingURL=text-response-format.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"text-response-format.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/text-response-format.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAUlC,IAAiB,kBAAkB,CAuBlC;AAvBD,WAAiB,kBAAkB;IACjC,SAAgB,UAAU,CAAC,GAAuB,EAAE,eAAyB;QAC3E,MAAM,OAAO,mCACR,CAAC,eAAe,CAAC,CAAC,CAAC,GAAG,CAAC,CAAC,CAAE,KAAK,CAAC,cAAc,CAAC,UAAU,CAAC,GAAG,CAAwB,CAAC,GACtF,EAAE,CACN,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAPe,6BAAU,aAOzB,CAAA;IACY,uBAAI,GAAG,MAAM,CAAC;IAC3B,SAAgB,sBAAsB,CACpC,GAAuB,EACvB,eAAyB;QAEzB,MAAM,OAAO,mCACR,CAAC,eAAe;YACjB,CAAC,CAAC,GAAG;YACL,CAAC,CAAE,KAAK,CAAC,cAAc,CAAC,sBAAsB,CAAC,GAAG,CAAwB,CAAC,GAC1E,EAAE,CACN,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAZe,yCAAsB,yBAYrC,CAAA;AACH,CAAC,EAvBgB,kBAAkB,GAAlB,0BAAkB,KAAlB,0BAAkB,QAuBlC"}
@@ -19,7 +19,7 @@ To learn more about the service, see the [Generative AI documentation](https://d
19
19
  */
20
20
  import * as model from "../model";
21
21
  /**
22
- * The tool choice for a function.
22
+ * The tool choice for a function. The model will generate the parameters for user to call the function.
23
23
  */
24
24
  export interface ToolChoiceFunction extends model.ToolChoice {
25
25
  /**
@@ -23,7 +23,7 @@ import * as model from "../model";
23
23
  */
24
24
  export interface ToolMessage extends model.Message {
25
25
  /**
26
- * Tool call that this message is responding to.
26
+ * Tool call that this message is responding to. The ID is the unique string generated by the model.
27
27
  */
28
28
  "toolCallId"?: string;
29
29
  "role": string;
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "oci-generativeaiinference",
3
- "version": "2.115.0",
3
+ "version": "2.116.0",
4
4
  "description": "OCI NodeJS client for Generative Ai Inference Service",
5
5
  "repository": {
6
6
  "type": "git",
@@ -15,8 +15,8 @@
15
15
  },
16
16
  "license": "(UPL-1.0 OR Apache-2.0)",
17
17
  "dependencies": {
18
- "oci-common": "2.115.0",
19
- "oci-workrequests": "2.115.0"
18
+ "oci-common": "2.116.0",
19
+ "oci-workrequests": "2.116.0"
20
20
  },
21
21
  "publishConfig": {
22
22
  "registry": "https://registry.npmjs.org"