oci-generativeaiinference 2.123.0 → 2.123.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (99) hide show
  1. package/lib/model/assistant-message.d.ts +4 -0
  2. package/lib/model/assistant-message.js.map +1 -1
  3. package/lib/model/base-chat-request.js +4 -0
  4. package/lib/model/base-chat-request.js.map +1 -1
  5. package/lib/model/base-chat-response.js +4 -0
  6. package/lib/model/base-chat-response.js.map +1 -1
  7. package/lib/model/chat-content.js +4 -0
  8. package/lib/model/chat-content.js.map +1 -1
  9. package/lib/model/chat-details.d.ts +1 -1
  10. package/lib/model/chat-result.d.ts +1 -1
  11. package/lib/model/citation-options-v2.d.ts +38 -0
  12. package/lib/model/citation-options-v2.js +42 -0
  13. package/lib/model/citation-options-v2.js.map +1 -0
  14. package/lib/model/cohere-assistant-message-v2.d.ts +43 -0
  15. package/lib/model/cohere-assistant-message-v2.js +83 -0
  16. package/lib/model/cohere-assistant-message-v2.js.map +1 -0
  17. package/lib/model/cohere-chat-request-v2.d.ts +142 -0
  18. package/lib/model/cohere-chat-request-v2.js +114 -0
  19. package/lib/model/cohere-chat-request-v2.js.map +1 -0
  20. package/lib/model/cohere-chat-response-v2.d.ts +61 -0
  21. package/lib/model/cohere-chat-response-v2.js +90 -0
  22. package/lib/model/cohere-chat-response-v2.js.map +1 -0
  23. package/lib/model/cohere-citation-document.d.ts +36 -0
  24. package/lib/model/cohere-citation-document.js +36 -0
  25. package/lib/model/cohere-citation-document.js.map +1 -0
  26. package/lib/model/cohere-citation-source-v2.d.ts +44 -0
  27. package/lib/model/cohere-citation-source-v2.js +74 -0
  28. package/lib/model/cohere-citation-source-v2.js.map +1 -0
  29. package/lib/model/cohere-citation-tool-v2.d.ts +36 -0
  30. package/lib/model/cohere-citation-tool-v2.js +36 -0
  31. package/lib/model/cohere-citation-tool-v2.js.map +1 -0
  32. package/lib/model/cohere-citation-v2.d.ts +63 -0
  33. package/lib/model/cohere-citation-v2.js +79 -0
  34. package/lib/model/cohere-citation-v2.js.map +1 -0
  35. package/lib/model/cohere-content-v2.d.ts +29 -0
  36. package/lib/model/cohere-content-v2.js +87 -0
  37. package/lib/model/cohere-content-v2.js.map +1 -0
  38. package/lib/model/cohere-document-content-v2.d.ts +35 -0
  39. package/lib/model/cohere-document-content-v2.js +61 -0
  40. package/lib/model/cohere-document-content-v2.js.map +1 -0
  41. package/lib/model/cohere-image-content-v2.d.ts +32 -0
  42. package/lib/model/cohere-image-content-v2.js +65 -0
  43. package/lib/model/cohere-image-content-v2.js.map +1 -0
  44. package/lib/model/cohere-image-url-v2.d.ts +46 -0
  45. package/lib/model/cohere-image-url-v2.js +47 -0
  46. package/lib/model/cohere-image-url-v2.js.map +1 -0
  47. package/lib/model/cohere-message-v2.d.ts +34 -0
  48. package/lib/model/cohere-message-v2.js +99 -0
  49. package/lib/model/cohere-message-v2.js.map +1 -0
  50. package/lib/model/cohere-system-message-v2.d.ts +31 -0
  51. package/lib/model/cohere-system-message-v2.js +59 -0
  52. package/lib/model/cohere-system-message-v2.js.map +1 -0
  53. package/lib/model/cohere-text-content-v2.d.ts +35 -0
  54. package/lib/model/cohere-text-content-v2.js +59 -0
  55. package/lib/model/cohere-text-content-v2.js.map +1 -0
  56. package/lib/model/cohere-thinking-content-v2.d.ts +35 -0
  57. package/lib/model/cohere-thinking-content-v2.js +61 -0
  58. package/lib/model/cohere-thinking-content-v2.js.map +1 -0
  59. package/lib/model/cohere-thinking-v2.d.ts +40 -0
  60. package/lib/model/cohere-thinking-v2.js +41 -0
  61. package/lib/model/cohere-thinking-v2.js.map +1 -0
  62. package/lib/model/cohere-tool-call-v2.d.ts +48 -0
  63. package/lib/model/cohere-tool-call-v2.js +45 -0
  64. package/lib/model/cohere-tool-call-v2.js.map +1 -0
  65. package/lib/model/cohere-tool-message-v2.d.ts +35 -0
  66. package/lib/model/cohere-tool-message-v2.js +59 -0
  67. package/lib/model/cohere-tool-message-v2.js.map +1 -0
  68. package/lib/model/cohere-tool-v2.d.ts +37 -0
  69. package/lib/model/cohere-tool-v2.js +64 -0
  70. package/lib/model/cohere-tool-v2.js.map +1 -0
  71. package/lib/model/cohere-user-message-v2.d.ts +31 -0
  72. package/lib/model/cohere-user-message-v2.js +59 -0
  73. package/lib/model/cohere-user-message-v2.js.map +1 -0
  74. package/lib/model/document-content.d.ts +32 -0
  75. package/lib/model/document-content.js +65 -0
  76. package/lib/model/document-content.js.map +1 -0
  77. package/lib/model/document-url.d.ts +63 -0
  78. package/lib/model/document-url.js +47 -0
  79. package/lib/model/document-url.js.map +1 -0
  80. package/lib/model/embed-text-details.d.ts +16 -0
  81. package/lib/model/embed-text-details.js +9 -0
  82. package/lib/model/embed-text-details.js.map +1 -1
  83. package/lib/model/embed-text-result.d.ts +5 -1
  84. package/lib/model/embed-text-result.js.map +1 -1
  85. package/lib/model/function.d.ts +40 -0
  86. package/lib/model/function.js +36 -0
  87. package/lib/model/function.js.map +1 -0
  88. package/lib/model/generic-chat-request.d.ts +1 -0
  89. package/lib/model/generic-chat-request.js +1 -0
  90. package/lib/model/generic-chat-request.js.map +1 -1
  91. package/lib/model/index.d.ts +50 -0
  92. package/lib/model/index.js +53 -3
  93. package/lib/model/index.js.map +1 -1
  94. package/lib/model/log-probability.d.ts +40 -0
  95. package/lib/model/log-probability.js +36 -0
  96. package/lib/model/log-probability.js.map +1 -0
  97. package/lib/model/rerank-text-details.d.ts +4 -0
  98. package/lib/model/rerank-text-details.js.map +1 -1
  99. package/package.json +3 -3
@@ -0,0 +1,59 @@
1
+ "use strict";
2
+ /**
3
+ * Generative AI Service Inference API
4
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
5
+
6
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
7
+
8
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
9
+
10
+ To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
11
+
12
+ * OpenAPI spec version: 20231130
13
+ *
14
+ *
15
+ * NOTE: This class is auto generated by OracleSDKGenerator.
16
+ * Do not edit the class manually.
17
+ *
18
+ * Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
19
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
20
+ */
21
+ var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
22
+ if (k2 === undefined) k2 = k;
23
+ Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
24
+ }) : (function(o, m, k, k2) {
25
+ if (k2 === undefined) k2 = k;
26
+ o[k2] = m[k];
27
+ }));
28
+ var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
29
+ Object.defineProperty(o, "default", { enumerable: true, value: v });
30
+ }) : function(o, v) {
31
+ o["default"] = v;
32
+ });
33
+ var __importStar = (this && this.__importStar) || function (mod) {
34
+ if (mod && mod.__esModule) return mod;
35
+ var result = {};
36
+ if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
37
+ __setModuleDefault(result, mod);
38
+ return result;
39
+ };
40
+ Object.defineProperty(exports, "__esModule", { value: true });
41
+ exports.CohereSystemMessageV2 = void 0;
42
+ const model = __importStar(require("../model"));
43
+ var CohereSystemMessageV2;
44
+ (function (CohereSystemMessageV2) {
45
+ function getJsonObj(obj, isParentJsonObj) {
46
+ const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj ? obj : model.CohereMessageV2.getJsonObj(obj))), {});
47
+ return jsonObj;
48
+ }
49
+ CohereSystemMessageV2.getJsonObj = getJsonObj;
50
+ CohereSystemMessageV2.role = "SYSTEM";
51
+ function getDeserializedJsonObj(obj, isParentJsonObj) {
52
+ const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj
53
+ ? obj
54
+ : model.CohereMessageV2.getDeserializedJsonObj(obj))), {});
55
+ return jsonObj;
56
+ }
57
+ CohereSystemMessageV2.getDeserializedJsonObj = getDeserializedJsonObj;
58
+ })(CohereSystemMessageV2 = exports.CohereSystemMessageV2 || (exports.CohereSystemMessageV2 = {}));
59
+ //# sourceMappingURL=cohere-system-message-v2.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"cohere-system-message-v2.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/cohere-system-message-v2.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAUlC,IAAiB,qBAAqB,CAuBrC;AAvBD,WAAiB,qBAAqB;IACpC,SAAgB,UAAU,CAAC,GAA0B,EAAE,eAAyB;QAC9E,MAAM,OAAO,mCACR,CAAC,eAAe,CAAC,CAAC,CAAC,GAAG,CAAC,CAAC,CAAE,KAAK,CAAC,eAAe,CAAC,UAAU,CAAC,GAAG,CAA2B,CAAC,GAC1F,EAAE,CACN,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAPe,gCAAU,aAOzB,CAAA;IACY,0BAAI,GAAG,QAAQ,CAAC;IAC7B,SAAgB,sBAAsB,CACpC,GAA0B,EAC1B,eAAyB;QAEzB,MAAM,OAAO,mCACR,CAAC,eAAe;YACjB,CAAC,CAAC,GAAG;YACL,CAAC,CAAE,KAAK,CAAC,eAAe,CAAC,sBAAsB,CAAC,GAAG,CAA2B,CAAC,GAC9E,EAAE,CACN,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAZe,4CAAsB,yBAYrC,CAAA;AACH,CAAC,EAvBgB,qBAAqB,GAArB,6BAAqB,KAArB,6BAAqB,QAuBrC"}
@@ -0,0 +1,35 @@
1
+ /**
2
+ * Generative AI Service Inference API
3
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
4
+
5
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
6
+
7
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
8
+
9
+ To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
10
+
11
+ * OpenAPI spec version: 20231130
12
+ *
13
+ *
14
+ * NOTE: This class is auto generated by OracleSDKGenerator.
15
+ * Do not edit the class manually.
16
+ *
17
+ * Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
18
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
19
+ */
20
+ import * as model from "../model";
21
+ /**
22
+ * Represents a single instance of text in the chat content.
23
+ */
24
+ export interface CohereTextContentV2 extends model.CohereContentV2 {
25
+ /**
26
+ * The text content.
27
+ */
28
+ "text"?: string;
29
+ "type": string;
30
+ }
31
+ export declare namespace CohereTextContentV2 {
32
+ function getJsonObj(obj: CohereTextContentV2, isParentJsonObj?: boolean): object;
33
+ const type = "TEXT";
34
+ function getDeserializedJsonObj(obj: CohereTextContentV2, isParentJsonObj?: boolean): object;
35
+ }
@@ -0,0 +1,59 @@
1
+ "use strict";
2
+ /**
3
+ * Generative AI Service Inference API
4
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
5
+
6
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
7
+
8
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
9
+
10
+ To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
11
+
12
+ * OpenAPI spec version: 20231130
13
+ *
14
+ *
15
+ * NOTE: This class is auto generated by OracleSDKGenerator.
16
+ * Do not edit the class manually.
17
+ *
18
+ * Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
19
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
20
+ */
21
+ var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
22
+ if (k2 === undefined) k2 = k;
23
+ Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
24
+ }) : (function(o, m, k, k2) {
25
+ if (k2 === undefined) k2 = k;
26
+ o[k2] = m[k];
27
+ }));
28
+ var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
29
+ Object.defineProperty(o, "default", { enumerable: true, value: v });
30
+ }) : function(o, v) {
31
+ o["default"] = v;
32
+ });
33
+ var __importStar = (this && this.__importStar) || function (mod) {
34
+ if (mod && mod.__esModule) return mod;
35
+ var result = {};
36
+ if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
37
+ __setModuleDefault(result, mod);
38
+ return result;
39
+ };
40
+ Object.defineProperty(exports, "__esModule", { value: true });
41
+ exports.CohereTextContentV2 = void 0;
42
+ const model = __importStar(require("../model"));
43
+ var CohereTextContentV2;
44
+ (function (CohereTextContentV2) {
45
+ function getJsonObj(obj, isParentJsonObj) {
46
+ const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj ? obj : model.CohereContentV2.getJsonObj(obj))), {});
47
+ return jsonObj;
48
+ }
49
+ CohereTextContentV2.getJsonObj = getJsonObj;
50
+ CohereTextContentV2.type = "TEXT";
51
+ function getDeserializedJsonObj(obj, isParentJsonObj) {
52
+ const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj
53
+ ? obj
54
+ : model.CohereContentV2.getDeserializedJsonObj(obj))), {});
55
+ return jsonObj;
56
+ }
57
+ CohereTextContentV2.getDeserializedJsonObj = getDeserializedJsonObj;
58
+ })(CohereTextContentV2 = exports.CohereTextContentV2 || (exports.CohereTextContentV2 = {}));
59
+ //# sourceMappingURL=cohere-text-content-v2.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"cohere-text-content-v2.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/cohere-text-content-v2.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAelC,IAAiB,mBAAmB,CAuBnC;AAvBD,WAAiB,mBAAmB;IAClC,SAAgB,UAAU,CAAC,GAAwB,EAAE,eAAyB;QAC5E,MAAM,OAAO,mCACR,CAAC,eAAe,CAAC,CAAC,CAAC,GAAG,CAAC,CAAC,CAAE,KAAK,CAAC,eAAe,CAAC,UAAU,CAAC,GAAG,CAAyB,CAAC,GACxF,EAAE,CACN,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAPe,8BAAU,aAOzB,CAAA;IACY,wBAAI,GAAG,MAAM,CAAC;IAC3B,SAAgB,sBAAsB,CACpC,GAAwB,EACxB,eAAyB;QAEzB,MAAM,OAAO,mCACR,CAAC,eAAe;YACjB,CAAC,CAAC,GAAG;YACL,CAAC,CAAE,KAAK,CAAC,eAAe,CAAC,sBAAsB,CAAC,GAAG,CAAyB,CAAC,GAC5E,EAAE,CACN,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAZe,0CAAsB,yBAYrC,CAAA;AACH,CAAC,EAvBgB,mBAAmB,GAAnB,2BAAmB,KAAnB,2BAAmB,QAuBnC"}
@@ -0,0 +1,35 @@
1
+ /**
2
+ * Generative AI Service Inference API
3
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
4
+
5
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
6
+
7
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
8
+
9
+ To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
10
+
11
+ * OpenAPI spec version: 20231130
12
+ *
13
+ *
14
+ * NOTE: This class is auto generated by OracleSDKGenerator.
15
+ * Do not edit the class manually.
16
+ *
17
+ * Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
18
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
19
+ */
20
+ import * as model from "../model";
21
+ /**
22
+ * Represents a single instance of Chat Doument content
23
+ */
24
+ export interface CohereThinkingContentV2 extends model.CohereContentV2 {
25
+ /**
26
+ * The thinking content.
27
+ */
28
+ "thinking"?: string;
29
+ "type": string;
30
+ }
31
+ export declare namespace CohereThinkingContentV2 {
32
+ function getJsonObj(obj: CohereThinkingContentV2, isParentJsonObj?: boolean): object;
33
+ const type = "THINKING";
34
+ function getDeserializedJsonObj(obj: CohereThinkingContentV2, isParentJsonObj?: boolean): object;
35
+ }
@@ -0,0 +1,61 @@
1
+ "use strict";
2
+ /**
3
+ * Generative AI Service Inference API
4
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
5
+
6
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
7
+
8
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
9
+
10
+ To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
11
+
12
+ * OpenAPI spec version: 20231130
13
+ *
14
+ *
15
+ * NOTE: This class is auto generated by OracleSDKGenerator.
16
+ * Do not edit the class manually.
17
+ *
18
+ * Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
19
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
20
+ */
21
+ var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
22
+ if (k2 === undefined) k2 = k;
23
+ Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
24
+ }) : (function(o, m, k, k2) {
25
+ if (k2 === undefined) k2 = k;
26
+ o[k2] = m[k];
27
+ }));
28
+ var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
29
+ Object.defineProperty(o, "default", { enumerable: true, value: v });
30
+ }) : function(o, v) {
31
+ o["default"] = v;
32
+ });
33
+ var __importStar = (this && this.__importStar) || function (mod) {
34
+ if (mod && mod.__esModule) return mod;
35
+ var result = {};
36
+ if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
37
+ __setModuleDefault(result, mod);
38
+ return result;
39
+ };
40
+ Object.defineProperty(exports, "__esModule", { value: true });
41
+ exports.CohereThinkingContentV2 = void 0;
42
+ const model = __importStar(require("../model"));
43
+ var CohereThinkingContentV2;
44
+ (function (CohereThinkingContentV2) {
45
+ function getJsonObj(obj, isParentJsonObj) {
46
+ const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj
47
+ ? obj
48
+ : model.CohereContentV2.getJsonObj(obj))), {});
49
+ return jsonObj;
50
+ }
51
+ CohereThinkingContentV2.getJsonObj = getJsonObj;
52
+ CohereThinkingContentV2.type = "THINKING";
53
+ function getDeserializedJsonObj(obj, isParentJsonObj) {
54
+ const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj
55
+ ? obj
56
+ : model.CohereContentV2.getDeserializedJsonObj(obj))), {});
57
+ return jsonObj;
58
+ }
59
+ CohereThinkingContentV2.getDeserializedJsonObj = getDeserializedJsonObj;
60
+ })(CohereThinkingContentV2 = exports.CohereThinkingContentV2 || (exports.CohereThinkingContentV2 = {}));
61
+ //# sourceMappingURL=cohere-thinking-content-v2.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"cohere-thinking-content-v2.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/cohere-thinking-content-v2.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAelC,IAAiB,uBAAuB,CAyBvC;AAzBD,WAAiB,uBAAuB;IACtC,SAAgB,UAAU,CAAC,GAA4B,EAAE,eAAyB;QAChF,MAAM,OAAO,mCACR,CAAC,eAAe;YACjB,CAAC,CAAC,GAAG;YACL,CAAC,CAAE,KAAK,CAAC,eAAe,CAAC,UAAU,CAAC,GAAG,CAA6B,CAAC,GACpE,EAAE,CACN,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IATe,kCAAU,aASzB,CAAA;IACY,4BAAI,GAAG,UAAU,CAAC;IAC/B,SAAgB,sBAAsB,CACpC,GAA4B,EAC5B,eAAyB;QAEzB,MAAM,OAAO,mCACR,CAAC,eAAe;YACjB,CAAC,CAAC,GAAG;YACL,CAAC,CAAE,KAAK,CAAC,eAAe,CAAC,sBAAsB,CAAC,GAAG,CAA6B,CAAC,GAChF,EAAE,CACN,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAZe,8CAAsB,yBAYrC,CAAA;AACH,CAAC,EAzBgB,uBAAuB,GAAvB,+BAAuB,KAAvB,+BAAuB,QAyBvC"}
@@ -0,0 +1,40 @@
1
+ /**
2
+ * Generative AI Service Inference API
3
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
4
+
5
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
6
+
7
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
8
+
9
+ To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
10
+
11
+ * OpenAPI spec version: 20231130
12
+ *
13
+ *
14
+ * NOTE: This class is auto generated by OracleSDKGenerator.
15
+ * Do not edit the class manually.
16
+ *
17
+ * Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
18
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
19
+ */
20
+ /**
21
+ * Configuration for reasoning features.
22
+ */
23
+ export interface CohereThinkingV2 {
24
+ /**
25
+ * Reasoning is enabled by default for models that support it, but can be turned off by setting type = disbaled.
26
+ */
27
+ "type": CohereThinkingV2.Type;
28
+ /**
29
+ * The maximum number of tokens the model can use for thinking, which must be set to a positive integer. The model will stop thinking if it reaches the thinking token budget and will proceed with the response. Note: Numbers greater than Number.MAX_SAFE_INTEGER will result in rounding issues.
30
+ */
31
+ "tokenBudget"?: number;
32
+ }
33
+ export declare namespace CohereThinkingV2 {
34
+ enum Type {
35
+ Enabled = "ENABLED",
36
+ Disabled = "DISABLED"
37
+ }
38
+ function getJsonObj(obj: CohereThinkingV2): object;
39
+ function getDeserializedJsonObj(obj: CohereThinkingV2): object;
40
+ }
@@ -0,0 +1,41 @@
1
+ "use strict";
2
+ /**
3
+ * Generative AI Service Inference API
4
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
5
+
6
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
7
+
8
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
9
+
10
+ To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
11
+
12
+ * OpenAPI spec version: 20231130
13
+ *
14
+ *
15
+ * NOTE: This class is auto generated by OracleSDKGenerator.
16
+ * Do not edit the class manually.
17
+ *
18
+ * Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
19
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
20
+ */
21
+ Object.defineProperty(exports, "__esModule", { value: true });
22
+ exports.CohereThinkingV2 = void 0;
23
+ var CohereThinkingV2;
24
+ (function (CohereThinkingV2) {
25
+ let Type;
26
+ (function (Type) {
27
+ Type["Enabled"] = "ENABLED";
28
+ Type["Disabled"] = "DISABLED";
29
+ })(Type = CohereThinkingV2.Type || (CohereThinkingV2.Type = {}));
30
+ function getJsonObj(obj) {
31
+ const jsonObj = Object.assign(Object.assign({}, obj), {});
32
+ return jsonObj;
33
+ }
34
+ CohereThinkingV2.getJsonObj = getJsonObj;
35
+ function getDeserializedJsonObj(obj) {
36
+ const jsonObj = Object.assign(Object.assign({}, obj), {});
37
+ return jsonObj;
38
+ }
39
+ CohereThinkingV2.getDeserializedJsonObj = getDeserializedJsonObj;
40
+ })(CohereThinkingV2 = exports.CohereThinkingV2 || (exports.CohereThinkingV2 = {}));
41
+ //# sourceMappingURL=cohere-thinking-v2.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"cohere-thinking-v2.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/cohere-thinking-v2.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;AAmBH,IAAiB,gBAAgB,CAgBhC;AAhBD,WAAiB,gBAAgB;IAC/B,IAAY,IAGX;IAHD,WAAY,IAAI;QACd,2BAAmB,CAAA;QACnB,6BAAqB,CAAA;IACvB,CAAC,EAHW,IAAI,GAAJ,qBAAI,KAAJ,qBAAI,QAGf;IAED,SAAgB,UAAU,CAAC,GAAqB;QAC9C,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,2BAAU,aAIzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAqB;QAC1D,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,uCAAsB,yBAIrC,CAAA;AACH,CAAC,EAhBgB,gBAAgB,GAAhB,wBAAgB,KAAhB,wBAAgB,QAgBhC"}
@@ -0,0 +1,48 @@
1
+ /**
2
+ * Generative AI Service Inference API
3
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
4
+
5
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
6
+
7
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
8
+
9
+ To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
10
+
11
+ * OpenAPI spec version: 20231130
12
+ *
13
+ *
14
+ * NOTE: This class is auto generated by OracleSDKGenerator.
15
+ * Do not edit the class manually.
16
+ *
17
+ * Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
18
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
19
+ */
20
+ /**
21
+ * A tool call generated by the model.
22
+ */
23
+ export interface CohereToolCallV2 {
24
+ /**
25
+ * Id of the tool call.
26
+ */
27
+ "id"?: string;
28
+ /**
29
+ * Type of a tool call.
30
+ */
31
+ "type"?: CohereToolCallV2.Type;
32
+ /**
33
+ * Function of a tool call.
34
+ */
35
+ "function"?: any;
36
+ }
37
+ export declare namespace CohereToolCallV2 {
38
+ enum Type {
39
+ Function = "FUNCTION",
40
+ /**
41
+ * This value is used if a service returns a value for this enum that is not recognized by this
42
+ * version of the SDK.
43
+ */
44
+ UnknownValue = "UNKNOWN_VALUE"
45
+ }
46
+ function getJsonObj(obj: CohereToolCallV2): object;
47
+ function getDeserializedJsonObj(obj: CohereToolCallV2): object;
48
+ }
@@ -0,0 +1,45 @@
1
+ "use strict";
2
+ /**
3
+ * Generative AI Service Inference API
4
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
5
+
6
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
7
+
8
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
9
+
10
+ To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
11
+
12
+ * OpenAPI spec version: 20231130
13
+ *
14
+ *
15
+ * NOTE: This class is auto generated by OracleSDKGenerator.
16
+ * Do not edit the class manually.
17
+ *
18
+ * Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
19
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
20
+ */
21
+ Object.defineProperty(exports, "__esModule", { value: true });
22
+ exports.CohereToolCallV2 = void 0;
23
+ var CohereToolCallV2;
24
+ (function (CohereToolCallV2) {
25
+ let Type;
26
+ (function (Type) {
27
+ Type["Function"] = "FUNCTION";
28
+ /**
29
+ * This value is used if a service returns a value for this enum that is not recognized by this
30
+ * version of the SDK.
31
+ */
32
+ Type["UnknownValue"] = "UNKNOWN_VALUE";
33
+ })(Type = CohereToolCallV2.Type || (CohereToolCallV2.Type = {}));
34
+ function getJsonObj(obj) {
35
+ const jsonObj = Object.assign(Object.assign({}, obj), {});
36
+ return jsonObj;
37
+ }
38
+ CohereToolCallV2.getJsonObj = getJsonObj;
39
+ function getDeserializedJsonObj(obj) {
40
+ const jsonObj = Object.assign(Object.assign({}, obj), {});
41
+ return jsonObj;
42
+ }
43
+ CohereToolCallV2.getDeserializedJsonObj = getDeserializedJsonObj;
44
+ })(CohereToolCallV2 = exports.CohereToolCallV2 || (exports.CohereToolCallV2 = {}));
45
+ //# sourceMappingURL=cohere-tool-call-v2.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"cohere-tool-call-v2.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/cohere-tool-call-v2.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;AAuBH,IAAiB,gBAAgB,CAoBhC;AApBD,WAAiB,gBAAgB;IAC/B,IAAY,IAOX;IAPD,WAAY,IAAI;QACd,6BAAqB,CAAA;QACrB;;;WAGG;QACH,sCAA8B,CAAA;IAChC,CAAC,EAPW,IAAI,GAAJ,qBAAI,KAAJ,qBAAI,QAOf;IAED,SAAgB,UAAU,CAAC,GAAqB;QAC9C,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,2BAAU,aAIzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAqB;QAC1D,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,uCAAsB,yBAIrC,CAAA;AACH,CAAC,EApBgB,gBAAgB,GAAhB,wBAAgB,KAAhB,wBAAgB,QAoBhC"}
@@ -0,0 +1,35 @@
1
+ /**
2
+ * Generative AI Service Inference API
3
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
4
+
5
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
6
+
7
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
8
+
9
+ To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
10
+
11
+ * OpenAPI spec version: 20231130
12
+ *
13
+ *
14
+ * NOTE: This class is auto generated by OracleSDKGenerator.
15
+ * Do not edit the class manually.
16
+ *
17
+ * Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
18
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
19
+ */
20
+ import * as model from "../model";
21
+ /**
22
+ * A message that represents a single chat dialog as TOOL role.
23
+ */
24
+ export interface CohereToolMessageV2 extends model.CohereMessageV2 {
25
+ /**
26
+ * The id of the associated tool call that has provided the given content
27
+ */
28
+ "toolCallId": string;
29
+ "role": string;
30
+ }
31
+ export declare namespace CohereToolMessageV2 {
32
+ function getJsonObj(obj: CohereToolMessageV2, isParentJsonObj?: boolean): object;
33
+ const role = "TOOL";
34
+ function getDeserializedJsonObj(obj: CohereToolMessageV2, isParentJsonObj?: boolean): object;
35
+ }
@@ -0,0 +1,59 @@
1
+ "use strict";
2
+ /**
3
+ * Generative AI Service Inference API
4
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
5
+
6
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
7
+
8
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
9
+
10
+ To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
11
+
12
+ * OpenAPI spec version: 20231130
13
+ *
14
+ *
15
+ * NOTE: This class is auto generated by OracleSDKGenerator.
16
+ * Do not edit the class manually.
17
+ *
18
+ * Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
19
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
20
+ */
21
+ var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
22
+ if (k2 === undefined) k2 = k;
23
+ Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
24
+ }) : (function(o, m, k, k2) {
25
+ if (k2 === undefined) k2 = k;
26
+ o[k2] = m[k];
27
+ }));
28
+ var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
29
+ Object.defineProperty(o, "default", { enumerable: true, value: v });
30
+ }) : function(o, v) {
31
+ o["default"] = v;
32
+ });
33
+ var __importStar = (this && this.__importStar) || function (mod) {
34
+ if (mod && mod.__esModule) return mod;
35
+ var result = {};
36
+ if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
37
+ __setModuleDefault(result, mod);
38
+ return result;
39
+ };
40
+ Object.defineProperty(exports, "__esModule", { value: true });
41
+ exports.CohereToolMessageV2 = void 0;
42
+ const model = __importStar(require("../model"));
43
+ var CohereToolMessageV2;
44
+ (function (CohereToolMessageV2) {
45
+ function getJsonObj(obj, isParentJsonObj) {
46
+ const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj ? obj : model.CohereMessageV2.getJsonObj(obj))), {});
47
+ return jsonObj;
48
+ }
49
+ CohereToolMessageV2.getJsonObj = getJsonObj;
50
+ CohereToolMessageV2.role = "TOOL";
51
+ function getDeserializedJsonObj(obj, isParentJsonObj) {
52
+ const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj
53
+ ? obj
54
+ : model.CohereMessageV2.getDeserializedJsonObj(obj))), {});
55
+ return jsonObj;
56
+ }
57
+ CohereToolMessageV2.getDeserializedJsonObj = getDeserializedJsonObj;
58
+ })(CohereToolMessageV2 = exports.CohereToolMessageV2 || (exports.CohereToolMessageV2 = {}));
59
+ //# sourceMappingURL=cohere-tool-message-v2.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"cohere-tool-message-v2.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/cohere-tool-message-v2.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAelC,IAAiB,mBAAmB,CAuBnC;AAvBD,WAAiB,mBAAmB;IAClC,SAAgB,UAAU,CAAC,GAAwB,EAAE,eAAyB;QAC5E,MAAM,OAAO,mCACR,CAAC,eAAe,CAAC,CAAC,CAAC,GAAG,CAAC,CAAC,CAAE,KAAK,CAAC,eAAe,CAAC,UAAU,CAAC,GAAG,CAAyB,CAAC,GACxF,EAAE,CACN,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAPe,8BAAU,aAOzB,CAAA;IACY,wBAAI,GAAG,MAAM,CAAC;IAC3B,SAAgB,sBAAsB,CACpC,GAAwB,EACxB,eAAyB;QAEzB,MAAM,OAAO,mCACR,CAAC,eAAe;YACjB,CAAC,CAAC,GAAG;YACL,CAAC,CAAE,KAAK,CAAC,eAAe,CAAC,sBAAsB,CAAC,GAAG,CAAyB,CAAC,GAC5E,EAAE,CACN,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAZe,0CAAsB,yBAYrC,CAAA;AACH,CAAC,EAvBgB,mBAAmB,GAAnB,2BAAmB,KAAnB,2BAAmB,QAuBnC"}
@@ -0,0 +1,37 @@
1
+ /**
2
+ * Generative AI Service Inference API
3
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
4
+
5
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
6
+
7
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
8
+
9
+ To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
10
+
11
+ * OpenAPI spec version: 20231130
12
+ *
13
+ *
14
+ * NOTE: This class is auto generated by OracleSDKGenerator.
15
+ * Do not edit the class manually.
16
+ *
17
+ * Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
18
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
19
+ */
20
+ import * as model from "../model";
21
+ /**
22
+ * A definition of tool (function).
23
+ */
24
+ export interface CohereToolV2 {
25
+ /**
26
+ * The name of the tool to be called. Valid names contain only the characters a-z, A-Z, 0-9, _ and must not begin with a digit.
27
+ */
28
+ "type": CohereToolV2.Type;
29
+ "function": model.Function;
30
+ }
31
+ export declare namespace CohereToolV2 {
32
+ enum Type {
33
+ Function = "FUNCTION"
34
+ }
35
+ function getJsonObj(obj: CohereToolV2): object;
36
+ function getDeserializedJsonObj(obj: CohereToolV2): object;
37
+ }