oci-generativeaiinference 2.123.0 → 2.123.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/lib/model/assistant-message.d.ts +4 -0
- package/lib/model/assistant-message.js.map +1 -1
- package/lib/model/base-chat-request.js +4 -0
- package/lib/model/base-chat-request.js.map +1 -1
- package/lib/model/base-chat-response.js +4 -0
- package/lib/model/base-chat-response.js.map +1 -1
- package/lib/model/chat-content.js +4 -0
- package/lib/model/chat-content.js.map +1 -1
- package/lib/model/chat-details.d.ts +1 -1
- package/lib/model/chat-result.d.ts +1 -1
- package/lib/model/citation-options-v2.d.ts +38 -0
- package/lib/model/citation-options-v2.js +42 -0
- package/lib/model/citation-options-v2.js.map +1 -0
- package/lib/model/cohere-assistant-message-v2.d.ts +43 -0
- package/lib/model/cohere-assistant-message-v2.js +83 -0
- package/lib/model/cohere-assistant-message-v2.js.map +1 -0
- package/lib/model/cohere-chat-request-v2.d.ts +142 -0
- package/lib/model/cohere-chat-request-v2.js +114 -0
- package/lib/model/cohere-chat-request-v2.js.map +1 -0
- package/lib/model/cohere-chat-response-v2.d.ts +61 -0
- package/lib/model/cohere-chat-response-v2.js +90 -0
- package/lib/model/cohere-chat-response-v2.js.map +1 -0
- package/lib/model/cohere-citation-document.d.ts +36 -0
- package/lib/model/cohere-citation-document.js +36 -0
- package/lib/model/cohere-citation-document.js.map +1 -0
- package/lib/model/cohere-citation-source-v2.d.ts +44 -0
- package/lib/model/cohere-citation-source-v2.js +74 -0
- package/lib/model/cohere-citation-source-v2.js.map +1 -0
- package/lib/model/cohere-citation-tool-v2.d.ts +36 -0
- package/lib/model/cohere-citation-tool-v2.js +36 -0
- package/lib/model/cohere-citation-tool-v2.js.map +1 -0
- package/lib/model/cohere-citation-v2.d.ts +63 -0
- package/lib/model/cohere-citation-v2.js +79 -0
- package/lib/model/cohere-citation-v2.js.map +1 -0
- package/lib/model/cohere-content-v2.d.ts +29 -0
- package/lib/model/cohere-content-v2.js +87 -0
- package/lib/model/cohere-content-v2.js.map +1 -0
- package/lib/model/cohere-document-content-v2.d.ts +35 -0
- package/lib/model/cohere-document-content-v2.js +61 -0
- package/lib/model/cohere-document-content-v2.js.map +1 -0
- package/lib/model/cohere-image-content-v2.d.ts +32 -0
- package/lib/model/cohere-image-content-v2.js +65 -0
- package/lib/model/cohere-image-content-v2.js.map +1 -0
- package/lib/model/cohere-image-url-v2.d.ts +46 -0
- package/lib/model/cohere-image-url-v2.js +47 -0
- package/lib/model/cohere-image-url-v2.js.map +1 -0
- package/lib/model/cohere-message-v2.d.ts +34 -0
- package/lib/model/cohere-message-v2.js +99 -0
- package/lib/model/cohere-message-v2.js.map +1 -0
- package/lib/model/cohere-system-message-v2.d.ts +31 -0
- package/lib/model/cohere-system-message-v2.js +59 -0
- package/lib/model/cohere-system-message-v2.js.map +1 -0
- package/lib/model/cohere-text-content-v2.d.ts +35 -0
- package/lib/model/cohere-text-content-v2.js +59 -0
- package/lib/model/cohere-text-content-v2.js.map +1 -0
- package/lib/model/cohere-thinking-content-v2.d.ts +35 -0
- package/lib/model/cohere-thinking-content-v2.js +61 -0
- package/lib/model/cohere-thinking-content-v2.js.map +1 -0
- package/lib/model/cohere-thinking-v2.d.ts +40 -0
- package/lib/model/cohere-thinking-v2.js +41 -0
- package/lib/model/cohere-thinking-v2.js.map +1 -0
- package/lib/model/cohere-tool-call-v2.d.ts +48 -0
- package/lib/model/cohere-tool-call-v2.js +45 -0
- package/lib/model/cohere-tool-call-v2.js.map +1 -0
- package/lib/model/cohere-tool-message-v2.d.ts +35 -0
- package/lib/model/cohere-tool-message-v2.js +59 -0
- package/lib/model/cohere-tool-message-v2.js.map +1 -0
- package/lib/model/cohere-tool-v2.d.ts +37 -0
- package/lib/model/cohere-tool-v2.js +64 -0
- package/lib/model/cohere-tool-v2.js.map +1 -0
- package/lib/model/cohere-user-message-v2.d.ts +31 -0
- package/lib/model/cohere-user-message-v2.js +59 -0
- package/lib/model/cohere-user-message-v2.js.map +1 -0
- package/lib/model/document-content.d.ts +32 -0
- package/lib/model/document-content.js +65 -0
- package/lib/model/document-content.js.map +1 -0
- package/lib/model/document-url.d.ts +63 -0
- package/lib/model/document-url.js +47 -0
- package/lib/model/document-url.js.map +1 -0
- package/lib/model/embed-text-details.d.ts +16 -0
- package/lib/model/embed-text-details.js +9 -0
- package/lib/model/embed-text-details.js.map +1 -1
- package/lib/model/embed-text-result.d.ts +5 -1
- package/lib/model/embed-text-result.js.map +1 -1
- package/lib/model/function.d.ts +40 -0
- package/lib/model/function.js +36 -0
- package/lib/model/function.js.map +1 -0
- package/lib/model/generic-chat-request.d.ts +1 -0
- package/lib/model/generic-chat-request.js +1 -0
- package/lib/model/generic-chat-request.js.map +1 -1
- package/lib/model/index.d.ts +50 -0
- package/lib/model/index.js +53 -3
- package/lib/model/index.js.map +1 -1
- package/lib/model/log-probability.d.ts +40 -0
- package/lib/model/log-probability.js +36 -0
- package/lib/model/log-probability.js.map +1 -0
- package/lib/model/rerank-text-details.d.ts +4 -0
- package/lib/model/rerank-text-details.js.map +1 -1
- package/package.json +3 -3
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"cohere-chat-request-v2.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/cohere-chat-request-v2.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAiHlC,IAAiB,mBAAmB,CAsFnC;AAtFD,WAAiB,mBAAmB;IAClC,IAAY,WAGX;IAHD,WAAY,WAAW;QACrB,oCAAqB,CAAA;QACrB,4BAAa,CAAA;IACf,CAAC,EAHW,WAAW,GAAX,+BAAW,KAAX,+BAAW,QAGtB;IAED,IAAY,UAIX;IAJD,WAAY,UAAU;QACpB,uCAAyB,CAAA;QACzB,+BAAiB,CAAA;QACjB,yBAAW,CAAA;IACb,CAAC,EAJW,UAAU,GAAV,8BAAU,KAAV,8BAAU,QAIrB;IAED,SAAgB,UAAU,CAAC,GAAwB,EAAE,eAAyB;QAC5E,MAAM,OAAO,mCACR,CAAC,eAAe,CAAC,CAAC,CAAC,GAAG,CAAC,CAAC,CAAE,KAAK,CAAC,eAAe,CAAC,UAAU,CAAC,GAAG,CAAyB,CAAC,GACxF;YACD,UAAU,EAAE,GAAG,CAAC,QAAQ;gBACtB,CAAC,CAAC,GAAG,CAAC,QAAQ,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACtB,OAAO,KAAK,CAAC,eAAe,CAAC,UAAU,CAAC,IAAI,CAAC,CAAC;gBAChD,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;YAEb,iBAAiB,EAAE,GAAG,CAAC,eAAe;gBACpC,CAAC,CAAC,KAAK,CAAC,iBAAiB,CAAC,UAAU,CAAC,GAAG,CAAC,eAAe,CAAC;gBACzD,CAAC,CAAC,SAAS;YAEb,OAAO,EAAE,GAAG,CAAC,KAAK;gBAChB,CAAC,CAAC,GAAG,CAAC,KAAK,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACnB,OAAO,KAAK,CAAC,YAAY,CAAC,UAAU,CAAC,IAAI,CAAC,CAAC;gBAC7C,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;YAEb,UAAU,EAAE,GAAG,CAAC,QAAQ,CAAC,CAAC,CAAC,KAAK,CAAC,gBAAgB,CAAC,UAAU,CAAC,GAAG,CAAC,QAAQ,CAAC,CAAC,CAAC,CAAC,SAAS;YACtF,gBAAgB,EAAE,GAAG,CAAC,cAAc;gBAClC,CAAC,CAAC,KAAK,CAAC,oBAAoB,CAAC,UAAU,CAAC,GAAG,CAAC,cAAc,CAAC;gBAC3D,CAAC,CAAC,SAAS;YAEb,eAAe,EAAE,GAAG,CAAC,aAAa;gBAChC,CAAC,CAAC,KAAK,CAAC,aAAa,CAAC,UAAU,CAAC,GAAG,CAAC,aAAa,CAAC;gBACnD,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAhCe,8BAAU,aAgCzB,CAAA;IACY,6BAAS,GAAG,UAAU,CAAC;IACpC,SAAgB,sBAAsB,CACpC,GAAwB,EACxB,eAAyB;QAEzB,MAAM,OAAO,mCACR,CAAC,eAAe;YACjB,CAAC,CAAC,GAAG;YACL,CAAC,CAAE,KAAK,CAAC,eAAe,CAAC,sBAAsB,CAAC,GAAG,CAAyB,CAAC,GAC5E;YACD,UAAU,EAAE,GAAG,CAAC,QAAQ;gBACtB,CAAC,CAAC,GAAG,CAAC,QAAQ,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACtB,OAAO,KAAK,CAAC,eAAe,CAAC,sBAAsB,CAAC,IAAI,CAAC,CAAC;gBAC5D,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;YAEb,iBAAiB,EAAE,GAAG,CAAC,eAAe;gBACpC,CAAC,CAAC,KAAK,CAAC,iBAAiB,CAAC,sBAAsB,CAAC,GAAG,CAAC,eAAe,CAAC;gBACrE,CAAC,CAAC,SAAS;YAEb,OAAO,EAAE,GAAG,CAAC,KAAK;gBAChB,CAAC,CAAC,GAAG,CAAC,KAAK,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACnB,OAAO,KAAK,CAAC,YAAY,CAAC,sBAAsB,CAAC,IAAI,CAAC,CAAC;gBACzD,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;YAEb,UAAU,EAAE,GAAG,CAAC,QAAQ;gBACtB,CAAC,CAAC,KAAK,CAAC,gBAAgB,CAAC,sBAAsB,CAAC,GAAG,CAAC,QAAQ,CAAC;gBAC7D,CAAC,CAAC,SAAS;YACb,gBAAgB,EAAE,GAAG,CAAC,cAAc;gBAClC,CAAC,CAAC,KAAK,CAAC,oBAAoB,CAAC,sBAAsB,CAAC,GAAG,CAAC,cAAc,CAAC;gBACvE,CAAC,CAAC,SAAS;YAEb,eAAe,EAAE,GAAG,CAAC,aAAa;gBAChC,CAAC,CAAC,KAAK,CAAC,aAAa,CAAC,sBAAsB,CAAC,GAAG,CAAC,aAAa,CAAC;gBAC/D,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAvCe,0CAAsB,yBAuCrC,CAAA;AACH,CAAC,EAtFgB,mBAAmB,GAAnB,2BAAmB,KAAnB,2BAAmB,QAsFnC"}
|
|
@@ -0,0 +1,61 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
import * as model from "../model";
|
|
21
|
+
/**
|
|
22
|
+
* The response to the chat conversation.
|
|
23
|
+
*/
|
|
24
|
+
export interface CohereChatResponseV2 extends model.BaseChatResponse {
|
|
25
|
+
/**
|
|
26
|
+
* Unique identifier for the generated reply
|
|
27
|
+
*/
|
|
28
|
+
"id": string;
|
|
29
|
+
"message": model.CohereAssistantMessageV2;
|
|
30
|
+
/**
|
|
31
|
+
* Why the generation stopped.
|
|
32
|
+
*/
|
|
33
|
+
"finishReason": CohereChatResponseV2.FinishReason;
|
|
34
|
+
/**
|
|
35
|
+
* The log probabilities of the generated tokens.
|
|
36
|
+
*/
|
|
37
|
+
"logProbabilities"?: Array<model.LogProbability>;
|
|
38
|
+
/**
|
|
39
|
+
* If there is an error during the streaming scenario, then the {@code errorMessage} parameter contains details for the error.
|
|
40
|
+
*/
|
|
41
|
+
"errorMessage"?: string;
|
|
42
|
+
"usage"?: model.Usage;
|
|
43
|
+
"apiFormat": string;
|
|
44
|
+
}
|
|
45
|
+
export declare namespace CohereChatResponseV2 {
|
|
46
|
+
enum FinishReason {
|
|
47
|
+
Complete = "COMPLETE",
|
|
48
|
+
StopSequence = "STOP_SEQUENCE",
|
|
49
|
+
MaxTokens = "MAX_TOKENS",
|
|
50
|
+
ToolCall = "TOOL_CALL",
|
|
51
|
+
Error = "ERROR",
|
|
52
|
+
/**
|
|
53
|
+
* This value is used if a service returns a value for this enum that is not recognized by this
|
|
54
|
+
* version of the SDK.
|
|
55
|
+
*/
|
|
56
|
+
UnknownValue = "UNKNOWN_VALUE"
|
|
57
|
+
}
|
|
58
|
+
function getJsonObj(obj: CohereChatResponseV2, isParentJsonObj?: boolean): object;
|
|
59
|
+
const apiFormat = "COHEREV2";
|
|
60
|
+
function getDeserializedJsonObj(obj: CohereChatResponseV2, isParentJsonObj?: boolean): object;
|
|
61
|
+
}
|
|
@@ -0,0 +1,90 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
22
|
+
if (k2 === undefined) k2 = k;
|
|
23
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
24
|
+
}) : (function(o, m, k, k2) {
|
|
25
|
+
if (k2 === undefined) k2 = k;
|
|
26
|
+
o[k2] = m[k];
|
|
27
|
+
}));
|
|
28
|
+
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
29
|
+
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
30
|
+
}) : function(o, v) {
|
|
31
|
+
o["default"] = v;
|
|
32
|
+
});
|
|
33
|
+
var __importStar = (this && this.__importStar) || function (mod) {
|
|
34
|
+
if (mod && mod.__esModule) return mod;
|
|
35
|
+
var result = {};
|
|
36
|
+
if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
37
|
+
__setModuleDefault(result, mod);
|
|
38
|
+
return result;
|
|
39
|
+
};
|
|
40
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
41
|
+
exports.CohereChatResponseV2 = void 0;
|
|
42
|
+
const model = __importStar(require("../model"));
|
|
43
|
+
var CohereChatResponseV2;
|
|
44
|
+
(function (CohereChatResponseV2) {
|
|
45
|
+
let FinishReason;
|
|
46
|
+
(function (FinishReason) {
|
|
47
|
+
FinishReason["Complete"] = "COMPLETE";
|
|
48
|
+
FinishReason["StopSequence"] = "STOP_SEQUENCE";
|
|
49
|
+
FinishReason["MaxTokens"] = "MAX_TOKENS";
|
|
50
|
+
FinishReason["ToolCall"] = "TOOL_CALL";
|
|
51
|
+
FinishReason["Error"] = "ERROR";
|
|
52
|
+
/**
|
|
53
|
+
* This value is used if a service returns a value for this enum that is not recognized by this
|
|
54
|
+
* version of the SDK.
|
|
55
|
+
*/
|
|
56
|
+
FinishReason["UnknownValue"] = "UNKNOWN_VALUE";
|
|
57
|
+
})(FinishReason = CohereChatResponseV2.FinishReason || (CohereChatResponseV2.FinishReason = {}));
|
|
58
|
+
function getJsonObj(obj, isParentJsonObj) {
|
|
59
|
+
const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj ? obj : model.BaseChatResponse.getJsonObj(obj))), {
|
|
60
|
+
"message": obj.message ? model.CohereAssistantMessageV2.getJsonObj(obj.message) : undefined,
|
|
61
|
+
"logProbabilities": obj.logProbabilities
|
|
62
|
+
? obj.logProbabilities.map(item => {
|
|
63
|
+
return model.LogProbability.getJsonObj(item);
|
|
64
|
+
})
|
|
65
|
+
: undefined,
|
|
66
|
+
"usage": obj.usage ? model.Usage.getJsonObj(obj.usage) : undefined
|
|
67
|
+
});
|
|
68
|
+
return jsonObj;
|
|
69
|
+
}
|
|
70
|
+
CohereChatResponseV2.getJsonObj = getJsonObj;
|
|
71
|
+
CohereChatResponseV2.apiFormat = "COHEREV2";
|
|
72
|
+
function getDeserializedJsonObj(obj, isParentJsonObj) {
|
|
73
|
+
const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj
|
|
74
|
+
? obj
|
|
75
|
+
: model.BaseChatResponse.getDeserializedJsonObj(obj))), {
|
|
76
|
+
"message": obj.message
|
|
77
|
+
? model.CohereAssistantMessageV2.getDeserializedJsonObj(obj.message)
|
|
78
|
+
: undefined,
|
|
79
|
+
"logProbabilities": obj.logProbabilities
|
|
80
|
+
? obj.logProbabilities.map(item => {
|
|
81
|
+
return model.LogProbability.getDeserializedJsonObj(item);
|
|
82
|
+
})
|
|
83
|
+
: undefined,
|
|
84
|
+
"usage": obj.usage ? model.Usage.getDeserializedJsonObj(obj.usage) : undefined
|
|
85
|
+
});
|
|
86
|
+
return jsonObj;
|
|
87
|
+
}
|
|
88
|
+
CohereChatResponseV2.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
89
|
+
})(CohereChatResponseV2 = exports.CohereChatResponseV2 || (exports.CohereChatResponseV2 = {}));
|
|
90
|
+
//# sourceMappingURL=cohere-chat-response-v2.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"cohere-chat-response-v2.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/cohere-chat-response-v2.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AA6BlC,IAAiB,oBAAoB,CA0DpC;AA1DD,WAAiB,oBAAoB;IACnC,IAAY,YAWX;IAXD,WAAY,YAAY;QACtB,qCAAqB,CAAA;QACrB,8CAA8B,CAAA;QAC9B,wCAAwB,CAAA;QACxB,sCAAsB,CAAA;QACtB,+BAAe,CAAA;QACf;;;WAGG;QACH,8CAA8B,CAAA;IAChC,CAAC,EAXW,YAAY,GAAZ,iCAAY,KAAZ,iCAAY,QAWvB;IAED,SAAgB,UAAU,CAAC,GAAyB,EAAE,eAAyB;QAC7E,MAAM,OAAO,mCACR,CAAC,eAAe,CAAC,CAAC,CAAC,GAAG,CAAC,CAAC,CAAE,KAAK,CAAC,gBAAgB,CAAC,UAAU,CAAC,GAAG,CAA0B,CAAC,GAC1F;YACD,SAAS,EAAE,GAAG,CAAC,OAAO,CAAC,CAAC,CAAC,KAAK,CAAC,wBAAwB,CAAC,UAAU,CAAC,GAAG,CAAC,OAAO,CAAC,CAAC,CAAC,CAAC,SAAS;YAE3F,kBAAkB,EAAE,GAAG,CAAC,gBAAgB;gBACtC,CAAC,CAAC,GAAG,CAAC,gBAAgB,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBAC9B,OAAO,KAAK,CAAC,cAAc,CAAC,UAAU,CAAC,IAAI,CAAC,CAAC;gBAC/C,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;YAEb,OAAO,EAAE,GAAG,CAAC,KAAK,CAAC,CAAC,CAAC,KAAK,CAAC,KAAK,CAAC,UAAU,CAAC,GAAG,CAAC,KAAK,CAAC,CAAC,CAAC,CAAC,SAAS;SACnE,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAjBe,+BAAU,aAiBzB,CAAA;IACY,8BAAS,GAAG,UAAU,CAAC;IACpC,SAAgB,sBAAsB,CACpC,GAAyB,EACzB,eAAyB;QAEzB,MAAM,OAAO,mCACR,CAAC,eAAe;YACjB,CAAC,CAAC,GAAG;YACL,CAAC,CAAE,KAAK,CAAC,gBAAgB,CAAC,sBAAsB,CAAC,GAAG,CAA0B,CAAC,GAC9E;YACD,SAAS,EAAE,GAAG,CAAC,OAAO;gBACpB,CAAC,CAAC,KAAK,CAAC,wBAAwB,CAAC,sBAAsB,CAAC,GAAG,CAAC,OAAO,CAAC;gBACpE,CAAC,CAAC,SAAS;YAEb,kBAAkB,EAAE,GAAG,CAAC,gBAAgB;gBACtC,CAAC,CAAC,GAAG,CAAC,gBAAgB,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBAC9B,OAAO,KAAK,CAAC,cAAc,CAAC,sBAAsB,CAAC,IAAI,CAAC,CAAC;gBAC3D,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;YAEb,OAAO,EAAE,GAAG,CAAC,KAAK,CAAC,CAAC,CAAC,KAAK,CAAC,KAAK,CAAC,sBAAsB,CAAC,GAAG,CAAC,KAAK,CAAC,CAAC,CAAC,CAAC,SAAS;SAC/E,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAxBe,2CAAsB,yBAwBrC,CAAA;AACH,CAAC,EA1DgB,oBAAoB,GAApB,4BAAoB,KAApB,4BAAoB,QA0DpC"}
|
|
@@ -0,0 +1,36 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
/**
|
|
21
|
+
* A citation source referencing a document.
|
|
22
|
+
*/
|
|
23
|
+
export interface CohereCitationDocument {
|
|
24
|
+
/**
|
|
25
|
+
* Unique identifier for the document.
|
|
26
|
+
*/
|
|
27
|
+
"id"?: string;
|
|
28
|
+
/**
|
|
29
|
+
* The actual document content or reference.
|
|
30
|
+
*/
|
|
31
|
+
"document"?: any;
|
|
32
|
+
}
|
|
33
|
+
export declare namespace CohereCitationDocument {
|
|
34
|
+
function getJsonObj(obj: CohereCitationDocument): object;
|
|
35
|
+
function getDeserializedJsonObj(obj: CohereCitationDocument): object;
|
|
36
|
+
}
|
|
@@ -0,0 +1,36 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
22
|
+
exports.CohereCitationDocument = void 0;
|
|
23
|
+
var CohereCitationDocument;
|
|
24
|
+
(function (CohereCitationDocument) {
|
|
25
|
+
function getJsonObj(obj) {
|
|
26
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
27
|
+
return jsonObj;
|
|
28
|
+
}
|
|
29
|
+
CohereCitationDocument.getJsonObj = getJsonObj;
|
|
30
|
+
function getDeserializedJsonObj(obj) {
|
|
31
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
32
|
+
return jsonObj;
|
|
33
|
+
}
|
|
34
|
+
CohereCitationDocument.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
35
|
+
})(CohereCitationDocument = exports.CohereCitationDocument || (exports.CohereCitationDocument = {}));
|
|
36
|
+
//# sourceMappingURL=cohere-citation-document.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"cohere-citation-document.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/cohere-citation-document.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;AAmBH,IAAiB,sBAAsB,CAWtC;AAXD,WAAiB,sBAAsB;IACrC,SAAgB,UAAU,CAAC,GAA2B;QACpD,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,iCAAU,aAIzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAA2B;QAChE,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,6CAAsB,yBAIrC,CAAA;AACH,CAAC,EAXgB,sBAAsB,GAAtB,8BAAsB,KAAtB,8BAAsB,QAWtC"}
|
|
@@ -0,0 +1,44 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
import * as model from "../model";
|
|
21
|
+
/**
|
|
22
|
+
* A citation source object.
|
|
23
|
+
*/
|
|
24
|
+
export interface CohereCitationSourceV2 {
|
|
25
|
+
/**
|
|
26
|
+
* The source type.
|
|
27
|
+
*/
|
|
28
|
+
"type": CohereCitationSourceV2.Type;
|
|
29
|
+
"tool"?: model.CohereCitationToolV2;
|
|
30
|
+
"document"?: model.CohereCitationDocument;
|
|
31
|
+
}
|
|
32
|
+
export declare namespace CohereCitationSourceV2 {
|
|
33
|
+
enum Type {
|
|
34
|
+
Tool = "TOOL",
|
|
35
|
+
Document = "DOCUMENT",
|
|
36
|
+
/**
|
|
37
|
+
* This value is used if a service returns a value for this enum that is not recognized by this
|
|
38
|
+
* version of the SDK.
|
|
39
|
+
*/
|
|
40
|
+
UnknownValue = "UNKNOWN_VALUE"
|
|
41
|
+
}
|
|
42
|
+
function getJsonObj(obj: CohereCitationSourceV2): object;
|
|
43
|
+
function getDeserializedJsonObj(obj: CohereCitationSourceV2): object;
|
|
44
|
+
}
|
|
@@ -0,0 +1,74 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
22
|
+
if (k2 === undefined) k2 = k;
|
|
23
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
24
|
+
}) : (function(o, m, k, k2) {
|
|
25
|
+
if (k2 === undefined) k2 = k;
|
|
26
|
+
o[k2] = m[k];
|
|
27
|
+
}));
|
|
28
|
+
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
29
|
+
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
30
|
+
}) : function(o, v) {
|
|
31
|
+
o["default"] = v;
|
|
32
|
+
});
|
|
33
|
+
var __importStar = (this && this.__importStar) || function (mod) {
|
|
34
|
+
if (mod && mod.__esModule) return mod;
|
|
35
|
+
var result = {};
|
|
36
|
+
if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
37
|
+
__setModuleDefault(result, mod);
|
|
38
|
+
return result;
|
|
39
|
+
};
|
|
40
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
41
|
+
exports.CohereCitationSourceV2 = void 0;
|
|
42
|
+
const model = __importStar(require("../model"));
|
|
43
|
+
var CohereCitationSourceV2;
|
|
44
|
+
(function (CohereCitationSourceV2) {
|
|
45
|
+
let Type;
|
|
46
|
+
(function (Type) {
|
|
47
|
+
Type["Tool"] = "TOOL";
|
|
48
|
+
Type["Document"] = "DOCUMENT";
|
|
49
|
+
/**
|
|
50
|
+
* This value is used if a service returns a value for this enum that is not recognized by this
|
|
51
|
+
* version of the SDK.
|
|
52
|
+
*/
|
|
53
|
+
Type["UnknownValue"] = "UNKNOWN_VALUE";
|
|
54
|
+
})(Type = CohereCitationSourceV2.Type || (CohereCitationSourceV2.Type = {}));
|
|
55
|
+
function getJsonObj(obj) {
|
|
56
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
57
|
+
"tool": obj.tool ? model.CohereCitationToolV2.getJsonObj(obj.tool) : undefined,
|
|
58
|
+
"document": obj.document ? model.CohereCitationDocument.getJsonObj(obj.document) : undefined
|
|
59
|
+
});
|
|
60
|
+
return jsonObj;
|
|
61
|
+
}
|
|
62
|
+
CohereCitationSourceV2.getJsonObj = getJsonObj;
|
|
63
|
+
function getDeserializedJsonObj(obj) {
|
|
64
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
65
|
+
"tool": obj.tool ? model.CohereCitationToolV2.getDeserializedJsonObj(obj.tool) : undefined,
|
|
66
|
+
"document": obj.document
|
|
67
|
+
? model.CohereCitationDocument.getDeserializedJsonObj(obj.document)
|
|
68
|
+
: undefined
|
|
69
|
+
});
|
|
70
|
+
return jsonObj;
|
|
71
|
+
}
|
|
72
|
+
CohereCitationSourceV2.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
73
|
+
})(CohereCitationSourceV2 = exports.CohereCitationSourceV2 || (exports.CohereCitationSourceV2 = {}));
|
|
74
|
+
//# sourceMappingURL=cohere-citation-source-v2.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"cohere-citation-source-v2.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/cohere-citation-source-v2.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAelC,IAAiB,sBAAsB,CAmCtC;AAnCD,WAAiB,sBAAsB;IACrC,IAAY,IAQX;IARD,WAAY,IAAI;QACd,qBAAa,CAAA;QACb,6BAAqB,CAAA;QACrB;;;WAGG;QACH,sCAA8B,CAAA;IAChC,CAAC,EARW,IAAI,GAAJ,2BAAI,KAAJ,2BAAI,QAQf;IAED,SAAgB,UAAU,CAAC,GAA2B;QACpD,MAAM,OAAO,mCACR,GAAG,GACH;YACD,MAAM,EAAE,GAAG,CAAC,IAAI,CAAC,CAAC,CAAC,KAAK,CAAC,oBAAoB,CAAC,UAAU,CAAC,GAAG,CAAC,IAAI,CAAC,CAAC,CAAC,CAAC,SAAS;YAC9E,UAAU,EAAE,GAAG,CAAC,QAAQ,CAAC,CAAC,CAAC,KAAK,CAAC,sBAAsB,CAAC,UAAU,CAAC,GAAG,CAAC,QAAQ,CAAC,CAAC,CAAC,CAAC,SAAS;SAC7F,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAVe,iCAAU,aAUzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAA2B;QAChE,MAAM,OAAO,mCACR,GAAG,GACH;YACD,MAAM,EAAE,GAAG,CAAC,IAAI,CAAC,CAAC,CAAC,KAAK,CAAC,oBAAoB,CAAC,sBAAsB,CAAC,GAAG,CAAC,IAAI,CAAC,CAAC,CAAC,CAAC,SAAS;YAC1F,UAAU,EAAE,GAAG,CAAC,QAAQ;gBACtB,CAAC,CAAC,KAAK,CAAC,sBAAsB,CAAC,sBAAsB,CAAC,GAAG,CAAC,QAAQ,CAAC;gBACnE,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAZe,6CAAsB,yBAYrC,CAAA;AACH,CAAC,EAnCgB,sBAAsB,GAAtB,8BAAsB,KAAtB,8BAAsB,QAmCtC"}
|
|
@@ -0,0 +1,36 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
/**
|
|
21
|
+
* A citation source referencing a tool result.
|
|
22
|
+
*/
|
|
23
|
+
export interface CohereCitationToolV2 {
|
|
24
|
+
/**
|
|
25
|
+
* Unique identifier of the tool call.
|
|
26
|
+
*/
|
|
27
|
+
"id"?: string;
|
|
28
|
+
/**
|
|
29
|
+
* Output from the tool.
|
|
30
|
+
*/
|
|
31
|
+
"toolOutput"?: any;
|
|
32
|
+
}
|
|
33
|
+
export declare namespace CohereCitationToolV2 {
|
|
34
|
+
function getJsonObj(obj: CohereCitationToolV2): object;
|
|
35
|
+
function getDeserializedJsonObj(obj: CohereCitationToolV2): object;
|
|
36
|
+
}
|
|
@@ -0,0 +1,36 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
22
|
+
exports.CohereCitationToolV2 = void 0;
|
|
23
|
+
var CohereCitationToolV2;
|
|
24
|
+
(function (CohereCitationToolV2) {
|
|
25
|
+
function getJsonObj(obj) {
|
|
26
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
27
|
+
return jsonObj;
|
|
28
|
+
}
|
|
29
|
+
CohereCitationToolV2.getJsonObj = getJsonObj;
|
|
30
|
+
function getDeserializedJsonObj(obj) {
|
|
31
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
32
|
+
return jsonObj;
|
|
33
|
+
}
|
|
34
|
+
CohereCitationToolV2.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
35
|
+
})(CohereCitationToolV2 = exports.CohereCitationToolV2 || (exports.CohereCitationToolV2 = {}));
|
|
36
|
+
//# sourceMappingURL=cohere-citation-tool-v2.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"cohere-citation-tool-v2.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/cohere-citation-tool-v2.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;AAmBH,IAAiB,oBAAoB,CAWpC;AAXD,WAAiB,oBAAoB;IACnC,SAAgB,UAAU,CAAC,GAAyB;QAClD,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,+BAAU,aAIzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAyB;QAC9D,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,2CAAsB,yBAIrC,CAAA;AACH,CAAC,EAXgB,oBAAoB,GAApB,4BAAoB,KAApB,4BAAoB,QAWpC"}
|
|
@@ -0,0 +1,63 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
import * as model from "../model";
|
|
21
|
+
/**
|
|
22
|
+
* A source reference or citation for a piece of content.
|
|
23
|
+
*/
|
|
24
|
+
export interface CohereCitationV2 {
|
|
25
|
+
/**
|
|
26
|
+
* Start index of the cited snippet in the original source text. Note: Numbers greater than Number.MAX_SAFE_INTEGER will result in rounding issues.
|
|
27
|
+
*/
|
|
28
|
+
"start"?: number;
|
|
29
|
+
/**
|
|
30
|
+
* End index of the cited snippet in the original source text. Note: Numbers greater than Number.MAX_SAFE_INTEGER will result in rounding issues.
|
|
31
|
+
*/
|
|
32
|
+
"end"?: number;
|
|
33
|
+
/**
|
|
34
|
+
* Text snippet that is being cited.
|
|
35
|
+
*/
|
|
36
|
+
"text"?: string;
|
|
37
|
+
/**
|
|
38
|
+
* Index of the content block in which this citation appears. Note: Numbers greater than Number.MAX_SAFE_INTEGER will result in rounding issues.
|
|
39
|
+
*/
|
|
40
|
+
"contentIndex"?: number;
|
|
41
|
+
/**
|
|
42
|
+
* The type of citation indicating what part of the response it is for.
|
|
43
|
+
*/
|
|
44
|
+
"type"?: CohereCitationV2.Type;
|
|
45
|
+
/**
|
|
46
|
+
* List of source objects for this citation.
|
|
47
|
+
*/
|
|
48
|
+
"sources"?: Array<model.CohereCitationSourceV2>;
|
|
49
|
+
}
|
|
50
|
+
export declare namespace CohereCitationV2 {
|
|
51
|
+
enum Type {
|
|
52
|
+
TextContent = "TEXT_CONTENT",
|
|
53
|
+
ThinkingContent = "THINKING_CONTENT",
|
|
54
|
+
Plan = "PLAN",
|
|
55
|
+
/**
|
|
56
|
+
* This value is used if a service returns a value for this enum that is not recognized by this
|
|
57
|
+
* version of the SDK.
|
|
58
|
+
*/
|
|
59
|
+
UnknownValue = "UNKNOWN_VALUE"
|
|
60
|
+
}
|
|
61
|
+
function getJsonObj(obj: CohereCitationV2): object;
|
|
62
|
+
function getDeserializedJsonObj(obj: CohereCitationV2): object;
|
|
63
|
+
}
|
|
@@ -0,0 +1,79 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
22
|
+
if (k2 === undefined) k2 = k;
|
|
23
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
24
|
+
}) : (function(o, m, k, k2) {
|
|
25
|
+
if (k2 === undefined) k2 = k;
|
|
26
|
+
o[k2] = m[k];
|
|
27
|
+
}));
|
|
28
|
+
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
29
|
+
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
30
|
+
}) : function(o, v) {
|
|
31
|
+
o["default"] = v;
|
|
32
|
+
});
|
|
33
|
+
var __importStar = (this && this.__importStar) || function (mod) {
|
|
34
|
+
if (mod && mod.__esModule) return mod;
|
|
35
|
+
var result = {};
|
|
36
|
+
if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
37
|
+
__setModuleDefault(result, mod);
|
|
38
|
+
return result;
|
|
39
|
+
};
|
|
40
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
41
|
+
exports.CohereCitationV2 = void 0;
|
|
42
|
+
const model = __importStar(require("../model"));
|
|
43
|
+
var CohereCitationV2;
|
|
44
|
+
(function (CohereCitationV2) {
|
|
45
|
+
let Type;
|
|
46
|
+
(function (Type) {
|
|
47
|
+
Type["TextContent"] = "TEXT_CONTENT";
|
|
48
|
+
Type["ThinkingContent"] = "THINKING_CONTENT";
|
|
49
|
+
Type["Plan"] = "PLAN";
|
|
50
|
+
/**
|
|
51
|
+
* This value is used if a service returns a value for this enum that is not recognized by this
|
|
52
|
+
* version of the SDK.
|
|
53
|
+
*/
|
|
54
|
+
Type["UnknownValue"] = "UNKNOWN_VALUE";
|
|
55
|
+
})(Type = CohereCitationV2.Type || (CohereCitationV2.Type = {}));
|
|
56
|
+
function getJsonObj(obj) {
|
|
57
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
58
|
+
"sources": obj.sources
|
|
59
|
+
? obj.sources.map(item => {
|
|
60
|
+
return model.CohereCitationSourceV2.getJsonObj(item);
|
|
61
|
+
})
|
|
62
|
+
: undefined
|
|
63
|
+
});
|
|
64
|
+
return jsonObj;
|
|
65
|
+
}
|
|
66
|
+
CohereCitationV2.getJsonObj = getJsonObj;
|
|
67
|
+
function getDeserializedJsonObj(obj) {
|
|
68
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
69
|
+
"sources": obj.sources
|
|
70
|
+
? obj.sources.map(item => {
|
|
71
|
+
return model.CohereCitationSourceV2.getDeserializedJsonObj(item);
|
|
72
|
+
})
|
|
73
|
+
: undefined
|
|
74
|
+
});
|
|
75
|
+
return jsonObj;
|
|
76
|
+
}
|
|
77
|
+
CohereCitationV2.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
78
|
+
})(CohereCitationV2 = exports.CohereCitationV2 || (exports.CohereCitationV2 = {}));
|
|
79
|
+
//# sourceMappingURL=cohere-citation-v2.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"cohere-citation-v2.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/cohere-citation-v2.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAiClC,IAAiB,gBAAgB,CAwChC;AAxCD,WAAiB,gBAAgB;IAC/B,IAAY,IASX;IATD,WAAY,IAAI;QACd,oCAA4B,CAAA;QAC5B,4CAAoC,CAAA;QACpC,qBAAa,CAAA;QACb;;;WAGG;QACH,sCAA8B,CAAA;IAChC,CAAC,EATW,IAAI,GAAJ,qBAAI,KAAJ,qBAAI,QASf;IAED,SAAgB,UAAU,CAAC,GAAqB;QAC9C,MAAM,OAAO,mCACR,GAAG,GACH;YACD,SAAS,EAAE,GAAG,CAAC,OAAO;gBACpB,CAAC,CAAC,GAAG,CAAC,OAAO,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACrB,OAAO,KAAK,CAAC,sBAAsB,CAAC,UAAU,CAAC,IAAI,CAAC,CAAC;gBACvD,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAbe,2BAAU,aAazB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAqB;QAC1D,MAAM,OAAO,mCACR,GAAG,GACH;YACD,SAAS,EAAE,GAAG,CAAC,OAAO;gBACpB,CAAC,CAAC,GAAG,CAAC,OAAO,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACrB,OAAO,KAAK,CAAC,sBAAsB,CAAC,sBAAsB,CAAC,IAAI,CAAC,CAAC;gBACnE,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAbe,uCAAsB,yBAarC,CAAA;AACH,CAAC,EAxCgB,gBAAgB,GAAhB,wBAAgB,KAAhB,wBAAgB,QAwChC"}
|