oci-generativeaiinference 2.123.0 → 2.123.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/lib/model/assistant-message.d.ts +4 -0
- package/lib/model/assistant-message.js.map +1 -1
- package/lib/model/base-chat-request.js +4 -0
- package/lib/model/base-chat-request.js.map +1 -1
- package/lib/model/base-chat-response.js +4 -0
- package/lib/model/base-chat-response.js.map +1 -1
- package/lib/model/chat-content.js +4 -0
- package/lib/model/chat-content.js.map +1 -1
- package/lib/model/chat-details.d.ts +1 -1
- package/lib/model/chat-result.d.ts +1 -1
- package/lib/model/citation-options-v2.d.ts +38 -0
- package/lib/model/citation-options-v2.js +42 -0
- package/lib/model/citation-options-v2.js.map +1 -0
- package/lib/model/cohere-assistant-message-v2.d.ts +43 -0
- package/lib/model/cohere-assistant-message-v2.js +83 -0
- package/lib/model/cohere-assistant-message-v2.js.map +1 -0
- package/lib/model/cohere-chat-request-v2.d.ts +142 -0
- package/lib/model/cohere-chat-request-v2.js +114 -0
- package/lib/model/cohere-chat-request-v2.js.map +1 -0
- package/lib/model/cohere-chat-response-v2.d.ts +61 -0
- package/lib/model/cohere-chat-response-v2.js +90 -0
- package/lib/model/cohere-chat-response-v2.js.map +1 -0
- package/lib/model/cohere-citation-document.d.ts +36 -0
- package/lib/model/cohere-citation-document.js +36 -0
- package/lib/model/cohere-citation-document.js.map +1 -0
- package/lib/model/cohere-citation-source-v2.d.ts +44 -0
- package/lib/model/cohere-citation-source-v2.js +74 -0
- package/lib/model/cohere-citation-source-v2.js.map +1 -0
- package/lib/model/cohere-citation-tool-v2.d.ts +36 -0
- package/lib/model/cohere-citation-tool-v2.js +36 -0
- package/lib/model/cohere-citation-tool-v2.js.map +1 -0
- package/lib/model/cohere-citation-v2.d.ts +63 -0
- package/lib/model/cohere-citation-v2.js +79 -0
- package/lib/model/cohere-citation-v2.js.map +1 -0
- package/lib/model/cohere-content-v2.d.ts +29 -0
- package/lib/model/cohere-content-v2.js +87 -0
- package/lib/model/cohere-content-v2.js.map +1 -0
- package/lib/model/cohere-document-content-v2.d.ts +35 -0
- package/lib/model/cohere-document-content-v2.js +61 -0
- package/lib/model/cohere-document-content-v2.js.map +1 -0
- package/lib/model/cohere-image-content-v2.d.ts +32 -0
- package/lib/model/cohere-image-content-v2.js +65 -0
- package/lib/model/cohere-image-content-v2.js.map +1 -0
- package/lib/model/cohere-image-url-v2.d.ts +46 -0
- package/lib/model/cohere-image-url-v2.js +47 -0
- package/lib/model/cohere-image-url-v2.js.map +1 -0
- package/lib/model/cohere-message-v2.d.ts +34 -0
- package/lib/model/cohere-message-v2.js +99 -0
- package/lib/model/cohere-message-v2.js.map +1 -0
- package/lib/model/cohere-system-message-v2.d.ts +31 -0
- package/lib/model/cohere-system-message-v2.js +59 -0
- package/lib/model/cohere-system-message-v2.js.map +1 -0
- package/lib/model/cohere-text-content-v2.d.ts +35 -0
- package/lib/model/cohere-text-content-v2.js +59 -0
- package/lib/model/cohere-text-content-v2.js.map +1 -0
- package/lib/model/cohere-thinking-content-v2.d.ts +35 -0
- package/lib/model/cohere-thinking-content-v2.js +61 -0
- package/lib/model/cohere-thinking-content-v2.js.map +1 -0
- package/lib/model/cohere-thinking-v2.d.ts +40 -0
- package/lib/model/cohere-thinking-v2.js +41 -0
- package/lib/model/cohere-thinking-v2.js.map +1 -0
- package/lib/model/cohere-tool-call-v2.d.ts +48 -0
- package/lib/model/cohere-tool-call-v2.js +45 -0
- package/lib/model/cohere-tool-call-v2.js.map +1 -0
- package/lib/model/cohere-tool-message-v2.d.ts +35 -0
- package/lib/model/cohere-tool-message-v2.js +59 -0
- package/lib/model/cohere-tool-message-v2.js.map +1 -0
- package/lib/model/cohere-tool-v2.d.ts +37 -0
- package/lib/model/cohere-tool-v2.js +64 -0
- package/lib/model/cohere-tool-v2.js.map +1 -0
- package/lib/model/cohere-user-message-v2.d.ts +31 -0
- package/lib/model/cohere-user-message-v2.js +59 -0
- package/lib/model/cohere-user-message-v2.js.map +1 -0
- package/lib/model/document-content.d.ts +32 -0
- package/lib/model/document-content.js +65 -0
- package/lib/model/document-content.js.map +1 -0
- package/lib/model/document-url.d.ts +63 -0
- package/lib/model/document-url.js +47 -0
- package/lib/model/document-url.js.map +1 -0
- package/lib/model/embed-text-details.d.ts +16 -0
- package/lib/model/embed-text-details.js +9 -0
- package/lib/model/embed-text-details.js.map +1 -1
- package/lib/model/embed-text-result.d.ts +5 -1
- package/lib/model/embed-text-result.js.map +1 -1
- package/lib/model/function.d.ts +40 -0
- package/lib/model/function.js +36 -0
- package/lib/model/function.js.map +1 -0
- package/lib/model/generic-chat-request.d.ts +1 -0
- package/lib/model/generic-chat-request.js +1 -0
- package/lib/model/generic-chat-request.js.map +1 -1
- package/lib/model/index.d.ts +50 -0
- package/lib/model/index.js +53 -3
- package/lib/model/index.js.map +1 -1
- package/lib/model/log-probability.d.ts +40 -0
- package/lib/model/log-probability.js +36 -0
- package/lib/model/log-probability.js.map +1 -0
- package/lib/model/rerank-text-details.d.ts +4 -0
- package/lib/model/rerank-text-details.js.map +1 -1
- package/package.json +3 -3
|
@@ -0,0 +1,64 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
22
|
+
if (k2 === undefined) k2 = k;
|
|
23
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
24
|
+
}) : (function(o, m, k, k2) {
|
|
25
|
+
if (k2 === undefined) k2 = k;
|
|
26
|
+
o[k2] = m[k];
|
|
27
|
+
}));
|
|
28
|
+
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
29
|
+
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
30
|
+
}) : function(o, v) {
|
|
31
|
+
o["default"] = v;
|
|
32
|
+
});
|
|
33
|
+
var __importStar = (this && this.__importStar) || function (mod) {
|
|
34
|
+
if (mod && mod.__esModule) return mod;
|
|
35
|
+
var result = {};
|
|
36
|
+
if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
37
|
+
__setModuleDefault(result, mod);
|
|
38
|
+
return result;
|
|
39
|
+
};
|
|
40
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
41
|
+
exports.CohereToolV2 = void 0;
|
|
42
|
+
const model = __importStar(require("../model"));
|
|
43
|
+
var CohereToolV2;
|
|
44
|
+
(function (CohereToolV2) {
|
|
45
|
+
let Type;
|
|
46
|
+
(function (Type) {
|
|
47
|
+
Type["Function"] = "FUNCTION";
|
|
48
|
+
})(Type = CohereToolV2.Type || (CohereToolV2.Type = {}));
|
|
49
|
+
function getJsonObj(obj) {
|
|
50
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
51
|
+
"function": obj.function ? model.Function.getJsonObj(obj.function) : undefined
|
|
52
|
+
});
|
|
53
|
+
return jsonObj;
|
|
54
|
+
}
|
|
55
|
+
CohereToolV2.getJsonObj = getJsonObj;
|
|
56
|
+
function getDeserializedJsonObj(obj) {
|
|
57
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
58
|
+
"function": obj.function ? model.Function.getDeserializedJsonObj(obj.function) : undefined
|
|
59
|
+
});
|
|
60
|
+
return jsonObj;
|
|
61
|
+
}
|
|
62
|
+
CohereToolV2.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
63
|
+
})(CohereToolV2 = exports.CohereToolV2 || (exports.CohereToolV2 = {}));
|
|
64
|
+
//# sourceMappingURL=cohere-tool-v2.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"cohere-tool-v2.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/cohere-tool-v2.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAclC,IAAiB,YAAY,CAyB5B;AAzBD,WAAiB,YAAY;IAC3B,IAAY,IAEX;IAFD,WAAY,IAAI;QACd,6BAAqB,CAAA;IACvB,CAAC,EAFW,IAAI,GAAJ,iBAAI,KAAJ,iBAAI,QAEf;IAED,SAAgB,UAAU,CAAC,GAAiB;QAC1C,MAAM,OAAO,mCACR,GAAG,GACH;YACD,UAAU,EAAE,GAAG,CAAC,QAAQ,CAAC,CAAC,CAAC,KAAK,CAAC,QAAQ,CAAC,UAAU,CAAC,GAAG,CAAC,QAAQ,CAAC,CAAC,CAAC,CAAC,SAAS;SAC/E,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IATe,uBAAU,aASzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAiB;QACtD,MAAM,OAAO,mCACR,GAAG,GACH;YACD,UAAU,EAAE,GAAG,CAAC,QAAQ,CAAC,CAAC,CAAC,KAAK,CAAC,QAAQ,CAAC,sBAAsB,CAAC,GAAG,CAAC,QAAQ,CAAC,CAAC,CAAC,CAAC,SAAS;SAC3F,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IATe,mCAAsB,yBASrC,CAAA;AACH,CAAC,EAzBgB,YAAY,GAAZ,oBAAY,KAAZ,oBAAY,QAyB5B"}
|
|
@@ -0,0 +1,31 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
import * as model from "../model";
|
|
21
|
+
/**
|
|
22
|
+
* A message that represents a single chat dialog as USER role.
|
|
23
|
+
*/
|
|
24
|
+
export interface CohereUserMessageV2 extends model.CohereMessageV2 {
|
|
25
|
+
"role": string;
|
|
26
|
+
}
|
|
27
|
+
export declare namespace CohereUserMessageV2 {
|
|
28
|
+
function getJsonObj(obj: CohereUserMessageV2, isParentJsonObj?: boolean): object;
|
|
29
|
+
const role = "USER";
|
|
30
|
+
function getDeserializedJsonObj(obj: CohereUserMessageV2, isParentJsonObj?: boolean): object;
|
|
31
|
+
}
|
|
@@ -0,0 +1,59 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
22
|
+
if (k2 === undefined) k2 = k;
|
|
23
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
24
|
+
}) : (function(o, m, k, k2) {
|
|
25
|
+
if (k2 === undefined) k2 = k;
|
|
26
|
+
o[k2] = m[k];
|
|
27
|
+
}));
|
|
28
|
+
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
29
|
+
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
30
|
+
}) : function(o, v) {
|
|
31
|
+
o["default"] = v;
|
|
32
|
+
});
|
|
33
|
+
var __importStar = (this && this.__importStar) || function (mod) {
|
|
34
|
+
if (mod && mod.__esModule) return mod;
|
|
35
|
+
var result = {};
|
|
36
|
+
if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
37
|
+
__setModuleDefault(result, mod);
|
|
38
|
+
return result;
|
|
39
|
+
};
|
|
40
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
41
|
+
exports.CohereUserMessageV2 = void 0;
|
|
42
|
+
const model = __importStar(require("../model"));
|
|
43
|
+
var CohereUserMessageV2;
|
|
44
|
+
(function (CohereUserMessageV2) {
|
|
45
|
+
function getJsonObj(obj, isParentJsonObj) {
|
|
46
|
+
const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj ? obj : model.CohereMessageV2.getJsonObj(obj))), {});
|
|
47
|
+
return jsonObj;
|
|
48
|
+
}
|
|
49
|
+
CohereUserMessageV2.getJsonObj = getJsonObj;
|
|
50
|
+
CohereUserMessageV2.role = "USER";
|
|
51
|
+
function getDeserializedJsonObj(obj, isParentJsonObj) {
|
|
52
|
+
const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj
|
|
53
|
+
? obj
|
|
54
|
+
: model.CohereMessageV2.getDeserializedJsonObj(obj))), {});
|
|
55
|
+
return jsonObj;
|
|
56
|
+
}
|
|
57
|
+
CohereUserMessageV2.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
58
|
+
})(CohereUserMessageV2 = exports.CohereUserMessageV2 || (exports.CohereUserMessageV2 = {}));
|
|
59
|
+
//# sourceMappingURL=cohere-user-message-v2.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"cohere-user-message-v2.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/cohere-user-message-v2.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAUlC,IAAiB,mBAAmB,CAuBnC;AAvBD,WAAiB,mBAAmB;IAClC,SAAgB,UAAU,CAAC,GAAwB,EAAE,eAAyB;QAC5E,MAAM,OAAO,mCACR,CAAC,eAAe,CAAC,CAAC,CAAC,GAAG,CAAC,CAAC,CAAE,KAAK,CAAC,eAAe,CAAC,UAAU,CAAC,GAAG,CAAyB,CAAC,GACxF,EAAE,CACN,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAPe,8BAAU,aAOzB,CAAA;IACY,wBAAI,GAAG,MAAM,CAAC;IAC3B,SAAgB,sBAAsB,CACpC,GAAwB,EACxB,eAAyB;QAEzB,MAAM,OAAO,mCACR,CAAC,eAAe;YACjB,CAAC,CAAC,GAAG;YACL,CAAC,CAAE,KAAK,CAAC,eAAe,CAAC,sBAAsB,CAAC,GAAG,CAAyB,CAAC,GAC5E,EAAE,CACN,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAZe,0CAAsB,yBAYrC,CAAA;AACH,CAAC,EAvBgB,mBAAmB,GAAnB,2BAAmB,KAAnB,2BAAmB,QAuBnC"}
|
|
@@ -0,0 +1,32 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
import * as model from "../model";
|
|
21
|
+
/**
|
|
22
|
+
* Represents a single instance of chat document content.
|
|
23
|
+
*/
|
|
24
|
+
export interface DocumentContent extends model.ChatContent {
|
|
25
|
+
"documentUrl"?: model.DocumentUrl;
|
|
26
|
+
"type": string;
|
|
27
|
+
}
|
|
28
|
+
export declare namespace DocumentContent {
|
|
29
|
+
function getJsonObj(obj: DocumentContent, isParentJsonObj?: boolean): object;
|
|
30
|
+
const type = "DOCUMENT";
|
|
31
|
+
function getDeserializedJsonObj(obj: DocumentContent, isParentJsonObj?: boolean): object;
|
|
32
|
+
}
|
|
@@ -0,0 +1,65 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
22
|
+
if (k2 === undefined) k2 = k;
|
|
23
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
24
|
+
}) : (function(o, m, k, k2) {
|
|
25
|
+
if (k2 === undefined) k2 = k;
|
|
26
|
+
o[k2] = m[k];
|
|
27
|
+
}));
|
|
28
|
+
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
29
|
+
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
30
|
+
}) : function(o, v) {
|
|
31
|
+
o["default"] = v;
|
|
32
|
+
});
|
|
33
|
+
var __importStar = (this && this.__importStar) || function (mod) {
|
|
34
|
+
if (mod && mod.__esModule) return mod;
|
|
35
|
+
var result = {};
|
|
36
|
+
if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
37
|
+
__setModuleDefault(result, mod);
|
|
38
|
+
return result;
|
|
39
|
+
};
|
|
40
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
41
|
+
exports.DocumentContent = void 0;
|
|
42
|
+
const model = __importStar(require("../model"));
|
|
43
|
+
var DocumentContent;
|
|
44
|
+
(function (DocumentContent) {
|
|
45
|
+
function getJsonObj(obj, isParentJsonObj) {
|
|
46
|
+
const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj ? obj : model.ChatContent.getJsonObj(obj))), {
|
|
47
|
+
"documentUrl": obj.documentUrl ? model.DocumentUrl.getJsonObj(obj.documentUrl) : undefined
|
|
48
|
+
});
|
|
49
|
+
return jsonObj;
|
|
50
|
+
}
|
|
51
|
+
DocumentContent.getJsonObj = getJsonObj;
|
|
52
|
+
DocumentContent.type = "DOCUMENT";
|
|
53
|
+
function getDeserializedJsonObj(obj, isParentJsonObj) {
|
|
54
|
+
const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj
|
|
55
|
+
? obj
|
|
56
|
+
: model.ChatContent.getDeserializedJsonObj(obj))), {
|
|
57
|
+
"documentUrl": obj.documentUrl
|
|
58
|
+
? model.DocumentUrl.getDeserializedJsonObj(obj.documentUrl)
|
|
59
|
+
: undefined
|
|
60
|
+
});
|
|
61
|
+
return jsonObj;
|
|
62
|
+
}
|
|
63
|
+
DocumentContent.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
64
|
+
})(DocumentContent = exports.DocumentContent || (exports.DocumentContent = {}));
|
|
65
|
+
//# sourceMappingURL=document-content.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"document-content.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/document-content.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAYlC,IAAiB,eAAe,CA0B/B;AA1BD,WAAiB,eAAe;IAC9B,SAAgB,UAAU,CAAC,GAAoB,EAAE,eAAyB;QACxE,MAAM,OAAO,mCACR,CAAC,eAAe,CAAC,CAAC,CAAC,GAAG,CAAC,CAAC,CAAE,KAAK,CAAC,WAAW,CAAC,UAAU,CAAC,GAAG,CAAqB,CAAC,GAChF;YACD,aAAa,EAAE,GAAG,CAAC,WAAW,CAAC,CAAC,CAAC,KAAK,CAAC,WAAW,CAAC,UAAU,CAAC,GAAG,CAAC,WAAW,CAAC,CAAC,CAAC,CAAC,SAAS;SAC3F,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IATe,0BAAU,aASzB,CAAA;IACY,oBAAI,GAAG,UAAU,CAAC;IAC/B,SAAgB,sBAAsB,CAAC,GAAoB,EAAE,eAAyB;QACpF,MAAM,OAAO,mCACR,CAAC,eAAe;YACjB,CAAC,CAAC,GAAG;YACL,CAAC,CAAE,KAAK,CAAC,WAAW,CAAC,sBAAsB,CAAC,GAAG,CAAqB,CAAC,GACpE;YACD,aAAa,EAAE,GAAG,CAAC,WAAW;gBAC5B,CAAC,CAAC,KAAK,CAAC,WAAW,CAAC,sBAAsB,CAAC,GAAG,CAAC,WAAW,CAAC;gBAC3D,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAbe,sCAAsB,yBAarC,CAAA;AACH,CAAC,EA1BgB,eAAe,GAAf,uBAAe,KAAf,uBAAe,QA0B/B"}
|
|
@@ -0,0 +1,63 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
/**
|
|
21
|
+
* The base64 encoded document data or a document uri if it's supported.
|
|
22
|
+
*/
|
|
23
|
+
export interface DocumentUrl {
|
|
24
|
+
/**
|
|
25
|
+
* The base64 encoded document data or a document uri if it's supported.
|
|
26
|
+
* <p>
|
|
27
|
+
Example for a pdf document:
|
|
28
|
+
* {@code {
|
|
29
|
+
* \"type\": \"DOCUMENT\",
|
|
30
|
+
* \"documentUrl\": {
|
|
31
|
+
* \"url\": \"data:application/pdf;base64,<base64 encoded document content>\"
|
|
32
|
+
* }
|
|
33
|
+
* }}
|
|
34
|
+
* <p>
|
|
35
|
+
Example with a document uri:
|
|
36
|
+
* {@code {
|
|
37
|
+
* \"type\": \"DOCUMENT\",
|
|
38
|
+
* \"videoUrl\": {
|
|
39
|
+
* \"url\": \"data:application/pdf;uri,<document uri>\"
|
|
40
|
+
* }
|
|
41
|
+
* }}
|
|
42
|
+
*
|
|
43
|
+
*/
|
|
44
|
+
"url": string;
|
|
45
|
+
/**
|
|
46
|
+
* The default value is AUTO and only AUTO is supported. This option controls how to convert the base64 encoded document to tokens.
|
|
47
|
+
*/
|
|
48
|
+
"detail"?: DocumentUrl.Detail;
|
|
49
|
+
}
|
|
50
|
+
export declare namespace DocumentUrl {
|
|
51
|
+
enum Detail {
|
|
52
|
+
Auto = "AUTO",
|
|
53
|
+
High = "HIGH",
|
|
54
|
+
Low = "LOW",
|
|
55
|
+
/**
|
|
56
|
+
* This value is used if a service returns a value for this enum that is not recognized by this
|
|
57
|
+
* version of the SDK.
|
|
58
|
+
*/
|
|
59
|
+
UnknownValue = "UNKNOWN_VALUE"
|
|
60
|
+
}
|
|
61
|
+
function getJsonObj(obj: DocumentUrl): object;
|
|
62
|
+
function getDeserializedJsonObj(obj: DocumentUrl): object;
|
|
63
|
+
}
|
|
@@ -0,0 +1,47 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
22
|
+
exports.DocumentUrl = void 0;
|
|
23
|
+
var DocumentUrl;
|
|
24
|
+
(function (DocumentUrl) {
|
|
25
|
+
let Detail;
|
|
26
|
+
(function (Detail) {
|
|
27
|
+
Detail["Auto"] = "AUTO";
|
|
28
|
+
Detail["High"] = "HIGH";
|
|
29
|
+
Detail["Low"] = "LOW";
|
|
30
|
+
/**
|
|
31
|
+
* This value is used if a service returns a value for this enum that is not recognized by this
|
|
32
|
+
* version of the SDK.
|
|
33
|
+
*/
|
|
34
|
+
Detail["UnknownValue"] = "UNKNOWN_VALUE";
|
|
35
|
+
})(Detail = DocumentUrl.Detail || (DocumentUrl.Detail = {}));
|
|
36
|
+
function getJsonObj(obj) {
|
|
37
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
38
|
+
return jsonObj;
|
|
39
|
+
}
|
|
40
|
+
DocumentUrl.getJsonObj = getJsonObj;
|
|
41
|
+
function getDeserializedJsonObj(obj) {
|
|
42
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
43
|
+
return jsonObj;
|
|
44
|
+
}
|
|
45
|
+
DocumentUrl.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
46
|
+
})(DocumentUrl = exports.DocumentUrl || (exports.DocumentUrl = {}));
|
|
47
|
+
//# sourceMappingURL=document-url.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"document-url.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/document-url.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;AAoCH,IAAiB,WAAW,CAsB3B;AAtBD,WAAiB,WAAW;IAC1B,IAAY,MASX;IATD,WAAY,MAAM;QAChB,uBAAa,CAAA;QACb,uBAAa,CAAA;QACb,qBAAW,CAAA;QACX;;;WAGG;QACH,wCAA8B,CAAA;IAChC,CAAC,EATW,MAAM,GAAN,kBAAM,KAAN,kBAAM,QASjB;IAED,SAAgB,UAAU,CAAC,GAAgB;QACzC,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,sBAAU,aAIzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAgB;QACrD,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,kCAAsB,yBAIrC,CAAA;AACH,CAAC,EAtBgB,WAAW,GAAX,mBAAW,KAAX,mBAAW,QAsB3B"}
|
|
@@ -35,6 +35,14 @@ export interface EmbedTextDetails {
|
|
|
35
35
|
* Whether or not to include the original inputs in the response. Results are index-based.
|
|
36
36
|
*/
|
|
37
37
|
"isEcho"?: boolean;
|
|
38
|
+
/**
|
|
39
|
+
* Specifies the types of embeddings you want to get back. Supports list of enums. Supported values :float, int8, uint8, binary, ubinary, base64. If nothing is passed default will be considered as float.
|
|
40
|
+
*/
|
|
41
|
+
"embeddingTypes"?: Array<EmbedTextDetails.EmbeddingTypes>;
|
|
42
|
+
/**
|
|
43
|
+
* The number of dimensions of the output embedding. This is only available for embed-v4 and newer models. Possible values are 256, 512, 1024, and 1536. Note: Numbers greater than Number.MAX_SAFE_INTEGER will result in rounding issues.
|
|
44
|
+
*/
|
|
45
|
+
"outputDimensions"?: number;
|
|
38
46
|
/**
|
|
39
47
|
* For an input that's longer than the maximum token length, specifies which part of the input text will be truncated.
|
|
40
48
|
*/
|
|
@@ -45,6 +53,14 @@ export interface EmbedTextDetails {
|
|
|
45
53
|
"inputType"?: EmbedTextDetails.InputType;
|
|
46
54
|
}
|
|
47
55
|
export declare namespace EmbedTextDetails {
|
|
56
|
+
enum EmbeddingTypes {
|
|
57
|
+
Float = "float",
|
|
58
|
+
Int8 = "int8",
|
|
59
|
+
Uint8 = "uint8",
|
|
60
|
+
Binary = "binary",
|
|
61
|
+
Ubinary = "ubinary",
|
|
62
|
+
Base64 = "base64"
|
|
63
|
+
}
|
|
48
64
|
enum Truncate {
|
|
49
65
|
None = "NONE",
|
|
50
66
|
Start = "START",
|
|
@@ -42,6 +42,15 @@ exports.EmbedTextDetails = void 0;
|
|
|
42
42
|
const model = __importStar(require("../model"));
|
|
43
43
|
var EmbedTextDetails;
|
|
44
44
|
(function (EmbedTextDetails) {
|
|
45
|
+
let EmbeddingTypes;
|
|
46
|
+
(function (EmbeddingTypes) {
|
|
47
|
+
EmbeddingTypes["Float"] = "float";
|
|
48
|
+
EmbeddingTypes["Int8"] = "int8";
|
|
49
|
+
EmbeddingTypes["Uint8"] = "uint8";
|
|
50
|
+
EmbeddingTypes["Binary"] = "binary";
|
|
51
|
+
EmbeddingTypes["Ubinary"] = "ubinary";
|
|
52
|
+
EmbeddingTypes["Base64"] = "base64";
|
|
53
|
+
})(EmbeddingTypes = EmbedTextDetails.EmbeddingTypes || (EmbedTextDetails.EmbeddingTypes = {}));
|
|
45
54
|
let Truncate;
|
|
46
55
|
(function (Truncate) {
|
|
47
56
|
Truncate["None"] = "NONE";
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"embed-text-details.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/embed-text-details.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;
|
|
1
|
+
{"version":3,"file":"embed-text-details.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/embed-text-details.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAsClC,IAAiB,gBAAgB,CA8ChC;AA9CD,WAAiB,gBAAgB;IAC/B,IAAY,cAOX;IAPD,WAAY,cAAc;QACxB,iCAAe,CAAA;QACf,+BAAa,CAAA;QACb,iCAAe,CAAA;QACf,mCAAiB,CAAA;QACjB,qCAAmB,CAAA;QACnB,mCAAiB,CAAA;IACnB,CAAC,EAPW,cAAc,GAAd,+BAAc,KAAd,+BAAc,QAOzB;IAED,IAAY,QAIX;IAJD,WAAY,QAAQ;QAClB,yBAAa,CAAA;QACb,2BAAe,CAAA;QACf,uBAAW,CAAA;IACb,CAAC,EAJW,QAAQ,GAAR,yBAAQ,KAAR,yBAAQ,QAInB;IAED,IAAY,SAMX;IAND,WAAY,SAAS;QACnB,+CAAkC,CAAA;QAClC,yCAA4B,CAAA;QAC5B,8CAAiC,CAAA;QACjC,sCAAyB,CAAA;QACzB,4BAAe,CAAA;IACjB,CAAC,EANW,SAAS,GAAT,0BAAS,KAAT,0BAAS,QAMpB;IAED,SAAgB,UAAU,CAAC,GAAqB;QAC9C,MAAM,OAAO,mCACR,GAAG,GACH;YACD,aAAa,EAAE,GAAG,CAAC,WAAW,CAAC,CAAC,CAAC,KAAK,CAAC,WAAW,CAAC,UAAU,CAAC,GAAG,CAAC,WAAW,CAAC,CAAC,CAAC,CAAC,SAAS;SAC3F,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IATe,2BAAU,aASzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAqB;QAC1D,MAAM,OAAO,mCACR,GAAG,GACH;YACD,aAAa,EAAE,GAAG,CAAC,WAAW;gBAC5B,CAAC,CAAC,KAAK,CAAC,WAAW,CAAC,sBAAsB,CAAC,GAAG,CAAC,WAAW,CAAC;gBAC3D,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAXe,uCAAsB,yBAWrC,CAAA;AACH,CAAC,EA9CgB,gBAAgB,GAAhB,wBAAgB,KAAhB,wBAAgB,QA8ChC"}
|
|
@@ -31,9 +31,13 @@ export interface EmbedTextResult {
|
|
|
31
31
|
*/
|
|
32
32
|
"inputs"?: Array<string>;
|
|
33
33
|
/**
|
|
34
|
-
* The embeddings corresponding to
|
|
34
|
+
* The embeddings corresponding to float.
|
|
35
35
|
*/
|
|
36
36
|
"embeddings": Array<Array<number>>;
|
|
37
|
+
/**
|
|
38
|
+
* The embeddings corresponding to embedding types input.
|
|
39
|
+
*/
|
|
40
|
+
"embeddingsByType"?: any;
|
|
37
41
|
/**
|
|
38
42
|
* The OCID of the model used in this inference request.
|
|
39
43
|
*/
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"embed-text-result.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/embed-text-result.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;
|
|
1
|
+
{"version":3,"file":"embed-text-result.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/embed-text-result.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAkClC,IAAiB,eAAe,CAqB/B;AArBD,WAAiB,eAAe;IAC9B,SAAgB,UAAU,CAAC,GAAoB;QAC7C,MAAM,OAAO,mCACR,GAAG,GACH;YACD,OAAO,EAAE,GAAG,CAAC,KAAK,CAAC,CAAC,CAAC,KAAK,CAAC,KAAK,CAAC,UAAU,CAAC,GAAG,CAAC,KAAK,CAAC,CAAC,CAAC,CAAC,SAAS;SACnE,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IATe,0BAAU,aASzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAoB;QACzD,MAAM,OAAO,mCACR,GAAG,GACH;YACD,OAAO,EAAE,GAAG,CAAC,KAAK,CAAC,CAAC,CAAC,KAAK,CAAC,KAAK,CAAC,sBAAsB,CAAC,GAAG,CAAC,KAAK,CAAC,CAAC,CAAC,CAAC,SAAS;SAC/E,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IATe,sCAAsB,yBASrC,CAAA;AACH,CAAC,EArBgB,eAAe,GAAf,uBAAe,KAAf,uBAAe,QAqB/B"}
|
|
@@ -0,0 +1,40 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
/**
|
|
21
|
+
* The function to be executed.
|
|
22
|
+
*/
|
|
23
|
+
export interface Function {
|
|
24
|
+
/**
|
|
25
|
+
* The name of the function.
|
|
26
|
+
*/
|
|
27
|
+
"name": string;
|
|
28
|
+
/**
|
|
29
|
+
* The description of the function.
|
|
30
|
+
*/
|
|
31
|
+
"description"?: string;
|
|
32
|
+
/**
|
|
33
|
+
* The parameters of the function as JSON schema.
|
|
34
|
+
*/
|
|
35
|
+
"parameters": any;
|
|
36
|
+
}
|
|
37
|
+
export declare namespace Function {
|
|
38
|
+
function getJsonObj(obj: Function): object;
|
|
39
|
+
function getDeserializedJsonObj(obj: Function): object;
|
|
40
|
+
}
|
|
@@ -0,0 +1,36 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
22
|
+
exports.Function = void 0;
|
|
23
|
+
var Function;
|
|
24
|
+
(function (Function) {
|
|
25
|
+
function getJsonObj(obj) {
|
|
26
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
27
|
+
return jsonObj;
|
|
28
|
+
}
|
|
29
|
+
Function.getJsonObj = getJsonObj;
|
|
30
|
+
function getDeserializedJsonObj(obj) {
|
|
31
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
32
|
+
return jsonObj;
|
|
33
|
+
}
|
|
34
|
+
Function.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
35
|
+
})(Function = exports.Function || (exports.Function = {}));
|
|
36
|
+
//# sourceMappingURL=function.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"function.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/function.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;AAuBH,IAAiB,QAAQ,CAWxB;AAXD,WAAiB,QAAQ;IACvB,SAAgB,UAAU,CAAC,GAAa;QACtC,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,mBAAU,aAIzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAa;QAClD,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,+BAAsB,yBAIrC,CAAA;AACH,CAAC,EAXgB,QAAQ,GAAR,gBAAQ,KAAR,gBAAQ,QAWxB"}
|
|
@@ -44,6 +44,7 @@ var GenericChatRequest;
|
|
|
44
44
|
(function (GenericChatRequest) {
|
|
45
45
|
let ReasoningEffort;
|
|
46
46
|
(function (ReasoningEffort) {
|
|
47
|
+
ReasoningEffort["None"] = "NONE";
|
|
47
48
|
ReasoningEffort["Minimal"] = "MINIMAL";
|
|
48
49
|
ReasoningEffort["Low"] = "LOW";
|
|
49
50
|
ReasoningEffort["Medium"] = "MEDIUM";
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"generic-chat-request.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/generic-chat-request.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAqIlC,IAAiB,kBAAkB,
|
|
1
|
+
{"version":3,"file":"generic-chat-request.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/generic-chat-request.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAqIlC,IAAiB,kBAAkB,CAiGlC;AAjGD,WAAiB,kBAAkB;IACjC,IAAY,eAMX;IAND,WAAY,eAAe;QACzB,gCAAa,CAAA;QACb,sCAAmB,CAAA;QACnB,8BAAW,CAAA;QACX,oCAAiB,CAAA;QACjB,gCAAa,CAAA;IACf,CAAC,EANW,eAAe,GAAf,kCAAe,KAAf,kCAAe,QAM1B;IAED,IAAY,SAIX;IAJD,WAAY,SAAS;QACnB,wBAAW,CAAA;QACX,8BAAiB,CAAA;QACjB,0BAAa,CAAA;IACf,CAAC,EAJW,SAAS,GAAT,4BAAS,KAAT,4BAAS,QAIpB;IAED,IAAY,WAIX;IAJD,WAAY,WAAW;QACrB,4BAAa,CAAA;QACb,kCAAmB,CAAA;QACnB,oCAAqB,CAAA;IACvB,CAAC,EAJW,WAAW,GAAX,8BAAW,KAAX,8BAAW,QAItB;IAED,SAAgB,UAAU,CAAC,GAAuB,EAAE,eAAyB;QAC3E,MAAM,OAAO,mCACR,CAAC,eAAe,CAAC,CAAC,CAAC,GAAG,CAAC,CAAC,CAAE,KAAK,CAAC,eAAe,CAAC,UAAU,CAAC,GAAG,CAAwB,CAAC,GACvF;YACD,UAAU,EAAE,GAAG,CAAC,QAAQ;gBACtB,CAAC,CAAC,GAAG,CAAC,QAAQ,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACtB,OAAO,KAAK,CAAC,OAAO,CAAC,UAAU,CAAC,IAAI,CAAC,CAAC;gBACxC,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;YAEb,eAAe,EAAE,GAAG,CAAC,aAAa;gBAChC,CAAC,CAAC,KAAK,CAAC,aAAa,CAAC,UAAU,CAAC,GAAG,CAAC,aAAa,CAAC;gBACnD,CAAC,CAAC,SAAS;YAEb,YAAY,EAAE,GAAG,CAAC,UAAU,CAAC,CAAC,CAAC,KAAK,CAAC,UAAU,CAAC,UAAU,CAAC,GAAG,CAAC,UAAU,CAAC,CAAC,CAAC,CAAC,SAAS;YACtF,gBAAgB,EAAE,GAAG,CAAC,cAAc;gBAClC,CAAC,CAAC,KAAK,CAAC,cAAc,CAAC,UAAU,CAAC,GAAG,CAAC,cAAc,CAAC;gBACrD,CAAC,CAAC,SAAS;YACb,YAAY,EAAE,GAAG,CAAC,UAAU,CAAC,CAAC,CAAC,KAAK,CAAC,UAAU,CAAC,UAAU,CAAC,GAAG,CAAC,UAAU,CAAC,CAAC,CAAC,CAAC,SAAS;YAEtF,OAAO,EAAE,GAAG,CAAC,KAAK;gBAChB,CAAC,CAAC,GAAG,CAAC,KAAK,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACnB,OAAO,KAAK,CAAC,cAAc,CAAC,UAAU,CAAC,IAAI,CAAC,CAAC;gBAC/C,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;YACb,kBAAkB,EAAE,GAAG,CAAC,gBAAgB;gBACtC,CAAC,CAAC,KAAK,CAAC,gBAAgB,CAAC,UAAU,CAAC,GAAG,CAAC,gBAAgB,CAAC;gBACzD,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAhCe,6BAAU,aAgCzB,CAAA;IACY,4BAAS,GAAG,SAAS,CAAC;IACnC,SAAgB,sBAAsB,CACpC,GAAuB,EACvB,eAAyB;QAEzB,MAAM,OAAO,mCACR,CAAC,eAAe;YACjB,CAAC,CAAC,GAAG;YACL,CAAC,CAAE,KAAK,CAAC,eAAe,CAAC,sBAAsB,CAAC,GAAG,CAAwB,CAAC,GAC3E;YACD,UAAU,EAAE,GAAG,CAAC,QAAQ;gBACtB,CAAC,CAAC,GAAG,CAAC,QAAQ,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACtB,OAAO,KAAK,CAAC,OAAO,CAAC,sBAAsB,CAAC,IAAI,CAAC,CAAC;gBACpD,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;YAEb,eAAe,EAAE,GAAG,CAAC,aAAa;gBAChC,CAAC,CAAC,KAAK,CAAC,aAAa,CAAC,sBAAsB,CAAC,GAAG,CAAC,aAAa,CAAC;gBAC/D,CAAC,CAAC,SAAS;YAEb,YAAY,EAAE,GAAG,CAAC,UAAU;gBAC1B,CAAC,CAAC,KAAK,CAAC,UAAU,CAAC,sBAAsB,CAAC,GAAG,CAAC,UAAU,CAAC;gBACzD,CAAC,CAAC,SAAS;YACb,gBAAgB,EAAE,GAAG,CAAC,cAAc;gBAClC,CAAC,CAAC,KAAK,CAAC,cAAc,CAAC,sBAAsB,CAAC,GAAG,CAAC,cAAc,CAAC;gBACjE,CAAC,CAAC,SAAS;YACb,YAAY,EAAE,GAAG,CAAC,UAAU;gBAC1B,CAAC,CAAC,KAAK,CAAC,UAAU,CAAC,sBAAsB,CAAC,GAAG,CAAC,UAAU,CAAC;gBACzD,CAAC,CAAC,SAAS;YAEb,OAAO,EAAE,GAAG,CAAC,KAAK;gBAChB,CAAC,CAAC,GAAG,CAAC,KAAK,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACnB,OAAO,KAAK,CAAC,cAAc,CAAC,sBAAsB,CAAC,IAAI,CAAC,CAAC;gBAC3D,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;YACb,kBAAkB,EAAE,GAAG,CAAC,gBAAgB;gBACtC,CAAC,CAAC,KAAK,CAAC,gBAAgB,CAAC,sBAAsB,CAAC,GAAG,CAAC,gBAAgB,CAAC;gBACrE,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAzCe,yCAAsB,yBAyCrC,CAAA;AACH,CAAC,EAjGgB,kBAAkB,GAAlB,0BAAkB,KAAlB,0BAAkB,QAiGlC"}
|