oci-generativeaiinference 2.118.0 → 2.118.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/lib/model/audio-content.d.ts +32 -0
- package/lib/model/audio-content.js +61 -0
- package/lib/model/audio-content.js.map +1 -0
- package/lib/model/audio-url.d.ts +63 -0
- package/lib/model/audio-url.js +47 -0
- package/lib/model/audio-url.js.map +1 -0
- package/lib/model/chat-choice.d.ts +1 -0
- package/lib/model/chat-choice.js +8 -2
- package/lib/model/chat-choice.js.map +1 -1
- package/lib/model/chat-content.js +8 -0
- package/lib/model/chat-content.js.map +1 -1
- package/lib/model/grounding-chunk.d.ts +30 -0
- package/lib/model/grounding-chunk.js +60 -0
- package/lib/model/grounding-chunk.js.map +1 -0
- package/lib/model/grounding-metadata.d.ts +42 -0
- package/lib/model/grounding-metadata.js +84 -0
- package/lib/model/grounding-metadata.js.map +1 -0
- package/lib/model/grounding-support-segment.d.ts +40 -0
- package/lib/model/grounding-support-segment.js +36 -0
- package/lib/model/grounding-support-segment.js.map +1 -0
- package/lib/model/grounding-support.d.ts +34 -0
- package/lib/model/grounding-support.js +62 -0
- package/lib/model/grounding-support.js.map +1 -0
- package/lib/model/grounding-web-chunk.d.ts +40 -0
- package/lib/model/grounding-web-chunk.js +36 -0
- package/lib/model/grounding-web-chunk.js.map +1 -0
- package/lib/model/image-url.d.ts +10 -2
- package/lib/model/image-url.js.map +1 -1
- package/lib/model/index.d.ts +20 -0
- package/lib/model/index.js +23 -2
- package/lib/model/index.js.map +1 -1
- package/lib/model/search-entry-point.d.ts +32 -0
- package/lib/model/search-entry-point.js +36 -0
- package/lib/model/search-entry-point.js.map +1 -0
- package/lib/model/video-content.d.ts +32 -0
- package/lib/model/video-content.js +61 -0
- package/lib/model/video-content.js.map +1 -0
- package/lib/model/video-url.d.ts +63 -0
- package/lib/model/video-url.js +47 -0
- package/lib/model/video-url.js.map +1 -0
- package/package.json +3 -3
|
@@ -0,0 +1,32 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
import * as model from "../model";
|
|
21
|
+
/**
|
|
22
|
+
* Represents a single instance of chat audio content.
|
|
23
|
+
*/
|
|
24
|
+
export interface AudioContent extends model.ChatContent {
|
|
25
|
+
"audioUrl"?: model.AudioUrl;
|
|
26
|
+
"type": string;
|
|
27
|
+
}
|
|
28
|
+
export declare namespace AudioContent {
|
|
29
|
+
function getJsonObj(obj: AudioContent, isParentJsonObj?: boolean): object;
|
|
30
|
+
const type = "AUDIO";
|
|
31
|
+
function getDeserializedJsonObj(obj: AudioContent, isParentJsonObj?: boolean): object;
|
|
32
|
+
}
|
|
@@ -0,0 +1,61 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
22
|
+
if (k2 === undefined) k2 = k;
|
|
23
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
24
|
+
}) : (function(o, m, k, k2) {
|
|
25
|
+
if (k2 === undefined) k2 = k;
|
|
26
|
+
o[k2] = m[k];
|
|
27
|
+
}));
|
|
28
|
+
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
29
|
+
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
30
|
+
}) : function(o, v) {
|
|
31
|
+
o["default"] = v;
|
|
32
|
+
});
|
|
33
|
+
var __importStar = (this && this.__importStar) || function (mod) {
|
|
34
|
+
if (mod && mod.__esModule) return mod;
|
|
35
|
+
var result = {};
|
|
36
|
+
if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
37
|
+
__setModuleDefault(result, mod);
|
|
38
|
+
return result;
|
|
39
|
+
};
|
|
40
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
41
|
+
exports.AudioContent = void 0;
|
|
42
|
+
const model = __importStar(require("../model"));
|
|
43
|
+
var AudioContent;
|
|
44
|
+
(function (AudioContent) {
|
|
45
|
+
function getJsonObj(obj, isParentJsonObj) {
|
|
46
|
+
const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj ? obj : model.ChatContent.getJsonObj(obj))), {
|
|
47
|
+
"audioUrl": obj.audioUrl ? model.AudioUrl.getJsonObj(obj.audioUrl) : undefined
|
|
48
|
+
});
|
|
49
|
+
return jsonObj;
|
|
50
|
+
}
|
|
51
|
+
AudioContent.getJsonObj = getJsonObj;
|
|
52
|
+
AudioContent.type = "AUDIO";
|
|
53
|
+
function getDeserializedJsonObj(obj, isParentJsonObj) {
|
|
54
|
+
const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj ? obj : model.ChatContent.getDeserializedJsonObj(obj))), {
|
|
55
|
+
"audioUrl": obj.audioUrl ? model.AudioUrl.getDeserializedJsonObj(obj.audioUrl) : undefined
|
|
56
|
+
});
|
|
57
|
+
return jsonObj;
|
|
58
|
+
}
|
|
59
|
+
AudioContent.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
60
|
+
})(AudioContent = exports.AudioContent || (exports.AudioContent = {}));
|
|
61
|
+
//# sourceMappingURL=audio-content.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"audio-content.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/audio-content.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAYlC,IAAiB,YAAY,CAsB5B;AAtBD,WAAiB,YAAY;IAC3B,SAAgB,UAAU,CAAC,GAAiB,EAAE,eAAyB;QACrE,MAAM,OAAO,mCACR,CAAC,eAAe,CAAC,CAAC,CAAC,GAAG,CAAC,CAAC,CAAE,KAAK,CAAC,WAAW,CAAC,UAAU,CAAC,GAAG,CAAkB,CAAC,GAC7E;YACD,UAAU,EAAE,GAAG,CAAC,QAAQ,CAAC,CAAC,CAAC,KAAK,CAAC,QAAQ,CAAC,UAAU,CAAC,GAAG,CAAC,QAAQ,CAAC,CAAC,CAAC,CAAC,SAAS;SAC/E,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IATe,uBAAU,aASzB,CAAA;IACY,iBAAI,GAAG,OAAO,CAAC;IAC5B,SAAgB,sBAAsB,CAAC,GAAiB,EAAE,eAAyB;QACjF,MAAM,OAAO,mCACR,CAAC,eAAe,CAAC,CAAC,CAAC,GAAG,CAAC,CAAC,CAAE,KAAK,CAAC,WAAW,CAAC,sBAAsB,CAAC,GAAG,CAAkB,CAAC,GACzF;YACD,UAAU,EAAE,GAAG,CAAC,QAAQ,CAAC,CAAC,CAAC,KAAK,CAAC,QAAQ,CAAC,sBAAsB,CAAC,GAAG,CAAC,QAAQ,CAAC,CAAC,CAAC,CAAC,SAAS;SAC3F,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IATe,mCAAsB,yBASrC,CAAA;AACH,CAAC,EAtBgB,YAAY,GAAZ,oBAAY,KAAZ,oBAAY,QAsB5B"}
|
|
@@ -0,0 +1,63 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
/**
|
|
21
|
+
* Provide a base64 encoded audio or an audio uri if it's supported.
|
|
22
|
+
*/
|
|
23
|
+
export interface AudioUrl {
|
|
24
|
+
/**
|
|
25
|
+
* The base64 encoded audio data or an audio uri if it's supported.
|
|
26
|
+
* <p>
|
|
27
|
+
Example for an mp3 audio:
|
|
28
|
+
* {@code {
|
|
29
|
+
* \"type\": \"AUDIO\",
|
|
30
|
+
* \"audioUrl\": {
|
|
31
|
+
* \"url\": \"data:audio/mp3;base64,<base64 encoded audio content>\"
|
|
32
|
+
* }
|
|
33
|
+
* }}
|
|
34
|
+
* <p>
|
|
35
|
+
Example with an audio uri:
|
|
36
|
+
* {@code {
|
|
37
|
+
* \"type\": \"AUDIO\",
|
|
38
|
+
* \"audioUrl\": {
|
|
39
|
+
* \"url\": \"data:audio/mp3;uri,<audio uri>\"
|
|
40
|
+
* }
|
|
41
|
+
* }}
|
|
42
|
+
*
|
|
43
|
+
*/
|
|
44
|
+
"url": string;
|
|
45
|
+
/**
|
|
46
|
+
* The default value is AUTO and only AUTO is supported. This option controls how to convert the base64 encoded audio to tokens.
|
|
47
|
+
*/
|
|
48
|
+
"detail"?: AudioUrl.Detail;
|
|
49
|
+
}
|
|
50
|
+
export declare namespace AudioUrl {
|
|
51
|
+
enum Detail {
|
|
52
|
+
Auto = "AUTO",
|
|
53
|
+
High = "HIGH",
|
|
54
|
+
Low = "LOW",
|
|
55
|
+
/**
|
|
56
|
+
* This value is used if a service returns a value for this enum that is not recognized by this
|
|
57
|
+
* version of the SDK.
|
|
58
|
+
*/
|
|
59
|
+
UnknownValue = "UNKNOWN_VALUE"
|
|
60
|
+
}
|
|
61
|
+
function getJsonObj(obj: AudioUrl): object;
|
|
62
|
+
function getDeserializedJsonObj(obj: AudioUrl): object;
|
|
63
|
+
}
|
|
@@ -0,0 +1,47 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
22
|
+
exports.AudioUrl = void 0;
|
|
23
|
+
var AudioUrl;
|
|
24
|
+
(function (AudioUrl) {
|
|
25
|
+
let Detail;
|
|
26
|
+
(function (Detail) {
|
|
27
|
+
Detail["Auto"] = "AUTO";
|
|
28
|
+
Detail["High"] = "HIGH";
|
|
29
|
+
Detail["Low"] = "LOW";
|
|
30
|
+
/**
|
|
31
|
+
* This value is used if a service returns a value for this enum that is not recognized by this
|
|
32
|
+
* version of the SDK.
|
|
33
|
+
*/
|
|
34
|
+
Detail["UnknownValue"] = "UNKNOWN_VALUE";
|
|
35
|
+
})(Detail = AudioUrl.Detail || (AudioUrl.Detail = {}));
|
|
36
|
+
function getJsonObj(obj) {
|
|
37
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
38
|
+
return jsonObj;
|
|
39
|
+
}
|
|
40
|
+
AudioUrl.getJsonObj = getJsonObj;
|
|
41
|
+
function getDeserializedJsonObj(obj) {
|
|
42
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
43
|
+
return jsonObj;
|
|
44
|
+
}
|
|
45
|
+
AudioUrl.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
46
|
+
})(AudioUrl = exports.AudioUrl || (exports.AudioUrl = {}));
|
|
47
|
+
//# sourceMappingURL=audio-url.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"audio-url.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/audio-url.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;AAoCH,IAAiB,QAAQ,CAsBxB;AAtBD,WAAiB,QAAQ;IACvB,IAAY,MASX;IATD,WAAY,MAAM;QAChB,uBAAa,CAAA;QACb,uBAAa,CAAA;QACb,qBAAW,CAAA;QACX;;;WAGG;QACH,wCAA8B,CAAA;IAChC,CAAC,EATW,MAAM,GAAN,eAAM,KAAN,eAAM,QASjB;IAED,SAAgB,UAAU,CAAC,GAAa;QACtC,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,mBAAU,aAIzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAa;QAClD,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,+BAAsB,yBAIrC,CAAA;AACH,CAAC,EAtBgB,QAAQ,GAAR,gBAAQ,KAAR,gBAAQ,QAsBxB"}
|
package/lib/model/chat-choice.js
CHANGED
|
@@ -46,7 +46,10 @@ var ChatChoice;
|
|
|
46
46
|
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
47
47
|
"message": obj.message ? model.Message.getJsonObj(obj.message) : undefined,
|
|
48
48
|
"logprobs": obj.logprobs ? model.Logprobs.getJsonObj(obj.logprobs) : undefined,
|
|
49
|
-
"usage": obj.usage ? model.Usage.getJsonObj(obj.usage) : undefined
|
|
49
|
+
"usage": obj.usage ? model.Usage.getJsonObj(obj.usage) : undefined,
|
|
50
|
+
"groundingMetadata": obj.groundingMetadata
|
|
51
|
+
? model.GroundingMetadata.getJsonObj(obj.groundingMetadata)
|
|
52
|
+
: undefined
|
|
50
53
|
});
|
|
51
54
|
return jsonObj;
|
|
52
55
|
}
|
|
@@ -55,7 +58,10 @@ var ChatChoice;
|
|
|
55
58
|
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
56
59
|
"message": obj.message ? model.Message.getDeserializedJsonObj(obj.message) : undefined,
|
|
57
60
|
"logprobs": obj.logprobs ? model.Logprobs.getDeserializedJsonObj(obj.logprobs) : undefined,
|
|
58
|
-
"usage": obj.usage ? model.Usage.getDeserializedJsonObj(obj.usage) : undefined
|
|
61
|
+
"usage": obj.usage ? model.Usage.getDeserializedJsonObj(obj.usage) : undefined,
|
|
62
|
+
"groundingMetadata": obj.groundingMetadata
|
|
63
|
+
? model.GroundingMetadata.getDeserializedJsonObj(obj.groundingMetadata)
|
|
64
|
+
: undefined
|
|
59
65
|
});
|
|
60
66
|
return jsonObj;
|
|
61
67
|
}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"chat-choice.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/chat-choice.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;
|
|
1
|
+
{"version":3,"file":"chat-choice.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/chat-choice.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AA6BlC,IAAiB,UAAU,CAiC1B;AAjCD,WAAiB,UAAU;IACzB,SAAgB,UAAU,CAAC,GAAe;QACxC,MAAM,OAAO,mCACR,GAAG,GACH;YACD,SAAS,EAAE,GAAG,CAAC,OAAO,CAAC,CAAC,CAAC,KAAK,CAAC,OAAO,CAAC,UAAU,CAAC,GAAG,CAAC,OAAO,CAAC,CAAC,CAAC,CAAC,SAAS;YAE1E,UAAU,EAAE,GAAG,CAAC,QAAQ,CAAC,CAAC,CAAC,KAAK,CAAC,QAAQ,CAAC,UAAU,CAAC,GAAG,CAAC,QAAQ,CAAC,CAAC,CAAC,CAAC,SAAS;YAC9E,OAAO,EAAE,GAAG,CAAC,KAAK,CAAC,CAAC,CAAC,KAAK,CAAC,KAAK,CAAC,UAAU,CAAC,GAAG,CAAC,KAAK,CAAC,CAAC,CAAC,CAAC,SAAS;YAClE,mBAAmB,EAAE,GAAG,CAAC,iBAAiB;gBACxC,CAAC,CAAC,KAAK,CAAC,iBAAiB,CAAC,UAAU,CAAC,GAAG,CAAC,iBAAiB,CAAC;gBAC3D,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAfe,qBAAU,aAezB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAe;QACpD,MAAM,OAAO,mCACR,GAAG,GACH;YACD,SAAS,EAAE,GAAG,CAAC,OAAO,CAAC,CAAC,CAAC,KAAK,CAAC,OAAO,CAAC,sBAAsB,CAAC,GAAG,CAAC,OAAO,CAAC,CAAC,CAAC,CAAC,SAAS;YAEtF,UAAU,EAAE,GAAG,CAAC,QAAQ,CAAC,CAAC,CAAC,KAAK,CAAC,QAAQ,CAAC,sBAAsB,CAAC,GAAG,CAAC,QAAQ,CAAC,CAAC,CAAC,CAAC,SAAS;YAC1F,OAAO,EAAE,GAAG,CAAC,KAAK,CAAC,CAAC,CAAC,KAAK,CAAC,KAAK,CAAC,sBAAsB,CAAC,GAAG,CAAC,KAAK,CAAC,CAAC,CAAC,CAAC,SAAS;YAC9E,mBAAmB,EAAE,GAAG,CAAC,iBAAiB;gBACxC,CAAC,CAAC,KAAK,CAAC,iBAAiB,CAAC,sBAAsB,CAAC,GAAG,CAAC,iBAAiB,CAAC;gBACvE,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAfe,iCAAsB,yBAerC,CAAA;AACH,CAAC,EAjCgB,UAAU,GAAV,kBAAU,KAAV,kBAAU,QAiC1B"}
|
|
@@ -47,8 +47,12 @@ var ChatContent;
|
|
|
47
47
|
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
48
48
|
if (obj && "type" in obj && obj.type) {
|
|
49
49
|
switch (obj.type) {
|
|
50
|
+
case "VIDEO":
|
|
51
|
+
return model.VideoContent.getJsonObj(jsonObj, true);
|
|
50
52
|
case "IMAGE":
|
|
51
53
|
return model.ImageContent.getJsonObj(jsonObj, true);
|
|
54
|
+
case "AUDIO":
|
|
55
|
+
return model.AudioContent.getJsonObj(jsonObj, true);
|
|
52
56
|
case "TEXT":
|
|
53
57
|
return model.TextContent.getJsonObj(jsonObj, true);
|
|
54
58
|
default:
|
|
@@ -63,8 +67,12 @@ var ChatContent;
|
|
|
63
67
|
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
64
68
|
if (obj && "type" in obj && obj.type) {
|
|
65
69
|
switch (obj.type) {
|
|
70
|
+
case "VIDEO":
|
|
71
|
+
return model.VideoContent.getDeserializedJsonObj(jsonObj, true);
|
|
66
72
|
case "IMAGE":
|
|
67
73
|
return model.ImageContent.getDeserializedJsonObj(jsonObj, true);
|
|
74
|
+
case "AUDIO":
|
|
75
|
+
return model.AudioContent.getDeserializedJsonObj(jsonObj, true);
|
|
68
76
|
case "TEXT":
|
|
69
77
|
return model.TextContent.getDeserializedJsonObj(jsonObj, true);
|
|
70
78
|
default:
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"chat-content.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/chat-content.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAClC,qCAAsC;AAStC,IAAiB,WAAW,
|
|
1
|
+
{"version":3,"file":"chat-content.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/chat-content.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAClC,qCAAsC;AAStC,IAAiB,WAAW,CAmD3B;AAnDD,WAAiB,WAAW;IAC1B,SAAgB,UAAU,CAAC,GAAgB;QACzC,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,IAAI,GAAG,IAAI,MAAM,IAAI,GAAG,IAAI,GAAG,CAAC,IAAI,EAAE;YACpC,QAAQ,GAAG,CAAC,IAAI,EAAE;gBAChB,KAAK,OAAO;oBACV,OAAO,KAAK,CAAC,YAAY,CAAC,UAAU,CAA8B,OAAQ,EAAE,IAAI,CAAC,CAAC;gBACpF,KAAK,OAAO;oBACV,OAAO,KAAK,CAAC,YAAY,CAAC,UAAU,CAA8B,OAAQ,EAAE,IAAI,CAAC,CAAC;gBACpF,KAAK,OAAO;oBACV,OAAO,KAAK,CAAC,YAAY,CAAC,UAAU,CAA8B,OAAQ,EAAE,IAAI,CAAC,CAAC;gBACpF,KAAK,MAAM;oBACT,OAAO,KAAK,CAAC,WAAW,CAAC,UAAU,CAA6B,OAAQ,EAAE,IAAI,CAAC,CAAC;gBAClF;oBACE,IAAI,MAAM,CAAC,GAAG,CAAC,MAAM;wBAAE,MAAM,CAAC,GAAG,CAAC,MAAM,CAAC,IAAI,CAAC,sBAAsB,GAAG,CAAC,IAAI,EAAE,CAAC,CAAC;aACnF;SACF;QACD,OAAO,OAAO,CAAC;IACjB,CAAC;IAlBe,sBAAU,aAkBzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAgB;QACrD,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,IAAI,GAAG,IAAI,MAAM,IAAI,GAAG,IAAI,GAAG,CAAC,IAAI,EAAE;YACpC,QAAQ,GAAG,CAAC,IAAI,EAAE;gBAChB,KAAK,OAAO;oBACV,OAAO,KAAK,CAAC,YAAY,CAAC,sBAAsB,CACjB,OAAQ,EACrC,IAAI,CACL,CAAC;gBACJ,KAAK,OAAO;oBACV,OAAO,KAAK,CAAC,YAAY,CAAC,sBAAsB,CACjB,OAAQ,EACrC,IAAI,CACL,CAAC;gBACJ,KAAK,OAAO;oBACV,OAAO,KAAK,CAAC,YAAY,CAAC,sBAAsB,CACjB,OAAQ,EACrC,IAAI,CACL,CAAC;gBACJ,KAAK,MAAM;oBACT,OAAO,KAAK,CAAC,WAAW,CAAC,sBAAsB,CACjB,OAAQ,EACpC,IAAI,CACL,CAAC;gBACJ;oBACE,IAAI,MAAM,CAAC,GAAG,CAAC,MAAM;wBAAE,MAAM,CAAC,GAAG,CAAC,MAAM,CAAC,IAAI,CAAC,sBAAsB,GAAG,CAAC,IAAI,EAAE,CAAC,CAAC;aACnF;SACF;QACD,OAAO,OAAO,CAAC;IACjB,CAAC;IA9Be,kCAAsB,yBA8BrC,CAAA;AACH,CAAC,EAnDgB,WAAW,GAAX,mBAAW,KAAX,mBAAW,QAmD3B"}
|
|
@@ -0,0 +1,30 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
import * as model from "../model";
|
|
21
|
+
/**
|
|
22
|
+
* object containing the source.
|
|
23
|
+
*/
|
|
24
|
+
export interface GroundingChunk {
|
|
25
|
+
"web"?: model.GroundingWebChunk;
|
|
26
|
+
}
|
|
27
|
+
export declare namespace GroundingChunk {
|
|
28
|
+
function getJsonObj(obj: GroundingChunk): object;
|
|
29
|
+
function getDeserializedJsonObj(obj: GroundingChunk): object;
|
|
30
|
+
}
|
|
@@ -0,0 +1,60 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
22
|
+
if (k2 === undefined) k2 = k;
|
|
23
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
24
|
+
}) : (function(o, m, k, k2) {
|
|
25
|
+
if (k2 === undefined) k2 = k;
|
|
26
|
+
o[k2] = m[k];
|
|
27
|
+
}));
|
|
28
|
+
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
29
|
+
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
30
|
+
}) : function(o, v) {
|
|
31
|
+
o["default"] = v;
|
|
32
|
+
});
|
|
33
|
+
var __importStar = (this && this.__importStar) || function (mod) {
|
|
34
|
+
if (mod && mod.__esModule) return mod;
|
|
35
|
+
var result = {};
|
|
36
|
+
if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
37
|
+
__setModuleDefault(result, mod);
|
|
38
|
+
return result;
|
|
39
|
+
};
|
|
40
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
41
|
+
exports.GroundingChunk = void 0;
|
|
42
|
+
const model = __importStar(require("../model"));
|
|
43
|
+
var GroundingChunk;
|
|
44
|
+
(function (GroundingChunk) {
|
|
45
|
+
function getJsonObj(obj) {
|
|
46
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
47
|
+
"web": obj.web ? model.GroundingWebChunk.getJsonObj(obj.web) : undefined
|
|
48
|
+
});
|
|
49
|
+
return jsonObj;
|
|
50
|
+
}
|
|
51
|
+
GroundingChunk.getJsonObj = getJsonObj;
|
|
52
|
+
function getDeserializedJsonObj(obj) {
|
|
53
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
54
|
+
"web": obj.web ? model.GroundingWebChunk.getDeserializedJsonObj(obj.web) : undefined
|
|
55
|
+
});
|
|
56
|
+
return jsonObj;
|
|
57
|
+
}
|
|
58
|
+
GroundingChunk.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
59
|
+
})(GroundingChunk = exports.GroundingChunk || (exports.GroundingChunk = {}));
|
|
60
|
+
//# sourceMappingURL=grounding-chunk.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"grounding-chunk.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/grounding-chunk.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAUlC,IAAiB,cAAc,CAqB9B;AArBD,WAAiB,cAAc;IAC7B,SAAgB,UAAU,CAAC,GAAmB;QAC5C,MAAM,OAAO,mCACR,GAAG,GACH;YACD,KAAK,EAAE,GAAG,CAAC,GAAG,CAAC,CAAC,CAAC,KAAK,CAAC,iBAAiB,CAAC,UAAU,CAAC,GAAG,CAAC,GAAG,CAAC,CAAC,CAAC,CAAC,SAAS;SACzE,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IATe,yBAAU,aASzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAmB;QACxD,MAAM,OAAO,mCACR,GAAG,GACH;YACD,KAAK,EAAE,GAAG,CAAC,GAAG,CAAC,CAAC,CAAC,KAAK,CAAC,iBAAiB,CAAC,sBAAsB,CAAC,GAAG,CAAC,GAAG,CAAC,CAAC,CAAC,CAAC,SAAS;SACrF,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IATe,qCAAsB,yBASrC,CAAA;AACH,CAAC,EArBgB,cAAc,GAAd,sBAAc,KAAd,sBAAc,QAqB9B"}
|
|
@@ -0,0 +1,42 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
import * as model from "../model";
|
|
21
|
+
/**
|
|
22
|
+
* Grounding metadata.
|
|
23
|
+
*/
|
|
24
|
+
export interface GroundingMetadata {
|
|
25
|
+
/**
|
|
26
|
+
* The queries to be used for Search suggestions.
|
|
27
|
+
*/
|
|
28
|
+
"webSearchQueries"?: Array<string>;
|
|
29
|
+
"searchEntryPoint"?: model.SearchEntryPoint;
|
|
30
|
+
/**
|
|
31
|
+
* Array of objects containing the web sources.
|
|
32
|
+
*/
|
|
33
|
+
"groundingChunks"?: Array<model.GroundingChunk>;
|
|
34
|
+
/**
|
|
35
|
+
* Array of chunks to connect model response text to the sources in groundingChunks.
|
|
36
|
+
*/
|
|
37
|
+
"groundingSupports"?: Array<model.GroundingSupport>;
|
|
38
|
+
}
|
|
39
|
+
export declare namespace GroundingMetadata {
|
|
40
|
+
function getJsonObj(obj: GroundingMetadata): object;
|
|
41
|
+
function getDeserializedJsonObj(obj: GroundingMetadata): object;
|
|
42
|
+
}
|
|
@@ -0,0 +1,84 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
22
|
+
if (k2 === undefined) k2 = k;
|
|
23
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
24
|
+
}) : (function(o, m, k, k2) {
|
|
25
|
+
if (k2 === undefined) k2 = k;
|
|
26
|
+
o[k2] = m[k];
|
|
27
|
+
}));
|
|
28
|
+
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
29
|
+
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
30
|
+
}) : function(o, v) {
|
|
31
|
+
o["default"] = v;
|
|
32
|
+
});
|
|
33
|
+
var __importStar = (this && this.__importStar) || function (mod) {
|
|
34
|
+
if (mod && mod.__esModule) return mod;
|
|
35
|
+
var result = {};
|
|
36
|
+
if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
37
|
+
__setModuleDefault(result, mod);
|
|
38
|
+
return result;
|
|
39
|
+
};
|
|
40
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
41
|
+
exports.GroundingMetadata = void 0;
|
|
42
|
+
const model = __importStar(require("../model"));
|
|
43
|
+
var GroundingMetadata;
|
|
44
|
+
(function (GroundingMetadata) {
|
|
45
|
+
function getJsonObj(obj) {
|
|
46
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
47
|
+
"searchEntryPoint": obj.searchEntryPoint
|
|
48
|
+
? model.SearchEntryPoint.getJsonObj(obj.searchEntryPoint)
|
|
49
|
+
: undefined,
|
|
50
|
+
"groundingChunks": obj.groundingChunks
|
|
51
|
+
? obj.groundingChunks.map(item => {
|
|
52
|
+
return model.GroundingChunk.getJsonObj(item);
|
|
53
|
+
})
|
|
54
|
+
: undefined,
|
|
55
|
+
"groundingSupports": obj.groundingSupports
|
|
56
|
+
? obj.groundingSupports.map(item => {
|
|
57
|
+
return model.GroundingSupport.getJsonObj(item);
|
|
58
|
+
})
|
|
59
|
+
: undefined
|
|
60
|
+
});
|
|
61
|
+
return jsonObj;
|
|
62
|
+
}
|
|
63
|
+
GroundingMetadata.getJsonObj = getJsonObj;
|
|
64
|
+
function getDeserializedJsonObj(obj) {
|
|
65
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
66
|
+
"searchEntryPoint": obj.searchEntryPoint
|
|
67
|
+
? model.SearchEntryPoint.getDeserializedJsonObj(obj.searchEntryPoint)
|
|
68
|
+
: undefined,
|
|
69
|
+
"groundingChunks": obj.groundingChunks
|
|
70
|
+
? obj.groundingChunks.map(item => {
|
|
71
|
+
return model.GroundingChunk.getDeserializedJsonObj(item);
|
|
72
|
+
})
|
|
73
|
+
: undefined,
|
|
74
|
+
"groundingSupports": obj.groundingSupports
|
|
75
|
+
? obj.groundingSupports.map(item => {
|
|
76
|
+
return model.GroundingSupport.getDeserializedJsonObj(item);
|
|
77
|
+
})
|
|
78
|
+
: undefined
|
|
79
|
+
});
|
|
80
|
+
return jsonObj;
|
|
81
|
+
}
|
|
82
|
+
GroundingMetadata.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
83
|
+
})(GroundingMetadata = exports.GroundingMetadata || (exports.GroundingMetadata = {}));
|
|
84
|
+
//# sourceMappingURL=grounding-metadata.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"grounding-metadata.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/grounding-metadata.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAsBlC,IAAiB,iBAAiB,CA6CjC;AA7CD,WAAiB,iBAAiB;IAChC,SAAgB,UAAU,CAAC,GAAsB;QAC/C,MAAM,OAAO,mCACR,GAAG,GACH;YACD,kBAAkB,EAAE,GAAG,CAAC,gBAAgB;gBACtC,CAAC,CAAC,KAAK,CAAC,gBAAgB,CAAC,UAAU,CAAC,GAAG,CAAC,gBAAgB,CAAC;gBACzD,CAAC,CAAC,SAAS;YACb,iBAAiB,EAAE,GAAG,CAAC,eAAe;gBACpC,CAAC,CAAC,GAAG,CAAC,eAAe,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBAC7B,OAAO,KAAK,CAAC,cAAc,CAAC,UAAU,CAAC,IAAI,CAAC,CAAC;gBAC/C,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;YACb,mBAAmB,EAAE,GAAG,CAAC,iBAAiB;gBACxC,CAAC,CAAC,GAAG,CAAC,iBAAiB,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBAC/B,OAAO,KAAK,CAAC,gBAAgB,CAAC,UAAU,CAAC,IAAI,CAAC,CAAC;gBACjD,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IArBe,4BAAU,aAqBzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAsB;QAC3D,MAAM,OAAO,mCACR,GAAG,GACH;YACD,kBAAkB,EAAE,GAAG,CAAC,gBAAgB;gBACtC,CAAC,CAAC,KAAK,CAAC,gBAAgB,CAAC,sBAAsB,CAAC,GAAG,CAAC,gBAAgB,CAAC;gBACrE,CAAC,CAAC,SAAS;YACb,iBAAiB,EAAE,GAAG,CAAC,eAAe;gBACpC,CAAC,CAAC,GAAG,CAAC,eAAe,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBAC7B,OAAO,KAAK,CAAC,cAAc,CAAC,sBAAsB,CAAC,IAAI,CAAC,CAAC;gBAC3D,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;YACb,mBAAmB,EAAE,GAAG,CAAC,iBAAiB;gBACxC,CAAC,CAAC,GAAG,CAAC,iBAAiB,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBAC/B,OAAO,KAAK,CAAC,gBAAgB,CAAC,sBAAsB,CAAC,IAAI,CAAC,CAAC;gBAC7D,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IArBe,wCAAsB,yBAqBrC,CAAA;AACH,CAAC,EA7CgB,iBAAiB,GAAjB,yBAAiB,KAAjB,yBAAiB,QA6CjC"}
|
|
@@ -0,0 +1,40 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
/**
|
|
21
|
+
* segment within groundingSupport.
|
|
22
|
+
*/
|
|
23
|
+
export interface GroundingSupportSegment {
|
|
24
|
+
/**
|
|
25
|
+
* The start index Note: Numbers greater than Number.MAX_SAFE_INTEGER will result in rounding issues.
|
|
26
|
+
*/
|
|
27
|
+
"startIndex"?: number;
|
|
28
|
+
/**
|
|
29
|
+
* The end index Note: Numbers greater than Number.MAX_SAFE_INTEGER will result in rounding issues.
|
|
30
|
+
*/
|
|
31
|
+
"endIndex"?: number;
|
|
32
|
+
/**
|
|
33
|
+
* the text in the segment
|
|
34
|
+
*/
|
|
35
|
+
"text"?: string;
|
|
36
|
+
}
|
|
37
|
+
export declare namespace GroundingSupportSegment {
|
|
38
|
+
function getJsonObj(obj: GroundingSupportSegment): object;
|
|
39
|
+
function getDeserializedJsonObj(obj: GroundingSupportSegment): object;
|
|
40
|
+
}
|
|
@@ -0,0 +1,36 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
22
|
+
exports.GroundingSupportSegment = void 0;
|
|
23
|
+
var GroundingSupportSegment;
|
|
24
|
+
(function (GroundingSupportSegment) {
|
|
25
|
+
function getJsonObj(obj) {
|
|
26
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
27
|
+
return jsonObj;
|
|
28
|
+
}
|
|
29
|
+
GroundingSupportSegment.getJsonObj = getJsonObj;
|
|
30
|
+
function getDeserializedJsonObj(obj) {
|
|
31
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
32
|
+
return jsonObj;
|
|
33
|
+
}
|
|
34
|
+
GroundingSupportSegment.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
35
|
+
})(GroundingSupportSegment = exports.GroundingSupportSegment || (exports.GroundingSupportSegment = {}));
|
|
36
|
+
//# sourceMappingURL=grounding-support-segment.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"grounding-support-segment.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/grounding-support-segment.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;AAuBH,IAAiB,uBAAuB,CAWvC;AAXD,WAAiB,uBAAuB;IACtC,SAAgB,UAAU,CAAC,GAA4B;QACrD,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,kCAAU,aAIzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAA4B;QACjE,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,8CAAsB,yBAIrC,CAAA;AACH,CAAC,EAXgB,uBAAuB,GAAvB,+BAAuB,KAAvB,+BAAuB,QAWvC"}
|