oci-generativeaiinference 2.117.1 → 2.118.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/lib/model/audio-content.d.ts +32 -0
- package/lib/model/audio-content.js +61 -0
- package/lib/model/audio-content.js.map +1 -0
- package/lib/model/audio-url.d.ts +63 -0
- package/lib/model/audio-url.js +47 -0
- package/lib/model/audio-url.js.map +1 -0
- package/lib/model/chat-choice.d.ts +1 -0
- package/lib/model/chat-choice.js +8 -2
- package/lib/model/chat-choice.js.map +1 -1
- package/lib/model/chat-content.js +8 -0
- package/lib/model/chat-content.js.map +1 -1
- package/lib/model/grounding-chunk.d.ts +30 -0
- package/lib/model/grounding-chunk.js +60 -0
- package/lib/model/grounding-chunk.js.map +1 -0
- package/lib/model/grounding-metadata.d.ts +42 -0
- package/lib/model/grounding-metadata.js +84 -0
- package/lib/model/grounding-metadata.js.map +1 -0
- package/lib/model/grounding-support-segment.d.ts +40 -0
- package/lib/model/grounding-support-segment.js +36 -0
- package/lib/model/grounding-support-segment.js.map +1 -0
- package/lib/model/grounding-support.d.ts +34 -0
- package/lib/model/grounding-support.js +62 -0
- package/lib/model/grounding-support.js.map +1 -0
- package/lib/model/grounding-web-chunk.d.ts +40 -0
- package/lib/model/grounding-web-chunk.js +36 -0
- package/lib/model/grounding-web-chunk.js.map +1 -0
- package/lib/model/image-url.d.ts +10 -2
- package/lib/model/image-url.js.map +1 -1
- package/lib/model/index.d.ts +20 -0
- package/lib/model/index.js +23 -2
- package/lib/model/index.js.map +1 -1
- package/lib/model/search-entry-point.d.ts +32 -0
- package/lib/model/search-entry-point.js +36 -0
- package/lib/model/search-entry-point.js.map +1 -0
- package/lib/model/video-content.d.ts +32 -0
- package/lib/model/video-content.js +61 -0
- package/lib/model/video-content.js.map +1 -0
- package/lib/model/video-url.d.ts +63 -0
- package/lib/model/video-url.js +47 -0
- package/lib/model/video-url.js.map +1 -0
- package/package.json +3 -3
|
@@ -0,0 +1,34 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
import * as model from "../model";
|
|
21
|
+
/**
|
|
22
|
+
* chunk to connect model response text to the source in groundingChunk
|
|
23
|
+
*/
|
|
24
|
+
export interface GroundingSupport {
|
|
25
|
+
"segment"?: model.GroundingSupportSegment;
|
|
26
|
+
/**
|
|
27
|
+
* The grounding chunk indices
|
|
28
|
+
*/
|
|
29
|
+
"groundingChunkIndices"?: Array<number>;
|
|
30
|
+
}
|
|
31
|
+
export declare namespace GroundingSupport {
|
|
32
|
+
function getJsonObj(obj: GroundingSupport): object;
|
|
33
|
+
function getDeserializedJsonObj(obj: GroundingSupport): object;
|
|
34
|
+
}
|
|
@@ -0,0 +1,62 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
22
|
+
if (k2 === undefined) k2 = k;
|
|
23
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
24
|
+
}) : (function(o, m, k, k2) {
|
|
25
|
+
if (k2 === undefined) k2 = k;
|
|
26
|
+
o[k2] = m[k];
|
|
27
|
+
}));
|
|
28
|
+
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
29
|
+
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
30
|
+
}) : function(o, v) {
|
|
31
|
+
o["default"] = v;
|
|
32
|
+
});
|
|
33
|
+
var __importStar = (this && this.__importStar) || function (mod) {
|
|
34
|
+
if (mod && mod.__esModule) return mod;
|
|
35
|
+
var result = {};
|
|
36
|
+
if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
37
|
+
__setModuleDefault(result, mod);
|
|
38
|
+
return result;
|
|
39
|
+
};
|
|
40
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
41
|
+
exports.GroundingSupport = void 0;
|
|
42
|
+
const model = __importStar(require("../model"));
|
|
43
|
+
var GroundingSupport;
|
|
44
|
+
(function (GroundingSupport) {
|
|
45
|
+
function getJsonObj(obj) {
|
|
46
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
47
|
+
"segment": obj.segment ? model.GroundingSupportSegment.getJsonObj(obj.segment) : undefined
|
|
48
|
+
});
|
|
49
|
+
return jsonObj;
|
|
50
|
+
}
|
|
51
|
+
GroundingSupport.getJsonObj = getJsonObj;
|
|
52
|
+
function getDeserializedJsonObj(obj) {
|
|
53
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
54
|
+
"segment": obj.segment
|
|
55
|
+
? model.GroundingSupportSegment.getDeserializedJsonObj(obj.segment)
|
|
56
|
+
: undefined
|
|
57
|
+
});
|
|
58
|
+
return jsonObj;
|
|
59
|
+
}
|
|
60
|
+
GroundingSupport.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
61
|
+
})(GroundingSupport = exports.GroundingSupport || (exports.GroundingSupport = {}));
|
|
62
|
+
//# sourceMappingURL=grounding-support.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"grounding-support.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/grounding-support.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAclC,IAAiB,gBAAgB,CAuBhC;AAvBD,WAAiB,gBAAgB;IAC/B,SAAgB,UAAU,CAAC,GAAqB;QAC9C,MAAM,OAAO,mCACR,GAAG,GACH;YACD,SAAS,EAAE,GAAG,CAAC,OAAO,CAAC,CAAC,CAAC,KAAK,CAAC,uBAAuB,CAAC,UAAU,CAAC,GAAG,CAAC,OAAO,CAAC,CAAC,CAAC,CAAC,SAAS;SAC3F,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IATe,2BAAU,aASzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAqB;QAC1D,MAAM,OAAO,mCACR,GAAG,GACH;YACD,SAAS,EAAE,GAAG,CAAC,OAAO;gBACpB,CAAC,CAAC,KAAK,CAAC,uBAAuB,CAAC,sBAAsB,CAAC,GAAG,CAAC,OAAO,CAAC;gBACnE,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAXe,uCAAsB,yBAWrC,CAAA;AACH,CAAC,EAvBgB,gBAAgB,GAAhB,wBAAgB,KAAhB,wBAAgB,QAuBhC"}
|
|
@@ -0,0 +1,40 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
/**
|
|
21
|
+
* object containing the web source.
|
|
22
|
+
*/
|
|
23
|
+
export interface GroundingWebChunk {
|
|
24
|
+
/**
|
|
25
|
+
* The web source's uri
|
|
26
|
+
*/
|
|
27
|
+
"uri"?: string;
|
|
28
|
+
/**
|
|
29
|
+
* The title of the web source
|
|
30
|
+
*/
|
|
31
|
+
"title"?: string;
|
|
32
|
+
/**
|
|
33
|
+
* the domain of the web source
|
|
34
|
+
*/
|
|
35
|
+
"domain"?: string;
|
|
36
|
+
}
|
|
37
|
+
export declare namespace GroundingWebChunk {
|
|
38
|
+
function getJsonObj(obj: GroundingWebChunk): object;
|
|
39
|
+
function getDeserializedJsonObj(obj: GroundingWebChunk): object;
|
|
40
|
+
}
|
|
@@ -0,0 +1,36 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
22
|
+
exports.GroundingWebChunk = void 0;
|
|
23
|
+
var GroundingWebChunk;
|
|
24
|
+
(function (GroundingWebChunk) {
|
|
25
|
+
function getJsonObj(obj) {
|
|
26
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
27
|
+
return jsonObj;
|
|
28
|
+
}
|
|
29
|
+
GroundingWebChunk.getJsonObj = getJsonObj;
|
|
30
|
+
function getDeserializedJsonObj(obj) {
|
|
31
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
32
|
+
return jsonObj;
|
|
33
|
+
}
|
|
34
|
+
GroundingWebChunk.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
35
|
+
})(GroundingWebChunk = exports.GroundingWebChunk || (exports.GroundingWebChunk = {}));
|
|
36
|
+
//# sourceMappingURL=grounding-web-chunk.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"grounding-web-chunk.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/grounding-web-chunk.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;AAuBH,IAAiB,iBAAiB,CAWjC;AAXD,WAAiB,iBAAiB;IAChC,SAAgB,UAAU,CAAC,GAAsB;QAC/C,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,4BAAU,aAIzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAsB;QAC3D,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,wCAAsB,yBAIrC,CAAA;AACH,CAAC,EAXgB,iBAAiB,GAAjB,yBAAiB,KAAjB,yBAAiB,QAWjC"}
|
package/lib/model/image-url.d.ts
CHANGED
|
@@ -18,11 +18,11 @@ To learn more about the service, see the [Generative AI documentation](https://d
|
|
|
18
18
|
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
19
|
*/
|
|
20
20
|
/**
|
|
21
|
-
* Provide a base64 encoded image.
|
|
21
|
+
* Provide a base64 encoded image or an image uri if it's supported.
|
|
22
22
|
*/
|
|
23
23
|
export interface ImageUrl {
|
|
24
24
|
/**
|
|
25
|
-
* The base64 encoded image data.
|
|
25
|
+
* The base64 encoded image data or an image uri if it's supported.
|
|
26
26
|
* <p>
|
|
27
27
|
Example for a png image:
|
|
28
28
|
* {@code {
|
|
@@ -31,6 +31,14 @@ export interface ImageUrl {
|
|
|
31
31
|
* \"url\": \"data:image/png;base64,<base64 encoded image content>\"
|
|
32
32
|
* }
|
|
33
33
|
* }}
|
|
34
|
+
* <p>
|
|
35
|
+
Example with an image uri:
|
|
36
|
+
* {@code {
|
|
37
|
+
* \"type\": \"IMAGE\",
|
|
38
|
+
* \"imageUrl\": {
|
|
39
|
+
* \"url\": \"data:image/png;uri,<image uri>\"
|
|
40
|
+
* }
|
|
41
|
+
* }}
|
|
34
42
|
*
|
|
35
43
|
*/
|
|
36
44
|
"url": string;
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"image-url.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/image-url.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;
|
|
1
|
+
{"version":3,"file":"image-url.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/image-url.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;AAoCH,IAAiB,QAAQ,CAsBxB;AAtBD,WAAiB,QAAQ;IACvB,IAAY,MASX;IATD,WAAY,MAAM;QAChB,uBAAa,CAAA;QACb,uBAAa,CAAA;QACb,qBAAW,CAAA;QACX;;;WAGG;QACH,wCAA8B,CAAA;IAChC,CAAC,EATW,MAAM,GAAN,eAAM,KAAN,eAAM,QASjB;IAED,SAAgB,UAAU,CAAC,GAAa;QACtC,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,mBAAU,aAIzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAa;QAClD,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,+BAAsB,yBAIrC,CAAA;AACH,CAAC,EAtBgB,QAAQ,GAAR,gBAAQ,KAAR,gBAAQ,QAsBxB"}
|
package/lib/model/index.d.ts
CHANGED
|
@@ -25,6 +25,8 @@ import * as ApplyGuardrailsResult from "./apply-guardrails-result";
|
|
|
25
25
|
export import ApplyGuardrailsResult = ApplyGuardrailsResult.ApplyGuardrailsResult;
|
|
26
26
|
import * as ApproximateLocation from "./approximate-location";
|
|
27
27
|
export import ApproximateLocation = ApproximateLocation.ApproximateLocation;
|
|
28
|
+
import * as AudioUrl from "./audio-url";
|
|
29
|
+
export import AudioUrl = AudioUrl.AudioUrl;
|
|
28
30
|
import * as BaseChatRequest from "./base-chat-request";
|
|
29
31
|
export import BaseChatRequest = BaseChatRequest.BaseChatRequest;
|
|
30
32
|
import * as BaseChatResponse from "./base-chat-response";
|
|
@@ -75,6 +77,16 @@ import * as GenerateTextResult from "./generate-text-result";
|
|
|
75
77
|
export import GenerateTextResult = GenerateTextResult.GenerateTextResult;
|
|
76
78
|
import * as GeneratedText from "./generated-text";
|
|
77
79
|
export import GeneratedText = GeneratedText.GeneratedText;
|
|
80
|
+
import * as GroundingChunk from "./grounding-chunk";
|
|
81
|
+
export import GroundingChunk = GroundingChunk.GroundingChunk;
|
|
82
|
+
import * as GroundingMetadata from "./grounding-metadata";
|
|
83
|
+
export import GroundingMetadata = GroundingMetadata.GroundingMetadata;
|
|
84
|
+
import * as GroundingSupport from "./grounding-support";
|
|
85
|
+
export import GroundingSupport = GroundingSupport.GroundingSupport;
|
|
86
|
+
import * as GroundingSupportSegment from "./grounding-support-segment";
|
|
87
|
+
export import GroundingSupportSegment = GroundingSupportSegment.GroundingSupportSegment;
|
|
88
|
+
import * as GroundingWebChunk from "./grounding-web-chunk";
|
|
89
|
+
export import GroundingWebChunk = GroundingWebChunk.GroundingWebChunk;
|
|
78
90
|
import * as GuardrailConfigs from "./guardrail-configs";
|
|
79
91
|
export import GuardrailConfigs = GuardrailConfigs.GuardrailConfigs;
|
|
80
92
|
import * as GuardrailsInput from "./guardrails-input";
|
|
@@ -111,6 +123,8 @@ import * as ResponseFormat from "./response-format";
|
|
|
111
123
|
export import ResponseFormat = ResponseFormat.ResponseFormat;
|
|
112
124
|
import * as ResponseJsonSchema from "./response-json-schema";
|
|
113
125
|
export import ResponseJsonSchema = ResponseJsonSchema.ResponseJsonSchema;
|
|
126
|
+
import * as SearchEntryPoint from "./search-entry-point";
|
|
127
|
+
export import SearchEntryPoint = SearchEntryPoint.SearchEntryPoint;
|
|
114
128
|
import * as SearchQuery from "./search-query";
|
|
115
129
|
export import SearchQuery = SearchQuery.SearchQuery;
|
|
116
130
|
import * as ServingMode from "./serving-mode";
|
|
@@ -133,10 +147,14 @@ import * as UrlCitation from "./url-citation";
|
|
|
133
147
|
export import UrlCitation = UrlCitation.UrlCitation;
|
|
134
148
|
import * as Usage from "./usage";
|
|
135
149
|
export import Usage = Usage.Usage;
|
|
150
|
+
import * as VideoUrl from "./video-url";
|
|
151
|
+
export import VideoUrl = VideoUrl.VideoUrl;
|
|
136
152
|
import * as WebSearchOptions from "./web-search-options";
|
|
137
153
|
export import WebSearchOptions = WebSearchOptions.WebSearchOptions;
|
|
138
154
|
import * as AssistantMessage from "./assistant-message";
|
|
139
155
|
export import AssistantMessage = AssistantMessage.AssistantMessage;
|
|
156
|
+
import * as AudioContent from "./audio-content";
|
|
157
|
+
export import AudioContent = AudioContent.AudioContent;
|
|
140
158
|
import * as CohereChatBotMessage from "./cohere-chat-bot-message";
|
|
141
159
|
export import CohereChatBotMessage = CohereChatBotMessage.CohereChatBotMessage;
|
|
142
160
|
import * as CohereChatRequest from "./cohere-chat-request";
|
|
@@ -203,3 +221,5 @@ import * as ToolMessage from "./tool-message";
|
|
|
203
221
|
export import ToolMessage = ToolMessage.ToolMessage;
|
|
204
222
|
import * as UserMessage from "./user-message";
|
|
205
223
|
export import UserMessage = UserMessage.UserMessage;
|
|
224
|
+
import * as VideoContent from "./video-content";
|
|
225
|
+
export import VideoContent = VideoContent.VideoContent;
|
package/lib/model/index.js
CHANGED
|
@@ -38,8 +38,9 @@ var __importStar = (this && this.__importStar) || function (mod) {
|
|
|
38
38
|
return result;
|
|
39
39
|
};
|
|
40
40
|
Object.defineProperty(exports, "__esModule", { value: true });
|
|
41
|
-
exports.
|
|
42
|
-
exports.
|
|
41
|
+
exports.RerankTextDetails = exports.PromptTokensDetails = exports.PromptInjectionProtectionResult = exports.PromptInjectionConfiguration = exports.Prediction = exports.PersonallyIdentifiableInformationResult = exports.PersonallyIdentifiableInformationConfiguration = exports.Message = exports.Logprobs = exports.LlmInferenceResponse = exports.LlmInferenceRequest = exports.ImageUrl = exports.GuardrailsResults = exports.GuardrailsInput = exports.GuardrailConfigs = exports.GroundingWebChunk = exports.GroundingSupportSegment = exports.GroundingSupport = exports.GroundingMetadata = exports.GroundingChunk = exports.GeneratedText = exports.GenerateTextResult = exports.GenerateTextDetails = exports.EmbedTextResult = exports.EmbedTextDetails = exports.DocumentRank = exports.Document = exports.ContentModerationResult = exports.ContentModerationConfiguration = exports.CompletionTokensDetails = exports.CohereToolResult = exports.CohereToolCall = exports.CohereTool = exports.CohereResponseFormat = exports.CohereParameterDefinition = exports.CohereMessage = exports.Citation = exports.Choice = exports.ChatResult = exports.ChatDetails = exports.ChatContent = exports.ChatChoice = exports.CategoryScore = exports.BaseChatResponse = exports.BaseChatRequest = exports.AudioUrl = exports.ApproximateLocation = exports.ApplyGuardrailsResult = exports.ApplyGuardrailsDetails = exports.Annotation = void 0;
|
|
42
|
+
exports.ToolChoiceRequired = exports.ToolChoiceNone = exports.ToolChoiceFunction = exports.ToolChoiceAuto = exports.TextResponseFormat = exports.TextContent = exports.SystemMessage = exports.StaticContent = exports.OnDemandServingMode = exports.LlamaLlmInferenceResponse = exports.LlamaLlmInferenceRequest = exports.JsonSchemaResponseFormat = exports.JsonObjectResponseFormat = exports.ImageContent = exports.GuardrailsTextInput = exports.GenericChatResponse = exports.GenericChatRequest = exports.FunctionDefinition = exports.FunctionCall = exports.DeveloperMessage = exports.DedicatedServingMode = exports.CohereUserMessage = exports.CohereToolMessage = exports.CohereSystemMessage = exports.CohereResponseTextFormat = exports.CohereResponseJsonFormat = exports.CohereLlmInferenceResponse = exports.CohereLlmInferenceRequest = exports.CohereChatResponse = exports.CohereChatRequest = exports.CohereChatBotMessage = exports.AudioContent = exports.AssistantMessage = exports.WebSearchOptions = exports.VideoUrl = exports.Usage = exports.UrlCitation = exports.ToolDefinition = exports.ToolChoice = exports.ToolCall = exports.TokenLikelihood = exports.SummarizeTextResult = exports.SummarizeTextDetails = exports.StreamOptions = exports.ServingMode = exports.SearchQuery = exports.SearchEntryPoint = exports.ResponseJsonSchema = exports.ResponseFormat = exports.RerankTextResult = void 0;
|
|
43
|
+
exports.VideoContent = exports.UserMessage = exports.ToolMessage = void 0;
|
|
43
44
|
const Annotation = __importStar(require("./annotation"));
|
|
44
45
|
exports.Annotation = Annotation.Annotation;
|
|
45
46
|
const ApplyGuardrailsDetails = __importStar(require("./apply-guardrails-details"));
|
|
@@ -48,6 +49,8 @@ const ApplyGuardrailsResult = __importStar(require("./apply-guardrails-result"))
|
|
|
48
49
|
exports.ApplyGuardrailsResult = ApplyGuardrailsResult.ApplyGuardrailsResult;
|
|
49
50
|
const ApproximateLocation = __importStar(require("./approximate-location"));
|
|
50
51
|
exports.ApproximateLocation = ApproximateLocation.ApproximateLocation;
|
|
52
|
+
const AudioUrl = __importStar(require("./audio-url"));
|
|
53
|
+
exports.AudioUrl = AudioUrl.AudioUrl;
|
|
51
54
|
const BaseChatRequest = __importStar(require("./base-chat-request"));
|
|
52
55
|
exports.BaseChatRequest = BaseChatRequest.BaseChatRequest;
|
|
53
56
|
const BaseChatResponse = __importStar(require("./base-chat-response"));
|
|
@@ -98,6 +101,16 @@ const GenerateTextResult = __importStar(require("./generate-text-result"));
|
|
|
98
101
|
exports.GenerateTextResult = GenerateTextResult.GenerateTextResult;
|
|
99
102
|
const GeneratedText = __importStar(require("./generated-text"));
|
|
100
103
|
exports.GeneratedText = GeneratedText.GeneratedText;
|
|
104
|
+
const GroundingChunk = __importStar(require("./grounding-chunk"));
|
|
105
|
+
exports.GroundingChunk = GroundingChunk.GroundingChunk;
|
|
106
|
+
const GroundingMetadata = __importStar(require("./grounding-metadata"));
|
|
107
|
+
exports.GroundingMetadata = GroundingMetadata.GroundingMetadata;
|
|
108
|
+
const GroundingSupport = __importStar(require("./grounding-support"));
|
|
109
|
+
exports.GroundingSupport = GroundingSupport.GroundingSupport;
|
|
110
|
+
const GroundingSupportSegment = __importStar(require("./grounding-support-segment"));
|
|
111
|
+
exports.GroundingSupportSegment = GroundingSupportSegment.GroundingSupportSegment;
|
|
112
|
+
const GroundingWebChunk = __importStar(require("./grounding-web-chunk"));
|
|
113
|
+
exports.GroundingWebChunk = GroundingWebChunk.GroundingWebChunk;
|
|
101
114
|
const GuardrailConfigs = __importStar(require("./guardrail-configs"));
|
|
102
115
|
exports.GuardrailConfigs = GuardrailConfigs.GuardrailConfigs;
|
|
103
116
|
const GuardrailsInput = __importStar(require("./guardrails-input"));
|
|
@@ -134,6 +147,8 @@ const ResponseFormat = __importStar(require("./response-format"));
|
|
|
134
147
|
exports.ResponseFormat = ResponseFormat.ResponseFormat;
|
|
135
148
|
const ResponseJsonSchema = __importStar(require("./response-json-schema"));
|
|
136
149
|
exports.ResponseJsonSchema = ResponseJsonSchema.ResponseJsonSchema;
|
|
150
|
+
const SearchEntryPoint = __importStar(require("./search-entry-point"));
|
|
151
|
+
exports.SearchEntryPoint = SearchEntryPoint.SearchEntryPoint;
|
|
137
152
|
const SearchQuery = __importStar(require("./search-query"));
|
|
138
153
|
exports.SearchQuery = SearchQuery.SearchQuery;
|
|
139
154
|
const ServingMode = __importStar(require("./serving-mode"));
|
|
@@ -156,10 +171,14 @@ const UrlCitation = __importStar(require("./url-citation"));
|
|
|
156
171
|
exports.UrlCitation = UrlCitation.UrlCitation;
|
|
157
172
|
const Usage = __importStar(require("./usage"));
|
|
158
173
|
exports.Usage = Usage.Usage;
|
|
174
|
+
const VideoUrl = __importStar(require("./video-url"));
|
|
175
|
+
exports.VideoUrl = VideoUrl.VideoUrl;
|
|
159
176
|
const WebSearchOptions = __importStar(require("./web-search-options"));
|
|
160
177
|
exports.WebSearchOptions = WebSearchOptions.WebSearchOptions;
|
|
161
178
|
const AssistantMessage = __importStar(require("./assistant-message"));
|
|
162
179
|
exports.AssistantMessage = AssistantMessage.AssistantMessage;
|
|
180
|
+
const AudioContent = __importStar(require("./audio-content"));
|
|
181
|
+
exports.AudioContent = AudioContent.AudioContent;
|
|
163
182
|
const CohereChatBotMessage = __importStar(require("./cohere-chat-bot-message"));
|
|
164
183
|
exports.CohereChatBotMessage = CohereChatBotMessage.CohereChatBotMessage;
|
|
165
184
|
const CohereChatRequest = __importStar(require("./cohere-chat-request"));
|
|
@@ -226,4 +245,6 @@ const ToolMessage = __importStar(require("./tool-message"));
|
|
|
226
245
|
exports.ToolMessage = ToolMessage.ToolMessage;
|
|
227
246
|
const UserMessage = __importStar(require("./user-message"));
|
|
228
247
|
exports.UserMessage = UserMessage.UserMessage;
|
|
248
|
+
const VideoContent = __importStar(require("./video-content"));
|
|
249
|
+
exports.VideoContent = VideoContent.VideoContent;
|
|
229
250
|
//# sourceMappingURL=index.js.map
|
package/lib/model/index.js.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"index.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/index.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG
|
|
1
|
+
{"version":3,"file":"index.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/index.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;;;AAEH,yDAA2C;AAC7B,QAAA,UAAU,GAAG,UAAU,CAAC,UAAU,CAAC;AACjD,mFAAqE;AACvD,QAAA,sBAAsB,GAAG,sBAAsB,CAAC,sBAAsB,CAAC;AACrF,iFAAmE;AACrD,QAAA,qBAAqB,GAAG,qBAAqB,CAAC,qBAAqB,CAAC;AAClF,4EAA8D;AAChD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,sDAAwC;AAC1B,QAAA,QAAQ,GAAG,QAAQ,CAAC,QAAQ,CAAC;AAC3C,qEAAuD;AACzC,QAAA,eAAe,GAAG,eAAe,CAAC,eAAe,CAAC;AAChE,uEAAyD;AAC3C,QAAA,gBAAgB,GAAG,gBAAgB,CAAC,gBAAgB,CAAC;AACnE,gEAAkD;AACpC,QAAA,aAAa,GAAG,aAAa,CAAC,aAAa,CAAC;AAC1D,0DAA4C;AAC9B,QAAA,UAAU,GAAG,UAAU,CAAC,UAAU,CAAC;AACjD,4DAA8C;AAChC,QAAA,WAAW,GAAG,WAAW,CAAC,WAAW,CAAC;AACpD,4DAA8C;AAChC,QAAA,WAAW,GAAG,WAAW,CAAC,WAAW,CAAC;AACpD,0DAA4C;AAC9B,QAAA,UAAU,GAAG,UAAU,CAAC,UAAU,CAAC;AACjD,iDAAmC;AACrB,QAAA,MAAM,GAAG,MAAM,CAAC,MAAM,CAAC;AACrC,qDAAuC;AACzB,QAAA,QAAQ,GAAG,QAAQ,CAAC,QAAQ,CAAC;AAC3C,gEAAkD;AACpC,QAAA,aAAa,GAAG,aAAa,CAAC,aAAa,CAAC;AAC1D,yFAA2E;AAC7D,QAAA,yBAAyB,GAAG,yBAAyB,CAAC,yBAAyB,CAAC;AAC9F,+EAAiE;AACnD,QAAA,oBAAoB,GAAG,oBAAoB,CAAC,oBAAoB,CAAC;AAC/E,0DAA4C;AAC9B,QAAA,UAAU,GAAG,UAAU,CAAC,UAAU,CAAC;AACjD,mEAAqD;AACvC,QAAA,cAAc,GAAG,cAAc,CAAC,cAAc,CAAC;AAC7D,uEAAyD;AAC3C,QAAA,gBAAgB,GAAG,gBAAgB,CAAC,gBAAgB,CAAC;AACnE,qFAAuE;AACzD,QAAA,uBAAuB,GAAG,uBAAuB,CAAC,uBAAuB,CAAC;AACxF,mGAAqF;AACvE,QAAA,8BAA8B,GAAG,8BAA8B,CAAC,8BAA8B,CAAC;AAC7G,qFAAuE;AACzD,QAAA,uBAAuB,GAAG,uBAAuB,CAAC,uBAAuB,CAAC;AACxF,qDAAuC;AACzB,QAAA,QAAQ,GAAG,QAAQ,CAAC,QAAQ,CAAC;AAC3C,8DAAgD;AAClC,QAAA,YAAY,GAAG,YAAY,CAAC,YAAY,CAAC;AACvD,uEAAyD;AAC3C,QAAA,gBAAgB,GAAG,gBAAgB,CAAC,gBAAgB,CAAC;AACnE,qEAAuD;AACzC,QAAA,eAAe,GAAG,eAAe,CAAC,eAAe,CAAC;AAChE,6EAA+D;AACjD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,2EAA6D;AAC/C,QAAA,kBAAkB,GAAG,kBAAkB,CAAC,kBAAkB,CAAC;AACzE,gEAAkD;AACpC,QAAA,aAAa,GAAG,aAAa,CAAC,aAAa,CAAC;AAC1D,kEAAoD;AACtC,QAAA,cAAc,GAAG,cAAc,CAAC,cAAc,CAAC;AAC7D,wEAA0D;AAC5C,QAAA,iBAAiB,GAAG,iBAAiB,CAAC,iBAAiB,CAAC;AACtE,sEAAwD;AAC1C,QAAA,gBAAgB,GAAG,gBAAgB,CAAC,gBAAgB,CAAC;AACnE,qFAAuE;AACzD,QAAA,uBAAuB,GAAG,uBAAuB,CAAC,uBAAuB,CAAC;AACxF,yEAA2D;AAC7C,QAAA,iBAAiB,GAAG,iBAAiB,CAAC,iBAAiB,CAAC;AACtE,sEAAwD;AAC1C,QAAA,gBAAgB,GAAG,gBAAgB,CAAC,gBAAgB,CAAC;AACnE,oEAAsD;AACxC,QAAA,eAAe,GAAG,eAAe,CAAC,eAAe,CAAC;AAChE,wEAA0D;AAC5C,QAAA,iBAAiB,GAAG,iBAAiB,CAAC,iBAAiB,CAAC;AACtE,sDAAwC;AAC1B,QAAA,QAAQ,GAAG,QAAQ,CAAC,QAAQ,CAAC;AAC3C,6EAA+D;AACjD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,+EAAiE;AACnD,QAAA,oBAAoB,GAAG,oBAAoB,CAAC,oBAAoB,CAAC;AAC/E,qDAAuC;AACzB,QAAA,QAAQ,GAAG,QAAQ,CAAC,QAAQ,CAAC;AAC3C,mDAAqC;AACvB,QAAA,OAAO,GAAG,OAAO,CAAC,OAAO,CAAC;AACxC,oIAAsH;AACxG,QAAA,8CAA8C,GAAG,8CAA8C,CAAC,8CAA8C,CAAC;AAC7J,sHAAwG;AAC1F,QAAA,uCAAuC,GAAG,uCAAuC,CAAC,uCAAuC,CAAC;AACxI,yDAA2C;AAC7B,QAAA,UAAU,GAAG,UAAU,CAAC,UAAU,CAAC;AACjD,+FAAiF;AACnE,QAAA,4BAA4B,GAAG,4BAA4B,CAAC,4BAA4B,CAAC;AACvG,sGAAwF;AAC1E,QAAA,+BAA+B,GAAG,+BAA+B,CAAC,+BAA+B,CAAC;AAChH,6EAA+D;AACjD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,yEAA2D;AAC7C,QAAA,iBAAiB,GAAG,iBAAiB,CAAC,iBAAiB,CAAC;AACtE,uEAAyD;AAC3C,QAAA,gBAAgB,GAAG,gBAAgB,CAAC,gBAAgB,CAAC;AACnE,kEAAoD;AACtC,QAAA,cAAc,GAAG,cAAc,CAAC,cAAc,CAAC;AAC7D,2EAA6D;AAC/C,QAAA,kBAAkB,GAAG,kBAAkB,CAAC,kBAAkB,CAAC;AACzE,uEAAyD;AAC3C,QAAA,gBAAgB,GAAG,gBAAgB,CAAC,gBAAgB,CAAC;AACnE,4DAA8C;AAChC,QAAA,WAAW,GAAG,WAAW,CAAC,WAAW,CAAC;AACpD,4DAA8C;AAChC,QAAA,WAAW,GAAG,WAAW,CAAC,WAAW,CAAC;AACpD,gEAAkD;AACpC,QAAA,aAAa,GAAG,aAAa,CAAC,aAAa,CAAC;AAC1D,+EAAiE;AACnD,QAAA,oBAAoB,GAAG,oBAAoB,CAAC,oBAAoB,CAAC;AAC/E,6EAA+D;AACjD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,oEAAsD;AACxC,QAAA,eAAe,GAAG,eAAe,CAAC,eAAe,CAAC;AAChE,sDAAwC;AAC1B,QAAA,QAAQ,GAAG,QAAQ,CAAC,QAAQ,CAAC;AAC3C,0DAA4C;AAC9B,QAAA,UAAU,GAAG,UAAU,CAAC,UAAU,CAAC;AACjD,kEAAoD;AACtC,QAAA,cAAc,GAAG,cAAc,CAAC,cAAc,CAAC;AAC7D,4DAA8C;AAChC,QAAA,WAAW,GAAG,WAAW,CAAC,WAAW,CAAC;AACpD,+CAAiC;AACnB,QAAA,KAAK,GAAG,KAAK,CAAC,KAAK,CAAC;AAClC,sDAAwC;AAC1B,QAAA,QAAQ,GAAG,QAAQ,CAAC,QAAQ,CAAC;AAC3C,uEAAyD;AAC3C,QAAA,gBAAgB,GAAG,gBAAgB,CAAC,gBAAgB,CAAC;AAEnE,sEAAwD;AAC1C,QAAA,gBAAgB,GAAG,gBAAgB,CAAC,gBAAgB,CAAC;AACnE,8DAAgD;AAClC,QAAA,YAAY,GAAG,YAAY,CAAC,YAAY,CAAC;AACvD,gFAAkE;AACpD,QAAA,oBAAoB,GAAG,oBAAoB,CAAC,oBAAoB,CAAC;AAC/E,yEAA2D;AAC7C,QAAA,iBAAiB,GAAG,iBAAiB,CAAC,iBAAiB,CAAC;AACtE,2EAA6D;AAC/C,QAAA,kBAAkB,GAAG,kBAAkB,CAAC,kBAAkB,CAAC;AACzE,0FAA4E;AAC9D,QAAA,yBAAyB,GAAG,yBAAyB,CAAC,yBAAyB,CAAC;AAC9F,4FAA8E;AAChE,QAAA,0BAA0B,GAAG,0BAA0B,CAAC,0BAA0B,CAAC;AACjG,wFAA0E;AAC5D,QAAA,wBAAwB,GAAG,wBAAwB,CAAC,wBAAwB,CAAC;AAC3F,wFAA0E;AAC5D,QAAA,wBAAwB,GAAG,wBAAwB,CAAC,wBAAwB,CAAC;AAC3F,6EAA+D;AACjD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,yEAA2D;AAC7C,QAAA,iBAAiB,GAAG,iBAAiB,CAAC,iBAAiB,CAAC;AACtE,yEAA2D;AAC7C,QAAA,iBAAiB,GAAG,iBAAiB,CAAC,iBAAiB,CAAC;AACtE,+EAAiE;AACnD,QAAA,oBAAoB,GAAG,oBAAoB,CAAC,oBAAoB,CAAC;AAC/E,sEAAwD;AAC1C,QAAA,gBAAgB,GAAG,gBAAgB,CAAC,gBAAgB,CAAC;AACnE,8DAAgD;AAClC,QAAA,YAAY,GAAG,YAAY,CAAC,YAAY,CAAC;AACvD,0EAA4D;AAC9C,QAAA,kBAAkB,GAAG,kBAAkB,CAAC,kBAAkB,CAAC;AACzE,2EAA6D;AAC/C,QAAA,kBAAkB,GAAG,kBAAkB,CAAC,kBAAkB,CAAC;AACzE,6EAA+D;AACjD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,6EAA+D;AACjD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,8DAAgD;AAClC,QAAA,YAAY,GAAG,YAAY,CAAC,YAAY,CAAC;AACvD,wFAA0E;AAC5D,QAAA,wBAAwB,GAAG,wBAAwB,CAAC,wBAAwB,CAAC;AAC3F,wFAA0E;AAC5D,QAAA,wBAAwB,GAAG,wBAAwB,CAAC,wBAAwB,CAAC;AAC3F,wFAA0E;AAC5D,QAAA,wBAAwB,GAAG,wBAAwB,CAAC,wBAAwB,CAAC;AAC3F,0FAA4E;AAC9D,QAAA,yBAAyB,GAAG,yBAAyB,CAAC,yBAAyB,CAAC;AAC9F,8EAAgE;AAClD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,gEAAkD;AACpC,QAAA,aAAa,GAAG,aAAa,CAAC,aAAa,CAAC;AAC1D,gEAAkD;AACpC,QAAA,aAAa,GAAG,aAAa,CAAC,aAAa,CAAC;AAC1D,4DAA8C;AAChC,QAAA,WAAW,GAAG,WAAW,CAAC,WAAW,CAAC;AACpD,2EAA6D;AAC/C,QAAA,kBAAkB,GAAG,kBAAkB,CAAC,kBAAkB,CAAC;AACzE,mEAAqD;AACvC,QAAA,cAAc,GAAG,cAAc,CAAC,cAAc,CAAC;AAC7D,2EAA6D;AAC/C,QAAA,kBAAkB,GAAG,kBAAkB,CAAC,kBAAkB,CAAC;AACzE,mEAAqD;AACvC,QAAA,cAAc,GAAG,cAAc,CAAC,cAAc,CAAC;AAC7D,2EAA6D;AAC/C,QAAA,kBAAkB,GAAG,kBAAkB,CAAC,kBAAkB,CAAC;AACzE,4DAA8C;AAChC,QAAA,WAAW,GAAG,WAAW,CAAC,WAAW,CAAC;AACpD,4DAA8C;AAChC,QAAA,WAAW,GAAG,WAAW,CAAC,WAAW,CAAC;AACpD,8DAAgD;AAClC,QAAA,YAAY,GAAG,YAAY,CAAC,YAAY,CAAC"}
|
|
@@ -0,0 +1,32 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
/**
|
|
21
|
+
* Contains the HTML and CSS to render the required Search Suggestions.
|
|
22
|
+
*/
|
|
23
|
+
export interface SearchEntryPoint {
|
|
24
|
+
/**
|
|
25
|
+
* The rendered content
|
|
26
|
+
*/
|
|
27
|
+
"renderedContent"?: string;
|
|
28
|
+
}
|
|
29
|
+
export declare namespace SearchEntryPoint {
|
|
30
|
+
function getJsonObj(obj: SearchEntryPoint): object;
|
|
31
|
+
function getDeserializedJsonObj(obj: SearchEntryPoint): object;
|
|
32
|
+
}
|
|
@@ -0,0 +1,36 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
22
|
+
exports.SearchEntryPoint = void 0;
|
|
23
|
+
var SearchEntryPoint;
|
|
24
|
+
(function (SearchEntryPoint) {
|
|
25
|
+
function getJsonObj(obj) {
|
|
26
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
27
|
+
return jsonObj;
|
|
28
|
+
}
|
|
29
|
+
SearchEntryPoint.getJsonObj = getJsonObj;
|
|
30
|
+
function getDeserializedJsonObj(obj) {
|
|
31
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
32
|
+
return jsonObj;
|
|
33
|
+
}
|
|
34
|
+
SearchEntryPoint.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
35
|
+
})(SearchEntryPoint = exports.SearchEntryPoint || (exports.SearchEntryPoint = {}));
|
|
36
|
+
//# sourceMappingURL=search-entry-point.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"search-entry-point.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/search-entry-point.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;AAeH,IAAiB,gBAAgB,CAWhC;AAXD,WAAiB,gBAAgB;IAC/B,SAAgB,UAAU,CAAC,GAAqB;QAC9C,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,2BAAU,aAIzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAqB;QAC1D,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,uCAAsB,yBAIrC,CAAA;AACH,CAAC,EAXgB,gBAAgB,GAAhB,wBAAgB,KAAhB,wBAAgB,QAWhC"}
|
|
@@ -0,0 +1,32 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
import * as model from "../model";
|
|
21
|
+
/**
|
|
22
|
+
* Represents a single instance of chat video content.
|
|
23
|
+
*/
|
|
24
|
+
export interface VideoContent extends model.ChatContent {
|
|
25
|
+
"videoUrl"?: model.VideoUrl;
|
|
26
|
+
"type": string;
|
|
27
|
+
}
|
|
28
|
+
export declare namespace VideoContent {
|
|
29
|
+
function getJsonObj(obj: VideoContent, isParentJsonObj?: boolean): object;
|
|
30
|
+
const type = "VIDEO";
|
|
31
|
+
function getDeserializedJsonObj(obj: VideoContent, isParentJsonObj?: boolean): object;
|
|
32
|
+
}
|
|
@@ -0,0 +1,61 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
22
|
+
if (k2 === undefined) k2 = k;
|
|
23
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
24
|
+
}) : (function(o, m, k, k2) {
|
|
25
|
+
if (k2 === undefined) k2 = k;
|
|
26
|
+
o[k2] = m[k];
|
|
27
|
+
}));
|
|
28
|
+
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
29
|
+
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
30
|
+
}) : function(o, v) {
|
|
31
|
+
o["default"] = v;
|
|
32
|
+
});
|
|
33
|
+
var __importStar = (this && this.__importStar) || function (mod) {
|
|
34
|
+
if (mod && mod.__esModule) return mod;
|
|
35
|
+
var result = {};
|
|
36
|
+
if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
37
|
+
__setModuleDefault(result, mod);
|
|
38
|
+
return result;
|
|
39
|
+
};
|
|
40
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
41
|
+
exports.VideoContent = void 0;
|
|
42
|
+
const model = __importStar(require("../model"));
|
|
43
|
+
var VideoContent;
|
|
44
|
+
(function (VideoContent) {
|
|
45
|
+
function getJsonObj(obj, isParentJsonObj) {
|
|
46
|
+
const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj ? obj : model.ChatContent.getJsonObj(obj))), {
|
|
47
|
+
"videoUrl": obj.videoUrl ? model.VideoUrl.getJsonObj(obj.videoUrl) : undefined
|
|
48
|
+
});
|
|
49
|
+
return jsonObj;
|
|
50
|
+
}
|
|
51
|
+
VideoContent.getJsonObj = getJsonObj;
|
|
52
|
+
VideoContent.type = "VIDEO";
|
|
53
|
+
function getDeserializedJsonObj(obj, isParentJsonObj) {
|
|
54
|
+
const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj ? obj : model.ChatContent.getDeserializedJsonObj(obj))), {
|
|
55
|
+
"videoUrl": obj.videoUrl ? model.VideoUrl.getDeserializedJsonObj(obj.videoUrl) : undefined
|
|
56
|
+
});
|
|
57
|
+
return jsonObj;
|
|
58
|
+
}
|
|
59
|
+
VideoContent.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
60
|
+
})(VideoContent = exports.VideoContent || (exports.VideoContent = {}));
|
|
61
|
+
//# sourceMappingURL=video-content.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"video-content.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/video-content.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAYlC,IAAiB,YAAY,CAsB5B;AAtBD,WAAiB,YAAY;IAC3B,SAAgB,UAAU,CAAC,GAAiB,EAAE,eAAyB;QACrE,MAAM,OAAO,mCACR,CAAC,eAAe,CAAC,CAAC,CAAC,GAAG,CAAC,CAAC,CAAE,KAAK,CAAC,WAAW,CAAC,UAAU,CAAC,GAAG,CAAkB,CAAC,GAC7E;YACD,UAAU,EAAE,GAAG,CAAC,QAAQ,CAAC,CAAC,CAAC,KAAK,CAAC,QAAQ,CAAC,UAAU,CAAC,GAAG,CAAC,QAAQ,CAAC,CAAC,CAAC,CAAC,SAAS;SAC/E,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IATe,uBAAU,aASzB,CAAA;IACY,iBAAI,GAAG,OAAO,CAAC;IAC5B,SAAgB,sBAAsB,CAAC,GAAiB,EAAE,eAAyB;QACjF,MAAM,OAAO,mCACR,CAAC,eAAe,CAAC,CAAC,CAAC,GAAG,CAAC,CAAC,CAAE,KAAK,CAAC,WAAW,CAAC,sBAAsB,CAAC,GAAG,CAAkB,CAAC,GACzF;YACD,UAAU,EAAE,GAAG,CAAC,QAAQ,CAAC,CAAC,CAAC,KAAK,CAAC,QAAQ,CAAC,sBAAsB,CAAC,GAAG,CAAC,QAAQ,CAAC,CAAC,CAAC,CAAC,SAAS;SAC3F,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IATe,mCAAsB,yBASrC,CAAA;AACH,CAAC,EAtBgB,YAAY,GAAZ,oBAAY,KAAZ,oBAAY,QAsB5B"}
|