oci-generativeaiinference 2.114.1 → 2.116.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/lib/model/assistant-message.d.ts +4 -0
- package/lib/model/assistant-message.js.map +1 -1
- package/lib/model/chat-choice.d.ts +2 -1
- package/lib/model/chat-choice.js +4 -2
- package/lib/model/chat-choice.js.map +1 -1
- package/lib/model/cohere-chat-request.d.ts +2 -1
- package/lib/model/cohere-chat-request.js.map +1 -1
- package/lib/model/developer-message.d.ts +35 -0
- package/lib/model/developer-message.js +57 -0
- package/lib/model/developer-message.js.map +1 -0
- package/lib/model/function-call.d.ts +1 -1
- package/lib/model/generic-chat-request.d.ts +41 -2
- package/lib/model/generic-chat-request.js +29 -0
- package/lib/model/generic-chat-request.js.map +1 -1
- package/lib/model/generic-chat-response.d.ts +1 -0
- package/lib/model/generic-chat-response.js +4 -2
- package/lib/model/generic-chat-response.js.map +1 -1
- package/lib/model/index.d.ts +16 -0
- package/lib/model/index.js +18 -2
- package/lib/model/index.js.map +1 -1
- package/lib/model/json-object-response-format.d.ts +31 -0
- package/lib/model/json-object-response-format.js +61 -0
- package/lib/model/json-object-response-format.js.map +1 -0
- package/lib/model/json-schema-response-format.d.ts +32 -0
- package/lib/model/json-schema-response-format.js +69 -0
- package/lib/model/json-schema-response-format.js.map +1 -0
- package/lib/model/message.js +4 -0
- package/lib/model/message.js.map +1 -1
- package/lib/model/prediction.d.ts +30 -0
- package/lib/model/prediction.js +75 -0
- package/lib/model/prediction.js.map +1 -0
- package/lib/model/response-format.d.ts +29 -0
- package/lib/model/response-format.js +83 -0
- package/lib/model/response-format.js.map +1 -0
- package/lib/model/response-json-schema.d.ts +45 -0
- package/lib/model/response-json-schema.js +36 -0
- package/lib/model/response-json-schema.js.map +1 -0
- package/lib/model/static-content.d.ts +36 -0
- package/lib/model/static-content.js +69 -0
- package/lib/model/static-content.js.map +1 -0
- package/lib/model/text-response-format.d.ts +31 -0
- package/lib/model/text-response-format.js +59 -0
- package/lib/model/text-response-format.js.map +1 -0
- package/lib/model/tool-choice-function.d.ts +1 -1
- package/lib/model/tool-message.d.ts +1 -1
- package/package.json +3 -3
|
@@ -0,0 +1,61 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
22
|
+
if (k2 === undefined) k2 = k;
|
|
23
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
24
|
+
}) : (function(o, m, k, k2) {
|
|
25
|
+
if (k2 === undefined) k2 = k;
|
|
26
|
+
o[k2] = m[k];
|
|
27
|
+
}));
|
|
28
|
+
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
29
|
+
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
30
|
+
}) : function(o, v) {
|
|
31
|
+
o["default"] = v;
|
|
32
|
+
});
|
|
33
|
+
var __importStar = (this && this.__importStar) || function (mod) {
|
|
34
|
+
if (mod && mod.__esModule) return mod;
|
|
35
|
+
var result = {};
|
|
36
|
+
if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
37
|
+
__setModuleDefault(result, mod);
|
|
38
|
+
return result;
|
|
39
|
+
};
|
|
40
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
41
|
+
exports.JsonObjectResponseFormat = void 0;
|
|
42
|
+
const model = __importStar(require("../model"));
|
|
43
|
+
var JsonObjectResponseFormat;
|
|
44
|
+
(function (JsonObjectResponseFormat) {
|
|
45
|
+
function getJsonObj(obj, isParentJsonObj) {
|
|
46
|
+
const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj
|
|
47
|
+
? obj
|
|
48
|
+
: model.ResponseFormat.getJsonObj(obj))), {});
|
|
49
|
+
return jsonObj;
|
|
50
|
+
}
|
|
51
|
+
JsonObjectResponseFormat.getJsonObj = getJsonObj;
|
|
52
|
+
JsonObjectResponseFormat.type = "JSON_OBJECT";
|
|
53
|
+
function getDeserializedJsonObj(obj, isParentJsonObj) {
|
|
54
|
+
const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj
|
|
55
|
+
? obj
|
|
56
|
+
: model.ResponseFormat.getDeserializedJsonObj(obj))), {});
|
|
57
|
+
return jsonObj;
|
|
58
|
+
}
|
|
59
|
+
JsonObjectResponseFormat.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
60
|
+
})(JsonObjectResponseFormat = exports.JsonObjectResponseFormat || (exports.JsonObjectResponseFormat = {}));
|
|
61
|
+
//# sourceMappingURL=json-object-response-format.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"json-object-response-format.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/json-object-response-format.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAUlC,IAAiB,wBAAwB,CAyBxC;AAzBD,WAAiB,wBAAwB;IACvC,SAAgB,UAAU,CAAC,GAA6B,EAAE,eAAyB;QACjF,MAAM,OAAO,mCACR,CAAC,eAAe;YACjB,CAAC,CAAC,GAAG;YACL,CAAC,CAAE,KAAK,CAAC,cAAc,CAAC,UAAU,CAAC,GAAG,CAA8B,CAAC,GACpE,EAAE,CACN,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IATe,mCAAU,aASzB,CAAA;IACY,6BAAI,GAAG,aAAa,CAAC;IAClC,SAAgB,sBAAsB,CACpC,GAA6B,EAC7B,eAAyB;QAEzB,MAAM,OAAO,mCACR,CAAC,eAAe;YACjB,CAAC,CAAC,GAAG;YACL,CAAC,CAAE,KAAK,CAAC,cAAc,CAAC,sBAAsB,CAAC,GAAG,CAA8B,CAAC,GAChF,EAAE,CACN,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAZe,+CAAsB,yBAYrC,CAAA;AACH,CAAC,EAzBgB,wBAAwB,GAAxB,gCAAwB,KAAxB,gCAAwB,QAyBxC"}
|
|
@@ -0,0 +1,32 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
import * as model from "../model";
|
|
21
|
+
/**
|
|
22
|
+
* Enables Structured Outputs which ensures the model will match your supplied JSON schema.
|
|
23
|
+
*/
|
|
24
|
+
export interface JsonSchemaResponseFormat extends model.ResponseFormat {
|
|
25
|
+
"jsonSchema"?: model.ResponseJsonSchema;
|
|
26
|
+
"type": string;
|
|
27
|
+
}
|
|
28
|
+
export declare namespace JsonSchemaResponseFormat {
|
|
29
|
+
function getJsonObj(obj: JsonSchemaResponseFormat, isParentJsonObj?: boolean): object;
|
|
30
|
+
const type = "JSON_SCHEMA";
|
|
31
|
+
function getDeserializedJsonObj(obj: JsonSchemaResponseFormat, isParentJsonObj?: boolean): object;
|
|
32
|
+
}
|
|
@@ -0,0 +1,69 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
22
|
+
if (k2 === undefined) k2 = k;
|
|
23
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
24
|
+
}) : (function(o, m, k, k2) {
|
|
25
|
+
if (k2 === undefined) k2 = k;
|
|
26
|
+
o[k2] = m[k];
|
|
27
|
+
}));
|
|
28
|
+
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
29
|
+
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
30
|
+
}) : function(o, v) {
|
|
31
|
+
o["default"] = v;
|
|
32
|
+
});
|
|
33
|
+
var __importStar = (this && this.__importStar) || function (mod) {
|
|
34
|
+
if (mod && mod.__esModule) return mod;
|
|
35
|
+
var result = {};
|
|
36
|
+
if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
37
|
+
__setModuleDefault(result, mod);
|
|
38
|
+
return result;
|
|
39
|
+
};
|
|
40
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
41
|
+
exports.JsonSchemaResponseFormat = void 0;
|
|
42
|
+
const model = __importStar(require("../model"));
|
|
43
|
+
var JsonSchemaResponseFormat;
|
|
44
|
+
(function (JsonSchemaResponseFormat) {
|
|
45
|
+
function getJsonObj(obj, isParentJsonObj) {
|
|
46
|
+
const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj
|
|
47
|
+
? obj
|
|
48
|
+
: model.ResponseFormat.getJsonObj(obj))), {
|
|
49
|
+
"jsonSchema": obj.jsonSchema
|
|
50
|
+
? model.ResponseJsonSchema.getJsonObj(obj.jsonSchema)
|
|
51
|
+
: undefined
|
|
52
|
+
});
|
|
53
|
+
return jsonObj;
|
|
54
|
+
}
|
|
55
|
+
JsonSchemaResponseFormat.getJsonObj = getJsonObj;
|
|
56
|
+
JsonSchemaResponseFormat.type = "JSON_SCHEMA";
|
|
57
|
+
function getDeserializedJsonObj(obj, isParentJsonObj) {
|
|
58
|
+
const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj
|
|
59
|
+
? obj
|
|
60
|
+
: model.ResponseFormat.getDeserializedJsonObj(obj))), {
|
|
61
|
+
"jsonSchema": obj.jsonSchema
|
|
62
|
+
? model.ResponseJsonSchema.getDeserializedJsonObj(obj.jsonSchema)
|
|
63
|
+
: undefined
|
|
64
|
+
});
|
|
65
|
+
return jsonObj;
|
|
66
|
+
}
|
|
67
|
+
JsonSchemaResponseFormat.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
68
|
+
})(JsonSchemaResponseFormat = exports.JsonSchemaResponseFormat || (exports.JsonSchemaResponseFormat = {}));
|
|
69
|
+
//# sourceMappingURL=json-schema-response-format.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"json-schema-response-format.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/json-schema-response-format.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAYlC,IAAiB,wBAAwB,CAiCxC;AAjCD,WAAiB,wBAAwB;IACvC,SAAgB,UAAU,CAAC,GAA6B,EAAE,eAAyB;QACjF,MAAM,OAAO,mCACR,CAAC,eAAe;YACjB,CAAC,CAAC,GAAG;YACL,CAAC,CAAE,KAAK,CAAC,cAAc,CAAC,UAAU,CAAC,GAAG,CAA8B,CAAC,GACpE;YACD,YAAY,EAAE,GAAG,CAAC,UAAU;gBAC1B,CAAC,CAAC,KAAK,CAAC,kBAAkB,CAAC,UAAU,CAAC,GAAG,CAAC,UAAU,CAAC;gBACrD,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAbe,mCAAU,aAazB,CAAA;IACY,6BAAI,GAAG,aAAa,CAAC;IAClC,SAAgB,sBAAsB,CACpC,GAA6B,EAC7B,eAAyB;QAEzB,MAAM,OAAO,mCACR,CAAC,eAAe;YACjB,CAAC,CAAC,GAAG;YACL,CAAC,CAAE,KAAK,CAAC,cAAc,CAAC,sBAAsB,CAAC,GAAG,CAA8B,CAAC,GAChF;YACD,YAAY,EAAE,GAAG,CAAC,UAAU;gBAC1B,CAAC,CAAC,KAAK,CAAC,kBAAkB,CAAC,sBAAsB,CAAC,GAAG,CAAC,UAAU,CAAC;gBACjE,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAhBe,+CAAsB,yBAgBrC,CAAA;AACH,CAAC,EAjCgB,wBAAwB,GAAxB,gCAAwB,KAAxB,gCAAwB,QAiCxC"}
|
package/lib/model/message.js
CHANGED
|
@@ -61,6 +61,8 @@ var Message;
|
|
|
61
61
|
return model.UserMessage.getJsonObj(jsonObj, true);
|
|
62
62
|
case "TOOL":
|
|
63
63
|
return model.ToolMessage.getJsonObj(jsonObj, true);
|
|
64
|
+
case "DEVELOPER":
|
|
65
|
+
return model.DeveloperMessage.getJsonObj(jsonObj, true);
|
|
64
66
|
default:
|
|
65
67
|
if (common.LOG.logger)
|
|
66
68
|
common.LOG.logger.info(`Unknown value for: ${obj.role}`);
|
|
@@ -87,6 +89,8 @@ var Message;
|
|
|
87
89
|
return model.UserMessage.getDeserializedJsonObj(jsonObj, true);
|
|
88
90
|
case "TOOL":
|
|
89
91
|
return model.ToolMessage.getDeserializedJsonObj(jsonObj, true);
|
|
92
|
+
case "DEVELOPER":
|
|
93
|
+
return model.DeveloperMessage.getDeserializedJsonObj(jsonObj, true);
|
|
90
94
|
default:
|
|
91
95
|
if (common.LOG.logger)
|
|
92
96
|
common.LOG.logger.info(`Unknown value for: ${obj.role}`);
|
package/lib/model/message.js.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"message.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/message.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAClC,qCAAsC;AActC,IAAiB,OAAO,
|
|
1
|
+
{"version":3,"file":"message.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/message.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAClC,qCAAsC;AActC,IAAiB,OAAO,CA4EvB;AA5ED,WAAiB,OAAO;IACtB,SAAgB,UAAU,CAAC,GAAY;QACrC,MAAM,OAAO,mCACR,GAAG,GACH;YACD,SAAS,EAAE,GAAG,CAAC,OAAO;gBACpB,CAAC,CAAC,GAAG,CAAC,OAAO,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACrB,OAAO,KAAK,CAAC,WAAW,CAAC,UAAU,CAAC,IAAI,CAAC,CAAC;gBAC5C,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,IAAI,GAAG,IAAI,MAAM,IAAI,GAAG,IAAI,GAAG,CAAC,IAAI,EAAE;YACpC,QAAQ,GAAG,CAAC,IAAI,EAAE;gBAChB,KAAK,QAAQ;oBACX,OAAO,KAAK,CAAC,aAAa,CAAC,UAAU,CAA+B,OAAQ,EAAE,IAAI,CAAC,CAAC;gBACtF,KAAK,WAAW;oBACd,OAAO,KAAK,CAAC,gBAAgB,CAAC,UAAU,CAAkC,OAAQ,EAAE,IAAI,CAAC,CAAC;gBAC5F,KAAK,MAAM;oBACT,OAAO,KAAK,CAAC,WAAW,CAAC,UAAU,CAA6B,OAAQ,EAAE,IAAI,CAAC,CAAC;gBAClF,KAAK,MAAM;oBACT,OAAO,KAAK,CAAC,WAAW,CAAC,UAAU,CAA6B,OAAQ,EAAE,IAAI,CAAC,CAAC;gBAClF,KAAK,WAAW;oBACd,OAAO,KAAK,CAAC,gBAAgB,CAAC,UAAU,CAAkC,OAAQ,EAAE,IAAI,CAAC,CAAC;gBAC5F;oBACE,IAAI,MAAM,CAAC,GAAG,CAAC,MAAM;wBAAE,MAAM,CAAC,GAAG,CAAC,MAAM,CAAC,IAAI,CAAC,sBAAsB,GAAG,CAAC,IAAI,EAAE,CAAC,CAAC;aACnF;SACF;QACD,OAAO,OAAO,CAAC;IACjB,CAAC;IA7Be,kBAAU,aA6BzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAY;QACjD,MAAM,OAAO,mCACR,GAAG,GACH;YACD,SAAS,EAAE,GAAG,CAAC,OAAO;gBACpB,CAAC,CAAC,GAAG,CAAC,OAAO,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACrB,OAAO,KAAK,CAAC,WAAW,CAAC,sBAAsB,CAAC,IAAI,CAAC,CAAC;gBACxD,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,IAAI,GAAG,IAAI,MAAM,IAAI,GAAG,IAAI,GAAG,CAAC,IAAI,EAAE;YACpC,QAAQ,GAAG,CAAC,IAAI,EAAE;gBAChB,KAAK,QAAQ;oBACX,OAAO,KAAK,CAAC,aAAa,CAAC,sBAAsB,CACjB,OAAQ,EACtC,IAAI,CACL,CAAC;gBACJ,KAAK,WAAW;oBACd,OAAO,KAAK,CAAC,gBAAgB,CAAC,sBAAsB,CACjB,OAAQ,EACzC,IAAI,CACL,CAAC;gBACJ,KAAK,MAAM;oBACT,OAAO,KAAK,CAAC,WAAW,CAAC,sBAAsB,CACjB,OAAQ,EACpC,IAAI,CACL,CAAC;gBACJ,KAAK,MAAM;oBACT,OAAO,KAAK,CAAC,WAAW,CAAC,sBAAsB,CACjB,OAAQ,EACpC,IAAI,CACL,CAAC;gBACJ,KAAK,WAAW;oBACd,OAAO,KAAK,CAAC,gBAAgB,CAAC,sBAAsB,CACjB,OAAQ,EACzC,IAAI,CACL,CAAC;gBACJ;oBACE,IAAI,MAAM,CAAC,GAAG,CAAC,MAAM;wBAAE,MAAM,CAAC,GAAG,CAAC,MAAM,CAAC,IAAI,CAAC,sBAAsB,GAAG,CAAC,IAAI,EAAE,CAAC,CAAC;aACnF;SACF;QACD,OAAO,OAAO,CAAC;IACjB,CAAC;IA5Ce,8BAAsB,yBA4CrC,CAAA;AACH,CAAC,EA5EgB,OAAO,GAAP,eAAO,KAAP,eAAO,QA4EvB"}
|
|
@@ -0,0 +1,30 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
/**
|
|
21
|
+
* Configuration for a Predicted Output, which can greatly improve response times when large parts of the model response are known ahead of time. This is most common when you are regenerating a file with only minor changes to most of the content.
|
|
22
|
+
*
|
|
23
|
+
*/
|
|
24
|
+
export interface Prediction {
|
|
25
|
+
"type": string;
|
|
26
|
+
}
|
|
27
|
+
export declare namespace Prediction {
|
|
28
|
+
function getJsonObj(obj: Prediction): object;
|
|
29
|
+
function getDeserializedJsonObj(obj: Prediction): object;
|
|
30
|
+
}
|
|
@@ -0,0 +1,75 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
22
|
+
if (k2 === undefined) k2 = k;
|
|
23
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
24
|
+
}) : (function(o, m, k, k2) {
|
|
25
|
+
if (k2 === undefined) k2 = k;
|
|
26
|
+
o[k2] = m[k];
|
|
27
|
+
}));
|
|
28
|
+
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
29
|
+
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
30
|
+
}) : function(o, v) {
|
|
31
|
+
o["default"] = v;
|
|
32
|
+
});
|
|
33
|
+
var __importStar = (this && this.__importStar) || function (mod) {
|
|
34
|
+
if (mod && mod.__esModule) return mod;
|
|
35
|
+
var result = {};
|
|
36
|
+
if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
37
|
+
__setModuleDefault(result, mod);
|
|
38
|
+
return result;
|
|
39
|
+
};
|
|
40
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
41
|
+
exports.Prediction = void 0;
|
|
42
|
+
const model = __importStar(require("../model"));
|
|
43
|
+
const common = require("oci-common");
|
|
44
|
+
var Prediction;
|
|
45
|
+
(function (Prediction) {
|
|
46
|
+
function getJsonObj(obj) {
|
|
47
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
48
|
+
if (obj && "type" in obj && obj.type) {
|
|
49
|
+
switch (obj.type) {
|
|
50
|
+
case "CONTENT":
|
|
51
|
+
return model.StaticContent.getJsonObj(jsonObj, true);
|
|
52
|
+
default:
|
|
53
|
+
if (common.LOG.logger)
|
|
54
|
+
common.LOG.logger.info(`Unknown value for: ${obj.type}`);
|
|
55
|
+
}
|
|
56
|
+
}
|
|
57
|
+
return jsonObj;
|
|
58
|
+
}
|
|
59
|
+
Prediction.getJsonObj = getJsonObj;
|
|
60
|
+
function getDeserializedJsonObj(obj) {
|
|
61
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
62
|
+
if (obj && "type" in obj && obj.type) {
|
|
63
|
+
switch (obj.type) {
|
|
64
|
+
case "CONTENT":
|
|
65
|
+
return model.StaticContent.getDeserializedJsonObj(jsonObj, true);
|
|
66
|
+
default:
|
|
67
|
+
if (common.LOG.logger)
|
|
68
|
+
common.LOG.logger.info(`Unknown value for: ${obj.type}`);
|
|
69
|
+
}
|
|
70
|
+
}
|
|
71
|
+
return jsonObj;
|
|
72
|
+
}
|
|
73
|
+
Prediction.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
74
|
+
})(Prediction = exports.Prediction || (exports.Prediction = {}));
|
|
75
|
+
//# sourceMappingURL=prediction.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"prediction.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/prediction.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAClC,qCAAsC;AAUtC,IAAiB,UAAU,CA8B1B;AA9BD,WAAiB,UAAU;IACzB,SAAgB,UAAU,CAAC,GAAe;QACxC,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,IAAI,GAAG,IAAI,MAAM,IAAI,GAAG,IAAI,GAAG,CAAC,IAAI,EAAE;YACpC,QAAQ,GAAG,CAAC,IAAI,EAAE;gBAChB,KAAK,SAAS;oBACZ,OAAO,KAAK,CAAC,aAAa,CAAC,UAAU,CAA+B,OAAQ,EAAE,IAAI,CAAC,CAAC;gBACtF;oBACE,IAAI,MAAM,CAAC,GAAG,CAAC,MAAM;wBAAE,MAAM,CAAC,GAAG,CAAC,MAAM,CAAC,IAAI,CAAC,sBAAsB,GAAG,CAAC,IAAI,EAAE,CAAC,CAAC;aACnF;SACF;QACD,OAAO,OAAO,CAAC;IACjB,CAAC;IAZe,qBAAU,aAYzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAe;QACpD,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,IAAI,GAAG,IAAI,MAAM,IAAI,GAAG,IAAI,GAAG,CAAC,IAAI,EAAE;YACpC,QAAQ,GAAG,CAAC,IAAI,EAAE;gBAChB,KAAK,SAAS;oBACZ,OAAO,KAAK,CAAC,aAAa,CAAC,sBAAsB,CACjB,OAAQ,EACtC,IAAI,CACL,CAAC;gBACJ;oBACE,IAAI,MAAM,CAAC,GAAG,CAAC,MAAM;wBAAE,MAAM,CAAC,GAAG,CAAC,MAAM,CAAC,IAAI,CAAC,sBAAsB,GAAG,CAAC,IAAI,EAAE,CAAC,CAAC;aACnF;SACF;QACD,OAAO,OAAO,CAAC;IACjB,CAAC;IAfe,iCAAsB,yBAerC,CAAA;AACH,CAAC,EA9BgB,UAAU,GAAV,kBAAU,KAAV,kBAAU,QA8B1B"}
|
|
@@ -0,0 +1,29 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
/**
|
|
21
|
+
* An object specifying the format that the model must output.
|
|
22
|
+
*/
|
|
23
|
+
export interface ResponseFormat {
|
|
24
|
+
"type": string;
|
|
25
|
+
}
|
|
26
|
+
export declare namespace ResponseFormat {
|
|
27
|
+
function getJsonObj(obj: ResponseFormat): object;
|
|
28
|
+
function getDeserializedJsonObj(obj: ResponseFormat): object;
|
|
29
|
+
}
|
|
@@ -0,0 +1,83 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
22
|
+
if (k2 === undefined) k2 = k;
|
|
23
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
24
|
+
}) : (function(o, m, k, k2) {
|
|
25
|
+
if (k2 === undefined) k2 = k;
|
|
26
|
+
o[k2] = m[k];
|
|
27
|
+
}));
|
|
28
|
+
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
29
|
+
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
30
|
+
}) : function(o, v) {
|
|
31
|
+
o["default"] = v;
|
|
32
|
+
});
|
|
33
|
+
var __importStar = (this && this.__importStar) || function (mod) {
|
|
34
|
+
if (mod && mod.__esModule) return mod;
|
|
35
|
+
var result = {};
|
|
36
|
+
if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
37
|
+
__setModuleDefault(result, mod);
|
|
38
|
+
return result;
|
|
39
|
+
};
|
|
40
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
41
|
+
exports.ResponseFormat = void 0;
|
|
42
|
+
const model = __importStar(require("../model"));
|
|
43
|
+
const common = require("oci-common");
|
|
44
|
+
var ResponseFormat;
|
|
45
|
+
(function (ResponseFormat) {
|
|
46
|
+
function getJsonObj(obj) {
|
|
47
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
48
|
+
if (obj && "type" in obj && obj.type) {
|
|
49
|
+
switch (obj.type) {
|
|
50
|
+
case "TEXT":
|
|
51
|
+
return model.TextResponseFormat.getJsonObj(jsonObj, true);
|
|
52
|
+
case "JSON_OBJECT":
|
|
53
|
+
return model.JsonObjectResponseFormat.getJsonObj(jsonObj, true);
|
|
54
|
+
case "JSON_SCHEMA":
|
|
55
|
+
return model.JsonSchemaResponseFormat.getJsonObj(jsonObj, true);
|
|
56
|
+
default:
|
|
57
|
+
if (common.LOG.logger)
|
|
58
|
+
common.LOG.logger.info(`Unknown value for: ${obj.type}`);
|
|
59
|
+
}
|
|
60
|
+
}
|
|
61
|
+
return jsonObj;
|
|
62
|
+
}
|
|
63
|
+
ResponseFormat.getJsonObj = getJsonObj;
|
|
64
|
+
function getDeserializedJsonObj(obj) {
|
|
65
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
66
|
+
if (obj && "type" in obj && obj.type) {
|
|
67
|
+
switch (obj.type) {
|
|
68
|
+
case "TEXT":
|
|
69
|
+
return model.TextResponseFormat.getDeserializedJsonObj(jsonObj, true);
|
|
70
|
+
case "JSON_OBJECT":
|
|
71
|
+
return model.JsonObjectResponseFormat.getDeserializedJsonObj(jsonObj, true);
|
|
72
|
+
case "JSON_SCHEMA":
|
|
73
|
+
return model.JsonSchemaResponseFormat.getDeserializedJsonObj(jsonObj, true);
|
|
74
|
+
default:
|
|
75
|
+
if (common.LOG.logger)
|
|
76
|
+
common.LOG.logger.info(`Unknown value for: ${obj.type}`);
|
|
77
|
+
}
|
|
78
|
+
}
|
|
79
|
+
return jsonObj;
|
|
80
|
+
}
|
|
81
|
+
ResponseFormat.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
82
|
+
})(ResponseFormat = exports.ResponseFormat || (exports.ResponseFormat = {}));
|
|
83
|
+
//# sourceMappingURL=response-format.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"response-format.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/response-format.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAClC,qCAAsC;AAStC,IAAiB,cAAc,CAqD9B;AArDD,WAAiB,cAAc;IAC7B,SAAgB,UAAU,CAAC,GAAmB;QAC5C,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,IAAI,GAAG,IAAI,MAAM,IAAI,GAAG,IAAI,GAAG,CAAC,IAAI,EAAE;YACpC,QAAQ,GAAG,CAAC,IAAI,EAAE;gBAChB,KAAK,MAAM;oBACT,OAAO,KAAK,CAAC,kBAAkB,CAAC,UAAU,CACL,OAAQ,EAC3C,IAAI,CACL,CAAC;gBACJ,KAAK,aAAa;oBAChB,OAAO,KAAK,CAAC,wBAAwB,CAAC,UAAU,CACL,OAAQ,EACjD,IAAI,CACL,CAAC;gBACJ,KAAK,aAAa;oBAChB,OAAO,KAAK,CAAC,wBAAwB,CAAC,UAAU,CACL,OAAQ,EACjD,IAAI,CACL,CAAC;gBACJ;oBACE,IAAI,MAAM,CAAC,GAAG,CAAC,MAAM;wBAAE,MAAM,CAAC,GAAG,CAAC,MAAM,CAAC,IAAI,CAAC,sBAAsB,GAAG,CAAC,IAAI,EAAE,CAAC,CAAC;aACnF;SACF;QACD,OAAO,OAAO,CAAC;IACjB,CAAC;IAzBe,yBAAU,aAyBzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAmB;QACxD,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,IAAI,GAAG,IAAI,MAAM,IAAI,GAAG,IAAI,GAAG,CAAC,IAAI,EAAE;YACpC,QAAQ,GAAG,CAAC,IAAI,EAAE;gBAChB,KAAK,MAAM;oBACT,OAAO,KAAK,CAAC,kBAAkB,CAAC,sBAAsB,CACjB,OAAQ,EAC3C,IAAI,CACL,CAAC;gBACJ,KAAK,aAAa;oBAChB,OAAO,KAAK,CAAC,wBAAwB,CAAC,sBAAsB,CACjB,OAAQ,EACjD,IAAI,CACL,CAAC;gBACJ,KAAK,aAAa;oBAChB,OAAO,KAAK,CAAC,wBAAwB,CAAC,sBAAsB,CACjB,OAAQ,EACjD,IAAI,CACL,CAAC;gBACJ;oBACE,IAAI,MAAM,CAAC,GAAG,CAAC,MAAM;wBAAE,MAAM,CAAC,GAAG,CAAC,MAAM,CAAC,IAAI,CAAC,sBAAsB,GAAG,CAAC,IAAI,EAAE,CAAC,CAAC;aACnF;SACF;QACD,OAAO,OAAO,CAAC;IACjB,CAAC;IAzBe,qCAAsB,yBAyBrC,CAAA;AACH,CAAC,EArDgB,cAAc,GAAd,sBAAc,KAAd,sBAAc,QAqD9B"}
|
|
@@ -0,0 +1,45 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
/**
|
|
21
|
+
* The JSON schema definition to be used in JSON_SCHEMA response format.
|
|
22
|
+
*/
|
|
23
|
+
export interface ResponseJsonSchema {
|
|
24
|
+
/**
|
|
25
|
+
* The name of the response format. Must be a-z, A-Z, 0-9, or contain underscores and dashes.
|
|
26
|
+
*/
|
|
27
|
+
"name": string;
|
|
28
|
+
/**
|
|
29
|
+
* A description of what the response format is for, used by the model to determine how to respond in the format.
|
|
30
|
+
*/
|
|
31
|
+
"description"?: string;
|
|
32
|
+
/**
|
|
33
|
+
* The schema used by the structured output, described as a JSON Schema object.
|
|
34
|
+
*/
|
|
35
|
+
"schema"?: any;
|
|
36
|
+
/**
|
|
37
|
+
* Whether to enable strict schema adherence when generating the output. If set to true, the model will always follow the exact schema defined in the schema field. Only a subset of JSON Schema is supported when strict is true.
|
|
38
|
+
*
|
|
39
|
+
*/
|
|
40
|
+
"isStrict"?: boolean;
|
|
41
|
+
}
|
|
42
|
+
export declare namespace ResponseJsonSchema {
|
|
43
|
+
function getJsonObj(obj: ResponseJsonSchema): object;
|
|
44
|
+
function getDeserializedJsonObj(obj: ResponseJsonSchema): object;
|
|
45
|
+
}
|
|
@@ -0,0 +1,36 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
22
|
+
exports.ResponseJsonSchema = void 0;
|
|
23
|
+
var ResponseJsonSchema;
|
|
24
|
+
(function (ResponseJsonSchema) {
|
|
25
|
+
function getJsonObj(obj) {
|
|
26
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
27
|
+
return jsonObj;
|
|
28
|
+
}
|
|
29
|
+
ResponseJsonSchema.getJsonObj = getJsonObj;
|
|
30
|
+
function getDeserializedJsonObj(obj) {
|
|
31
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
32
|
+
return jsonObj;
|
|
33
|
+
}
|
|
34
|
+
ResponseJsonSchema.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
35
|
+
})(ResponseJsonSchema = exports.ResponseJsonSchema || (exports.ResponseJsonSchema = {}));
|
|
36
|
+
//# sourceMappingURL=response-json-schema.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"response-json-schema.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/response-json-schema.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;AA4BH,IAAiB,kBAAkB,CAWlC;AAXD,WAAiB,kBAAkB;IACjC,SAAgB,UAAU,CAAC,GAAuB;QAChD,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,6BAAU,aAIzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAuB;QAC5D,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,yCAAsB,yBAIrC,CAAA;AACH,CAAC,EAXgB,kBAAkB,GAAlB,0BAAkB,KAAlB,0BAAkB,QAWlC"}
|
|
@@ -0,0 +1,36 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
import * as model from "../model";
|
|
21
|
+
/**
|
|
22
|
+
* Static predicted output content, such as the content of a text file that is being regenerated.
|
|
23
|
+
*/
|
|
24
|
+
export interface StaticContent extends model.Prediction {
|
|
25
|
+
/**
|
|
26
|
+
* The content that should be matched when generating a model response. If generated tokens would match this content, the entire model response can be returned much more quickly.
|
|
27
|
+
*
|
|
28
|
+
*/
|
|
29
|
+
"content"?: Array<model.TextContent>;
|
|
30
|
+
"type": string;
|
|
31
|
+
}
|
|
32
|
+
export declare namespace StaticContent {
|
|
33
|
+
function getJsonObj(obj: StaticContent, isParentJsonObj?: boolean): object;
|
|
34
|
+
const type = "CONTENT";
|
|
35
|
+
function getDeserializedJsonObj(obj: StaticContent, isParentJsonObj?: boolean): object;
|
|
36
|
+
}
|