oci-generativeaiinference 2.84.0 → 2.84.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/lib/client.d.ts +17 -3
- package/lib/client.js +78 -3
- package/lib/client.js.map +1 -1
- package/lib/model/base-chat-request.d.ts +29 -0
- package/lib/model/base-chat-request.js +79 -0
- package/lib/model/base-chat-request.js.map +1 -0
- package/lib/model/base-chat-response.d.ts +29 -0
- package/lib/model/base-chat-response.js +79 -0
- package/lib/model/base-chat-response.js.map +1 -0
- package/lib/model/chat-choice.d.ts +42 -0
- package/lib/model/chat-choice.js +62 -0
- package/lib/model/chat-choice.js.map +1 -0
- package/lib/model/chat-content.d.ts +29 -0
- package/lib/model/chat-content.js +75 -0
- package/lib/model/chat-content.js.map +1 -0
- package/lib/model/chat-details.d.ts +35 -0
- package/lib/model/chat-details.js +68 -0
- package/lib/model/chat-details.js.map +1 -0
- package/lib/model/chat-result.d.ts +38 -0
- package/lib/model/chat-result.js +64 -0
- package/lib/model/chat-result.js.map +1 -0
- package/lib/model/citation.d.ts +44 -0
- package/lib/model/citation.js +36 -0
- package/lib/model/citation.js.map +1 -0
- package/lib/model/cohere-chat-request.d.ts +95 -0
- package/lib/model/cohere-chat-request.js +71 -0
- package/lib/model/cohere-chat-request.js.map +1 -0
- package/lib/model/cohere-chat-response.d.ts +70 -0
- package/lib/model/cohere-chat-response.js +95 -0
- package/lib/model/cohere-chat-response.js.map +1 -0
- package/lib/model/cohere-message.d.ts +40 -0
- package/lib/model/cohere-message.js +41 -0
- package/lib/model/cohere-message.js.map +1 -0
- package/lib/model/embed-text-details.d.ts +1 -1
- package/lib/model/generic-chat-request.d.ts +98 -0
- package/lib/model/generic-chat-request.js +71 -0
- package/lib/model/generic-chat-request.js.map +1 -0
- package/lib/model/generic-chat-response.d.ts +39 -0
- package/lib/model/generic-chat-response.js +71 -0
- package/lib/model/generic-chat-response.js.map +1 -0
- package/lib/model/index.d.ts +30 -0
- package/lib/model/index.js +31 -1
- package/lib/model/index.js.map +1 -1
- package/lib/model/llama-llm-inference-request.d.ts +1 -1
- package/lib/model/message.d.ts +37 -0
- package/lib/model/message.js +68 -0
- package/lib/model/message.js.map +1 -0
- package/lib/model/search-query.d.ts +32 -0
- package/lib/model/search-query.js +36 -0
- package/lib/model/search-query.js.map +1 -0
- package/lib/model/text-content.d.ts +35 -0
- package/lib/model/text-content.js +57 -0
- package/lib/model/text-content.js.map +1 -0
- package/lib/request/chat-request.d.ts +36 -0
- package/lib/request/chat-request.js +15 -0
- package/lib/request/chat-request.js.map +1 -0
- package/lib/request/embed-text-request.d.ts +1 -1
- package/lib/request/generate-text-request.d.ts +1 -1
- package/lib/request/index.d.ts +2 -0
- package/lib/request/summarize-text-request.d.ts +1 -1
- package/lib/response/chat-response.d.ts +30 -0
- package/lib/response/chat-response.js +15 -0
- package/lib/response/chat-response.js.map +1 -0
- package/lib/response/index.d.ts +2 -0
- package/package.json +3 -3
package/lib/model/index.js.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"index.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/index.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,iDAAmC;AACrB,QAAA,MAAM,GAAG,MAAM,CAAC,MAAM,CAAC;AACrC,uEAAyD;AAC3C,QAAA,gBAAgB,GAAG,gBAAgB,CAAC,gBAAgB,CAAC;AACnE,qEAAuD;AACzC,QAAA,eAAe,GAAG,eAAe,CAAC,eAAe,CAAC;AAChE,6EAA+D;AACjD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,2EAA6D;AAC/C,QAAA,kBAAkB,GAAG,kBAAkB,CAAC,kBAAkB,CAAC;AACzE,gEAAkD;AACpC,QAAA,aAAa,GAAG,aAAa,CAAC,aAAa,CAAC;AAC1D,6EAA+D;AACjD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,+EAAiE;AACnD,QAAA,oBAAoB,GAAG,oBAAoB,CAAC,oBAAoB,CAAC;AAC/E,qDAAuC;AACzB,QAAA,QAAQ,GAAG,QAAQ,CAAC,QAAQ,CAAC;AAC3C,4DAA8C;AAChC,QAAA,WAAW,GAAG,WAAW,CAAC,WAAW,CAAC;AACpD,+EAAiE;AACnD,QAAA,oBAAoB,GAAG,oBAAoB,CAAC,oBAAoB,CAAC;AAC/E,6EAA+D;AACjD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,oEAAsD;AACxC,QAAA,eAAe,GAAG,eAAe,CAAC,eAAe,CAAC;AAEhE,0FAA4E;AAC9D,QAAA,yBAAyB,GAAG,yBAAyB,CAAC,yBAAyB,CAAC;AAC9F,4FAA8E;AAChE,QAAA,0BAA0B,GAAG,0BAA0B,CAAC,0BAA0B,CAAC;AACjG,+EAAiE;AACnD,QAAA,oBAAoB,GAAG,oBAAoB,CAAC,oBAAoB,CAAC;AAC/E,wFAA0E;AAC5D,QAAA,wBAAwB,GAAG,wBAAwB,CAAC,wBAAwB,CAAC;AAC3F,0FAA4E;AAC9D,QAAA,yBAAyB,GAAG,yBAAyB,CAAC,yBAAyB,CAAC;AAC9F,8EAAgE;AAClD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC"}
|
|
1
|
+
{"version":3,"file":"index.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/index.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,qEAAuD;AACzC,QAAA,eAAe,GAAG,eAAe,CAAC,eAAe,CAAC;AAChE,uEAAyD;AAC3C,QAAA,gBAAgB,GAAG,gBAAgB,CAAC,gBAAgB,CAAC;AACnE,0DAA4C;AAC9B,QAAA,UAAU,GAAG,UAAU,CAAC,UAAU,CAAC;AACjD,4DAA8C;AAChC,QAAA,WAAW,GAAG,WAAW,CAAC,WAAW,CAAC;AACpD,4DAA8C;AAChC,QAAA,WAAW,GAAG,WAAW,CAAC,WAAW,CAAC;AACpD,0DAA4C;AAC9B,QAAA,UAAU,GAAG,UAAU,CAAC,UAAU,CAAC;AACjD,iDAAmC;AACrB,QAAA,MAAM,GAAG,MAAM,CAAC,MAAM,CAAC;AACrC,qDAAuC;AACzB,QAAA,QAAQ,GAAG,QAAQ,CAAC,QAAQ,CAAC;AAC3C,gEAAkD;AACpC,QAAA,aAAa,GAAG,aAAa,CAAC,aAAa,CAAC;AAC1D,uEAAyD;AAC3C,QAAA,gBAAgB,GAAG,gBAAgB,CAAC,gBAAgB,CAAC;AACnE,qEAAuD;AACzC,QAAA,eAAe,GAAG,eAAe,CAAC,eAAe,CAAC;AAChE,6EAA+D;AACjD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,2EAA6D;AAC/C,QAAA,kBAAkB,GAAG,kBAAkB,CAAC,kBAAkB,CAAC;AACzE,gEAAkD;AACpC,QAAA,aAAa,GAAG,aAAa,CAAC,aAAa,CAAC;AAC1D,6EAA+D;AACjD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,+EAAiE;AACnD,QAAA,oBAAoB,GAAG,oBAAoB,CAAC,oBAAoB,CAAC;AAC/E,qDAAuC;AACzB,QAAA,QAAQ,GAAG,QAAQ,CAAC,QAAQ,CAAC;AAC3C,mDAAqC;AACvB,QAAA,OAAO,GAAG,OAAO,CAAC,OAAO,CAAC;AACxC,4DAA8C;AAChC,QAAA,WAAW,GAAG,WAAW,CAAC,WAAW,CAAC;AACpD,4DAA8C;AAChC,QAAA,WAAW,GAAG,WAAW,CAAC,WAAW,CAAC;AACpD,+EAAiE;AACnD,QAAA,oBAAoB,GAAG,oBAAoB,CAAC,oBAAoB,CAAC;AAC/E,6EAA+D;AACjD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,oEAAsD;AACxC,QAAA,eAAe,GAAG,eAAe,CAAC,eAAe,CAAC;AAEhE,yEAA2D;AAC7C,QAAA,iBAAiB,GAAG,iBAAiB,CAAC,iBAAiB,CAAC;AACtE,2EAA6D;AAC/C,QAAA,kBAAkB,GAAG,kBAAkB,CAAC,kBAAkB,CAAC;AACzE,0FAA4E;AAC9D,QAAA,yBAAyB,GAAG,yBAAyB,CAAC,yBAAyB,CAAC;AAC9F,4FAA8E;AAChE,QAAA,0BAA0B,GAAG,0BAA0B,CAAC,0BAA0B,CAAC;AACjG,+EAAiE;AACnD,QAAA,oBAAoB,GAAG,oBAAoB,CAAC,oBAAoB,CAAC;AAC/E,2EAA6D;AAC/C,QAAA,kBAAkB,GAAG,kBAAkB,CAAC,kBAAkB,CAAC;AACzE,6EAA+D;AACjD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,wFAA0E;AAC5D,QAAA,wBAAwB,GAAG,wBAAwB,CAAC,wBAAwB,CAAC;AAC3F,0FAA4E;AAC9D,QAAA,yBAAyB,GAAG,yBAAyB,CAAC,yBAAyB,CAAC;AAC9F,8EAAgE;AAClD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,4DAA8C;AAChC,QAAA,WAAW,GAAG,WAAW,CAAC,WAAW,CAAC"}
|
|
@@ -82,7 +82,7 @@ export interface LlamaLlmInferenceRequest extends model.LlmInferenceRequest {
|
|
|
82
82
|
*/
|
|
83
83
|
"logProbs"?: number;
|
|
84
84
|
/**
|
|
85
|
-
* The maximum number of tokens
|
|
85
|
+
* The maximum number of tokens that can be generated per output sequence. The token count of your prompt plus max_tokens cannot exceed the model's context length. Note: Numbers greater than Number.MAX_SAFE_INTEGER will result in rounding issues.
|
|
86
86
|
*/
|
|
87
87
|
"maxTokens"?: number;
|
|
88
88
|
"runtimeType": string;
|
|
@@ -0,0 +1,37 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [generate text](#/en/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/en/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/en/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](/#/en/generative-ai/latest/) to [create a custom model](#/en/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](/#/en/generative-ai/latest/).
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
import * as model from "../model";
|
|
21
|
+
/**
|
|
22
|
+
* An message that represents a single dialogue of chat
|
|
23
|
+
*/
|
|
24
|
+
export interface Message {
|
|
25
|
+
/**
|
|
26
|
+
* Indicates who is giving the current message.
|
|
27
|
+
*/
|
|
28
|
+
"role": string;
|
|
29
|
+
/**
|
|
30
|
+
* Contents of the chat message.
|
|
31
|
+
*/
|
|
32
|
+
"content": Array<model.ChatContent>;
|
|
33
|
+
}
|
|
34
|
+
export declare namespace Message {
|
|
35
|
+
function getJsonObj(obj: Message): object;
|
|
36
|
+
function getDeserializedJsonObj(obj: Message): object;
|
|
37
|
+
}
|
|
@@ -0,0 +1,68 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [generate text](#/en/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/en/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/en/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](/#/en/generative-ai/latest/) to [create a custom model](#/en/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](/#/en/generative-ai/latest/).
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
22
|
+
if (k2 === undefined) k2 = k;
|
|
23
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
24
|
+
}) : (function(o, m, k, k2) {
|
|
25
|
+
if (k2 === undefined) k2 = k;
|
|
26
|
+
o[k2] = m[k];
|
|
27
|
+
}));
|
|
28
|
+
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
29
|
+
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
30
|
+
}) : function(o, v) {
|
|
31
|
+
o["default"] = v;
|
|
32
|
+
});
|
|
33
|
+
var __importStar = (this && this.__importStar) || function (mod) {
|
|
34
|
+
if (mod && mod.__esModule) return mod;
|
|
35
|
+
var result = {};
|
|
36
|
+
if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
37
|
+
__setModuleDefault(result, mod);
|
|
38
|
+
return result;
|
|
39
|
+
};
|
|
40
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
41
|
+
exports.Message = void 0;
|
|
42
|
+
const model = __importStar(require("../model"));
|
|
43
|
+
var Message;
|
|
44
|
+
(function (Message) {
|
|
45
|
+
function getJsonObj(obj) {
|
|
46
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
47
|
+
"content": obj.content
|
|
48
|
+
? obj.content.map(item => {
|
|
49
|
+
return model.ChatContent.getJsonObj(item);
|
|
50
|
+
})
|
|
51
|
+
: undefined
|
|
52
|
+
});
|
|
53
|
+
return jsonObj;
|
|
54
|
+
}
|
|
55
|
+
Message.getJsonObj = getJsonObj;
|
|
56
|
+
function getDeserializedJsonObj(obj) {
|
|
57
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
58
|
+
"content": obj.content
|
|
59
|
+
? obj.content.map(item => {
|
|
60
|
+
return model.ChatContent.getDeserializedJsonObj(item);
|
|
61
|
+
})
|
|
62
|
+
: undefined
|
|
63
|
+
});
|
|
64
|
+
return jsonObj;
|
|
65
|
+
}
|
|
66
|
+
Message.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
67
|
+
})(Message = exports.Message || (exports.Message = {}));
|
|
68
|
+
//# sourceMappingURL=message.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"message.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/message.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAiBlC,IAAiB,OAAO,CA6BvB;AA7BD,WAAiB,OAAO;IACtB,SAAgB,UAAU,CAAC,GAAY;QACrC,MAAM,OAAO,mCACR,GAAG,GACH;YACD,SAAS,EAAE,GAAG,CAAC,OAAO;gBACpB,CAAC,CAAC,GAAG,CAAC,OAAO,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACrB,OAAO,KAAK,CAAC,WAAW,CAAC,UAAU,CAAC,IAAI,CAAC,CAAC;gBAC5C,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAbe,kBAAU,aAazB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAY;QACjD,MAAM,OAAO,mCACR,GAAG,GACH;YACD,SAAS,EAAE,GAAG,CAAC,OAAO;gBACpB,CAAC,CAAC,GAAG,CAAC,OAAO,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACrB,OAAO,KAAK,CAAC,WAAW,CAAC,sBAAsB,CAAC,IAAI,CAAC,CAAC;gBACxD,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAbe,8BAAsB,yBAarC,CAAA;AACH,CAAC,EA7BgB,OAAO,GAAP,eAAO,KAAP,eAAO,QA6BvB"}
|
|
@@ -0,0 +1,32 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [generate text](#/en/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/en/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/en/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](/#/en/generative-ai/latest/) to [create a custom model](#/en/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](/#/en/generative-ai/latest/).
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
/**
|
|
21
|
+
* The generated search query.
|
|
22
|
+
*/
|
|
23
|
+
export interface SearchQuery {
|
|
24
|
+
/**
|
|
25
|
+
* The text of the search query.
|
|
26
|
+
*/
|
|
27
|
+
"text": string;
|
|
28
|
+
}
|
|
29
|
+
export declare namespace SearchQuery {
|
|
30
|
+
function getJsonObj(obj: SearchQuery): object;
|
|
31
|
+
function getDeserializedJsonObj(obj: SearchQuery): object;
|
|
32
|
+
}
|
|
@@ -0,0 +1,36 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [generate text](#/en/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/en/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/en/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](/#/en/generative-ai/latest/) to [create a custom model](#/en/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](/#/en/generative-ai/latest/).
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
22
|
+
exports.SearchQuery = void 0;
|
|
23
|
+
var SearchQuery;
|
|
24
|
+
(function (SearchQuery) {
|
|
25
|
+
function getJsonObj(obj) {
|
|
26
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
27
|
+
return jsonObj;
|
|
28
|
+
}
|
|
29
|
+
SearchQuery.getJsonObj = getJsonObj;
|
|
30
|
+
function getDeserializedJsonObj(obj) {
|
|
31
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
32
|
+
return jsonObj;
|
|
33
|
+
}
|
|
34
|
+
SearchQuery.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
35
|
+
})(SearchQuery = exports.SearchQuery || (exports.SearchQuery = {}));
|
|
36
|
+
//# sourceMappingURL=search-query.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"search-query.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/search-query.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;AAeH,IAAiB,WAAW,CAW3B;AAXD,WAAiB,WAAW;IAC1B,SAAgB,UAAU,CAAC,GAAgB;QACzC,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,sBAAU,aAIzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAgB;QACrD,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,kCAAsB,yBAIrC,CAAA;AACH,CAAC,EAXgB,WAAW,GAAX,mBAAW,KAAX,mBAAW,QAW3B"}
|
|
@@ -0,0 +1,35 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [generate text](#/en/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/en/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/en/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](/#/en/generative-ai/latest/) to [create a custom model](#/en/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](/#/en/generative-ai/latest/).
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
import * as model from "../model";
|
|
21
|
+
/**
|
|
22
|
+
* Represents a single instance of text chat content.
|
|
23
|
+
*/
|
|
24
|
+
export interface TextContent extends model.ChatContent {
|
|
25
|
+
/**
|
|
26
|
+
* The text content.
|
|
27
|
+
*/
|
|
28
|
+
"text"?: string;
|
|
29
|
+
"type": string;
|
|
30
|
+
}
|
|
31
|
+
export declare namespace TextContent {
|
|
32
|
+
function getJsonObj(obj: TextContent, isParentJsonObj?: boolean): object;
|
|
33
|
+
const type = "TEXT";
|
|
34
|
+
function getDeserializedJsonObj(obj: TextContent, isParentJsonObj?: boolean): object;
|
|
35
|
+
}
|
|
@@ -0,0 +1,57 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [generate text](#/en/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/en/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/en/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](/#/en/generative-ai/latest/) to [create a custom model](#/en/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](/#/en/generative-ai/latest/).
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
22
|
+
if (k2 === undefined) k2 = k;
|
|
23
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
24
|
+
}) : (function(o, m, k, k2) {
|
|
25
|
+
if (k2 === undefined) k2 = k;
|
|
26
|
+
o[k2] = m[k];
|
|
27
|
+
}));
|
|
28
|
+
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
29
|
+
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
30
|
+
}) : function(o, v) {
|
|
31
|
+
o["default"] = v;
|
|
32
|
+
});
|
|
33
|
+
var __importStar = (this && this.__importStar) || function (mod) {
|
|
34
|
+
if (mod && mod.__esModule) return mod;
|
|
35
|
+
var result = {};
|
|
36
|
+
if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
37
|
+
__setModuleDefault(result, mod);
|
|
38
|
+
return result;
|
|
39
|
+
};
|
|
40
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
41
|
+
exports.TextContent = void 0;
|
|
42
|
+
const model = __importStar(require("../model"));
|
|
43
|
+
var TextContent;
|
|
44
|
+
(function (TextContent) {
|
|
45
|
+
function getJsonObj(obj, isParentJsonObj) {
|
|
46
|
+
const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj ? obj : model.ChatContent.getJsonObj(obj))), {});
|
|
47
|
+
return jsonObj;
|
|
48
|
+
}
|
|
49
|
+
TextContent.getJsonObj = getJsonObj;
|
|
50
|
+
TextContent.type = "TEXT";
|
|
51
|
+
function getDeserializedJsonObj(obj, isParentJsonObj) {
|
|
52
|
+
const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj ? obj : model.ChatContent.getDeserializedJsonObj(obj))), {});
|
|
53
|
+
return jsonObj;
|
|
54
|
+
}
|
|
55
|
+
TextContent.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
56
|
+
})(TextContent = exports.TextContent || (exports.TextContent = {}));
|
|
57
|
+
//# sourceMappingURL=text-content.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"text-content.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/text-content.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAelC,IAAiB,WAAW,CAkB3B;AAlBD,WAAiB,WAAW;IAC1B,SAAgB,UAAU,CAAC,GAAgB,EAAE,eAAyB;QACpE,MAAM,OAAO,mCACR,CAAC,eAAe,CAAC,CAAC,CAAC,GAAG,CAAC,CAAC,CAAE,KAAK,CAAC,WAAW,CAAC,UAAU,CAAC,GAAG,CAAiB,CAAC,GAC5E,EAAE,CACN,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAPe,sBAAU,aAOzB,CAAA;IACY,gBAAI,GAAG,MAAM,CAAC;IAC3B,SAAgB,sBAAsB,CAAC,GAAgB,EAAE,eAAyB;QAChF,MAAM,OAAO,mCACR,CAAC,eAAe,CAAC,CAAC,CAAC,GAAG,CAAC,CAAC,CAAE,KAAK,CAAC,WAAW,CAAC,sBAAsB,CAAC,GAAG,CAAiB,CAAC,GACxF,EAAE,CACN,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAPe,kCAAsB,yBAOrC,CAAA;AACH,CAAC,EAlBgB,WAAW,GAAX,mBAAW,KAAX,mBAAW,QAkB3B"}
|
|
@@ -0,0 +1,36 @@
|
|
|
1
|
+
/**
|
|
2
|
+
*
|
|
3
|
+
*
|
|
4
|
+
* OpenAPI spec version: 20231130
|
|
5
|
+
*
|
|
6
|
+
*
|
|
7
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
8
|
+
* Do not edit the class manually.
|
|
9
|
+
*
|
|
10
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
11
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
12
|
+
*/
|
|
13
|
+
import * as model from "../model";
|
|
14
|
+
import common = require("oci-common");
|
|
15
|
+
/**
|
|
16
|
+
* @example Click {@link https://docs.cloud.oracle.com/en-us/iaas/tools/typescript-sdk-examples/2.84.2/generativeaiinference/Chat.ts.html |here} to see how to use ChatRequest.
|
|
17
|
+
*/
|
|
18
|
+
export interface ChatRequest extends common.BaseRequest {
|
|
19
|
+
/**
|
|
20
|
+
* Details of the conversation for the model to respond.
|
|
21
|
+
*/
|
|
22
|
+
"chatDetails": model.ChatDetails;
|
|
23
|
+
/**
|
|
24
|
+
* A token that uniquely identifies a request so it can be retried in case of a timeout or
|
|
25
|
+
* server error without risk of executing that same action again. Retry tokens expire after 24
|
|
26
|
+
* hours, but can be invalidated before then due to conflicting operations. For example, if a resource
|
|
27
|
+
* has been deleted and purged from the system, then a retry of the original creation request
|
|
28
|
+
* might be rejected.
|
|
29
|
+
*
|
|
30
|
+
*/
|
|
31
|
+
"opcRetryToken"?: string;
|
|
32
|
+
/**
|
|
33
|
+
* The client request ID for tracing.
|
|
34
|
+
*/
|
|
35
|
+
"opcRequestId"?: string;
|
|
36
|
+
}
|
|
@@ -0,0 +1,15 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
*
|
|
4
|
+
*
|
|
5
|
+
* OpenAPI spec version: 20231130
|
|
6
|
+
*
|
|
7
|
+
*
|
|
8
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
9
|
+
* Do not edit the class manually.
|
|
10
|
+
*
|
|
11
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
12
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
13
|
+
*/
|
|
14
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
15
|
+
//# sourceMappingURL=chat-request.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"chat-request.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/request/chat-request.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;GAWG"}
|
|
@@ -13,7 +13,7 @@
|
|
|
13
13
|
import * as model from "../model";
|
|
14
14
|
import common = require("oci-common");
|
|
15
15
|
/**
|
|
16
|
-
* @example Click {@link https://docs.cloud.oracle.com/en-us/iaas/tools/typescript-sdk-examples/2.84.
|
|
16
|
+
* @example Click {@link https://docs.cloud.oracle.com/en-us/iaas/tools/typescript-sdk-examples/2.84.2/generativeaiinference/EmbedText.ts.html |here} to see how to use EmbedTextRequest.
|
|
17
17
|
*/
|
|
18
18
|
export interface EmbedTextRequest extends common.BaseRequest {
|
|
19
19
|
/**
|
|
@@ -13,7 +13,7 @@
|
|
|
13
13
|
import * as model from "../model";
|
|
14
14
|
import common = require("oci-common");
|
|
15
15
|
/**
|
|
16
|
-
* @example Click {@link https://docs.cloud.oracle.com/en-us/iaas/tools/typescript-sdk-examples/2.84.
|
|
16
|
+
* @example Click {@link https://docs.cloud.oracle.com/en-us/iaas/tools/typescript-sdk-examples/2.84.2/generativeaiinference/GenerateText.ts.html |here} to see how to use GenerateTextRequest.
|
|
17
17
|
*/
|
|
18
18
|
export interface GenerateTextRequest extends common.BaseRequest {
|
|
19
19
|
/**
|
package/lib/request/index.d.ts
CHANGED
|
@@ -17,6 +17,8 @@ To learn more about the service, see the [Generative AI documentation](/iaas/Con
|
|
|
17
17
|
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
18
18
|
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
19
|
*/
|
|
20
|
+
import * as ChatRequest from "./chat-request";
|
|
21
|
+
export import ChatRequest = ChatRequest.ChatRequest;
|
|
20
22
|
import * as EmbedTextRequest from "./embed-text-request";
|
|
21
23
|
export import EmbedTextRequest = EmbedTextRequest.EmbedTextRequest;
|
|
22
24
|
import * as GenerateTextRequest from "./generate-text-request";
|
|
@@ -13,7 +13,7 @@
|
|
|
13
13
|
import * as model from "../model";
|
|
14
14
|
import common = require("oci-common");
|
|
15
15
|
/**
|
|
16
|
-
* @example Click {@link https://docs.cloud.oracle.com/en-us/iaas/tools/typescript-sdk-examples/2.84.
|
|
16
|
+
* @example Click {@link https://docs.cloud.oracle.com/en-us/iaas/tools/typescript-sdk-examples/2.84.2/generativeaiinference/SummarizeText.ts.html |here} to see how to use SummarizeTextRequest.
|
|
17
17
|
*/
|
|
18
18
|
export interface SummarizeTextRequest extends common.BaseRequest {
|
|
19
19
|
/**
|
|
@@ -0,0 +1,30 @@
|
|
|
1
|
+
/**
|
|
2
|
+
*
|
|
3
|
+
*
|
|
4
|
+
* OpenAPI spec version: 20231130
|
|
5
|
+
*
|
|
6
|
+
*
|
|
7
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
8
|
+
* Do not edit the class manually.
|
|
9
|
+
*
|
|
10
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
11
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
12
|
+
*/
|
|
13
|
+
import * as model from "../model";
|
|
14
|
+
export interface ChatResponse {
|
|
15
|
+
/**
|
|
16
|
+
* For optimistic concurrency control. See {@code if-match}.
|
|
17
|
+
*
|
|
18
|
+
*/
|
|
19
|
+
"etag": string;
|
|
20
|
+
/**
|
|
21
|
+
* Unique Oracle-assigned identifier for the request. If you need to contact
|
|
22
|
+
* Oracle about a particular request, please provide the request ID.
|
|
23
|
+
*
|
|
24
|
+
*/
|
|
25
|
+
"opcRequestId": string;
|
|
26
|
+
/**
|
|
27
|
+
* The returned model.ChatResult instance.
|
|
28
|
+
*/
|
|
29
|
+
"chatResult": model.ChatResult;
|
|
30
|
+
}
|
|
@@ -0,0 +1,15 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
*
|
|
4
|
+
*
|
|
5
|
+
* OpenAPI spec version: 20231130
|
|
6
|
+
*
|
|
7
|
+
*
|
|
8
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
9
|
+
* Do not edit the class manually.
|
|
10
|
+
*
|
|
11
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
12
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
13
|
+
*/
|
|
14
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
15
|
+
//# sourceMappingURL=chat-response.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"chat-response.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/response/chat-response.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;GAWG"}
|
package/lib/response/index.d.ts
CHANGED
|
@@ -17,6 +17,8 @@ To learn more about the service, see the [Generative AI documentation](/iaas/Con
|
|
|
17
17
|
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
18
18
|
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
19
|
*/
|
|
20
|
+
import * as ChatResponse from "./chat-response";
|
|
21
|
+
export import ChatResponse = ChatResponse.ChatResponse;
|
|
20
22
|
import * as EmbedTextResponse from "./embed-text-response";
|
|
21
23
|
export import EmbedTextResponse = EmbedTextResponse.EmbedTextResponse;
|
|
22
24
|
import * as GenerateTextResponse from "./generate-text-response";
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "oci-generativeaiinference",
|
|
3
|
-
"version": "2.84.
|
|
3
|
+
"version": "2.84.2",
|
|
4
4
|
"description": "OCI NodeJS client for Generative Ai Inference Service",
|
|
5
5
|
"repository": {
|
|
6
6
|
"type": "git",
|
|
@@ -15,8 +15,8 @@
|
|
|
15
15
|
},
|
|
16
16
|
"license": "(UPL-1.0 OR Apache-2.0)",
|
|
17
17
|
"dependencies": {
|
|
18
|
-
"oci-common": "2.84.
|
|
19
|
-
"oci-workrequests": "2.84.
|
|
18
|
+
"oci-common": "2.84.2",
|
|
19
|
+
"oci-workrequests": "2.84.2"
|
|
20
20
|
},
|
|
21
21
|
"publishConfig": {
|
|
22
22
|
"registry": "https://registry.npmjs.org"
|