oci-generativeaiinference 2.107.0 → 2.107.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/lib/client.d.ts +10 -0
- package/lib/client.js +59 -0
- package/lib/client.js.map +1 -1
- package/lib/model/apply-guardrails-details.d.ts +35 -0
- package/lib/model/apply-guardrails-details.js +66 -0
- package/lib/model/apply-guardrails-details.js.map +1 -0
- package/lib/model/apply-guardrails-result.d.ts +30 -0
- package/lib/model/apply-guardrails-result.js +62 -0
- package/lib/model/apply-guardrails-result.js.map +1 -0
- package/lib/model/assistant-message.d.ts +6 -2
- package/lib/model/assistant-message.js.map +1 -1
- package/lib/model/category-score.d.ts +36 -0
- package/lib/model/category-score.js +36 -0
- package/lib/model/category-score.js.map +1 -0
- package/lib/model/cohere-chat-request.d.ts +10 -0
- package/lib/model/cohere-chat-request.js +6 -0
- package/lib/model/cohere-chat-request.js.map +1 -1
- package/lib/model/content-moderation-configuration.d.ts +32 -0
- package/lib/model/content-moderation-configuration.js +36 -0
- package/lib/model/content-moderation-configuration.js.map +1 -0
- package/lib/model/content-moderation-result.d.ts +33 -0
- package/lib/model/content-moderation-result.js +68 -0
- package/lib/model/content-moderation-result.js.map +1 -0
- package/lib/model/guardrail-configs.d.ts +32 -0
- package/lib/model/guardrail-configs.js +76 -0
- package/lib/model/guardrail-configs.js.map +1 -0
- package/lib/model/guardrails-input.d.ts +29 -0
- package/lib/model/guardrails-input.js +75 -0
- package/lib/model/guardrails-input.js.map +1 -0
- package/lib/model/guardrails-results.d.ts +35 -0
- package/lib/model/guardrails-results.js +80 -0
- package/lib/model/guardrails-results.js.map +1 -0
- package/lib/model/guardrails-text-input.d.ts +39 -0
- package/lib/model/guardrails-text-input.js +59 -0
- package/lib/model/guardrails-text-input.js.map +1 -0
- package/lib/model/index.d.ts +26 -0
- package/lib/model/index.js +28 -2
- package/lib/model/index.js.map +1 -1
- package/lib/model/personally-identifiable-information-configuration.d.ts +32 -0
- package/lib/model/personally-identifiable-information-configuration.js +36 -0
- package/lib/model/personally-identifiable-information-configuration.js.map +1 -0
- package/lib/model/personally-identifiable-information-result.d.ts +48 -0
- package/lib/model/personally-identifiable-information-result.js +36 -0
- package/lib/model/personally-identifiable-information-result.js.map +1 -0
- package/lib/model/prompt-injection-configuration.d.ts +28 -0
- package/lib/model/prompt-injection-configuration.js +36 -0
- package/lib/model/prompt-injection-configuration.js.map +1 -0
- package/lib/model/prompt-injection-protection-result.d.ts +32 -0
- package/lib/model/prompt-injection-protection-result.js +36 -0
- package/lib/model/prompt-injection-protection-result.js.map +1 -0
- package/lib/request/apply-guardrails-request.d.ts +35 -0
- package/lib/request/apply-guardrails-request.js +15 -0
- package/lib/request/apply-guardrails-request.js.map +1 -0
- package/lib/request/index.d.ts +2 -0
- package/lib/response/apply-guardrails-response.d.ts +25 -0
- package/lib/response/apply-guardrails-response.js +15 -0
- package/lib/response/apply-guardrails-response.js.map +1 -0
- package/lib/response/index.d.ts +2 -0
- package/package.json +3 -3
package/lib/model/index.d.ts
CHANGED
|
@@ -17,10 +17,16 @@ To learn more about the service, see the [Generative AI documentation](https://d
|
|
|
17
17
|
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
18
18
|
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
19
|
*/
|
|
20
|
+
import * as ApplyGuardrailsDetails from "./apply-guardrails-details";
|
|
21
|
+
export import ApplyGuardrailsDetails = ApplyGuardrailsDetails.ApplyGuardrailsDetails;
|
|
22
|
+
import * as ApplyGuardrailsResult from "./apply-guardrails-result";
|
|
23
|
+
export import ApplyGuardrailsResult = ApplyGuardrailsResult.ApplyGuardrailsResult;
|
|
20
24
|
import * as BaseChatRequest from "./base-chat-request";
|
|
21
25
|
export import BaseChatRequest = BaseChatRequest.BaseChatRequest;
|
|
22
26
|
import * as BaseChatResponse from "./base-chat-response";
|
|
23
27
|
export import BaseChatResponse = BaseChatResponse.BaseChatResponse;
|
|
28
|
+
import * as CategoryScore from "./category-score";
|
|
29
|
+
export import CategoryScore = CategoryScore.CategoryScore;
|
|
24
30
|
import * as ChatChoice from "./chat-choice";
|
|
25
31
|
export import ChatChoice = ChatChoice.ChatChoice;
|
|
26
32
|
import * as ChatContent from "./chat-content";
|
|
@@ -45,6 +51,10 @@ import * as CohereToolCall from "./cohere-tool-call";
|
|
|
45
51
|
export import CohereToolCall = CohereToolCall.CohereToolCall;
|
|
46
52
|
import * as CohereToolResult from "./cohere-tool-result";
|
|
47
53
|
export import CohereToolResult = CohereToolResult.CohereToolResult;
|
|
54
|
+
import * as ContentModerationConfiguration from "./content-moderation-configuration";
|
|
55
|
+
export import ContentModerationConfiguration = ContentModerationConfiguration.ContentModerationConfiguration;
|
|
56
|
+
import * as ContentModerationResult from "./content-moderation-result";
|
|
57
|
+
export import ContentModerationResult = ContentModerationResult.ContentModerationResult;
|
|
48
58
|
import * as Document from "./document";
|
|
49
59
|
export import Document = Document.Document;
|
|
50
60
|
import * as DocumentRank from "./document-rank";
|
|
@@ -59,6 +69,12 @@ import * as GenerateTextResult from "./generate-text-result";
|
|
|
59
69
|
export import GenerateTextResult = GenerateTextResult.GenerateTextResult;
|
|
60
70
|
import * as GeneratedText from "./generated-text";
|
|
61
71
|
export import GeneratedText = GeneratedText.GeneratedText;
|
|
72
|
+
import * as GuardrailConfigs from "./guardrail-configs";
|
|
73
|
+
export import GuardrailConfigs = GuardrailConfigs.GuardrailConfigs;
|
|
74
|
+
import * as GuardrailsInput from "./guardrails-input";
|
|
75
|
+
export import GuardrailsInput = GuardrailsInput.GuardrailsInput;
|
|
76
|
+
import * as GuardrailsResults from "./guardrails-results";
|
|
77
|
+
export import GuardrailsResults = GuardrailsResults.GuardrailsResults;
|
|
62
78
|
import * as ImageUrl from "./image-url";
|
|
63
79
|
export import ImageUrl = ImageUrl.ImageUrl;
|
|
64
80
|
import * as LlmInferenceRequest from "./llm-inference-request";
|
|
@@ -69,6 +85,14 @@ import * as Logprobs from "./logprobs";
|
|
|
69
85
|
export import Logprobs = Logprobs.Logprobs;
|
|
70
86
|
import * as Message from "./message";
|
|
71
87
|
export import Message = Message.Message;
|
|
88
|
+
import * as PersonallyIdentifiableInformationConfiguration from "./personally-identifiable-information-configuration";
|
|
89
|
+
export import PersonallyIdentifiableInformationConfiguration = PersonallyIdentifiableInformationConfiguration.PersonallyIdentifiableInformationConfiguration;
|
|
90
|
+
import * as PersonallyIdentifiableInformationResult from "./personally-identifiable-information-result";
|
|
91
|
+
export import PersonallyIdentifiableInformationResult = PersonallyIdentifiableInformationResult.PersonallyIdentifiableInformationResult;
|
|
92
|
+
import * as PromptInjectionConfiguration from "./prompt-injection-configuration";
|
|
93
|
+
export import PromptInjectionConfiguration = PromptInjectionConfiguration.PromptInjectionConfiguration;
|
|
94
|
+
import * as PromptInjectionProtectionResult from "./prompt-injection-protection-result";
|
|
95
|
+
export import PromptInjectionProtectionResult = PromptInjectionProtectionResult.PromptInjectionProtectionResult;
|
|
72
96
|
import * as RerankTextDetails from "./rerank-text-details";
|
|
73
97
|
export import RerankTextDetails = RerankTextDetails.RerankTextDetails;
|
|
74
98
|
import * as RerankTextResult from "./rerank-text-result";
|
|
@@ -121,6 +145,8 @@ import * as GenericChatRequest from "./generic-chat-request";
|
|
|
121
145
|
export import GenericChatRequest = GenericChatRequest.GenericChatRequest;
|
|
122
146
|
import * as GenericChatResponse from "./generic-chat-response";
|
|
123
147
|
export import GenericChatResponse = GenericChatResponse.GenericChatResponse;
|
|
148
|
+
import * as GuardrailsTextInput from "./guardrails-text-input";
|
|
149
|
+
export import GuardrailsTextInput = GuardrailsTextInput.GuardrailsTextInput;
|
|
124
150
|
import * as ImageContent from "./image-content";
|
|
125
151
|
export import ImageContent = ImageContent.ImageContent;
|
|
126
152
|
import * as LlamaLlmInferenceRequest from "./llama-llm-inference-request";
|
package/lib/model/index.js
CHANGED
|
@@ -38,12 +38,18 @@ var __importStar = (this && this.__importStar) || function (mod) {
|
|
|
38
38
|
return result;
|
|
39
39
|
};
|
|
40
40
|
Object.defineProperty(exports, "__esModule", { value: true });
|
|
41
|
-
exports.
|
|
42
|
-
exports.UserMessage = exports.ToolMessage = exports.ToolChoiceRequired = exports.ToolChoiceNone = exports.ToolChoiceFunction = exports.ToolChoiceAuto = exports.TextContent = exports.SystemMessage = exports.OnDemandServingMode = exports.LlamaLlmInferenceResponse = exports.LlamaLlmInferenceRequest = exports.ImageContent = exports.GenericChatResponse = exports.GenericChatRequest = void 0;
|
|
41
|
+
exports.CohereChatBotMessage = exports.AssistantMessage = exports.ToolDefinition = exports.ToolChoice = exports.ToolCall = exports.TokenLikelihood = exports.SummarizeTextResult = exports.SummarizeTextDetails = exports.ServingMode = exports.SearchQuery = exports.RerankTextResult = exports.RerankTextDetails = exports.PromptInjectionProtectionResult = exports.PromptInjectionConfiguration = exports.PersonallyIdentifiableInformationResult = exports.PersonallyIdentifiableInformationConfiguration = exports.Message = exports.Logprobs = exports.LlmInferenceResponse = exports.LlmInferenceRequest = exports.ImageUrl = exports.GuardrailsResults = exports.GuardrailsInput = exports.GuardrailConfigs = exports.GeneratedText = exports.GenerateTextResult = exports.GenerateTextDetails = exports.EmbedTextResult = exports.EmbedTextDetails = exports.DocumentRank = exports.Document = exports.ContentModerationResult = exports.ContentModerationConfiguration = exports.CohereToolResult = exports.CohereToolCall = exports.CohereTool = exports.CohereResponseFormat = exports.CohereParameterDefinition = exports.CohereMessage = exports.Citation = exports.Choice = exports.ChatResult = exports.ChatDetails = exports.ChatContent = exports.ChatChoice = exports.CategoryScore = exports.BaseChatResponse = exports.BaseChatRequest = exports.ApplyGuardrailsResult = exports.ApplyGuardrailsDetails = void 0;
|
|
42
|
+
exports.UserMessage = exports.ToolMessage = exports.ToolChoiceRequired = exports.ToolChoiceNone = exports.ToolChoiceFunction = exports.ToolChoiceAuto = exports.TextContent = exports.SystemMessage = exports.OnDemandServingMode = exports.LlamaLlmInferenceResponse = exports.LlamaLlmInferenceRequest = exports.ImageContent = exports.GuardrailsTextInput = exports.GenericChatResponse = exports.GenericChatRequest = exports.FunctionDefinition = exports.FunctionCall = exports.DedicatedServingMode = exports.CohereUserMessage = exports.CohereToolMessage = exports.CohereSystemMessage = exports.CohereResponseTextFormat = exports.CohereResponseJsonFormat = exports.CohereLlmInferenceResponse = exports.CohereLlmInferenceRequest = exports.CohereChatResponse = exports.CohereChatRequest = void 0;
|
|
43
|
+
const ApplyGuardrailsDetails = __importStar(require("./apply-guardrails-details"));
|
|
44
|
+
exports.ApplyGuardrailsDetails = ApplyGuardrailsDetails.ApplyGuardrailsDetails;
|
|
45
|
+
const ApplyGuardrailsResult = __importStar(require("./apply-guardrails-result"));
|
|
46
|
+
exports.ApplyGuardrailsResult = ApplyGuardrailsResult.ApplyGuardrailsResult;
|
|
43
47
|
const BaseChatRequest = __importStar(require("./base-chat-request"));
|
|
44
48
|
exports.BaseChatRequest = BaseChatRequest.BaseChatRequest;
|
|
45
49
|
const BaseChatResponse = __importStar(require("./base-chat-response"));
|
|
46
50
|
exports.BaseChatResponse = BaseChatResponse.BaseChatResponse;
|
|
51
|
+
const CategoryScore = __importStar(require("./category-score"));
|
|
52
|
+
exports.CategoryScore = CategoryScore.CategoryScore;
|
|
47
53
|
const ChatChoice = __importStar(require("./chat-choice"));
|
|
48
54
|
exports.ChatChoice = ChatChoice.ChatChoice;
|
|
49
55
|
const ChatContent = __importStar(require("./chat-content"));
|
|
@@ -68,6 +74,10 @@ const CohereToolCall = __importStar(require("./cohere-tool-call"));
|
|
|
68
74
|
exports.CohereToolCall = CohereToolCall.CohereToolCall;
|
|
69
75
|
const CohereToolResult = __importStar(require("./cohere-tool-result"));
|
|
70
76
|
exports.CohereToolResult = CohereToolResult.CohereToolResult;
|
|
77
|
+
const ContentModerationConfiguration = __importStar(require("./content-moderation-configuration"));
|
|
78
|
+
exports.ContentModerationConfiguration = ContentModerationConfiguration.ContentModerationConfiguration;
|
|
79
|
+
const ContentModerationResult = __importStar(require("./content-moderation-result"));
|
|
80
|
+
exports.ContentModerationResult = ContentModerationResult.ContentModerationResult;
|
|
71
81
|
const Document = __importStar(require("./document"));
|
|
72
82
|
exports.Document = Document.Document;
|
|
73
83
|
const DocumentRank = __importStar(require("./document-rank"));
|
|
@@ -82,6 +92,12 @@ const GenerateTextResult = __importStar(require("./generate-text-result"));
|
|
|
82
92
|
exports.GenerateTextResult = GenerateTextResult.GenerateTextResult;
|
|
83
93
|
const GeneratedText = __importStar(require("./generated-text"));
|
|
84
94
|
exports.GeneratedText = GeneratedText.GeneratedText;
|
|
95
|
+
const GuardrailConfigs = __importStar(require("./guardrail-configs"));
|
|
96
|
+
exports.GuardrailConfigs = GuardrailConfigs.GuardrailConfigs;
|
|
97
|
+
const GuardrailsInput = __importStar(require("./guardrails-input"));
|
|
98
|
+
exports.GuardrailsInput = GuardrailsInput.GuardrailsInput;
|
|
99
|
+
const GuardrailsResults = __importStar(require("./guardrails-results"));
|
|
100
|
+
exports.GuardrailsResults = GuardrailsResults.GuardrailsResults;
|
|
85
101
|
const ImageUrl = __importStar(require("./image-url"));
|
|
86
102
|
exports.ImageUrl = ImageUrl.ImageUrl;
|
|
87
103
|
const LlmInferenceRequest = __importStar(require("./llm-inference-request"));
|
|
@@ -92,6 +108,14 @@ const Logprobs = __importStar(require("./logprobs"));
|
|
|
92
108
|
exports.Logprobs = Logprobs.Logprobs;
|
|
93
109
|
const Message = __importStar(require("./message"));
|
|
94
110
|
exports.Message = Message.Message;
|
|
111
|
+
const PersonallyIdentifiableInformationConfiguration = __importStar(require("./personally-identifiable-information-configuration"));
|
|
112
|
+
exports.PersonallyIdentifiableInformationConfiguration = PersonallyIdentifiableInformationConfiguration.PersonallyIdentifiableInformationConfiguration;
|
|
113
|
+
const PersonallyIdentifiableInformationResult = __importStar(require("./personally-identifiable-information-result"));
|
|
114
|
+
exports.PersonallyIdentifiableInformationResult = PersonallyIdentifiableInformationResult.PersonallyIdentifiableInformationResult;
|
|
115
|
+
const PromptInjectionConfiguration = __importStar(require("./prompt-injection-configuration"));
|
|
116
|
+
exports.PromptInjectionConfiguration = PromptInjectionConfiguration.PromptInjectionConfiguration;
|
|
117
|
+
const PromptInjectionProtectionResult = __importStar(require("./prompt-injection-protection-result"));
|
|
118
|
+
exports.PromptInjectionProtectionResult = PromptInjectionProtectionResult.PromptInjectionProtectionResult;
|
|
95
119
|
const RerankTextDetails = __importStar(require("./rerank-text-details"));
|
|
96
120
|
exports.RerankTextDetails = RerankTextDetails.RerankTextDetails;
|
|
97
121
|
const RerankTextResult = __importStar(require("./rerank-text-result"));
|
|
@@ -144,6 +168,8 @@ const GenericChatRequest = __importStar(require("./generic-chat-request"));
|
|
|
144
168
|
exports.GenericChatRequest = GenericChatRequest.GenericChatRequest;
|
|
145
169
|
const GenericChatResponse = __importStar(require("./generic-chat-response"));
|
|
146
170
|
exports.GenericChatResponse = GenericChatResponse.GenericChatResponse;
|
|
171
|
+
const GuardrailsTextInput = __importStar(require("./guardrails-text-input"));
|
|
172
|
+
exports.GuardrailsTextInput = GuardrailsTextInput.GuardrailsTextInput;
|
|
147
173
|
const ImageContent = __importStar(require("./image-content"));
|
|
148
174
|
exports.ImageContent = ImageContent.ImageContent;
|
|
149
175
|
const LlamaLlmInferenceRequest = __importStar(require("./llama-llm-inference-request"));
|
package/lib/model/index.js.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"index.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/index.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;;AAEH,qEAAuD;AACzC,QAAA,eAAe,GAAG,eAAe,CAAC,eAAe,CAAC;AAChE,uEAAyD;AAC3C,QAAA,gBAAgB,GAAG,gBAAgB,CAAC,gBAAgB,CAAC;AACnE,0DAA4C;AAC9B,QAAA,UAAU,GAAG,UAAU,CAAC,UAAU,CAAC;AACjD,4DAA8C;AAChC,QAAA,WAAW,GAAG,WAAW,CAAC,WAAW,CAAC;AACpD,4DAA8C;AAChC,QAAA,WAAW,GAAG,WAAW,CAAC,WAAW,CAAC;AACpD,0DAA4C;AAC9B,QAAA,UAAU,GAAG,UAAU,CAAC,UAAU,CAAC;AACjD,iDAAmC;AACrB,QAAA,MAAM,GAAG,MAAM,CAAC,MAAM,CAAC;AACrC,qDAAuC;AACzB,QAAA,QAAQ,GAAG,QAAQ,CAAC,QAAQ,CAAC;AAC3C,gEAAkD;AACpC,QAAA,aAAa,GAAG,aAAa,CAAC,aAAa,CAAC;AAC1D,yFAA2E;AAC7D,QAAA,yBAAyB,GAAG,yBAAyB,CAAC,yBAAyB,CAAC;AAC9F,+EAAiE;AACnD,QAAA,oBAAoB,GAAG,oBAAoB,CAAC,oBAAoB,CAAC;AAC/E,0DAA4C;AAC9B,QAAA,UAAU,GAAG,UAAU,CAAC,UAAU,CAAC;AACjD,mEAAqD;AACvC,QAAA,cAAc,GAAG,cAAc,CAAC,cAAc,CAAC;AAC7D,uEAAyD;AAC3C,QAAA,gBAAgB,GAAG,gBAAgB,CAAC,gBAAgB,CAAC;AACnE,qDAAuC;AACzB,QAAA,QAAQ,GAAG,QAAQ,CAAC,QAAQ,CAAC;AAC3C,8DAAgD;AAClC,QAAA,YAAY,GAAG,YAAY,CAAC,YAAY,CAAC;AACvD,uEAAyD;AAC3C,QAAA,gBAAgB,GAAG,gBAAgB,CAAC,gBAAgB,CAAC;AACnE,qEAAuD;AACzC,QAAA,eAAe,GAAG,eAAe,CAAC,eAAe,CAAC;AAChE,6EAA+D;AACjD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,2EAA6D;AAC/C,QAAA,kBAAkB,GAAG,kBAAkB,CAAC,kBAAkB,CAAC;AACzE,gEAAkD;AACpC,QAAA,aAAa,GAAG,aAAa,CAAC,aAAa,CAAC;AAC1D,sDAAwC;AAC1B,QAAA,QAAQ,GAAG,QAAQ,CAAC,QAAQ,CAAC;AAC3C,6EAA+D;AACjD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,+EAAiE;AACnD,QAAA,oBAAoB,GAAG,oBAAoB,CAAC,oBAAoB,CAAC;AAC/E,qDAAuC;AACzB,QAAA,QAAQ,GAAG,QAAQ,CAAC,QAAQ,CAAC;AAC3C,mDAAqC;AACvB,QAAA,OAAO,GAAG,OAAO,CAAC,OAAO,CAAC;AACxC,yEAA2D;AAC7C,QAAA,iBAAiB,GAAG,iBAAiB,CAAC,iBAAiB,CAAC;AACtE,uEAAyD;AAC3C,QAAA,gBAAgB,GAAG,gBAAgB,CAAC,gBAAgB,CAAC;AACnE,4DAA8C;AAChC,QAAA,WAAW,GAAG,WAAW,CAAC,WAAW,CAAC;AACpD,4DAA8C;AAChC,QAAA,WAAW,GAAG,WAAW,CAAC,WAAW,CAAC;AACpD,+EAAiE;AACnD,QAAA,oBAAoB,GAAG,oBAAoB,CAAC,oBAAoB,CAAC;AAC/E,6EAA+D;AACjD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,oEAAsD;AACxC,QAAA,eAAe,GAAG,eAAe,CAAC,eAAe,CAAC;AAChE,sDAAwC;AAC1B,QAAA,QAAQ,GAAG,QAAQ,CAAC,QAAQ,CAAC;AAC3C,0DAA4C;AAC9B,QAAA,UAAU,GAAG,UAAU,CAAC,UAAU,CAAC;AACjD,kEAAoD;AACtC,QAAA,cAAc,GAAG,cAAc,CAAC,cAAc,CAAC;AAE7D,sEAAwD;AAC1C,QAAA,gBAAgB,GAAG,gBAAgB,CAAC,gBAAgB,CAAC;AACnE,gFAAkE;AACpD,QAAA,oBAAoB,GAAG,oBAAoB,CAAC,oBAAoB,CAAC;AAC/E,yEAA2D;AAC7C,QAAA,iBAAiB,GAAG,iBAAiB,CAAC,iBAAiB,CAAC;AACtE,2EAA6D;AAC/C,QAAA,kBAAkB,GAAG,kBAAkB,CAAC,kBAAkB,CAAC;AACzE,0FAA4E;AAC9D,QAAA,yBAAyB,GAAG,yBAAyB,CAAC,yBAAyB,CAAC;AAC9F,4FAA8E;AAChE,QAAA,0BAA0B,GAAG,0BAA0B,CAAC,0BAA0B,CAAC;AACjG,wFAA0E;AAC5D,QAAA,wBAAwB,GAAG,wBAAwB,CAAC,wBAAwB,CAAC;AAC3F,wFAA0E;AAC5D,QAAA,wBAAwB,GAAG,wBAAwB,CAAC,wBAAwB,CAAC;AAC3F,6EAA+D;AACjD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,yEAA2D;AAC7C,QAAA,iBAAiB,GAAG,iBAAiB,CAAC,iBAAiB,CAAC;AACtE,yEAA2D;AAC7C,QAAA,iBAAiB,GAAG,iBAAiB,CAAC,iBAAiB,CAAC;AACtE,+EAAiE;AACnD,QAAA,oBAAoB,GAAG,oBAAoB,CAAC,oBAAoB,CAAC;AAC/E,8DAAgD;AAClC,QAAA,YAAY,GAAG,YAAY,CAAC,YAAY,CAAC;AACvD,0EAA4D;AAC9C,QAAA,kBAAkB,GAAG,kBAAkB,CAAC,kBAAkB,CAAC;AACzE,2EAA6D;AAC/C,QAAA,kBAAkB,GAAG,kBAAkB,CAAC,kBAAkB,CAAC;AACzE,6EAA+D;AACjD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,8DAAgD;AAClC,QAAA,YAAY,GAAG,YAAY,CAAC,YAAY,CAAC;AACvD,wFAA0E;AAC5D,QAAA,wBAAwB,GAAG,wBAAwB,CAAC,wBAAwB,CAAC;AAC3F,0FAA4E;AAC9D,QAAA,yBAAyB,GAAG,yBAAyB,CAAC,yBAAyB,CAAC;AAC9F,8EAAgE;AAClD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,gEAAkD;AACpC,QAAA,aAAa,GAAG,aAAa,CAAC,aAAa,CAAC;AAC1D,4DAA8C;AAChC,QAAA,WAAW,GAAG,WAAW,CAAC,WAAW,CAAC;AACpD,mEAAqD;AACvC,QAAA,cAAc,GAAG,cAAc,CAAC,cAAc,CAAC;AAC7D,2EAA6D;AAC/C,QAAA,kBAAkB,GAAG,kBAAkB,CAAC,kBAAkB,CAAC;AACzE,mEAAqD;AACvC,QAAA,cAAc,GAAG,cAAc,CAAC,cAAc,CAAC;AAC7D,2EAA6D;AAC/C,QAAA,kBAAkB,GAAG,kBAAkB,CAAC,kBAAkB,CAAC;AACzE,4DAA8C;AAChC,QAAA,WAAW,GAAG,WAAW,CAAC,WAAW,CAAC;AACpD,4DAA8C;AAChC,QAAA,WAAW,GAAG,WAAW,CAAC,WAAW,CAAC"}
|
|
1
|
+
{"version":3,"file":"index.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/index.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;;AAEH,mFAAqE;AACvD,QAAA,sBAAsB,GAAG,sBAAsB,CAAC,sBAAsB,CAAC;AACrF,iFAAmE;AACrD,QAAA,qBAAqB,GAAG,qBAAqB,CAAC,qBAAqB,CAAC;AAClF,qEAAuD;AACzC,QAAA,eAAe,GAAG,eAAe,CAAC,eAAe,CAAC;AAChE,uEAAyD;AAC3C,QAAA,gBAAgB,GAAG,gBAAgB,CAAC,gBAAgB,CAAC;AACnE,gEAAkD;AACpC,QAAA,aAAa,GAAG,aAAa,CAAC,aAAa,CAAC;AAC1D,0DAA4C;AAC9B,QAAA,UAAU,GAAG,UAAU,CAAC,UAAU,CAAC;AACjD,4DAA8C;AAChC,QAAA,WAAW,GAAG,WAAW,CAAC,WAAW,CAAC;AACpD,4DAA8C;AAChC,QAAA,WAAW,GAAG,WAAW,CAAC,WAAW,CAAC;AACpD,0DAA4C;AAC9B,QAAA,UAAU,GAAG,UAAU,CAAC,UAAU,CAAC;AACjD,iDAAmC;AACrB,QAAA,MAAM,GAAG,MAAM,CAAC,MAAM,CAAC;AACrC,qDAAuC;AACzB,QAAA,QAAQ,GAAG,QAAQ,CAAC,QAAQ,CAAC;AAC3C,gEAAkD;AACpC,QAAA,aAAa,GAAG,aAAa,CAAC,aAAa,CAAC;AAC1D,yFAA2E;AAC7D,QAAA,yBAAyB,GAAG,yBAAyB,CAAC,yBAAyB,CAAC;AAC9F,+EAAiE;AACnD,QAAA,oBAAoB,GAAG,oBAAoB,CAAC,oBAAoB,CAAC;AAC/E,0DAA4C;AAC9B,QAAA,UAAU,GAAG,UAAU,CAAC,UAAU,CAAC;AACjD,mEAAqD;AACvC,QAAA,cAAc,GAAG,cAAc,CAAC,cAAc,CAAC;AAC7D,uEAAyD;AAC3C,QAAA,gBAAgB,GAAG,gBAAgB,CAAC,gBAAgB,CAAC;AACnE,mGAAqF;AACvE,QAAA,8BAA8B,GAAG,8BAA8B,CAAC,8BAA8B,CAAC;AAC7G,qFAAuE;AACzD,QAAA,uBAAuB,GAAG,uBAAuB,CAAC,uBAAuB,CAAC;AACxF,qDAAuC;AACzB,QAAA,QAAQ,GAAG,QAAQ,CAAC,QAAQ,CAAC;AAC3C,8DAAgD;AAClC,QAAA,YAAY,GAAG,YAAY,CAAC,YAAY,CAAC;AACvD,uEAAyD;AAC3C,QAAA,gBAAgB,GAAG,gBAAgB,CAAC,gBAAgB,CAAC;AACnE,qEAAuD;AACzC,QAAA,eAAe,GAAG,eAAe,CAAC,eAAe,CAAC;AAChE,6EAA+D;AACjD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,2EAA6D;AAC/C,QAAA,kBAAkB,GAAG,kBAAkB,CAAC,kBAAkB,CAAC;AACzE,gEAAkD;AACpC,QAAA,aAAa,GAAG,aAAa,CAAC,aAAa,CAAC;AAC1D,sEAAwD;AAC1C,QAAA,gBAAgB,GAAG,gBAAgB,CAAC,gBAAgB,CAAC;AACnE,oEAAsD;AACxC,QAAA,eAAe,GAAG,eAAe,CAAC,eAAe,CAAC;AAChE,wEAA0D;AAC5C,QAAA,iBAAiB,GAAG,iBAAiB,CAAC,iBAAiB,CAAC;AACtE,sDAAwC;AAC1B,QAAA,QAAQ,GAAG,QAAQ,CAAC,QAAQ,CAAC;AAC3C,6EAA+D;AACjD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,+EAAiE;AACnD,QAAA,oBAAoB,GAAG,oBAAoB,CAAC,oBAAoB,CAAC;AAC/E,qDAAuC;AACzB,QAAA,QAAQ,GAAG,QAAQ,CAAC,QAAQ,CAAC;AAC3C,mDAAqC;AACvB,QAAA,OAAO,GAAG,OAAO,CAAC,OAAO,CAAC;AACxC,oIAAsH;AACxG,QAAA,8CAA8C,GAAG,8CAA8C,CAAC,8CAA8C,CAAC;AAC7J,sHAAwG;AAC1F,QAAA,uCAAuC,GAAG,uCAAuC,CAAC,uCAAuC,CAAC;AACxI,+FAAiF;AACnE,QAAA,4BAA4B,GAAG,4BAA4B,CAAC,4BAA4B,CAAC;AACvG,sGAAwF;AAC1E,QAAA,+BAA+B,GAAG,+BAA+B,CAAC,+BAA+B,CAAC;AAChH,yEAA2D;AAC7C,QAAA,iBAAiB,GAAG,iBAAiB,CAAC,iBAAiB,CAAC;AACtE,uEAAyD;AAC3C,QAAA,gBAAgB,GAAG,gBAAgB,CAAC,gBAAgB,CAAC;AACnE,4DAA8C;AAChC,QAAA,WAAW,GAAG,WAAW,CAAC,WAAW,CAAC;AACpD,4DAA8C;AAChC,QAAA,WAAW,GAAG,WAAW,CAAC,WAAW,CAAC;AACpD,+EAAiE;AACnD,QAAA,oBAAoB,GAAG,oBAAoB,CAAC,oBAAoB,CAAC;AAC/E,6EAA+D;AACjD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,oEAAsD;AACxC,QAAA,eAAe,GAAG,eAAe,CAAC,eAAe,CAAC;AAChE,sDAAwC;AAC1B,QAAA,QAAQ,GAAG,QAAQ,CAAC,QAAQ,CAAC;AAC3C,0DAA4C;AAC9B,QAAA,UAAU,GAAG,UAAU,CAAC,UAAU,CAAC;AACjD,kEAAoD;AACtC,QAAA,cAAc,GAAG,cAAc,CAAC,cAAc,CAAC;AAE7D,sEAAwD;AAC1C,QAAA,gBAAgB,GAAG,gBAAgB,CAAC,gBAAgB,CAAC;AACnE,gFAAkE;AACpD,QAAA,oBAAoB,GAAG,oBAAoB,CAAC,oBAAoB,CAAC;AAC/E,yEAA2D;AAC7C,QAAA,iBAAiB,GAAG,iBAAiB,CAAC,iBAAiB,CAAC;AACtE,2EAA6D;AAC/C,QAAA,kBAAkB,GAAG,kBAAkB,CAAC,kBAAkB,CAAC;AACzE,0FAA4E;AAC9D,QAAA,yBAAyB,GAAG,yBAAyB,CAAC,yBAAyB,CAAC;AAC9F,4FAA8E;AAChE,QAAA,0BAA0B,GAAG,0BAA0B,CAAC,0BAA0B,CAAC;AACjG,wFAA0E;AAC5D,QAAA,wBAAwB,GAAG,wBAAwB,CAAC,wBAAwB,CAAC;AAC3F,wFAA0E;AAC5D,QAAA,wBAAwB,GAAG,wBAAwB,CAAC,wBAAwB,CAAC;AAC3F,6EAA+D;AACjD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,yEAA2D;AAC7C,QAAA,iBAAiB,GAAG,iBAAiB,CAAC,iBAAiB,CAAC;AACtE,yEAA2D;AAC7C,QAAA,iBAAiB,GAAG,iBAAiB,CAAC,iBAAiB,CAAC;AACtE,+EAAiE;AACnD,QAAA,oBAAoB,GAAG,oBAAoB,CAAC,oBAAoB,CAAC;AAC/E,8DAAgD;AAClC,QAAA,YAAY,GAAG,YAAY,CAAC,YAAY,CAAC;AACvD,0EAA4D;AAC9C,QAAA,kBAAkB,GAAG,kBAAkB,CAAC,kBAAkB,CAAC;AACzE,2EAA6D;AAC/C,QAAA,kBAAkB,GAAG,kBAAkB,CAAC,kBAAkB,CAAC;AACzE,6EAA+D;AACjD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,6EAA+D;AACjD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,8DAAgD;AAClC,QAAA,YAAY,GAAG,YAAY,CAAC,YAAY,CAAC;AACvD,wFAA0E;AAC5D,QAAA,wBAAwB,GAAG,wBAAwB,CAAC,wBAAwB,CAAC;AAC3F,0FAA4E;AAC9D,QAAA,yBAAyB,GAAG,yBAAyB,CAAC,yBAAyB,CAAC;AAC9F,8EAAgE;AAClD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,gEAAkD;AACpC,QAAA,aAAa,GAAG,aAAa,CAAC,aAAa,CAAC;AAC1D,4DAA8C;AAChC,QAAA,WAAW,GAAG,WAAW,CAAC,WAAW,CAAC;AACpD,mEAAqD;AACvC,QAAA,cAAc,GAAG,cAAc,CAAC,cAAc,CAAC;AAC7D,2EAA6D;AAC/C,QAAA,kBAAkB,GAAG,kBAAkB,CAAC,kBAAkB,CAAC;AACzE,mEAAqD;AACvC,QAAA,cAAc,GAAG,cAAc,CAAC,cAAc,CAAC;AAC7D,2EAA6D;AAC/C,QAAA,kBAAkB,GAAG,kBAAkB,CAAC,kBAAkB,CAAC;AACzE,4DAA8C;AAChC,QAAA,WAAW,GAAG,WAAW,CAAC,WAAW,CAAC;AACpD,4DAA8C;AAChC,QAAA,WAAW,GAAG,WAAW,CAAC,WAAW,CAAC"}
|
|
@@ -0,0 +1,32 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
/**
|
|
21
|
+
* Configuration for personally identifiable information detection.
|
|
22
|
+
*/
|
|
23
|
+
export interface PersonallyIdentifiableInformationConfiguration {
|
|
24
|
+
/**
|
|
25
|
+
* The types of personally identifiable information to detect.
|
|
26
|
+
*/
|
|
27
|
+
"types": Array<string>;
|
|
28
|
+
}
|
|
29
|
+
export declare namespace PersonallyIdentifiableInformationConfiguration {
|
|
30
|
+
function getJsonObj(obj: PersonallyIdentifiableInformationConfiguration): object;
|
|
31
|
+
function getDeserializedJsonObj(obj: PersonallyIdentifiableInformationConfiguration): object;
|
|
32
|
+
}
|
|
@@ -0,0 +1,36 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
22
|
+
exports.PersonallyIdentifiableInformationConfiguration = void 0;
|
|
23
|
+
var PersonallyIdentifiableInformationConfiguration;
|
|
24
|
+
(function (PersonallyIdentifiableInformationConfiguration) {
|
|
25
|
+
function getJsonObj(obj) {
|
|
26
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
27
|
+
return jsonObj;
|
|
28
|
+
}
|
|
29
|
+
PersonallyIdentifiableInformationConfiguration.getJsonObj = getJsonObj;
|
|
30
|
+
function getDeserializedJsonObj(obj) {
|
|
31
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
32
|
+
return jsonObj;
|
|
33
|
+
}
|
|
34
|
+
PersonallyIdentifiableInformationConfiguration.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
35
|
+
})(PersonallyIdentifiableInformationConfiguration = exports.PersonallyIdentifiableInformationConfiguration || (exports.PersonallyIdentifiableInformationConfiguration = {}));
|
|
36
|
+
//# sourceMappingURL=personally-identifiable-information-configuration.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"personally-identifiable-information-configuration.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/personally-identifiable-information-configuration.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;AAeH,IAAiB,8CAA8C,CAa9D;AAbD,WAAiB,8CAA8C;IAC7D,SAAgB,UAAU,CAAC,GAAmD;QAC5E,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,yDAAU,aAIzB,CAAA;IACD,SAAgB,sBAAsB,CACpC,GAAmD;QAEnD,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IANe,qEAAsB,yBAMrC,CAAA;AACH,CAAC,EAbgB,8CAA8C,GAA9C,sDAA8C,KAA9C,sDAA8C,QAa9D"}
|
|
@@ -0,0 +1,48 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
/**
|
|
21
|
+
* An item of personally identifiable information.
|
|
22
|
+
*/
|
|
23
|
+
export interface PersonallyIdentifiableInformationResult {
|
|
24
|
+
/**
|
|
25
|
+
* The length of the personally identifiable information. Note: Numbers greater than Number.MAX_SAFE_INTEGER will result in rounding issues.
|
|
26
|
+
*/
|
|
27
|
+
"length": number;
|
|
28
|
+
/**
|
|
29
|
+
* The offset of the personally identifiable information. Note: Numbers greater than Number.MAX_SAFE_INTEGER will result in rounding issues.
|
|
30
|
+
*/
|
|
31
|
+
"offset": number;
|
|
32
|
+
/**
|
|
33
|
+
* The text of the personally identifiable information.
|
|
34
|
+
*/
|
|
35
|
+
"text": string;
|
|
36
|
+
/**
|
|
37
|
+
* The label of the personally identifiable information.
|
|
38
|
+
*/
|
|
39
|
+
"label": string;
|
|
40
|
+
/**
|
|
41
|
+
* The score of the personally identifiable information. Note: Numbers greater than Number.MAX_SAFE_INTEGER will result in rounding issues.
|
|
42
|
+
*/
|
|
43
|
+
"score": number;
|
|
44
|
+
}
|
|
45
|
+
export declare namespace PersonallyIdentifiableInformationResult {
|
|
46
|
+
function getJsonObj(obj: PersonallyIdentifiableInformationResult): object;
|
|
47
|
+
function getDeserializedJsonObj(obj: PersonallyIdentifiableInformationResult): object;
|
|
48
|
+
}
|
|
@@ -0,0 +1,36 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
22
|
+
exports.PersonallyIdentifiableInformationResult = void 0;
|
|
23
|
+
var PersonallyIdentifiableInformationResult;
|
|
24
|
+
(function (PersonallyIdentifiableInformationResult) {
|
|
25
|
+
function getJsonObj(obj) {
|
|
26
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
27
|
+
return jsonObj;
|
|
28
|
+
}
|
|
29
|
+
PersonallyIdentifiableInformationResult.getJsonObj = getJsonObj;
|
|
30
|
+
function getDeserializedJsonObj(obj) {
|
|
31
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
32
|
+
return jsonObj;
|
|
33
|
+
}
|
|
34
|
+
PersonallyIdentifiableInformationResult.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
35
|
+
})(PersonallyIdentifiableInformationResult = exports.PersonallyIdentifiableInformationResult || (exports.PersonallyIdentifiableInformationResult = {}));
|
|
36
|
+
//# sourceMappingURL=personally-identifiable-information-result.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"personally-identifiable-information-result.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/personally-identifiable-information-result.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;AA+BH,IAAiB,uCAAuC,CAWvD;AAXD,WAAiB,uCAAuC;IACtD,SAAgB,UAAU,CAAC,GAA4C;QACrE,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,kDAAU,aAIzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAA4C;QACjF,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,8DAAsB,yBAIrC,CAAA;AACH,CAAC,EAXgB,uCAAuC,GAAvC,+CAAuC,KAAvC,+CAAuC,QAWvD"}
|
|
@@ -0,0 +1,28 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
/**
|
|
21
|
+
* Configuration for prompt injection
|
|
22
|
+
*/
|
|
23
|
+
export interface PromptInjectionConfiguration {
|
|
24
|
+
}
|
|
25
|
+
export declare namespace PromptInjectionConfiguration {
|
|
26
|
+
function getJsonObj(obj: PromptInjectionConfiguration): object;
|
|
27
|
+
function getDeserializedJsonObj(obj: PromptInjectionConfiguration): object;
|
|
28
|
+
}
|
|
@@ -0,0 +1,36 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
22
|
+
exports.PromptInjectionConfiguration = void 0;
|
|
23
|
+
var PromptInjectionConfiguration;
|
|
24
|
+
(function (PromptInjectionConfiguration) {
|
|
25
|
+
function getJsonObj(obj) {
|
|
26
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
27
|
+
return jsonObj;
|
|
28
|
+
}
|
|
29
|
+
PromptInjectionConfiguration.getJsonObj = getJsonObj;
|
|
30
|
+
function getDeserializedJsonObj(obj) {
|
|
31
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
32
|
+
return jsonObj;
|
|
33
|
+
}
|
|
34
|
+
PromptInjectionConfiguration.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
35
|
+
})(PromptInjectionConfiguration = exports.PromptInjectionConfiguration || (exports.PromptInjectionConfiguration = {}));
|
|
36
|
+
//# sourceMappingURL=prompt-injection-configuration.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"prompt-injection-configuration.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/prompt-injection-configuration.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;AAUH,IAAiB,4BAA4B,CAW5C;AAXD,WAAiB,4BAA4B;IAC3C,SAAgB,UAAU,CAAC,GAAiC;QAC1D,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,uCAAU,aAIzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAiC;QACtE,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,mDAAsB,yBAIrC,CAAA;AACH,CAAC,EAXgB,4BAA4B,GAA5B,oCAA4B,KAA5B,oCAA4B,QAW5C"}
|
|
@@ -0,0 +1,32 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
/**
|
|
21
|
+
* The result of prompt injection protection.
|
|
22
|
+
*/
|
|
23
|
+
export interface PromptInjectionProtectionResult {
|
|
24
|
+
/**
|
|
25
|
+
* The score indicating the likelihood of a prompt injection attack. Note: Numbers greater than Number.MAX_SAFE_INTEGER will result in rounding issues.
|
|
26
|
+
*/
|
|
27
|
+
"score": number;
|
|
28
|
+
}
|
|
29
|
+
export declare namespace PromptInjectionProtectionResult {
|
|
30
|
+
function getJsonObj(obj: PromptInjectionProtectionResult): object;
|
|
31
|
+
function getDeserializedJsonObj(obj: PromptInjectionProtectionResult): object;
|
|
32
|
+
}
|
|
@@ -0,0 +1,36 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
22
|
+
exports.PromptInjectionProtectionResult = void 0;
|
|
23
|
+
var PromptInjectionProtectionResult;
|
|
24
|
+
(function (PromptInjectionProtectionResult) {
|
|
25
|
+
function getJsonObj(obj) {
|
|
26
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
27
|
+
return jsonObj;
|
|
28
|
+
}
|
|
29
|
+
PromptInjectionProtectionResult.getJsonObj = getJsonObj;
|
|
30
|
+
function getDeserializedJsonObj(obj) {
|
|
31
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
32
|
+
return jsonObj;
|
|
33
|
+
}
|
|
34
|
+
PromptInjectionProtectionResult.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
35
|
+
})(PromptInjectionProtectionResult = exports.PromptInjectionProtectionResult || (exports.PromptInjectionProtectionResult = {}));
|
|
36
|
+
//# sourceMappingURL=prompt-injection-protection-result.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"prompt-injection-protection-result.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/prompt-injection-protection-result.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;AAeH,IAAiB,+BAA+B,CAW/C;AAXD,WAAiB,+BAA+B;IAC9C,SAAgB,UAAU,CAAC,GAAoC;QAC7D,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,0CAAU,aAIzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAoC;QACzE,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,sDAAsB,yBAIrC,CAAA;AACH,CAAC,EAXgB,+BAA+B,GAA/B,uCAA+B,KAA/B,uCAA+B,QAW/C"}
|
|
@@ -0,0 +1,35 @@
|
|
|
1
|
+
/**
|
|
2
|
+
*
|
|
3
|
+
*
|
|
4
|
+
* OpenAPI spec version: 20231130
|
|
5
|
+
*
|
|
6
|
+
*
|
|
7
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
8
|
+
* Do not edit the class manually.
|
|
9
|
+
*
|
|
10
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
11
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
12
|
+
*/
|
|
13
|
+
import * as model from "../model";
|
|
14
|
+
import common = require("oci-common");
|
|
15
|
+
/**
|
|
16
|
+
* @example Click {@link https://docs.oracle.com/en-us/iaas/tools/typescript-sdk-examples/latest/generativeaiinference/ApplyGuardrails.ts.html |here} to see how to use ApplyGuardrailsRequest.
|
|
17
|
+
*/
|
|
18
|
+
export interface ApplyGuardrailsRequest extends common.BaseRequest {
|
|
19
|
+
/**
|
|
20
|
+
* Details for applying guardrails to the input text.
|
|
21
|
+
*/
|
|
22
|
+
"applyGuardrailsDetails": model.ApplyGuardrailsDetails;
|
|
23
|
+
/**
|
|
24
|
+
* A token that uniquely identifies a request so it can be retried in case of a timeout or
|
|
25
|
+
* server error without risk of executing that same action again. Retry tokens expire after 24
|
|
26
|
+
* hours, but can be invalidated before that, in case of conflicting operations. For example, if a resource is deleted and purged from the system, then a retry of the original creation request
|
|
27
|
+
* is rejected.
|
|
28
|
+
*
|
|
29
|
+
*/
|
|
30
|
+
"opcRetryToken"?: string;
|
|
31
|
+
/**
|
|
32
|
+
* The client request ID for tracing.
|
|
33
|
+
*/
|
|
34
|
+
"opcRequestId"?: string;
|
|
35
|
+
}
|
|
@@ -0,0 +1,15 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
*
|
|
4
|
+
*
|
|
5
|
+
* OpenAPI spec version: 20231130
|
|
6
|
+
*
|
|
7
|
+
*
|
|
8
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
9
|
+
* Do not edit the class manually.
|
|
10
|
+
*
|
|
11
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
12
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
13
|
+
*/
|
|
14
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
15
|
+
//# sourceMappingURL=apply-guardrails-request.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"apply-guardrails-request.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/request/apply-guardrails-request.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;GAWG"}
|
package/lib/request/index.d.ts
CHANGED
|
@@ -17,6 +17,8 @@ To learn more about the service, see the [Generative AI documentation](https://d
|
|
|
17
17
|
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
18
18
|
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
19
|
*/
|
|
20
|
+
import * as ApplyGuardrailsRequest from "./apply-guardrails-request";
|
|
21
|
+
export import ApplyGuardrailsRequest = ApplyGuardrailsRequest.ApplyGuardrailsRequest;
|
|
20
22
|
import * as ChatRequest from "./chat-request";
|
|
21
23
|
export import ChatRequest = ChatRequest.ChatRequest;
|
|
22
24
|
import * as EmbedTextRequest from "./embed-text-request";
|
|
@@ -0,0 +1,25 @@
|
|
|
1
|
+
/**
|
|
2
|
+
*
|
|
3
|
+
*
|
|
4
|
+
* OpenAPI spec version: 20231130
|
|
5
|
+
*
|
|
6
|
+
*
|
|
7
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
8
|
+
* Do not edit the class manually.
|
|
9
|
+
*
|
|
10
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
11
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
12
|
+
*/
|
|
13
|
+
import * as model from "../model";
|
|
14
|
+
export interface ApplyGuardrailsResponse {
|
|
15
|
+
/**
|
|
16
|
+
* Unique Oracle-assigned identifier for the request. If you need to contact
|
|
17
|
+
* Oracle about a particular request, please provide the request ID.
|
|
18
|
+
*
|
|
19
|
+
*/
|
|
20
|
+
"opcRequestId": string;
|
|
21
|
+
/**
|
|
22
|
+
* The returned model.ApplyGuardrailsResult instance.
|
|
23
|
+
*/
|
|
24
|
+
"applyGuardrailsResult": model.ApplyGuardrailsResult;
|
|
25
|
+
}
|
|
@@ -0,0 +1,15 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
*
|
|
4
|
+
*
|
|
5
|
+
* OpenAPI spec version: 20231130
|
|
6
|
+
*
|
|
7
|
+
*
|
|
8
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
9
|
+
* Do not edit the class manually.
|
|
10
|
+
*
|
|
11
|
+
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
12
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
13
|
+
*/
|
|
14
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
15
|
+
//# sourceMappingURL=apply-guardrails-response.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"apply-guardrails-response.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/response/apply-guardrails-response.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;GAWG"}
|
package/lib/response/index.d.ts
CHANGED
|
@@ -17,6 +17,8 @@ To learn more about the service, see the [Generative AI documentation](https://d
|
|
|
17
17
|
* Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
|
|
18
18
|
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
19
|
*/
|
|
20
|
+
import * as ApplyGuardrailsResponse from "./apply-guardrails-response";
|
|
21
|
+
export import ApplyGuardrailsResponse = ApplyGuardrailsResponse.ApplyGuardrailsResponse;
|
|
20
22
|
import * as ChatResponse from "./chat-response";
|
|
21
23
|
export import ChatResponse = ChatResponse.ChatResponse;
|
|
22
24
|
import * as EmbedTextResponse from "./embed-text-response";
|