oci-generativeaiinference 2.106.2 → 2.107.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (59) hide show
  1. package/lib/client.d.ts +10 -0
  2. package/lib/client.js +59 -0
  3. package/lib/client.js.map +1 -1
  4. package/lib/model/apply-guardrails-details.d.ts +35 -0
  5. package/lib/model/apply-guardrails-details.js +66 -0
  6. package/lib/model/apply-guardrails-details.js.map +1 -0
  7. package/lib/model/apply-guardrails-result.d.ts +30 -0
  8. package/lib/model/apply-guardrails-result.js +62 -0
  9. package/lib/model/apply-guardrails-result.js.map +1 -0
  10. package/lib/model/assistant-message.d.ts +6 -2
  11. package/lib/model/assistant-message.js.map +1 -1
  12. package/lib/model/category-score.d.ts +36 -0
  13. package/lib/model/category-score.js +36 -0
  14. package/lib/model/category-score.js.map +1 -0
  15. package/lib/model/cohere-chat-request.d.ts +10 -0
  16. package/lib/model/cohere-chat-request.js +6 -0
  17. package/lib/model/cohere-chat-request.js.map +1 -1
  18. package/lib/model/content-moderation-configuration.d.ts +32 -0
  19. package/lib/model/content-moderation-configuration.js +36 -0
  20. package/lib/model/content-moderation-configuration.js.map +1 -0
  21. package/lib/model/content-moderation-result.d.ts +33 -0
  22. package/lib/model/content-moderation-result.js +68 -0
  23. package/lib/model/content-moderation-result.js.map +1 -0
  24. package/lib/model/guardrail-configs.d.ts +32 -0
  25. package/lib/model/guardrail-configs.js +76 -0
  26. package/lib/model/guardrail-configs.js.map +1 -0
  27. package/lib/model/guardrails-input.d.ts +29 -0
  28. package/lib/model/guardrails-input.js +75 -0
  29. package/lib/model/guardrails-input.js.map +1 -0
  30. package/lib/model/guardrails-results.d.ts +35 -0
  31. package/lib/model/guardrails-results.js +80 -0
  32. package/lib/model/guardrails-results.js.map +1 -0
  33. package/lib/model/guardrails-text-input.d.ts +39 -0
  34. package/lib/model/guardrails-text-input.js +59 -0
  35. package/lib/model/guardrails-text-input.js.map +1 -0
  36. package/lib/model/index.d.ts +26 -0
  37. package/lib/model/index.js +28 -2
  38. package/lib/model/index.js.map +1 -1
  39. package/lib/model/personally-identifiable-information-configuration.d.ts +32 -0
  40. package/lib/model/personally-identifiable-information-configuration.js +36 -0
  41. package/lib/model/personally-identifiable-information-configuration.js.map +1 -0
  42. package/lib/model/personally-identifiable-information-result.d.ts +48 -0
  43. package/lib/model/personally-identifiable-information-result.js +36 -0
  44. package/lib/model/personally-identifiable-information-result.js.map +1 -0
  45. package/lib/model/prompt-injection-configuration.d.ts +28 -0
  46. package/lib/model/prompt-injection-configuration.js +36 -0
  47. package/lib/model/prompt-injection-configuration.js.map +1 -0
  48. package/lib/model/prompt-injection-protection-result.d.ts +32 -0
  49. package/lib/model/prompt-injection-protection-result.js +36 -0
  50. package/lib/model/prompt-injection-protection-result.js.map +1 -0
  51. package/lib/request/apply-guardrails-request.d.ts +35 -0
  52. package/lib/request/apply-guardrails-request.js +15 -0
  53. package/lib/request/apply-guardrails-request.js.map +1 -0
  54. package/lib/request/index.d.ts +2 -0
  55. package/lib/response/apply-guardrails-response.d.ts +25 -0
  56. package/lib/response/apply-guardrails-response.js +15 -0
  57. package/lib/response/apply-guardrails-response.js.map +1 -0
  58. package/lib/response/index.d.ts +2 -0
  59. package/package.json +3 -3
@@ -0,0 +1,33 @@
1
+ /**
2
+ * Generative AI Service Inference API
3
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
4
+
5
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
6
+
7
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
8
+
9
+ To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
10
+
11
+ * OpenAPI spec version: 20231130
12
+ *
13
+ *
14
+ * NOTE: This class is auto generated by OracleSDKGenerator.
15
+ * Do not edit the class manually.
16
+ *
17
+ * Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
18
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
19
+ */
20
+ import * as model from "../model";
21
+ /**
22
+ * The result of content moderation.
23
+ */
24
+ export interface ContentModerationResult {
25
+ /**
26
+ * A list of categories with their respective scores.
27
+ */
28
+ "categories": Array<model.CategoryScore>;
29
+ }
30
+ export declare namespace ContentModerationResult {
31
+ function getJsonObj(obj: ContentModerationResult): object;
32
+ function getDeserializedJsonObj(obj: ContentModerationResult): object;
33
+ }
@@ -0,0 +1,68 @@
1
+ "use strict";
2
+ /**
3
+ * Generative AI Service Inference API
4
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
5
+
6
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
7
+
8
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
9
+
10
+ To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
11
+
12
+ * OpenAPI spec version: 20231130
13
+ *
14
+ *
15
+ * NOTE: This class is auto generated by OracleSDKGenerator.
16
+ * Do not edit the class manually.
17
+ *
18
+ * Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
19
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
20
+ */
21
+ var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
22
+ if (k2 === undefined) k2 = k;
23
+ Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
24
+ }) : (function(o, m, k, k2) {
25
+ if (k2 === undefined) k2 = k;
26
+ o[k2] = m[k];
27
+ }));
28
+ var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
29
+ Object.defineProperty(o, "default", { enumerable: true, value: v });
30
+ }) : function(o, v) {
31
+ o["default"] = v;
32
+ });
33
+ var __importStar = (this && this.__importStar) || function (mod) {
34
+ if (mod && mod.__esModule) return mod;
35
+ var result = {};
36
+ if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
37
+ __setModuleDefault(result, mod);
38
+ return result;
39
+ };
40
+ Object.defineProperty(exports, "__esModule", { value: true });
41
+ exports.ContentModerationResult = void 0;
42
+ const model = __importStar(require("../model"));
43
+ var ContentModerationResult;
44
+ (function (ContentModerationResult) {
45
+ function getJsonObj(obj) {
46
+ const jsonObj = Object.assign(Object.assign({}, obj), {
47
+ "categories": obj.categories
48
+ ? obj.categories.map(item => {
49
+ return model.CategoryScore.getJsonObj(item);
50
+ })
51
+ : undefined
52
+ });
53
+ return jsonObj;
54
+ }
55
+ ContentModerationResult.getJsonObj = getJsonObj;
56
+ function getDeserializedJsonObj(obj) {
57
+ const jsonObj = Object.assign(Object.assign({}, obj), {
58
+ "categories": obj.categories
59
+ ? obj.categories.map(item => {
60
+ return model.CategoryScore.getDeserializedJsonObj(item);
61
+ })
62
+ : undefined
63
+ });
64
+ return jsonObj;
65
+ }
66
+ ContentModerationResult.getDeserializedJsonObj = getDeserializedJsonObj;
67
+ })(ContentModerationResult = exports.ContentModerationResult || (exports.ContentModerationResult = {}));
68
+ //# sourceMappingURL=content-moderation-result.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"content-moderation-result.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/content-moderation-result.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAalC,IAAiB,uBAAuB,CA6BvC;AA7BD,WAAiB,uBAAuB;IACtC,SAAgB,UAAU,CAAC,GAA4B;QACrD,MAAM,OAAO,mCACR,GAAG,GACH;YACD,YAAY,EAAE,GAAG,CAAC,UAAU;gBAC1B,CAAC,CAAC,GAAG,CAAC,UAAU,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACxB,OAAO,KAAK,CAAC,aAAa,CAAC,UAAU,CAAC,IAAI,CAAC,CAAC;gBAC9C,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAbe,kCAAU,aAazB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAA4B;QACjE,MAAM,OAAO,mCACR,GAAG,GACH;YACD,YAAY,EAAE,GAAG,CAAC,UAAU;gBAC1B,CAAC,CAAC,GAAG,CAAC,UAAU,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACxB,OAAO,KAAK,CAAC,aAAa,CAAC,sBAAsB,CAAC,IAAI,CAAC,CAAC;gBAC1D,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAbe,8CAAsB,yBAarC,CAAA;AACH,CAAC,EA7BgB,uBAAuB,GAAvB,+BAAuB,KAAvB,+BAAuB,QA6BvC"}
@@ -0,0 +1,32 @@
1
+ /**
2
+ * Generative AI Service Inference API
3
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
4
+
5
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
6
+
7
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
8
+
9
+ To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
10
+
11
+ * OpenAPI spec version: 20231130
12
+ *
13
+ *
14
+ * NOTE: This class is auto generated by OracleSDKGenerator.
15
+ * Do not edit the class manually.
16
+ *
17
+ * Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
18
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
19
+ */
20
+ import * as model from "../model";
21
+ /**
22
+ * Additional configuration for each guardrail.
23
+ */
24
+ export interface GuardrailConfigs {
25
+ "contentModerationConfig"?: model.ContentModerationConfiguration;
26
+ "personallyIdentifiableInformationConfig"?: model.PersonallyIdentifiableInformationConfiguration;
27
+ "promptInjectionConfig"?: model.PromptInjectionConfiguration;
28
+ }
29
+ export declare namespace GuardrailConfigs {
30
+ function getJsonObj(obj: GuardrailConfigs): object;
31
+ function getDeserializedJsonObj(obj: GuardrailConfigs): object;
32
+ }
@@ -0,0 +1,76 @@
1
+ "use strict";
2
+ /**
3
+ * Generative AI Service Inference API
4
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
5
+
6
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
7
+
8
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
9
+
10
+ To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
11
+
12
+ * OpenAPI spec version: 20231130
13
+ *
14
+ *
15
+ * NOTE: This class is auto generated by OracleSDKGenerator.
16
+ * Do not edit the class manually.
17
+ *
18
+ * Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
19
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
20
+ */
21
+ var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
22
+ if (k2 === undefined) k2 = k;
23
+ Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
24
+ }) : (function(o, m, k, k2) {
25
+ if (k2 === undefined) k2 = k;
26
+ o[k2] = m[k];
27
+ }));
28
+ var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
29
+ Object.defineProperty(o, "default", { enumerable: true, value: v });
30
+ }) : function(o, v) {
31
+ o["default"] = v;
32
+ });
33
+ var __importStar = (this && this.__importStar) || function (mod) {
34
+ if (mod && mod.__esModule) return mod;
35
+ var result = {};
36
+ if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
37
+ __setModuleDefault(result, mod);
38
+ return result;
39
+ };
40
+ Object.defineProperty(exports, "__esModule", { value: true });
41
+ exports.GuardrailConfigs = void 0;
42
+ const model = __importStar(require("../model"));
43
+ var GuardrailConfigs;
44
+ (function (GuardrailConfigs) {
45
+ function getJsonObj(obj) {
46
+ const jsonObj = Object.assign(Object.assign({}, obj), {
47
+ "contentModerationConfig": obj.contentModerationConfig
48
+ ? model.ContentModerationConfiguration.getJsonObj(obj.contentModerationConfig)
49
+ : undefined,
50
+ "personallyIdentifiableInformationConfig": obj.personallyIdentifiableInformationConfig
51
+ ? model.PersonallyIdentifiableInformationConfiguration.getJsonObj(obj.personallyIdentifiableInformationConfig)
52
+ : undefined,
53
+ "promptInjectionConfig": obj.promptInjectionConfig
54
+ ? model.PromptInjectionConfiguration.getJsonObj(obj.promptInjectionConfig)
55
+ : undefined
56
+ });
57
+ return jsonObj;
58
+ }
59
+ GuardrailConfigs.getJsonObj = getJsonObj;
60
+ function getDeserializedJsonObj(obj) {
61
+ const jsonObj = Object.assign(Object.assign({}, obj), {
62
+ "contentModerationConfig": obj.contentModerationConfig
63
+ ? model.ContentModerationConfiguration.getDeserializedJsonObj(obj.contentModerationConfig)
64
+ : undefined,
65
+ "personallyIdentifiableInformationConfig": obj.personallyIdentifiableInformationConfig
66
+ ? model.PersonallyIdentifiableInformationConfiguration.getDeserializedJsonObj(obj.personallyIdentifiableInformationConfig)
67
+ : undefined,
68
+ "promptInjectionConfig": obj.promptInjectionConfig
69
+ ? model.PromptInjectionConfiguration.getDeserializedJsonObj(obj.promptInjectionConfig)
70
+ : undefined
71
+ });
72
+ return jsonObj;
73
+ }
74
+ GuardrailConfigs.getDeserializedJsonObj = getDeserializedJsonObj;
75
+ })(GuardrailConfigs = exports.GuardrailConfigs || (exports.GuardrailConfigs = {}));
76
+ //# sourceMappingURL=guardrail-configs.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"guardrail-configs.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/guardrail-configs.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAYlC,IAAiB,gBAAgB,CAyChC;AAzCD,WAAiB,gBAAgB;IAC/B,SAAgB,UAAU,CAAC,GAAqB;QAC9C,MAAM,OAAO,mCACR,GAAG,GACH;YACD,yBAAyB,EAAE,GAAG,CAAC,uBAAuB;gBACpD,CAAC,CAAC,KAAK,CAAC,8BAA8B,CAAC,UAAU,CAAC,GAAG,CAAC,uBAAuB,CAAC;gBAC9E,CAAC,CAAC,SAAS;YACb,yCAAyC,EAAE,GAAG,CAAC,uCAAuC;gBACpF,CAAC,CAAC,KAAK,CAAC,8CAA8C,CAAC,UAAU,CAC7D,GAAG,CAAC,uCAAuC,CAC5C;gBACH,CAAC,CAAC,SAAS;YACb,uBAAuB,EAAE,GAAG,CAAC,qBAAqB;gBAChD,CAAC,CAAC,KAAK,CAAC,4BAA4B,CAAC,UAAU,CAAC,GAAG,CAAC,qBAAqB,CAAC;gBAC1E,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAnBe,2BAAU,aAmBzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAqB;QAC1D,MAAM,OAAO,mCACR,GAAG,GACH;YACD,yBAAyB,EAAE,GAAG,CAAC,uBAAuB;gBACpD,CAAC,CAAC,KAAK,CAAC,8BAA8B,CAAC,sBAAsB,CAAC,GAAG,CAAC,uBAAuB,CAAC;gBAC1F,CAAC,CAAC,SAAS;YACb,yCAAyC,EAAE,GAAG,CAAC,uCAAuC;gBACpF,CAAC,CAAC,KAAK,CAAC,8CAA8C,CAAC,sBAAsB,CACzE,GAAG,CAAC,uCAAuC,CAC5C;gBACH,CAAC,CAAC,SAAS;YACb,uBAAuB,EAAE,GAAG,CAAC,qBAAqB;gBAChD,CAAC,CAAC,KAAK,CAAC,4BAA4B,CAAC,sBAAsB,CAAC,GAAG,CAAC,qBAAqB,CAAC;gBACtF,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAnBe,uCAAsB,yBAmBrC,CAAA;AACH,CAAC,EAzCgB,gBAAgB,GAAhB,wBAAgB,KAAhB,wBAAgB,QAyChC"}
@@ -0,0 +1,29 @@
1
+ /**
2
+ * Generative AI Service Inference API
3
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
4
+
5
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
6
+
7
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
8
+
9
+ To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
10
+
11
+ * OpenAPI spec version: 20231130
12
+ *
13
+ *
14
+ * NOTE: This class is auto generated by OracleSDKGenerator.
15
+ * Do not edit the class manually.
16
+ *
17
+ * Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
18
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
19
+ */
20
+ /**
21
+ * The input data for applying guardrails. This can be text, image, video, or audio content.
22
+ */
23
+ export interface GuardrailsInput {
24
+ "type": string;
25
+ }
26
+ export declare namespace GuardrailsInput {
27
+ function getJsonObj(obj: GuardrailsInput): object;
28
+ function getDeserializedJsonObj(obj: GuardrailsInput): object;
29
+ }
@@ -0,0 +1,75 @@
1
+ "use strict";
2
+ /**
3
+ * Generative AI Service Inference API
4
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
5
+
6
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
7
+
8
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
9
+
10
+ To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
11
+
12
+ * OpenAPI spec version: 20231130
13
+ *
14
+ *
15
+ * NOTE: This class is auto generated by OracleSDKGenerator.
16
+ * Do not edit the class manually.
17
+ *
18
+ * Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
19
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
20
+ */
21
+ var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
22
+ if (k2 === undefined) k2 = k;
23
+ Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
24
+ }) : (function(o, m, k, k2) {
25
+ if (k2 === undefined) k2 = k;
26
+ o[k2] = m[k];
27
+ }));
28
+ var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
29
+ Object.defineProperty(o, "default", { enumerable: true, value: v });
30
+ }) : function(o, v) {
31
+ o["default"] = v;
32
+ });
33
+ var __importStar = (this && this.__importStar) || function (mod) {
34
+ if (mod && mod.__esModule) return mod;
35
+ var result = {};
36
+ if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
37
+ __setModuleDefault(result, mod);
38
+ return result;
39
+ };
40
+ Object.defineProperty(exports, "__esModule", { value: true });
41
+ exports.GuardrailsInput = void 0;
42
+ const model = __importStar(require("../model"));
43
+ const common = require("oci-common");
44
+ var GuardrailsInput;
45
+ (function (GuardrailsInput) {
46
+ function getJsonObj(obj) {
47
+ const jsonObj = Object.assign(Object.assign({}, obj), {});
48
+ if (obj && "type" in obj && obj.type) {
49
+ switch (obj.type) {
50
+ case "TEXT":
51
+ return model.GuardrailsTextInput.getJsonObj(jsonObj, true);
52
+ default:
53
+ if (common.LOG.logger)
54
+ common.LOG.logger.info(`Unknown value for: ${obj.type}`);
55
+ }
56
+ }
57
+ return jsonObj;
58
+ }
59
+ GuardrailsInput.getJsonObj = getJsonObj;
60
+ function getDeserializedJsonObj(obj) {
61
+ const jsonObj = Object.assign(Object.assign({}, obj), {});
62
+ if (obj && "type" in obj && obj.type) {
63
+ switch (obj.type) {
64
+ case "TEXT":
65
+ return model.GuardrailsTextInput.getDeserializedJsonObj(jsonObj, true);
66
+ default:
67
+ if (common.LOG.logger)
68
+ common.LOG.logger.info(`Unknown value for: ${obj.type}`);
69
+ }
70
+ }
71
+ return jsonObj;
72
+ }
73
+ GuardrailsInput.getDeserializedJsonObj = getDeserializedJsonObj;
74
+ })(GuardrailsInput = exports.GuardrailsInput || (exports.GuardrailsInput = {}));
75
+ //# sourceMappingURL=guardrails-input.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"guardrails-input.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/guardrails-input.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAClC,qCAAsC;AAStC,IAAiB,eAAe,CAiC/B;AAjCD,WAAiB,eAAe;IAC9B,SAAgB,UAAU,CAAC,GAAoB;QAC7C,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,IAAI,GAAG,IAAI,MAAM,IAAI,GAAG,IAAI,GAAG,CAAC,IAAI,EAAE;YACpC,QAAQ,GAAG,CAAC,IAAI,EAAE;gBAChB,KAAK,MAAM;oBACT,OAAO,KAAK,CAAC,mBAAmB,CAAC,UAAU,CACL,OAAQ,EAC5C,IAAI,CACL,CAAC;gBACJ;oBACE,IAAI,MAAM,CAAC,GAAG,CAAC,MAAM;wBAAE,MAAM,CAAC,GAAG,CAAC,MAAM,CAAC,IAAI,CAAC,sBAAsB,GAAG,CAAC,IAAI,EAAE,CAAC,CAAC;aACnF;SACF;QACD,OAAO,OAAO,CAAC;IACjB,CAAC;IAfe,0BAAU,aAezB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAoB;QACzD,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,IAAI,GAAG,IAAI,MAAM,IAAI,GAAG,IAAI,GAAG,CAAC,IAAI,EAAE;YACpC,QAAQ,GAAG,CAAC,IAAI,EAAE;gBAChB,KAAK,MAAM;oBACT,OAAO,KAAK,CAAC,mBAAmB,CAAC,sBAAsB,CACjB,OAAQ,EAC5C,IAAI,CACL,CAAC;gBACJ;oBACE,IAAI,MAAM,CAAC,GAAG,CAAC,MAAM;wBAAE,MAAM,CAAC,GAAG,CAAC,MAAM,CAAC,IAAI,CAAC,sBAAsB,GAAG,CAAC,IAAI,EAAE,CAAC,CAAC;aACnF;SACF;QACD,OAAO,OAAO,CAAC;IACjB,CAAC;IAfe,sCAAsB,yBAerC,CAAA;AACH,CAAC,EAjCgB,eAAe,GAAf,uBAAe,KAAf,uBAAe,QAiC/B"}
@@ -0,0 +1,35 @@
1
+ /**
2
+ * Generative AI Service Inference API
3
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
4
+
5
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
6
+
7
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
8
+
9
+ To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
10
+
11
+ * OpenAPI spec version: 20231130
12
+ *
13
+ *
14
+ * NOTE: This class is auto generated by OracleSDKGenerator.
15
+ * Do not edit the class manually.
16
+ *
17
+ * Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
18
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
19
+ */
20
+ import * as model from "../model";
21
+ /**
22
+ * The results of applying each guardrail.
23
+ */
24
+ export interface GuardrailsResults {
25
+ "contentModeration"?: model.ContentModerationResult;
26
+ /**
27
+ * The result of PII detection
28
+ */
29
+ "personallyIdentifiableInformation"?: Array<model.PersonallyIdentifiableInformationResult>;
30
+ "promptInjection"?: model.PromptInjectionProtectionResult;
31
+ }
32
+ export declare namespace GuardrailsResults {
33
+ function getJsonObj(obj: GuardrailsResults): object;
34
+ function getDeserializedJsonObj(obj: GuardrailsResults): object;
35
+ }
@@ -0,0 +1,80 @@
1
+ "use strict";
2
+ /**
3
+ * Generative AI Service Inference API
4
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
5
+
6
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
7
+
8
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
9
+
10
+ To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
11
+
12
+ * OpenAPI spec version: 20231130
13
+ *
14
+ *
15
+ * NOTE: This class is auto generated by OracleSDKGenerator.
16
+ * Do not edit the class manually.
17
+ *
18
+ * Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
19
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
20
+ */
21
+ var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
22
+ if (k2 === undefined) k2 = k;
23
+ Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
24
+ }) : (function(o, m, k, k2) {
25
+ if (k2 === undefined) k2 = k;
26
+ o[k2] = m[k];
27
+ }));
28
+ var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
29
+ Object.defineProperty(o, "default", { enumerable: true, value: v });
30
+ }) : function(o, v) {
31
+ o["default"] = v;
32
+ });
33
+ var __importStar = (this && this.__importStar) || function (mod) {
34
+ if (mod && mod.__esModule) return mod;
35
+ var result = {};
36
+ if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
37
+ __setModuleDefault(result, mod);
38
+ return result;
39
+ };
40
+ Object.defineProperty(exports, "__esModule", { value: true });
41
+ exports.GuardrailsResults = void 0;
42
+ const model = __importStar(require("../model"));
43
+ var GuardrailsResults;
44
+ (function (GuardrailsResults) {
45
+ function getJsonObj(obj) {
46
+ const jsonObj = Object.assign(Object.assign({}, obj), {
47
+ "contentModeration": obj.contentModeration
48
+ ? model.ContentModerationResult.getJsonObj(obj.contentModeration)
49
+ : undefined,
50
+ "personallyIdentifiableInformation": obj.personallyIdentifiableInformation
51
+ ? obj.personallyIdentifiableInformation.map(item => {
52
+ return model.PersonallyIdentifiableInformationResult.getJsonObj(item);
53
+ })
54
+ : undefined,
55
+ "promptInjection": obj.promptInjection
56
+ ? model.PromptInjectionProtectionResult.getJsonObj(obj.promptInjection)
57
+ : undefined
58
+ });
59
+ return jsonObj;
60
+ }
61
+ GuardrailsResults.getJsonObj = getJsonObj;
62
+ function getDeserializedJsonObj(obj) {
63
+ const jsonObj = Object.assign(Object.assign({}, obj), {
64
+ "contentModeration": obj.contentModeration
65
+ ? model.ContentModerationResult.getDeserializedJsonObj(obj.contentModeration)
66
+ : undefined,
67
+ "personallyIdentifiableInformation": obj.personallyIdentifiableInformation
68
+ ? obj.personallyIdentifiableInformation.map(item => {
69
+ return model.PersonallyIdentifiableInformationResult.getDeserializedJsonObj(item);
70
+ })
71
+ : undefined,
72
+ "promptInjection": obj.promptInjection
73
+ ? model.PromptInjectionProtectionResult.getDeserializedJsonObj(obj.promptInjection)
74
+ : undefined
75
+ });
76
+ return jsonObj;
77
+ }
78
+ GuardrailsResults.getDeserializedJsonObj = getDeserializedJsonObj;
79
+ })(GuardrailsResults = exports.GuardrailsResults || (exports.GuardrailsResults = {}));
80
+ //# sourceMappingURL=guardrails-results.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"guardrails-results.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/guardrails-results.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAelC,IAAiB,iBAAiB,CAyCjC;AAzCD,WAAiB,iBAAiB;IAChC,SAAgB,UAAU,CAAC,GAAsB;QAC/C,MAAM,OAAO,mCACR,GAAG,GACH;YACD,mBAAmB,EAAE,GAAG,CAAC,iBAAiB;gBACxC,CAAC,CAAC,KAAK,CAAC,uBAAuB,CAAC,UAAU,CAAC,GAAG,CAAC,iBAAiB,CAAC;gBACjE,CAAC,CAAC,SAAS;YACb,mCAAmC,EAAE,GAAG,CAAC,iCAAiC;gBACxE,CAAC,CAAC,GAAG,CAAC,iCAAiC,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBAC/C,OAAO,KAAK,CAAC,uCAAuC,CAAC,UAAU,CAAC,IAAI,CAAC,CAAC;gBACxE,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;YACb,iBAAiB,EAAE,GAAG,CAAC,eAAe;gBACpC,CAAC,CAAC,KAAK,CAAC,+BAA+B,CAAC,UAAU,CAAC,GAAG,CAAC,eAAe,CAAC;gBACvE,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAnBe,4BAAU,aAmBzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAsB;QAC3D,MAAM,OAAO,mCACR,GAAG,GACH;YACD,mBAAmB,EAAE,GAAG,CAAC,iBAAiB;gBACxC,CAAC,CAAC,KAAK,CAAC,uBAAuB,CAAC,sBAAsB,CAAC,GAAG,CAAC,iBAAiB,CAAC;gBAC7E,CAAC,CAAC,SAAS;YACb,mCAAmC,EAAE,GAAG,CAAC,iCAAiC;gBACxE,CAAC,CAAC,GAAG,CAAC,iCAAiC,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBAC/C,OAAO,KAAK,CAAC,uCAAuC,CAAC,sBAAsB,CAAC,IAAI,CAAC,CAAC;gBACpF,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;YACb,iBAAiB,EAAE,GAAG,CAAC,eAAe;gBACpC,CAAC,CAAC,KAAK,CAAC,+BAA+B,CAAC,sBAAsB,CAAC,GAAG,CAAC,eAAe,CAAC;gBACnF,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAnBe,wCAAsB,yBAmBrC,CAAA;AACH,CAAC,EAzCgB,iBAAiB,GAAjB,yBAAiB,KAAjB,yBAAiB,QAyCjC"}
@@ -0,0 +1,39 @@
1
+ /**
2
+ * Generative AI Service Inference API
3
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
4
+
5
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
6
+
7
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
8
+
9
+ To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
10
+
11
+ * OpenAPI spec version: 20231130
12
+ *
13
+ *
14
+ * NOTE: This class is auto generated by OracleSDKGenerator.
15
+ * Do not edit the class manually.
16
+ *
17
+ * Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
18
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
19
+ */
20
+ import * as model from "../model";
21
+ /**
22
+ * Represents a single instance of text in the guardrails input.
23
+ */
24
+ export interface GuardrailsTextInput extends model.GuardrailsInput {
25
+ /**
26
+ * The actual input data.
27
+ */
28
+ "content"?: string;
29
+ /**
30
+ * The language code of the input text. example - en | es | en-US | zh-CN
31
+ */
32
+ "languageCode"?: string;
33
+ "type": string;
34
+ }
35
+ export declare namespace GuardrailsTextInput {
36
+ function getJsonObj(obj: GuardrailsTextInput, isParentJsonObj?: boolean): object;
37
+ const type = "TEXT";
38
+ function getDeserializedJsonObj(obj: GuardrailsTextInput, isParentJsonObj?: boolean): object;
39
+ }
@@ -0,0 +1,59 @@
1
+ "use strict";
2
+ /**
3
+ * Generative AI Service Inference API
4
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
5
+
6
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to {@link #eNGenerative-ai-inferenceLatestChatResultChat(ENGenerative-ai-inferenceLatestChatResultChatRequest) eNGenerative-ai-inferenceLatestChatResultChat}, {@link #eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText(ENGenerative-ai-inferenceLatestGenerateTextResultGenerateTextRequest) eNGenerative-ai-inferenceLatestGenerateTextResultGenerateText}, {@link #eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText(ENGenerative-ai-inferenceLatestSummarizeTextResultSummarizeTextRequest) eNGenerative-ai-inferenceLatestSummarizeTextResultSummarizeText}, and {@link #eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText(ENGenerative-ai-inferenceLatestEmbedTextResultEmbedTextRequest) eNGenerative-ai-inferenceLatestEmbedTextResultEmbedText}.
7
+
8
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest} to {@link #eNGenerative-aiLatestModel(ENGenerative-aiLatestModelRequest) eNGenerative-aiLatestModel} by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster}. Then, create a {@link #eNGenerative-aiLatestDedicatedAiCluster(ENGenerative-aiLatestDedicatedAiClusterRequest) eNGenerative-aiLatestDedicatedAiCluster} with an {@link Endpoint} to host your custom model. For resource management in the Generative AI service, use the {@link #eNGenerative-aiLatest(ENGenerative-aiLatestRequest) eNGenerative-aiLatest}.
9
+
10
+ To learn more about the service, see the [Generative AI documentation](https://docs.oracle.com/iaas/Content/generative-ai/home.htm).
11
+
12
+ * OpenAPI spec version: 20231130
13
+ *
14
+ *
15
+ * NOTE: This class is auto generated by OracleSDKGenerator.
16
+ * Do not edit the class manually.
17
+ *
18
+ * Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved.
19
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
20
+ */
21
+ var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
22
+ if (k2 === undefined) k2 = k;
23
+ Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
24
+ }) : (function(o, m, k, k2) {
25
+ if (k2 === undefined) k2 = k;
26
+ o[k2] = m[k];
27
+ }));
28
+ var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
29
+ Object.defineProperty(o, "default", { enumerable: true, value: v });
30
+ }) : function(o, v) {
31
+ o["default"] = v;
32
+ });
33
+ var __importStar = (this && this.__importStar) || function (mod) {
34
+ if (mod && mod.__esModule) return mod;
35
+ var result = {};
36
+ if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
37
+ __setModuleDefault(result, mod);
38
+ return result;
39
+ };
40
+ Object.defineProperty(exports, "__esModule", { value: true });
41
+ exports.GuardrailsTextInput = void 0;
42
+ const model = __importStar(require("../model"));
43
+ var GuardrailsTextInput;
44
+ (function (GuardrailsTextInput) {
45
+ function getJsonObj(obj, isParentJsonObj) {
46
+ const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj ? obj : model.GuardrailsInput.getJsonObj(obj))), {});
47
+ return jsonObj;
48
+ }
49
+ GuardrailsTextInput.getJsonObj = getJsonObj;
50
+ GuardrailsTextInput.type = "TEXT";
51
+ function getDeserializedJsonObj(obj, isParentJsonObj) {
52
+ const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj
53
+ ? obj
54
+ : model.GuardrailsInput.getDeserializedJsonObj(obj))), {});
55
+ return jsonObj;
56
+ }
57
+ GuardrailsTextInput.getDeserializedJsonObj = getDeserializedJsonObj;
58
+ })(GuardrailsTextInput = exports.GuardrailsTextInput || (exports.GuardrailsTextInput = {}));
59
+ //# sourceMappingURL=guardrails-text-input.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"guardrails-text-input.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/guardrails-text-input.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAmBlC,IAAiB,mBAAmB,CAuBnC;AAvBD,WAAiB,mBAAmB;IAClC,SAAgB,UAAU,CAAC,GAAwB,EAAE,eAAyB;QAC5E,MAAM,OAAO,mCACR,CAAC,eAAe,CAAC,CAAC,CAAC,GAAG,CAAC,CAAC,CAAE,KAAK,CAAC,eAAe,CAAC,UAAU,CAAC,GAAG,CAAyB,CAAC,GACxF,EAAE,CACN,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAPe,8BAAU,aAOzB,CAAA;IACY,wBAAI,GAAG,MAAM,CAAC;IAC3B,SAAgB,sBAAsB,CACpC,GAAwB,EACxB,eAAyB;QAEzB,MAAM,OAAO,mCACR,CAAC,eAAe;YACjB,CAAC,CAAC,GAAG;YACL,CAAC,CAAE,KAAK,CAAC,eAAe,CAAC,sBAAsB,CAAC,GAAG,CAAyB,CAAC,GAC5E,EAAE,CACN,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAZe,0CAAsB,yBAYrC,CAAA;AACH,CAAC,EAvBgB,mBAAmB,GAAnB,2BAAmB,KAAnB,2BAAmB,QAuBnC"}