oci-generativeaiinference 2.87.0 → 2.88.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (123) hide show
  1. package/index.d.ts +1 -1
  2. package/index.js +1 -1
  3. package/lib/client.d.ts +5 -5
  4. package/lib/client.js +5 -5
  5. package/lib/model/assistant-message.d.ts +35 -0
  6. package/lib/model/assistant-message.js +57 -0
  7. package/lib/model/assistant-message.js.map +1 -0
  8. package/lib/model/base-chat-request.d.ts +2 -2
  9. package/lib/model/base-chat-request.js +1 -1
  10. package/lib/model/base-chat-response.d.ts +2 -2
  11. package/lib/model/base-chat-response.js +1 -1
  12. package/lib/model/chat-choice.d.ts +2 -2
  13. package/lib/model/chat-choice.js +1 -1
  14. package/lib/model/chat-content.d.ts +1 -1
  15. package/lib/model/chat-content.js +1 -1
  16. package/lib/model/chat-details.d.ts +3 -3
  17. package/lib/model/chat-details.js +1 -1
  18. package/lib/model/chat-result.d.ts +2 -2
  19. package/lib/model/chat-result.js +1 -1
  20. package/lib/model/choice.d.ts +2 -2
  21. package/lib/model/choice.js +1 -1
  22. package/lib/model/citation.d.ts +6 -6
  23. package/lib/model/citation.js +1 -1
  24. package/lib/model/cohere-chat-bot-message.d.ts +39 -0
  25. package/lib/model/cohere-chat-bot-message.js +71 -0
  26. package/lib/model/cohere-chat-bot-message.js.map +1 -0
  27. package/lib/model/cohere-chat-request.d.ts +74 -18
  28. package/lib/model/cohere-chat-request.js +31 -1
  29. package/lib/model/cohere-chat-request.js.map +1 -1
  30. package/lib/model/cohere-chat-response.d.ts +39 -10
  31. package/lib/model/cohere-chat-response.js +21 -1
  32. package/lib/model/cohere-chat-response.js.map +1 -1
  33. package/lib/model/cohere-llm-inference-request.d.ts +1 -1
  34. package/lib/model/cohere-llm-inference-request.js +1 -1
  35. package/lib/model/cohere-llm-inference-response.d.ts +1 -1
  36. package/lib/model/cohere-llm-inference-response.js +1 -1
  37. package/lib/model/cohere-message.d.ts +3 -14
  38. package/lib/model/cohere-message.js +52 -6
  39. package/lib/model/cohere-message.js.map +1 -1
  40. package/lib/model/cohere-parameter-definition.d.ts +40 -0
  41. package/lib/model/cohere-parameter-definition.js +36 -0
  42. package/lib/model/cohere-parameter-definition.js.map +1 -0
  43. package/lib/model/cohere-system-message.d.ts +35 -0
  44. package/lib/model/cohere-system-message.js +59 -0
  45. package/lib/model/cohere-system-message.js.map +1 -0
  46. package/lib/model/cohere-tool-call.d.ts +36 -0
  47. package/lib/model/cohere-tool-call.js +36 -0
  48. package/lib/model/cohere-tool-call.js.map +1 -0
  49. package/lib/model/cohere-tool-message.d.ts +35 -0
  50. package/lib/model/cohere-tool-message.js +71 -0
  51. package/lib/model/cohere-tool-message.js.map +1 -0
  52. package/lib/model/cohere-tool-result.d.ts +34 -0
  53. package/lib/model/cohere-tool-result.js +60 -0
  54. package/lib/model/cohere-tool-result.js.map +1 -0
  55. package/lib/model/cohere-tool.d.ts +43 -0
  56. package/lib/model/cohere-tool.js +65 -0
  57. package/lib/model/cohere-tool.js.map +1 -0
  58. package/lib/model/cohere-user-message.d.ts +35 -0
  59. package/lib/model/cohere-user-message.js +59 -0
  60. package/lib/model/cohere-user-message.js.map +1 -0
  61. package/lib/model/dedicated-serving-mode.d.ts +1 -1
  62. package/lib/model/dedicated-serving-mode.js +1 -1
  63. package/lib/model/embed-text-details.d.ts +3 -3
  64. package/lib/model/embed-text-details.js +1 -1
  65. package/lib/model/embed-text-result.d.ts +1 -1
  66. package/lib/model/embed-text-result.js +1 -1
  67. package/lib/model/generate-text-details.d.ts +2 -2
  68. package/lib/model/generate-text-details.js +1 -1
  69. package/lib/model/generate-text-result.d.ts +1 -1
  70. package/lib/model/generate-text-result.js +1 -1
  71. package/lib/model/generated-text.d.ts +1 -1
  72. package/lib/model/generated-text.js +1 -1
  73. package/lib/model/generic-chat-request.d.ts +12 -7
  74. package/lib/model/generic-chat-request.js +1 -1
  75. package/lib/model/generic-chat-request.js.map +1 -1
  76. package/lib/model/generic-chat-response.d.ts +3 -3
  77. package/lib/model/generic-chat-response.js +1 -1
  78. package/lib/model/index.d.ts +23 -1
  79. package/lib/model/index.js +24 -2
  80. package/lib/model/index.js.map +1 -1
  81. package/lib/model/llama-llm-inference-request.d.ts +2 -2
  82. package/lib/model/llama-llm-inference-request.js +1 -1
  83. package/lib/model/llama-llm-inference-response.d.ts +1 -1
  84. package/lib/model/llama-llm-inference-response.js +1 -1
  85. package/lib/model/llm-inference-request.d.ts +1 -1
  86. package/lib/model/llm-inference-request.js +1 -1
  87. package/lib/model/llm-inference-response.d.ts +1 -1
  88. package/lib/model/llm-inference-response.js +1 -1
  89. package/lib/model/logprobs.d.ts +6 -3
  90. package/lib/model/logprobs.js +1 -1
  91. package/lib/model/logprobs.js.map +1 -1
  92. package/lib/model/message.d.ts +4 -7
  93. package/lib/model/message.js +28 -1
  94. package/lib/model/message.js.map +1 -1
  95. package/lib/model/on-demand-serving-mode.d.ts +2 -2
  96. package/lib/model/on-demand-serving-mode.js +1 -1
  97. package/lib/model/search-query.d.ts +1 -1
  98. package/lib/model/search-query.js +1 -1
  99. package/lib/model/serving-mode.d.ts +2 -2
  100. package/lib/model/serving-mode.js +1 -1
  101. package/lib/model/summarize-text-details.d.ts +2 -2
  102. package/lib/model/summarize-text-details.js +1 -1
  103. package/lib/model/summarize-text-result.d.ts +1 -1
  104. package/lib/model/summarize-text-result.js +1 -1
  105. package/lib/model/system-message.d.ts +35 -0
  106. package/lib/model/system-message.js +57 -0
  107. package/lib/model/system-message.js.map +1 -0
  108. package/lib/model/text-content.d.ts +2 -2
  109. package/lib/model/text-content.js +1 -1
  110. package/lib/model/token-likelihood.d.ts +1 -1
  111. package/lib/model/token-likelihood.js +1 -1
  112. package/lib/model/user-message.d.ts +35 -0
  113. package/lib/model/user-message.js +57 -0
  114. package/lib/model/user-message.js.map +1 -0
  115. package/lib/request/chat-request.d.ts +3 -4
  116. package/lib/request/embed-text-request.d.ts +3 -4
  117. package/lib/request/generate-text-request.d.ts +3 -4
  118. package/lib/request/index.d.ts +1 -1
  119. package/lib/request/index.js +1 -1
  120. package/lib/request/summarize-text-request.d.ts +3 -4
  121. package/lib/response/index.d.ts +1 -1
  122. package/lib/response/index.js +1 -1
  123. package/package.json +3 -3
@@ -3,7 +3,7 @@
3
3
  * Generative AI Service Inference API
4
4
  * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
5
5
 
6
- Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [generate text](#/en/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/en/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/en/generative-ai-inference/latest/EmbedTextResult/EmbedText).
6
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/en/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/en/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/en/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/en/generative-ai-inference/latest/EmbedTextResult/EmbedText).
7
7
 
8
8
  To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](/#/en/generative-ai/latest/) to [create a custom model](#/en/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](/#/en/generative-ai/latest/).
9
9
 
@@ -0,0 +1,35 @@
1
+ /**
2
+ * Generative AI Service Inference API
3
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
4
+
5
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/en/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/en/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/en/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/en/generative-ai-inference/latest/EmbedTextResult/EmbedText).
6
+
7
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](/#/en/generative-ai/latest/) to [create a custom model](#/en/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](/#/en/generative-ai/latest/).
8
+
9
+ To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
10
+
11
+ * OpenAPI spec version: 20231130
12
+ *
13
+ *
14
+ * NOTE: This class is auto generated by OracleSDKGenerator.
15
+ * Do not edit the class manually.
16
+ *
17
+ * Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
18
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
19
+ */
20
+ import * as model from "../model";
21
+ /**
22
+ * Represents a single instance of user message.
23
+ */
24
+ export interface UserMessage extends model.Message {
25
+ /**
26
+ * An optional name for the participant. Provides the model information to differentiate between participants of the same role.
27
+ */
28
+ "name"?: string;
29
+ "role": string;
30
+ }
31
+ export declare namespace UserMessage {
32
+ function getJsonObj(obj: UserMessage, isParentJsonObj?: boolean): object;
33
+ const role = "USER";
34
+ function getDeserializedJsonObj(obj: UserMessage, isParentJsonObj?: boolean): object;
35
+ }
@@ -0,0 +1,57 @@
1
+ "use strict";
2
+ /**
3
+ * Generative AI Service Inference API
4
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
5
+
6
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/en/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/en/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/en/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/en/generative-ai-inference/latest/EmbedTextResult/EmbedText).
7
+
8
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](/#/en/generative-ai/latest/) to [create a custom model](#/en/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](/#/en/generative-ai/latest/).
9
+
10
+ To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
11
+
12
+ * OpenAPI spec version: 20231130
13
+ *
14
+ *
15
+ * NOTE: This class is auto generated by OracleSDKGenerator.
16
+ * Do not edit the class manually.
17
+ *
18
+ * Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
19
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
20
+ */
21
+ var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
22
+ if (k2 === undefined) k2 = k;
23
+ Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
24
+ }) : (function(o, m, k, k2) {
25
+ if (k2 === undefined) k2 = k;
26
+ o[k2] = m[k];
27
+ }));
28
+ var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
29
+ Object.defineProperty(o, "default", { enumerable: true, value: v });
30
+ }) : function(o, v) {
31
+ o["default"] = v;
32
+ });
33
+ var __importStar = (this && this.__importStar) || function (mod) {
34
+ if (mod && mod.__esModule) return mod;
35
+ var result = {};
36
+ if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
37
+ __setModuleDefault(result, mod);
38
+ return result;
39
+ };
40
+ Object.defineProperty(exports, "__esModule", { value: true });
41
+ exports.UserMessage = void 0;
42
+ const model = __importStar(require("../model"));
43
+ var UserMessage;
44
+ (function (UserMessage) {
45
+ function getJsonObj(obj, isParentJsonObj) {
46
+ const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj ? obj : model.Message.getJsonObj(obj))), {});
47
+ return jsonObj;
48
+ }
49
+ UserMessage.getJsonObj = getJsonObj;
50
+ UserMessage.role = "USER";
51
+ function getDeserializedJsonObj(obj, isParentJsonObj) {
52
+ const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj ? obj : model.Message.getDeserializedJsonObj(obj))), {});
53
+ return jsonObj;
54
+ }
55
+ UserMessage.getDeserializedJsonObj = getDeserializedJsonObj;
56
+ })(UserMessage = exports.UserMessage || (exports.UserMessage = {}));
57
+ //# sourceMappingURL=user-message.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"user-message.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/user-message.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAelC,IAAiB,WAAW,CAkB3B;AAlBD,WAAiB,WAAW;IAC1B,SAAgB,UAAU,CAAC,GAAgB,EAAE,eAAyB;QACpE,MAAM,OAAO,mCACR,CAAC,eAAe,CAAC,CAAC,CAAC,GAAG,CAAC,CAAC,CAAE,KAAK,CAAC,OAAO,CAAC,UAAU,CAAC,GAAG,CAAiB,CAAC,GACxE,EAAE,CACN,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAPe,sBAAU,aAOzB,CAAA;IACY,gBAAI,GAAG,MAAM,CAAC;IAC3B,SAAgB,sBAAsB,CAAC,GAAgB,EAAE,eAAyB;QAChF,MAAM,OAAO,mCACR,CAAC,eAAe,CAAC,CAAC,CAAC,GAAG,CAAC,CAAC,CAAE,KAAK,CAAC,OAAO,CAAC,sBAAsB,CAAC,GAAG,CAAiB,CAAC,GACpF,EAAE,CACN,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAPe,kCAAsB,yBAOrC,CAAA;AACH,CAAC,EAlBgB,WAAW,GAAX,mBAAW,KAAX,mBAAW,QAkB3B"}
@@ -13,7 +13,7 @@
13
13
  import * as model from "../model";
14
14
  import common = require("oci-common");
15
15
  /**
16
- * @example Click {@link https://docs.cloud.oracle.com/en-us/iaas/tools/typescript-sdk-examples/2.87.0/generativeaiinference/Chat.ts.html |here} to see how to use ChatRequest.
16
+ * @example Click {@link https://docs.cloud.oracle.com/en-us/iaas/tools/typescript-sdk-examples/2.88.1/generativeaiinference/Chat.ts.html |here} to see how to use ChatRequest.
17
17
  */
18
18
  export interface ChatRequest extends common.BaseRequest {
19
19
  /**
@@ -23,9 +23,8 @@ export interface ChatRequest extends common.BaseRequest {
23
23
  /**
24
24
  * A token that uniquely identifies a request so it can be retried in case of a timeout or
25
25
  * server error without risk of executing that same action again. Retry tokens expire after 24
26
- * hours, but can be invalidated before then due to conflicting operations. For example, if a resource
27
- * has been deleted and purged from the system, then a retry of the original creation request
28
- * might be rejected.
26
+ * hours, but can be invalidated before that, in case of conflicting operations. For example, if a resource is deleted and purged from the system, then a retry of the original creation request
27
+ * is rejected.
29
28
  *
30
29
  */
31
30
  "opcRetryToken"?: string;
@@ -13,7 +13,7 @@
13
13
  import * as model from "../model";
14
14
  import common = require("oci-common");
15
15
  /**
16
- * @example Click {@link https://docs.cloud.oracle.com/en-us/iaas/tools/typescript-sdk-examples/2.87.0/generativeaiinference/EmbedText.ts.html |here} to see how to use EmbedTextRequest.
16
+ * @example Click {@link https://docs.cloud.oracle.com/en-us/iaas/tools/typescript-sdk-examples/2.88.1/generativeaiinference/EmbedText.ts.html |here} to see how to use EmbedTextRequest.
17
17
  */
18
18
  export interface EmbedTextRequest extends common.BaseRequest {
19
19
  /**
@@ -23,9 +23,8 @@ export interface EmbedTextRequest extends common.BaseRequest {
23
23
  /**
24
24
  * A token that uniquely identifies a request so it can be retried in case of a timeout or
25
25
  * server error without risk of executing that same action again. Retry tokens expire after 24
26
- * hours, but can be invalidated before then due to conflicting operations. For example, if a resource
27
- * has been deleted and purged from the system, then a retry of the original creation request
28
- * might be rejected.
26
+ * hours, but can be invalidated before that, in case of conflicting operations. For example, if a resource is deleted and purged from the system, then a retry of the original creation request
27
+ * is rejected.
29
28
  *
30
29
  */
31
30
  "opcRetryToken"?: string;
@@ -13,7 +13,7 @@
13
13
  import * as model from "../model";
14
14
  import common = require("oci-common");
15
15
  /**
16
- * @example Click {@link https://docs.cloud.oracle.com/en-us/iaas/tools/typescript-sdk-examples/2.87.0/generativeaiinference/GenerateText.ts.html |here} to see how to use GenerateTextRequest.
16
+ * @example Click {@link https://docs.cloud.oracle.com/en-us/iaas/tools/typescript-sdk-examples/2.88.1/generativeaiinference/GenerateText.ts.html |here} to see how to use GenerateTextRequest.
17
17
  */
18
18
  export interface GenerateTextRequest extends common.BaseRequest {
19
19
  /**
@@ -23,9 +23,8 @@ export interface GenerateTextRequest extends common.BaseRequest {
23
23
  /**
24
24
  * A token that uniquely identifies a request so it can be retried in case of a timeout or
25
25
  * server error without risk of executing that same action again. Retry tokens expire after 24
26
- * hours, but can be invalidated before then due to conflicting operations. For example, if a resource
27
- * has been deleted and purged from the system, then a retry of the original creation request
28
- * might be rejected.
26
+ * hours, but can be invalidated before that, in case of conflicting operations. For example, if a resource is deleted and purged from the system, then a retry of the original creation request
27
+ * is rejected.
29
28
  *
30
29
  */
31
30
  "opcRetryToken"?: string;
@@ -2,7 +2,7 @@
2
2
  * Generative AI Service Inference API
3
3
  * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
4
4
 
5
- Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [generate text](#/en/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/en/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/en/generative-ai-inference/latest/EmbedTextResult/EmbedText).
5
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/en/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/en/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/en/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/en/generative-ai-inference/latest/EmbedTextResult/EmbedText).
6
6
 
7
7
  To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](/#/en/generative-ai/latest/) to [create a custom model](#/en/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](/#/en/generative-ai/latest/).
8
8
 
@@ -3,7 +3,7 @@
3
3
  * Generative AI Service Inference API
4
4
  * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
5
5
 
6
- Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [generate text](#/en/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/en/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/en/generative-ai-inference/latest/EmbedTextResult/EmbedText).
6
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/en/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/en/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/en/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/en/generative-ai-inference/latest/EmbedTextResult/EmbedText).
7
7
 
8
8
  To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](/#/en/generative-ai/latest/) to [create a custom model](#/en/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](/#/en/generative-ai/latest/).
9
9
 
@@ -13,7 +13,7 @@
13
13
  import * as model from "../model";
14
14
  import common = require("oci-common");
15
15
  /**
16
- * @example Click {@link https://docs.cloud.oracle.com/en-us/iaas/tools/typescript-sdk-examples/2.87.0/generativeaiinference/SummarizeText.ts.html |here} to see how to use SummarizeTextRequest.
16
+ * @example Click {@link https://docs.cloud.oracle.com/en-us/iaas/tools/typescript-sdk-examples/2.88.1/generativeaiinference/SummarizeText.ts.html |here} to see how to use SummarizeTextRequest.
17
17
  */
18
18
  export interface SummarizeTextRequest extends common.BaseRequest {
19
19
  /**
@@ -23,9 +23,8 @@ export interface SummarizeTextRequest extends common.BaseRequest {
23
23
  /**
24
24
  * A token that uniquely identifies a request so it can be retried in case of a timeout or
25
25
  * server error without risk of executing that same action again. Retry tokens expire after 24
26
- * hours, but can be invalidated before then due to conflicting operations. For example, if a resource
27
- * has been deleted and purged from the system, then a retry of the original creation request
28
- * might be rejected.
26
+ * hours, but can be invalidated before that, in case of conflicting operations. For example, if a resource is deleted and purged from the system, then a retry of the original creation request
27
+ * is rejected.
29
28
  *
30
29
  */
31
30
  "opcRetryToken"?: string;
@@ -2,7 +2,7 @@
2
2
  * Generative AI Service Inference API
3
3
  * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
4
4
 
5
- Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [generate text](#/en/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/en/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/en/generative-ai-inference/latest/EmbedTextResult/EmbedText).
5
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/en/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/en/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/en/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/en/generative-ai-inference/latest/EmbedTextResult/EmbedText).
6
6
 
7
7
  To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](/#/en/generative-ai/latest/) to [create a custom model](#/en/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](/#/en/generative-ai/latest/).
8
8
 
@@ -3,7 +3,7 @@
3
3
  * Generative AI Service Inference API
4
4
  * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
5
5
 
6
- Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [generate text](#/en/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/en/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/en/generative-ai-inference/latest/EmbedTextResult/EmbedText).
6
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/en/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/en/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/en/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/en/generative-ai-inference/latest/EmbedTextResult/EmbedText).
7
7
 
8
8
  To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](/#/en/generative-ai/latest/) to [create a custom model](#/en/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](/#/en/generative-ai/latest/).
9
9
 
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "oci-generativeaiinference",
3
- "version": "2.87.0",
3
+ "version": "2.88.1",
4
4
  "description": "OCI NodeJS client for Generative Ai Inference Service",
5
5
  "repository": {
6
6
  "type": "git",
@@ -15,8 +15,8 @@
15
15
  },
16
16
  "license": "(UPL-1.0 OR Apache-2.0)",
17
17
  "dependencies": {
18
- "oci-common": "2.87.0",
19
- "oci-workrequests": "2.87.0"
18
+ "oci-common": "2.88.1",
19
+ "oci-workrequests": "2.88.1"
20
20
  },
21
21
  "publishConfig": {
22
22
  "registry": "https://registry.npmjs.org"