oci-generativeaiinference 2.98.0 → 2.98.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/index.d.ts +2 -2
- package/index.js +2 -2
- package/lib/client.d.ts +2 -2
- package/lib/client.js +2 -2
- package/lib/model/assistant-message.d.ts +2 -2
- package/lib/model/assistant-message.js +2 -2
- package/lib/model/base-chat-request.d.ts +2 -2
- package/lib/model/base-chat-request.js +2 -2
- package/lib/model/base-chat-response.d.ts +2 -2
- package/lib/model/base-chat-response.js +2 -2
- package/lib/model/chat-choice.d.ts +2 -2
- package/lib/model/chat-choice.js +2 -2
- package/lib/model/chat-content.d.ts +2 -2
- package/lib/model/chat-content.js +6 -2
- package/lib/model/chat-content.js.map +1 -1
- package/lib/model/chat-details.d.ts +2 -2
- package/lib/model/chat-details.js +2 -2
- package/lib/model/chat-result.d.ts +2 -2
- package/lib/model/chat-result.js +2 -2
- package/lib/model/choice.d.ts +2 -2
- package/lib/model/choice.js +2 -2
- package/lib/model/citation.d.ts +2 -2
- package/lib/model/citation.js +2 -2
- package/lib/model/cohere-chat-bot-message.d.ts +2 -2
- package/lib/model/cohere-chat-bot-message.js +2 -2
- package/lib/model/cohere-chat-request.d.ts +8 -3
- package/lib/model/cohere-chat-request.js +8 -2
- package/lib/model/cohere-chat-request.js.map +1 -1
- package/lib/model/cohere-chat-response.d.ts +2 -2
- package/lib/model/cohere-chat-response.js +2 -2
- package/lib/model/cohere-llm-inference-request.d.ts +2 -2
- package/lib/model/cohere-llm-inference-request.js +2 -2
- package/lib/model/cohere-llm-inference-response.d.ts +2 -2
- package/lib/model/cohere-llm-inference-response.js +2 -2
- package/lib/model/cohere-message.d.ts +2 -2
- package/lib/model/cohere-message.js +2 -2
- package/lib/model/cohere-parameter-definition.d.ts +2 -2
- package/lib/model/cohere-parameter-definition.js +2 -2
- package/lib/model/cohere-response-format.d.ts +29 -0
- package/lib/model/cohere-response-format.js +79 -0
- package/lib/model/cohere-response-format.js.map +1 -0
- package/lib/model/cohere-response-json-format.d.ts +35 -0
- package/lib/model/cohere-response-json-format.js +61 -0
- package/lib/model/cohere-response-json-format.js.map +1 -0
- package/lib/model/cohere-response-text-format.d.ts +31 -0
- package/lib/model/cohere-response-text-format.js +61 -0
- package/lib/model/cohere-response-text-format.js.map +1 -0
- package/lib/model/cohere-system-message.d.ts +2 -2
- package/lib/model/cohere-system-message.js +2 -2
- package/lib/model/cohere-tool-call.d.ts +2 -2
- package/lib/model/cohere-tool-call.js +2 -2
- package/lib/model/cohere-tool-message.d.ts +2 -2
- package/lib/model/cohere-tool-message.js +2 -2
- package/lib/model/cohere-tool-result.d.ts +2 -2
- package/lib/model/cohere-tool-result.js +2 -2
- package/lib/model/cohere-tool.d.ts +2 -2
- package/lib/model/cohere-tool.js +2 -2
- package/lib/model/cohere-user-message.d.ts +2 -2
- package/lib/model/cohere-user-message.js +2 -2
- package/lib/model/dedicated-serving-mode.d.ts +2 -2
- package/lib/model/dedicated-serving-mode.js +2 -2
- package/lib/model/embed-text-details.d.ts +2 -2
- package/lib/model/embed-text-details.js +2 -2
- package/lib/model/embed-text-result.d.ts +2 -2
- package/lib/model/embed-text-result.js +2 -2
- package/lib/model/generate-text-details.d.ts +2 -2
- package/lib/model/generate-text-details.js +2 -2
- package/lib/model/generate-text-result.d.ts +2 -2
- package/lib/model/generate-text-result.js +2 -2
- package/lib/model/generated-text.d.ts +2 -2
- package/lib/model/generated-text.js +2 -2
- package/lib/model/generic-chat-request.d.ts +2 -2
- package/lib/model/generic-chat-request.js +2 -2
- package/lib/model/generic-chat-response.d.ts +2 -2
- package/lib/model/generic-chat-response.js +2 -2
- package/lib/model/image-content.d.ts +32 -0
- package/lib/model/image-content.js +61 -0
- package/lib/model/image-content.js.map +1 -0
- package/lib/model/image-url.d.ts +46 -0
- package/lib/model/image-url.js +47 -0
- package/lib/model/image-url.js.map +1 -0
- package/lib/model/index.d.ts +12 -2
- package/lib/model/index.js +13 -3
- package/lib/model/index.js.map +1 -1
- package/lib/model/llama-llm-inference-request.d.ts +2 -2
- package/lib/model/llama-llm-inference-request.js +2 -2
- package/lib/model/llama-llm-inference-response.d.ts +2 -2
- package/lib/model/llama-llm-inference-response.js +2 -2
- package/lib/model/llm-inference-request.d.ts +2 -2
- package/lib/model/llm-inference-request.js +2 -2
- package/lib/model/llm-inference-response.d.ts +2 -2
- package/lib/model/llm-inference-response.js +2 -2
- package/lib/model/logprobs.d.ts +2 -2
- package/lib/model/logprobs.js +2 -2
- package/lib/model/message.d.ts +2 -2
- package/lib/model/message.js +2 -2
- package/lib/model/on-demand-serving-mode.d.ts +2 -2
- package/lib/model/on-demand-serving-mode.js +2 -2
- package/lib/model/search-query.d.ts +2 -2
- package/lib/model/search-query.js +2 -2
- package/lib/model/serving-mode.d.ts +2 -2
- package/lib/model/serving-mode.js +2 -2
- package/lib/model/summarize-text-details.d.ts +2 -2
- package/lib/model/summarize-text-details.js +2 -2
- package/lib/model/summarize-text-result.d.ts +2 -2
- package/lib/model/summarize-text-result.js +2 -2
- package/lib/model/system-message.d.ts +2 -2
- package/lib/model/system-message.js +2 -2
- package/lib/model/text-content.d.ts +2 -2
- package/lib/model/text-content.js +2 -2
- package/lib/model/token-likelihood.d.ts +2 -2
- package/lib/model/token-likelihood.js +2 -2
- package/lib/model/user-message.d.ts +2 -2
- package/lib/model/user-message.js +2 -2
- package/lib/request/index.d.ts +2 -2
- package/lib/request/index.js +2 -2
- package/lib/response/index.d.ts +2 -2
- package/lib/response/index.js +2 -2
- package/package.json +3 -3
package/index.d.ts
CHANGED
|
@@ -2,9 +2,9 @@
|
|
|
2
2
|
* Generative AI Service Inference API
|
|
3
3
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
4
|
|
|
5
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
6
6
|
|
|
7
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
8
8
|
|
|
9
9
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
10
10
|
|
package/index.js
CHANGED
|
@@ -3,9 +3,9 @@
|
|
|
3
3
|
* Generative AI Service Inference API
|
|
4
4
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
5
|
|
|
6
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
7
7
|
|
|
8
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
9
9
|
|
|
10
10
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
11
11
|
|
package/lib/client.d.ts
CHANGED
|
@@ -2,9 +2,9 @@
|
|
|
2
2
|
* Generative AI Service Inference API
|
|
3
3
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
4
|
|
|
5
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
6
6
|
|
|
7
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
8
8
|
|
|
9
9
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
10
10
|
|
package/lib/client.js
CHANGED
|
@@ -3,9 +3,9 @@
|
|
|
3
3
|
* Generative AI Service Inference API
|
|
4
4
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
5
|
|
|
6
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
7
7
|
|
|
8
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
9
9
|
|
|
10
10
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
11
11
|
|
|
@@ -2,9 +2,9 @@
|
|
|
2
2
|
* Generative AI Service Inference API
|
|
3
3
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
4
|
|
|
5
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
6
6
|
|
|
7
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
8
8
|
|
|
9
9
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
10
10
|
|
|
@@ -3,9 +3,9 @@
|
|
|
3
3
|
* Generative AI Service Inference API
|
|
4
4
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
5
|
|
|
6
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
7
7
|
|
|
8
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
9
9
|
|
|
10
10
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
11
11
|
|
|
@@ -2,9 +2,9 @@
|
|
|
2
2
|
* Generative AI Service Inference API
|
|
3
3
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
4
|
|
|
5
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
6
6
|
|
|
7
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
8
8
|
|
|
9
9
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
10
10
|
|
|
@@ -3,9 +3,9 @@
|
|
|
3
3
|
* Generative AI Service Inference API
|
|
4
4
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
5
|
|
|
6
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
7
7
|
|
|
8
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
9
9
|
|
|
10
10
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
11
11
|
|
|
@@ -2,9 +2,9 @@
|
|
|
2
2
|
* Generative AI Service Inference API
|
|
3
3
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
4
|
|
|
5
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
6
6
|
|
|
7
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
8
8
|
|
|
9
9
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
10
10
|
|
|
@@ -3,9 +3,9 @@
|
|
|
3
3
|
* Generative AI Service Inference API
|
|
4
4
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
5
|
|
|
6
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
7
7
|
|
|
8
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
9
9
|
|
|
10
10
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
11
11
|
|
|
@@ -2,9 +2,9 @@
|
|
|
2
2
|
* Generative AI Service Inference API
|
|
3
3
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
4
|
|
|
5
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
6
6
|
|
|
7
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
8
8
|
|
|
9
9
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
10
10
|
|
package/lib/model/chat-choice.js
CHANGED
|
@@ -3,9 +3,9 @@
|
|
|
3
3
|
* Generative AI Service Inference API
|
|
4
4
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
5
|
|
|
6
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
7
7
|
|
|
8
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
9
9
|
|
|
10
10
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
11
11
|
|
|
@@ -2,9 +2,9 @@
|
|
|
2
2
|
* Generative AI Service Inference API
|
|
3
3
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
4
|
|
|
5
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
6
6
|
|
|
7
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
8
8
|
|
|
9
9
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
10
10
|
|
|
@@ -3,9 +3,9 @@
|
|
|
3
3
|
* Generative AI Service Inference API
|
|
4
4
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
5
|
|
|
6
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
7
7
|
|
|
8
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
9
9
|
|
|
10
10
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
11
11
|
|
|
@@ -47,6 +47,8 @@ var ChatContent;
|
|
|
47
47
|
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
48
48
|
if (obj && "type" in obj && obj.type) {
|
|
49
49
|
switch (obj.type) {
|
|
50
|
+
case "IMAGE":
|
|
51
|
+
return model.ImageContent.getJsonObj(jsonObj, true);
|
|
50
52
|
case "TEXT":
|
|
51
53
|
return model.TextContent.getJsonObj(jsonObj, true);
|
|
52
54
|
default:
|
|
@@ -61,6 +63,8 @@ var ChatContent;
|
|
|
61
63
|
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
62
64
|
if (obj && "type" in obj && obj.type) {
|
|
63
65
|
switch (obj.type) {
|
|
66
|
+
case "IMAGE":
|
|
67
|
+
return model.ImageContent.getDeserializedJsonObj(jsonObj, true);
|
|
64
68
|
case "TEXT":
|
|
65
69
|
return model.TextContent.getDeserializedJsonObj(jsonObj, true);
|
|
66
70
|
default:
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"chat-content.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/chat-content.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAClC,qCAAsC;AAStC,IAAiB,WAAW,
|
|
1
|
+
{"version":3,"file":"chat-content.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/chat-content.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAClC,qCAAsC;AAStC,IAAiB,WAAW,CAqC3B;AArCD,WAAiB,WAAW;IAC1B,SAAgB,UAAU,CAAC,GAAgB;QACzC,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,IAAI,GAAG,IAAI,MAAM,IAAI,GAAG,IAAI,GAAG,CAAC,IAAI,EAAE;YACpC,QAAQ,GAAG,CAAC,IAAI,EAAE;gBAChB,KAAK,OAAO;oBACV,OAAO,KAAK,CAAC,YAAY,CAAC,UAAU,CAA8B,OAAQ,EAAE,IAAI,CAAC,CAAC;gBACpF,KAAK,MAAM;oBACT,OAAO,KAAK,CAAC,WAAW,CAAC,UAAU,CAA6B,OAAQ,EAAE,IAAI,CAAC,CAAC;gBAClF;oBACE,IAAI,MAAM,CAAC,GAAG,CAAC,MAAM;wBAAE,MAAM,CAAC,GAAG,CAAC,MAAM,CAAC,IAAI,CAAC,sBAAsB,GAAG,CAAC,IAAI,EAAE,CAAC,CAAC;aACnF;SACF;QACD,OAAO,OAAO,CAAC;IACjB,CAAC;IAde,sBAAU,aAczB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAgB;QACrD,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,IAAI,GAAG,IAAI,MAAM,IAAI,GAAG,IAAI,GAAG,CAAC,IAAI,EAAE;YACpC,QAAQ,GAAG,CAAC,IAAI,EAAE;gBAChB,KAAK,OAAO;oBACV,OAAO,KAAK,CAAC,YAAY,CAAC,sBAAsB,CACjB,OAAQ,EACrC,IAAI,CACL,CAAC;gBACJ,KAAK,MAAM;oBACT,OAAO,KAAK,CAAC,WAAW,CAAC,sBAAsB,CACjB,OAAQ,EACpC,IAAI,CACL,CAAC;gBACJ;oBACE,IAAI,MAAM,CAAC,GAAG,CAAC,MAAM;wBAAE,MAAM,CAAC,GAAG,CAAC,MAAM,CAAC,IAAI,CAAC,sBAAsB,GAAG,CAAC,IAAI,EAAE,CAAC,CAAC;aACnF;SACF;QACD,OAAO,OAAO,CAAC;IACjB,CAAC;IApBe,kCAAsB,yBAoBrC,CAAA;AACH,CAAC,EArCgB,WAAW,GAAX,mBAAW,KAAX,mBAAW,QAqC3B"}
|
|
@@ -2,9 +2,9 @@
|
|
|
2
2
|
* Generative AI Service Inference API
|
|
3
3
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
4
|
|
|
5
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
6
6
|
|
|
7
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
8
8
|
|
|
9
9
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
10
10
|
|
|
@@ -3,9 +3,9 @@
|
|
|
3
3
|
* Generative AI Service Inference API
|
|
4
4
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
5
|
|
|
6
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
7
7
|
|
|
8
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
9
9
|
|
|
10
10
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
11
11
|
|
|
@@ -2,9 +2,9 @@
|
|
|
2
2
|
* Generative AI Service Inference API
|
|
3
3
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
4
|
|
|
5
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
6
6
|
|
|
7
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
8
8
|
|
|
9
9
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
10
10
|
|
package/lib/model/chat-result.js
CHANGED
|
@@ -3,9 +3,9 @@
|
|
|
3
3
|
* Generative AI Service Inference API
|
|
4
4
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
5
|
|
|
6
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
7
7
|
|
|
8
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
9
9
|
|
|
10
10
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
11
11
|
|
package/lib/model/choice.d.ts
CHANGED
|
@@ -2,9 +2,9 @@
|
|
|
2
2
|
* Generative AI Service Inference API
|
|
3
3
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
4
|
|
|
5
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
6
6
|
|
|
7
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
8
8
|
|
|
9
9
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
10
10
|
|
package/lib/model/choice.js
CHANGED
|
@@ -3,9 +3,9 @@
|
|
|
3
3
|
* Generative AI Service Inference API
|
|
4
4
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
5
|
|
|
6
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
7
7
|
|
|
8
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
9
9
|
|
|
10
10
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
11
11
|
|
package/lib/model/citation.d.ts
CHANGED
|
@@ -2,9 +2,9 @@
|
|
|
2
2
|
* Generative AI Service Inference API
|
|
3
3
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
4
|
|
|
5
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
6
6
|
|
|
7
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
8
8
|
|
|
9
9
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
10
10
|
|
package/lib/model/citation.js
CHANGED
|
@@ -3,9 +3,9 @@
|
|
|
3
3
|
* Generative AI Service Inference API
|
|
4
4
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
5
|
|
|
6
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
7
7
|
|
|
8
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
9
9
|
|
|
10
10
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
11
11
|
|
|
@@ -2,9 +2,9 @@
|
|
|
2
2
|
* Generative AI Service Inference API
|
|
3
3
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
4
|
|
|
5
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
6
6
|
|
|
7
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
8
8
|
|
|
9
9
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
10
10
|
|
|
@@ -3,9 +3,9 @@
|
|
|
3
3
|
* Generative AI Service Inference API
|
|
4
4
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
5
|
|
|
6
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
7
7
|
|
|
8
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
9
9
|
|
|
10
10
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
11
11
|
|
|
@@ -2,9 +2,9 @@
|
|
|
2
2
|
* Generative AI Service Inference API
|
|
3
3
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
4
|
|
|
5
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
6
6
|
|
|
7
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
8
8
|
|
|
9
9
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
10
10
|
|
|
@@ -42,6 +42,7 @@ export interface CohereChatRequest extends model.BaseChatRequest {
|
|
|
42
42
|
*
|
|
43
43
|
*/
|
|
44
44
|
"documents"?: Array<any>;
|
|
45
|
+
"responseFormat"?: model.CohereResponseTextFormat | model.CohereResponseJsonFormat;
|
|
45
46
|
/**
|
|
46
47
|
* When set to true, the response contains only a list of generated search queries without the search results and the model will not respond to the user's message.
|
|
47
48
|
*
|
|
@@ -62,6 +63,10 @@ export interface CohereChatRequest extends model.BaseChatRequest {
|
|
|
62
63
|
* The maximum number of output tokens that the model will generate for the response. Note: Numbers greater than Number.MAX_SAFE_INTEGER will result in rounding issues.
|
|
63
64
|
*/
|
|
64
65
|
"maxTokens"?: number;
|
|
66
|
+
/**
|
|
67
|
+
* The maximum number of input tokens to send to the model. If not specified, max_input_tokens is the model's context length limit minus a small buffer. Note: Numbers greater than Number.MAX_SAFE_INTEGER will result in rounding issues.
|
|
68
|
+
*/
|
|
69
|
+
"maxInputTokens"?: number;
|
|
65
70
|
/**
|
|
66
71
|
* A number that sets the randomness of the generated output. A lower temperature means less random generations.
|
|
67
72
|
* Use lower numbers for tasks such as question answering or summarizing. High temperatures can generate hallucinations or factually incorrect information. Start with temperatures lower than 1.0 and increase the temperature for more creative outputs, as you regenerate the prompts to refine the outputs.
|
|
@@ -83,7 +88,7 @@ export interface CohereChatRequest extends model.BaseChatRequest {
|
|
|
83
88
|
*/
|
|
84
89
|
"topP"?: number;
|
|
85
90
|
/**
|
|
86
|
-
* Defaults to OFF. Dictates how the prompt will be constructed. With {@code
|
|
91
|
+
* Defaults to OFF. Dictates how the prompt will be constructed. With {@code promptTruncation} set to AUTO_PRESERVE_ORDER, some elements from {@code chatHistory} and {@code documents} will be dropped to construct a prompt that fits within the model's context length limit. During this process the order of the documents and chat history will be preserved. With {@code prompt_truncation} set to OFF, no elements will be dropped.
|
|
87
92
|
*
|
|
88
93
|
*/
|
|
89
94
|
"promptTruncation"?: CohereChatRequest.PromptTruncation;
|
|
@@ -3,9 +3,9 @@
|
|
|
3
3
|
* Generative AI Service Inference API
|
|
4
4
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
5
|
|
|
6
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
7
7
|
|
|
8
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
9
9
|
|
|
10
10
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
11
11
|
|
|
@@ -59,6 +59,9 @@ var CohereChatRequest;
|
|
|
59
59
|
return model.CohereMessage.getJsonObj(item);
|
|
60
60
|
})
|
|
61
61
|
: undefined,
|
|
62
|
+
"responseFormat": obj.responseFormat
|
|
63
|
+
? model.CohereResponseFormat.getJsonObj(obj.responseFormat)
|
|
64
|
+
: undefined,
|
|
62
65
|
"tools": obj.tools
|
|
63
66
|
? obj.tools.map(item => {
|
|
64
67
|
return model.CohereTool.getJsonObj(item);
|
|
@@ -83,6 +86,9 @@ var CohereChatRequest;
|
|
|
83
86
|
return model.CohereMessage.getDeserializedJsonObj(item);
|
|
84
87
|
})
|
|
85
88
|
: undefined,
|
|
89
|
+
"responseFormat": obj.responseFormat
|
|
90
|
+
? model.CohereResponseFormat.getDeserializedJsonObj(obj.responseFormat)
|
|
91
|
+
: undefined,
|
|
86
92
|
"tools": obj.tools
|
|
87
93
|
? obj.tools.map(item => {
|
|
88
94
|
return model.CohereTool.getDeserializedJsonObj(item);
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"cohere-chat-request.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/cohere-chat-request.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;
|
|
1
|
+
{"version":3,"file":"cohere-chat-request.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/cohere-chat-request.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAgIlC,IAAiB,iBAAiB,CA2EjC;AA3ED,WAAiB,iBAAiB;IAChC,IAAY,gBAGX;IAHD,WAAY,gBAAgB;QAC1B,+BAAW,CAAA;QACX,6DAAyC,CAAA;IAC3C,CAAC,EAHW,gBAAgB,GAAhB,kCAAgB,KAAhB,kCAAgB,QAG3B;IAED,IAAY,eAGX;IAHD,WAAY,eAAe;QACzB,wCAAqB,CAAA;QACrB,gCAAa,CAAA;IACf,CAAC,EAHW,eAAe,GAAf,iCAAe,KAAf,iCAAe,QAG1B;IAED,SAAgB,UAAU,CAAC,GAAsB,EAAE,eAAyB;QAC1E,MAAM,OAAO,mCACR,CAAC,eAAe,CAAC,CAAC,CAAC,GAAG,CAAC,CAAC,CAAE,KAAK,CAAC,eAAe,CAAC,UAAU,CAAC,GAAG,CAAuB,CAAC,GACtF;YACD,aAAa,EAAE,GAAG,CAAC,WAAW;gBAC5B,CAAC,CAAC,GAAG,CAAC,WAAW,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACzB,OAAO,KAAK,CAAC,aAAa,CAAC,UAAU,CAAC,IAAI,CAAC,CAAC;gBAC9C,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;YAEb,gBAAgB,EAAE,GAAG,CAAC,cAAc;gBAClC,CAAC,CAAC,KAAK,CAAC,oBAAoB,CAAC,UAAU,CAAC,GAAG,CAAC,cAAc,CAAC;gBAC3D,CAAC,CAAC,SAAS;YAEb,OAAO,EAAE,GAAG,CAAC,KAAK;gBAChB,CAAC,CAAC,GAAG,CAAC,KAAK,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACnB,OAAO,KAAK,CAAC,UAAU,CAAC,UAAU,CAAC,IAAI,CAAC,CAAC;gBAC3C,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;YACb,aAAa,EAAE,GAAG,CAAC,WAAW;gBAC5B,CAAC,CAAC,GAAG,CAAC,WAAW,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACzB,OAAO,KAAK,CAAC,gBAAgB,CAAC,UAAU,CAAC,IAAI,CAAC,CAAC;gBACjD,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IA5Be,4BAAU,aA4BzB,CAAA;IACY,2BAAS,GAAG,QAAQ,CAAC;IAClC,SAAgB,sBAAsB,CACpC,GAAsB,EACtB,eAAyB;QAEzB,MAAM,OAAO,mCACR,CAAC,eAAe;YACjB,CAAC,CAAC,GAAG;YACL,CAAC,CAAE,KAAK,CAAC,eAAe,CAAC,sBAAsB,CAAC,GAAG,CAAuB,CAAC,GAC1E;YACD,aAAa,EAAE,GAAG,CAAC,WAAW;gBAC5B,CAAC,CAAC,GAAG,CAAC,WAAW,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACzB,OAAO,KAAK,CAAC,aAAa,CAAC,sBAAsB,CAAC,IAAI,CAAC,CAAC;gBAC1D,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;YAEb,gBAAgB,EAAE,GAAG,CAAC,cAAc;gBAClC,CAAC,CAAC,KAAK,CAAC,oBAAoB,CAAC,sBAAsB,CAAC,GAAG,CAAC,cAAc,CAAC;gBACvE,CAAC,CAAC,SAAS;YAEb,OAAO,EAAE,GAAG,CAAC,KAAK;gBAChB,CAAC,CAAC,GAAG,CAAC,KAAK,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACnB,OAAO,KAAK,CAAC,UAAU,CAAC,sBAAsB,CAAC,IAAI,CAAC,CAAC;gBACvD,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;YACb,aAAa,EAAE,GAAG,CAAC,WAAW;gBAC5B,CAAC,CAAC,GAAG,CAAC,WAAW,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACzB,OAAO,KAAK,CAAC,gBAAgB,CAAC,sBAAsB,CAAC,IAAI,CAAC,CAAC;gBAC7D,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAjCe,wCAAsB,yBAiCrC,CAAA;AACH,CAAC,EA3EgB,iBAAiB,GAAjB,yBAAiB,KAAjB,yBAAiB,QA2EjC"}
|