oci-generativeaiinference 2.97.1 → 2.98.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/index.d.ts +2 -2
- package/index.js +2 -2
- package/lib/client.d.ts +2 -2
- package/lib/client.js +2 -2
- package/lib/model/assistant-message.d.ts +2 -2
- package/lib/model/assistant-message.js +2 -2
- package/lib/model/base-chat-request.d.ts +2 -2
- package/lib/model/base-chat-request.js +2 -2
- package/lib/model/base-chat-response.d.ts +2 -2
- package/lib/model/base-chat-response.js +2 -2
- package/lib/model/chat-choice.d.ts +2 -2
- package/lib/model/chat-choice.js +2 -2
- package/lib/model/chat-content.d.ts +2 -2
- package/lib/model/chat-content.js +6 -2
- package/lib/model/chat-content.js.map +1 -1
- package/lib/model/chat-details.d.ts +2 -2
- package/lib/model/chat-details.js +2 -2
- package/lib/model/chat-result.d.ts +2 -2
- package/lib/model/chat-result.js +2 -2
- package/lib/model/choice.d.ts +2 -2
- package/lib/model/choice.js +2 -2
- package/lib/model/citation.d.ts +2 -2
- package/lib/model/citation.js +2 -2
- package/lib/model/cohere-chat-bot-message.d.ts +2 -2
- package/lib/model/cohere-chat-bot-message.js +2 -2
- package/lib/model/cohere-chat-request.d.ts +8 -3
- package/lib/model/cohere-chat-request.js +8 -2
- package/lib/model/cohere-chat-request.js.map +1 -1
- package/lib/model/cohere-chat-response.d.ts +2 -2
- package/lib/model/cohere-chat-response.js +2 -2
- package/lib/model/cohere-llm-inference-request.d.ts +2 -2
- package/lib/model/cohere-llm-inference-request.js +2 -2
- package/lib/model/cohere-llm-inference-response.d.ts +2 -2
- package/lib/model/cohere-llm-inference-response.js +2 -2
- package/lib/model/cohere-message.d.ts +2 -2
- package/lib/model/cohere-message.js +2 -2
- package/lib/model/cohere-parameter-definition.d.ts +2 -2
- package/lib/model/cohere-parameter-definition.js +2 -2
- package/lib/model/cohere-response-format.d.ts +29 -0
- package/lib/model/cohere-response-format.js +79 -0
- package/lib/model/cohere-response-format.js.map +1 -0
- package/lib/model/cohere-response-json-format.d.ts +35 -0
- package/lib/model/cohere-response-json-format.js +61 -0
- package/lib/model/cohere-response-json-format.js.map +1 -0
- package/lib/model/cohere-response-text-format.d.ts +31 -0
- package/lib/model/cohere-response-text-format.js +61 -0
- package/lib/model/cohere-response-text-format.js.map +1 -0
- package/lib/model/cohere-system-message.d.ts +2 -2
- package/lib/model/cohere-system-message.js +2 -2
- package/lib/model/cohere-tool-call.d.ts +2 -2
- package/lib/model/cohere-tool-call.js +2 -2
- package/lib/model/cohere-tool-message.d.ts +2 -2
- package/lib/model/cohere-tool-message.js +2 -2
- package/lib/model/cohere-tool-result.d.ts +2 -2
- package/lib/model/cohere-tool-result.js +2 -2
- package/lib/model/cohere-tool.d.ts +2 -2
- package/lib/model/cohere-tool.js +2 -2
- package/lib/model/cohere-user-message.d.ts +2 -2
- package/lib/model/cohere-user-message.js +2 -2
- package/lib/model/dedicated-serving-mode.d.ts +2 -2
- package/lib/model/dedicated-serving-mode.js +2 -2
- package/lib/model/embed-text-details.d.ts +2 -2
- package/lib/model/embed-text-details.js +2 -2
- package/lib/model/embed-text-result.d.ts +2 -2
- package/lib/model/embed-text-result.js +2 -2
- package/lib/model/generate-text-details.d.ts +2 -2
- package/lib/model/generate-text-details.js +2 -2
- package/lib/model/generate-text-result.d.ts +2 -2
- package/lib/model/generate-text-result.js +2 -2
- package/lib/model/generated-text.d.ts +2 -2
- package/lib/model/generated-text.js +2 -2
- package/lib/model/generic-chat-request.d.ts +2 -2
- package/lib/model/generic-chat-request.js +2 -2
- package/lib/model/generic-chat-response.d.ts +2 -2
- package/lib/model/generic-chat-response.js +2 -2
- package/lib/model/image-content.d.ts +32 -0
- package/lib/model/image-content.js +61 -0
- package/lib/model/image-content.js.map +1 -0
- package/lib/model/image-url.d.ts +46 -0
- package/lib/model/image-url.js +47 -0
- package/lib/model/image-url.js.map +1 -0
- package/lib/model/index.d.ts +12 -2
- package/lib/model/index.js +13 -3
- package/lib/model/index.js.map +1 -1
- package/lib/model/llama-llm-inference-request.d.ts +2 -2
- package/lib/model/llama-llm-inference-request.js +2 -2
- package/lib/model/llama-llm-inference-response.d.ts +2 -2
- package/lib/model/llama-llm-inference-response.js +2 -2
- package/lib/model/llm-inference-request.d.ts +2 -2
- package/lib/model/llm-inference-request.js +2 -2
- package/lib/model/llm-inference-response.d.ts +2 -2
- package/lib/model/llm-inference-response.js +2 -2
- package/lib/model/logprobs.d.ts +2 -2
- package/lib/model/logprobs.js +2 -2
- package/lib/model/message.d.ts +2 -2
- package/lib/model/message.js +2 -2
- package/lib/model/on-demand-serving-mode.d.ts +2 -2
- package/lib/model/on-demand-serving-mode.js +2 -2
- package/lib/model/search-query.d.ts +2 -2
- package/lib/model/search-query.js +2 -2
- package/lib/model/serving-mode.d.ts +2 -2
- package/lib/model/serving-mode.js +2 -2
- package/lib/model/summarize-text-details.d.ts +2 -2
- package/lib/model/summarize-text-details.js +2 -2
- package/lib/model/summarize-text-result.d.ts +2 -2
- package/lib/model/summarize-text-result.js +2 -2
- package/lib/model/system-message.d.ts +2 -2
- package/lib/model/system-message.js +2 -2
- package/lib/model/text-content.d.ts +2 -2
- package/lib/model/text-content.js +2 -2
- package/lib/model/token-likelihood.d.ts +2 -2
- package/lib/model/token-likelihood.js +2 -2
- package/lib/model/user-message.d.ts +2 -2
- package/lib/model/user-message.js +2 -2
- package/lib/request/index.d.ts +2 -2
- package/lib/request/index.js +2 -2
- package/lib/response/index.d.ts +2 -2
- package/lib/response/index.js +2 -2
- package/package.json +3 -3
|
@@ -2,9 +2,9 @@
|
|
|
2
2
|
* Generative AI Service Inference API
|
|
3
3
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
4
|
|
|
5
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
6
6
|
|
|
7
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
8
8
|
|
|
9
9
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
10
10
|
|
|
@@ -3,9 +3,9 @@
|
|
|
3
3
|
* Generative AI Service Inference API
|
|
4
4
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
5
|
|
|
6
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
7
7
|
|
|
8
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
9
9
|
|
|
10
10
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
11
11
|
|
|
@@ -2,9 +2,9 @@
|
|
|
2
2
|
* Generative AI Service Inference API
|
|
3
3
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
4
|
|
|
5
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
6
6
|
|
|
7
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
8
8
|
|
|
9
9
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
10
10
|
|
|
@@ -3,9 +3,9 @@
|
|
|
3
3
|
* Generative AI Service Inference API
|
|
4
4
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
5
|
|
|
6
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
7
7
|
|
|
8
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
9
9
|
|
|
10
10
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
11
11
|
|
|
@@ -2,9 +2,9 @@
|
|
|
2
2
|
* Generative AI Service Inference API
|
|
3
3
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
4
|
|
|
5
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
6
6
|
|
|
7
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
8
8
|
|
|
9
9
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
10
10
|
|
|
@@ -3,9 +3,9 @@
|
|
|
3
3
|
* Generative AI Service Inference API
|
|
4
4
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
5
|
|
|
6
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
7
7
|
|
|
8
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
9
9
|
|
|
10
10
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
11
11
|
|
|
@@ -2,9 +2,9 @@
|
|
|
2
2
|
* Generative AI Service Inference API
|
|
3
3
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
4
|
|
|
5
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
6
6
|
|
|
7
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
8
8
|
|
|
9
9
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
10
10
|
|
|
@@ -3,9 +3,9 @@
|
|
|
3
3
|
* Generative AI Service Inference API
|
|
4
4
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
5
|
|
|
6
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
7
7
|
|
|
8
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
9
9
|
|
|
10
10
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
11
11
|
|
|
@@ -2,9 +2,9 @@
|
|
|
2
2
|
* Generative AI Service Inference API
|
|
3
3
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
4
|
|
|
5
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
6
6
|
|
|
7
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
8
8
|
|
|
9
9
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
10
10
|
|
|
@@ -3,9 +3,9 @@
|
|
|
3
3
|
* Generative AI Service Inference API
|
|
4
4
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
5
|
|
|
6
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
7
7
|
|
|
8
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
9
9
|
|
|
10
10
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
11
11
|
|
|
@@ -0,0 +1,29 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
/**
|
|
21
|
+
* Specify the format the model output is guaranteed to be of
|
|
22
|
+
*/
|
|
23
|
+
export interface CohereResponseFormat {
|
|
24
|
+
"type": string;
|
|
25
|
+
}
|
|
26
|
+
export declare namespace CohereResponseFormat {
|
|
27
|
+
function getJsonObj(obj: CohereResponseFormat): object;
|
|
28
|
+
function getDeserializedJsonObj(obj: CohereResponseFormat): object;
|
|
29
|
+
}
|
|
@@ -0,0 +1,79 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
22
|
+
if (k2 === undefined) k2 = k;
|
|
23
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
24
|
+
}) : (function(o, m, k, k2) {
|
|
25
|
+
if (k2 === undefined) k2 = k;
|
|
26
|
+
o[k2] = m[k];
|
|
27
|
+
}));
|
|
28
|
+
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
29
|
+
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
30
|
+
}) : function(o, v) {
|
|
31
|
+
o["default"] = v;
|
|
32
|
+
});
|
|
33
|
+
var __importStar = (this && this.__importStar) || function (mod) {
|
|
34
|
+
if (mod && mod.__esModule) return mod;
|
|
35
|
+
var result = {};
|
|
36
|
+
if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
37
|
+
__setModuleDefault(result, mod);
|
|
38
|
+
return result;
|
|
39
|
+
};
|
|
40
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
41
|
+
exports.CohereResponseFormat = void 0;
|
|
42
|
+
const model = __importStar(require("../model"));
|
|
43
|
+
const common = require("oci-common");
|
|
44
|
+
var CohereResponseFormat;
|
|
45
|
+
(function (CohereResponseFormat) {
|
|
46
|
+
function getJsonObj(obj) {
|
|
47
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
48
|
+
if (obj && "type" in obj && obj.type) {
|
|
49
|
+
switch (obj.type) {
|
|
50
|
+
case "TEXT":
|
|
51
|
+
return model.CohereResponseTextFormat.getJsonObj(jsonObj, true);
|
|
52
|
+
case "JSON_OBJECT":
|
|
53
|
+
return model.CohereResponseJsonFormat.getJsonObj(jsonObj, true);
|
|
54
|
+
default:
|
|
55
|
+
if (common.LOG.logger)
|
|
56
|
+
common.LOG.logger.info(`Unknown value for: ${obj.type}`);
|
|
57
|
+
}
|
|
58
|
+
}
|
|
59
|
+
return jsonObj;
|
|
60
|
+
}
|
|
61
|
+
CohereResponseFormat.getJsonObj = getJsonObj;
|
|
62
|
+
function getDeserializedJsonObj(obj) {
|
|
63
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
64
|
+
if (obj && "type" in obj && obj.type) {
|
|
65
|
+
switch (obj.type) {
|
|
66
|
+
case "TEXT":
|
|
67
|
+
return model.CohereResponseTextFormat.getDeserializedJsonObj(jsonObj, true);
|
|
68
|
+
case "JSON_OBJECT":
|
|
69
|
+
return model.CohereResponseJsonFormat.getDeserializedJsonObj(jsonObj, true);
|
|
70
|
+
default:
|
|
71
|
+
if (common.LOG.logger)
|
|
72
|
+
common.LOG.logger.info(`Unknown value for: ${obj.type}`);
|
|
73
|
+
}
|
|
74
|
+
}
|
|
75
|
+
return jsonObj;
|
|
76
|
+
}
|
|
77
|
+
CohereResponseFormat.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
78
|
+
})(CohereResponseFormat = exports.CohereResponseFormat || (exports.CohereResponseFormat = {}));
|
|
79
|
+
//# sourceMappingURL=cohere-response-format.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"cohere-response-format.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/cohere-response-format.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAClC,qCAAsC;AAStC,IAAiB,oBAAoB,CA2CpC;AA3CD,WAAiB,oBAAoB;IACnC,SAAgB,UAAU,CAAC,GAAyB;QAClD,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,IAAI,GAAG,IAAI,MAAM,IAAI,GAAG,IAAI,GAAG,CAAC,IAAI,EAAE;YACpC,QAAQ,GAAG,CAAC,IAAI,EAAE;gBAChB,KAAK,MAAM;oBACT,OAAO,KAAK,CAAC,wBAAwB,CAAC,UAAU,CACL,OAAQ,EACjD,IAAI,CACL,CAAC;gBACJ,KAAK,aAAa;oBAChB,OAAO,KAAK,CAAC,wBAAwB,CAAC,UAAU,CACL,OAAQ,EACjD,IAAI,CACL,CAAC;gBACJ;oBACE,IAAI,MAAM,CAAC,GAAG,CAAC,MAAM;wBAAE,MAAM,CAAC,GAAG,CAAC,MAAM,CAAC,IAAI,CAAC,sBAAsB,GAAG,CAAC,IAAI,EAAE,CAAC,CAAC;aACnF;SACF;QACD,OAAO,OAAO,CAAC;IACjB,CAAC;IApBe,+BAAU,aAoBzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAyB;QAC9D,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,IAAI,GAAG,IAAI,MAAM,IAAI,GAAG,IAAI,GAAG,CAAC,IAAI,EAAE;YACpC,QAAQ,GAAG,CAAC,IAAI,EAAE;gBAChB,KAAK,MAAM;oBACT,OAAO,KAAK,CAAC,wBAAwB,CAAC,sBAAsB,CACjB,OAAQ,EACjD,IAAI,CACL,CAAC;gBACJ,KAAK,aAAa;oBAChB,OAAO,KAAK,CAAC,wBAAwB,CAAC,sBAAsB,CACjB,OAAQ,EACjD,IAAI,CACL,CAAC;gBACJ;oBACE,IAAI,MAAM,CAAC,GAAG,CAAC,MAAM;wBAAE,MAAM,CAAC,GAAG,CAAC,MAAM,CAAC,IAAI,CAAC,sBAAsB,GAAG,CAAC,IAAI,EAAE,CAAC,CAAC;aACnF;SACF;QACD,OAAO,OAAO,CAAC;IACjB,CAAC;IApBe,2CAAsB,yBAoBrC,CAAA;AACH,CAAC,EA3CgB,oBAAoB,GAApB,4BAAoB,KAApB,4BAAoB,QA2CpC"}
|
|
@@ -0,0 +1,35 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
import * as model from "../model";
|
|
21
|
+
/**
|
|
22
|
+
* The json object format for the model structured output
|
|
23
|
+
*/
|
|
24
|
+
export interface CohereResponseJsonFormat extends model.CohereResponseFormat {
|
|
25
|
+
/**
|
|
26
|
+
* The schema used by the structured output, described as a JSON Schema object.
|
|
27
|
+
*/
|
|
28
|
+
"schema"?: any;
|
|
29
|
+
"type": string;
|
|
30
|
+
}
|
|
31
|
+
export declare namespace CohereResponseJsonFormat {
|
|
32
|
+
function getJsonObj(obj: CohereResponseJsonFormat, isParentJsonObj?: boolean): object;
|
|
33
|
+
const type = "JSON_OBJECT";
|
|
34
|
+
function getDeserializedJsonObj(obj: CohereResponseJsonFormat, isParentJsonObj?: boolean): object;
|
|
35
|
+
}
|
|
@@ -0,0 +1,61 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
22
|
+
if (k2 === undefined) k2 = k;
|
|
23
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
24
|
+
}) : (function(o, m, k, k2) {
|
|
25
|
+
if (k2 === undefined) k2 = k;
|
|
26
|
+
o[k2] = m[k];
|
|
27
|
+
}));
|
|
28
|
+
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
29
|
+
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
30
|
+
}) : function(o, v) {
|
|
31
|
+
o["default"] = v;
|
|
32
|
+
});
|
|
33
|
+
var __importStar = (this && this.__importStar) || function (mod) {
|
|
34
|
+
if (mod && mod.__esModule) return mod;
|
|
35
|
+
var result = {};
|
|
36
|
+
if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
37
|
+
__setModuleDefault(result, mod);
|
|
38
|
+
return result;
|
|
39
|
+
};
|
|
40
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
41
|
+
exports.CohereResponseJsonFormat = void 0;
|
|
42
|
+
const model = __importStar(require("../model"));
|
|
43
|
+
var CohereResponseJsonFormat;
|
|
44
|
+
(function (CohereResponseJsonFormat) {
|
|
45
|
+
function getJsonObj(obj, isParentJsonObj) {
|
|
46
|
+
const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj
|
|
47
|
+
? obj
|
|
48
|
+
: model.CohereResponseFormat.getJsonObj(obj))), {});
|
|
49
|
+
return jsonObj;
|
|
50
|
+
}
|
|
51
|
+
CohereResponseJsonFormat.getJsonObj = getJsonObj;
|
|
52
|
+
CohereResponseJsonFormat.type = "JSON_OBJECT";
|
|
53
|
+
function getDeserializedJsonObj(obj, isParentJsonObj) {
|
|
54
|
+
const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj
|
|
55
|
+
? obj
|
|
56
|
+
: model.CohereResponseFormat.getDeserializedJsonObj(obj))), {});
|
|
57
|
+
return jsonObj;
|
|
58
|
+
}
|
|
59
|
+
CohereResponseJsonFormat.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
60
|
+
})(CohereResponseJsonFormat = exports.CohereResponseJsonFormat || (exports.CohereResponseJsonFormat = {}));
|
|
61
|
+
//# sourceMappingURL=cohere-response-json-format.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"cohere-response-json-format.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/cohere-response-json-format.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAelC,IAAiB,wBAAwB,CAyBxC;AAzBD,WAAiB,wBAAwB;IACvC,SAAgB,UAAU,CAAC,GAA6B,EAAE,eAAyB;QACjF,MAAM,OAAO,mCACR,CAAC,eAAe;YACjB,CAAC,CAAC,GAAG;YACL,CAAC,CAAE,KAAK,CAAC,oBAAoB,CAAC,UAAU,CAAC,GAAG,CAA8B,CAAC,GAC1E,EAAE,CACN,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IATe,mCAAU,aASzB,CAAA;IACY,6BAAI,GAAG,aAAa,CAAC;IAClC,SAAgB,sBAAsB,CACpC,GAA6B,EAC7B,eAAyB;QAEzB,MAAM,OAAO,mCACR,CAAC,eAAe;YACjB,CAAC,CAAC,GAAG;YACL,CAAC,CAAE,KAAK,CAAC,oBAAoB,CAAC,sBAAsB,CAAC,GAAG,CAA8B,CAAC,GACtF,EAAE,CACN,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAZe,+CAAsB,yBAYrC,CAAA;AACH,CAAC,EAzBgB,wBAAwB,GAAxB,gCAAwB,KAAxB,gCAAwB,QAyBxC"}
|
|
@@ -0,0 +1,31 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generative AI Service Inference API
|
|
3
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
|
+
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
6
|
+
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
8
|
+
|
|
9
|
+
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
10
|
+
|
|
11
|
+
* OpenAPI spec version: 20231130
|
|
12
|
+
*
|
|
13
|
+
*
|
|
14
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
15
|
+
* Do not edit the class manually.
|
|
16
|
+
*
|
|
17
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
18
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
19
|
+
*/
|
|
20
|
+
import * as model from "../model";
|
|
21
|
+
/**
|
|
22
|
+
* The text format for cohere model response
|
|
23
|
+
*/
|
|
24
|
+
export interface CohereResponseTextFormat extends model.CohereResponseFormat {
|
|
25
|
+
"type": string;
|
|
26
|
+
}
|
|
27
|
+
export declare namespace CohereResponseTextFormat {
|
|
28
|
+
function getJsonObj(obj: CohereResponseTextFormat, isParentJsonObj?: boolean): object;
|
|
29
|
+
const type = "TEXT";
|
|
30
|
+
function getDeserializedJsonObj(obj: CohereResponseTextFormat, isParentJsonObj?: boolean): object;
|
|
31
|
+
}
|
|
@@ -0,0 +1,61 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Generative AI Service Inference API
|
|
4
|
+
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
|
+
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
7
|
+
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
9
|
+
|
|
10
|
+
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
11
|
+
|
|
12
|
+
* OpenAPI spec version: 20231130
|
|
13
|
+
*
|
|
14
|
+
*
|
|
15
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
16
|
+
* Do not edit the class manually.
|
|
17
|
+
*
|
|
18
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
19
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
20
|
+
*/
|
|
21
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
22
|
+
if (k2 === undefined) k2 = k;
|
|
23
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
24
|
+
}) : (function(o, m, k, k2) {
|
|
25
|
+
if (k2 === undefined) k2 = k;
|
|
26
|
+
o[k2] = m[k];
|
|
27
|
+
}));
|
|
28
|
+
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
29
|
+
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
30
|
+
}) : function(o, v) {
|
|
31
|
+
o["default"] = v;
|
|
32
|
+
});
|
|
33
|
+
var __importStar = (this && this.__importStar) || function (mod) {
|
|
34
|
+
if (mod && mod.__esModule) return mod;
|
|
35
|
+
var result = {};
|
|
36
|
+
if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
37
|
+
__setModuleDefault(result, mod);
|
|
38
|
+
return result;
|
|
39
|
+
};
|
|
40
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
41
|
+
exports.CohereResponseTextFormat = void 0;
|
|
42
|
+
const model = __importStar(require("../model"));
|
|
43
|
+
var CohereResponseTextFormat;
|
|
44
|
+
(function (CohereResponseTextFormat) {
|
|
45
|
+
function getJsonObj(obj, isParentJsonObj) {
|
|
46
|
+
const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj
|
|
47
|
+
? obj
|
|
48
|
+
: model.CohereResponseFormat.getJsonObj(obj))), {});
|
|
49
|
+
return jsonObj;
|
|
50
|
+
}
|
|
51
|
+
CohereResponseTextFormat.getJsonObj = getJsonObj;
|
|
52
|
+
CohereResponseTextFormat.type = "TEXT";
|
|
53
|
+
function getDeserializedJsonObj(obj, isParentJsonObj) {
|
|
54
|
+
const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj
|
|
55
|
+
? obj
|
|
56
|
+
: model.CohereResponseFormat.getDeserializedJsonObj(obj))), {});
|
|
57
|
+
return jsonObj;
|
|
58
|
+
}
|
|
59
|
+
CohereResponseTextFormat.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
60
|
+
})(CohereResponseTextFormat = exports.CohereResponseTextFormat || (exports.CohereResponseTextFormat = {}));
|
|
61
|
+
//# sourceMappingURL=cohere-response-text-format.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"cohere-response-text-format.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/cohere-response-text-format.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAUlC,IAAiB,wBAAwB,CAyBxC;AAzBD,WAAiB,wBAAwB;IACvC,SAAgB,UAAU,CAAC,GAA6B,EAAE,eAAyB;QACjF,MAAM,OAAO,mCACR,CAAC,eAAe;YACjB,CAAC,CAAC,GAAG;YACL,CAAC,CAAE,KAAK,CAAC,oBAAoB,CAAC,UAAU,CAAC,GAAG,CAA8B,CAAC,GAC1E,EAAE,CACN,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IATe,mCAAU,aASzB,CAAA;IACY,6BAAI,GAAG,MAAM,CAAC;IAC3B,SAAgB,sBAAsB,CACpC,GAA6B,EAC7B,eAAyB;QAEzB,MAAM,OAAO,mCACR,CAAC,eAAe;YACjB,CAAC,CAAC,GAAG;YACL,CAAC,CAAE,KAAK,CAAC,oBAAoB,CAAC,sBAAsB,CAAC,GAAG,CAA8B,CAAC,GACtF,EAAE,CACN,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAZe,+CAAsB,yBAYrC,CAAA;AACH,CAAC,EAzBgB,wBAAwB,GAAxB,gCAAwB,KAAxB,gCAAwB,QAyBxC"}
|
|
@@ -2,9 +2,9 @@
|
|
|
2
2
|
* Generative AI Service Inference API
|
|
3
3
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
4
|
|
|
5
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
6
6
|
|
|
7
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
8
8
|
|
|
9
9
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
10
10
|
|
|
@@ -3,9 +3,9 @@
|
|
|
3
3
|
* Generative AI Service Inference API
|
|
4
4
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
5
|
|
|
6
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
7
7
|
|
|
8
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
9
9
|
|
|
10
10
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
11
11
|
|
|
@@ -2,9 +2,9 @@
|
|
|
2
2
|
* Generative AI Service Inference API
|
|
3
3
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
4
|
|
|
5
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
6
6
|
|
|
7
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
8
8
|
|
|
9
9
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
10
10
|
|
|
@@ -3,9 +3,9 @@
|
|
|
3
3
|
* Generative AI Service Inference API
|
|
4
4
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
5
5
|
|
|
6
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
6
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
7
7
|
|
|
8
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
8
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
9
9
|
|
|
10
10
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
11
11
|
|
|
@@ -2,9 +2,9 @@
|
|
|
2
2
|
* Generative AI Service Inference API
|
|
3
3
|
* OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
|
|
4
4
|
|
|
5
|
-
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/
|
|
5
|
+
Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [chat](#/EN/generative-ai-inference/latest/ChatResult/Chat), [generate text](#/EN/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/EN/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/EN/generative-ai-inference/latest/EmbedTextResult/EmbedText).
|
|
6
6
|
|
|
7
|
-
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](
|
|
7
|
+
To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](#/EN/generative-ai/latest/) to [create a custom model](#/EN/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/EN/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](#/EN/generative-ai/latest/).
|
|
8
8
|
|
|
9
9
|
To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
|
|
10
10
|
|