oci-generativeaiinference 2.77.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (95) hide show
  1. package/LICENSE.txt +89 -0
  2. package/NOTICE.txt +1 -0
  3. package/README.md +22 -0
  4. package/THIRD_PARTY_LICENSES.txt +576 -0
  5. package/index.d.ts +27 -0
  6. package/index.js +50 -0
  7. package/index.js.map +1 -0
  8. package/lib/client.d.ts +113 -0
  9. package/lib/client.js +376 -0
  10. package/lib/client.js.map +1 -0
  11. package/lib/model/choice.d.ts +45 -0
  12. package/lib/model/choice.js +60 -0
  13. package/lib/model/choice.js.map +1 -0
  14. package/lib/model/cohere-llm-inference-request.d.ts +106 -0
  15. package/lib/model/cohere-llm-inference-request.js +73 -0
  16. package/lib/model/cohere-llm-inference-request.js.map +1 -0
  17. package/lib/model/cohere-llm-inference-response.d.ts +43 -0
  18. package/lib/model/cohere-llm-inference-response.js +73 -0
  19. package/lib/model/cohere-llm-inference-response.js.map +1 -0
  20. package/lib/model/dedicated-serving-mode.d.ts +35 -0
  21. package/lib/model/dedicated-serving-mode.js +59 -0
  22. package/lib/model/dedicated-serving-mode.js.map +1 -0
  23. package/lib/model/embed-text-details.d.ts +61 -0
  24. package/lib/model/embed-text-details.js +75 -0
  25. package/lib/model/embed-text-details.js.map +1 -0
  26. package/lib/model/embed-text-result.d.ts +48 -0
  27. package/lib/model/embed-text-result.js +36 -0
  28. package/lib/model/embed-text-result.js.map +1 -0
  29. package/lib/model/generate-text-details.d.ts +35 -0
  30. package/lib/model/generate-text-details.js +68 -0
  31. package/lib/model/generate-text-details.js.map +1 -0
  32. package/lib/model/generate-text-result.d.ts +38 -0
  33. package/lib/model/generate-text-result.js +64 -0
  34. package/lib/model/generate-text-result.js.map +1 -0
  35. package/lib/model/generated-text.d.ts +55 -0
  36. package/lib/model/generated-text.js +68 -0
  37. package/lib/model/generated-text.js.map +1 -0
  38. package/lib/model/index.d.ts +57 -0
  39. package/lib/model/index.js +80 -0
  40. package/lib/model/index.js.map +1 -0
  41. package/lib/model/llama-llm-inference-request.d.ts +94 -0
  42. package/lib/model/llama-llm-inference-request.js +61 -0
  43. package/lib/model/llama-llm-inference-request.js.map +1 -0
  44. package/lib/model/llama-llm-inference-response.d.ts +39 -0
  45. package/lib/model/llama-llm-inference-response.js +73 -0
  46. package/lib/model/llama-llm-inference-response.js.map +1 -0
  47. package/lib/model/llm-inference-request.d.ts +29 -0
  48. package/lib/model/llm-inference-request.js +79 -0
  49. package/lib/model/llm-inference-request.js.map +1 -0
  50. package/lib/model/llm-inference-response.d.ts +29 -0
  51. package/lib/model/llm-inference-response.js +79 -0
  52. package/lib/model/llm-inference-response.js.map +1 -0
  53. package/lib/model/logprobs.d.ts +46 -0
  54. package/lib/model/logprobs.js +36 -0
  55. package/lib/model/logprobs.js.map +1 -0
  56. package/lib/model/on-demand-serving-mode.d.ts +35 -0
  57. package/lib/model/on-demand-serving-mode.js +59 -0
  58. package/lib/model/on-demand-serving-mode.js.map +1 -0
  59. package/lib/model/serving-mode.d.ts +29 -0
  60. package/lib/model/serving-mode.js +79 -0
  61. package/lib/model/serving-mode.js.map +1 -0
  62. package/lib/model/summarize-text-details.d.ts +82 -0
  63. package/lib/model/summarize-text-details.js +82 -0
  64. package/lib/model/summarize-text-details.js.map +1 -0
  65. package/lib/model/summarize-text-result.d.ts +48 -0
  66. package/lib/model/summarize-text-result.js +36 -0
  67. package/lib/model/summarize-text-result.js.map +1 -0
  68. package/lib/model/token-likelihood.d.ts +39 -0
  69. package/lib/model/token-likelihood.js +36 -0
  70. package/lib/model/token-likelihood.js.map +1 -0
  71. package/lib/request/embed-text-request.d.ts +36 -0
  72. package/lib/request/embed-text-request.js +15 -0
  73. package/lib/request/embed-text-request.js.map +1 -0
  74. package/lib/request/generate-text-request.d.ts +36 -0
  75. package/lib/request/generate-text-request.js +15 -0
  76. package/lib/request/generate-text-request.js.map +1 -0
  77. package/lib/request/index.d.ts +25 -0
  78. package/lib/request/index.js +22 -0
  79. package/lib/request/index.js.map +1 -0
  80. package/lib/request/summarize-text-request.d.ts +36 -0
  81. package/lib/request/summarize-text-request.js +15 -0
  82. package/lib/request/summarize-text-request.js.map +1 -0
  83. package/lib/response/embed-text-response.d.ts +30 -0
  84. package/lib/response/embed-text-response.js +15 -0
  85. package/lib/response/embed-text-response.js.map +1 -0
  86. package/lib/response/generate-text-response.d.ts +30 -0
  87. package/lib/response/generate-text-response.js +15 -0
  88. package/lib/response/generate-text-response.js.map +1 -0
  89. package/lib/response/index.d.ts +25 -0
  90. package/lib/response/index.js +22 -0
  91. package/lib/response/index.js.map +1 -0
  92. package/lib/response/summarize-text-response.d.ts +30 -0
  93. package/lib/response/summarize-text-response.js +15 -0
  94. package/lib/response/summarize-text-response.js.map +1 -0
  95. package/package.json +29 -0
@@ -0,0 +1,36 @@
1
+ "use strict";
2
+ /**
3
+ * Generative AI Service Inference API
4
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
5
+
6
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [generate text](#/en/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/en/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/en/generative-ai-inference/latest/EmbedTextResult/EmbedText).
7
+
8
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](/#/en/generative-ai/latest/) to [create a custom model](#/en/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](/#/en/generative-ai/latest/).
9
+
10
+ To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
11
+
12
+ * OpenAPI spec version: 20231130
13
+ *
14
+ *
15
+ * NOTE: This class is auto generated by OracleSDKGenerator.
16
+ * Do not edit the class manually.
17
+ *
18
+ * Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
19
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
20
+ */
21
+ Object.defineProperty(exports, "__esModule", { value: true });
22
+ exports.EmbedTextResult = void 0;
23
+ var EmbedTextResult;
24
+ (function (EmbedTextResult) {
25
+ function getJsonObj(obj) {
26
+ const jsonObj = Object.assign(Object.assign({}, obj), {});
27
+ return jsonObj;
28
+ }
29
+ EmbedTextResult.getJsonObj = getJsonObj;
30
+ function getDeserializedJsonObj(obj) {
31
+ const jsonObj = Object.assign(Object.assign({}, obj), {});
32
+ return jsonObj;
33
+ }
34
+ EmbedTextResult.getDeserializedJsonObj = getDeserializedJsonObj;
35
+ })(EmbedTextResult = exports.EmbedTextResult || (exports.EmbedTextResult = {}));
36
+ //# sourceMappingURL=embed-text-result.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"embed-text-result.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/embed-text-result.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;AA+BH,IAAiB,eAAe,CAW/B;AAXD,WAAiB,eAAe;IAC9B,SAAgB,UAAU,CAAC,GAAoB;QAC7C,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,0BAAU,aAIzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAoB;QACzD,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,sCAAsB,yBAIrC,CAAA;AACH,CAAC,EAXgB,eAAe,GAAf,uBAAe,KAAf,uBAAe,QAW/B"}
@@ -0,0 +1,35 @@
1
+ /**
2
+ * Generative AI Service Inference API
3
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
4
+
5
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [generate text](#/en/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/en/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/en/generative-ai-inference/latest/EmbedTextResult/EmbedText).
6
+
7
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](/#/en/generative-ai/latest/) to [create a custom model](#/en/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](/#/en/generative-ai/latest/).
8
+
9
+ To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
10
+
11
+ * OpenAPI spec version: 20231130
12
+ *
13
+ *
14
+ * NOTE: This class is auto generated by OracleSDKGenerator.
15
+ * Do not edit the class manually.
16
+ *
17
+ * Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
18
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
19
+ */
20
+ import * as model from "../model";
21
+ /**
22
+ * Details for the request to generate text.
23
+ */
24
+ export interface GenerateTextDetails {
25
+ /**
26
+ * The OCID of compartment that the user is authorized to use to call into the Generative AI service.
27
+ */
28
+ "compartmentId": string;
29
+ "servingMode": model.DedicatedServingMode | model.OnDemandServingMode;
30
+ "inferenceRequest": model.LlamaLlmInferenceRequest | model.CohereLlmInferenceRequest;
31
+ }
32
+ export declare namespace GenerateTextDetails {
33
+ function getJsonObj(obj: GenerateTextDetails): object;
34
+ function getDeserializedJsonObj(obj: GenerateTextDetails): object;
35
+ }
@@ -0,0 +1,68 @@
1
+ "use strict";
2
+ /**
3
+ * Generative AI Service Inference API
4
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
5
+
6
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [generate text](#/en/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/en/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/en/generative-ai-inference/latest/EmbedTextResult/EmbedText).
7
+
8
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](/#/en/generative-ai/latest/) to [create a custom model](#/en/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](/#/en/generative-ai/latest/).
9
+
10
+ To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
11
+
12
+ * OpenAPI spec version: 20231130
13
+ *
14
+ *
15
+ * NOTE: This class is auto generated by OracleSDKGenerator.
16
+ * Do not edit the class manually.
17
+ *
18
+ * Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
19
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
20
+ */
21
+ var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
22
+ if (k2 === undefined) k2 = k;
23
+ Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
24
+ }) : (function(o, m, k, k2) {
25
+ if (k2 === undefined) k2 = k;
26
+ o[k2] = m[k];
27
+ }));
28
+ var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
29
+ Object.defineProperty(o, "default", { enumerable: true, value: v });
30
+ }) : function(o, v) {
31
+ o["default"] = v;
32
+ });
33
+ var __importStar = (this && this.__importStar) || function (mod) {
34
+ if (mod && mod.__esModule) return mod;
35
+ var result = {};
36
+ if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
37
+ __setModuleDefault(result, mod);
38
+ return result;
39
+ };
40
+ Object.defineProperty(exports, "__esModule", { value: true });
41
+ exports.GenerateTextDetails = void 0;
42
+ const model = __importStar(require("../model"));
43
+ var GenerateTextDetails;
44
+ (function (GenerateTextDetails) {
45
+ function getJsonObj(obj) {
46
+ const jsonObj = Object.assign(Object.assign({}, obj), {
47
+ "servingMode": obj.servingMode ? model.ServingMode.getJsonObj(obj.servingMode) : undefined,
48
+ "inferenceRequest": obj.inferenceRequest
49
+ ? model.LlmInferenceRequest.getJsonObj(obj.inferenceRequest)
50
+ : undefined
51
+ });
52
+ return jsonObj;
53
+ }
54
+ GenerateTextDetails.getJsonObj = getJsonObj;
55
+ function getDeserializedJsonObj(obj) {
56
+ const jsonObj = Object.assign(Object.assign({}, obj), {
57
+ "servingMode": obj.servingMode
58
+ ? model.ServingMode.getDeserializedJsonObj(obj.servingMode)
59
+ : undefined,
60
+ "inferenceRequest": obj.inferenceRequest
61
+ ? model.LlmInferenceRequest.getDeserializedJsonObj(obj.inferenceRequest)
62
+ : undefined
63
+ });
64
+ return jsonObj;
65
+ }
66
+ GenerateTextDetails.getDeserializedJsonObj = getDeserializedJsonObj;
67
+ })(GenerateTextDetails = exports.GenerateTextDetails || (exports.GenerateTextDetails = {}));
68
+ //# sourceMappingURL=generate-text-details.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"generate-text-details.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/generate-text-details.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAelC,IAAiB,mBAAmB,CA6BnC;AA7BD,WAAiB,mBAAmB;IAClC,SAAgB,UAAU,CAAC,GAAwB;QACjD,MAAM,OAAO,mCACR,GAAG,GACH;YACD,aAAa,EAAE,GAAG,CAAC,WAAW,CAAC,CAAC,CAAC,KAAK,CAAC,WAAW,CAAC,UAAU,CAAC,GAAG,CAAC,WAAW,CAAC,CAAC,CAAC,CAAC,SAAS;YAC1F,kBAAkB,EAAE,GAAG,CAAC,gBAAgB;gBACtC,CAAC,CAAC,KAAK,CAAC,mBAAmB,CAAC,UAAU,CAAC,GAAG,CAAC,gBAAgB,CAAC;gBAC5D,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAZe,8BAAU,aAYzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAwB;QAC7D,MAAM,OAAO,mCACR,GAAG,GACH;YACD,aAAa,EAAE,GAAG,CAAC,WAAW;gBAC5B,CAAC,CAAC,KAAK,CAAC,WAAW,CAAC,sBAAsB,CAAC,GAAG,CAAC,WAAW,CAAC;gBAC3D,CAAC,CAAC,SAAS;YACb,kBAAkB,EAAE,GAAG,CAAC,gBAAgB;gBACtC,CAAC,CAAC,KAAK,CAAC,mBAAmB,CAAC,sBAAsB,CAAC,GAAG,CAAC,gBAAgB,CAAC;gBACxE,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAde,0CAAsB,yBAcrC,CAAA;AACH,CAAC,EA7BgB,mBAAmB,GAAnB,2BAAmB,KAAnB,2BAAmB,QA6BnC"}
@@ -0,0 +1,38 @@
1
+ /**
2
+ * Generative AI Service Inference API
3
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
4
+
5
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [generate text](#/en/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/en/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/en/generative-ai-inference/latest/EmbedTextResult/EmbedText).
6
+
7
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](/#/en/generative-ai/latest/) to [create a custom model](#/en/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](/#/en/generative-ai/latest/).
8
+
9
+ To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
10
+
11
+ * OpenAPI spec version: 20231130
12
+ *
13
+ *
14
+ * NOTE: This class is auto generated by OracleSDKGenerator.
15
+ * Do not edit the class manually.
16
+ *
17
+ * Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
18
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
19
+ */
20
+ import * as model from "../model";
21
+ /**
22
+ * The generated text result to return.
23
+ */
24
+ export interface GenerateTextResult {
25
+ /**
26
+ * The OCID of the model used in this inference request.
27
+ */
28
+ "modelId": string;
29
+ /**
30
+ * The version of the model.
31
+ */
32
+ "modelVersion": string;
33
+ "inferenceResponse": model.LlamaLlmInferenceResponse | model.CohereLlmInferenceResponse;
34
+ }
35
+ export declare namespace GenerateTextResult {
36
+ function getJsonObj(obj: GenerateTextResult): object;
37
+ function getDeserializedJsonObj(obj: GenerateTextResult): object;
38
+ }
@@ -0,0 +1,64 @@
1
+ "use strict";
2
+ /**
3
+ * Generative AI Service Inference API
4
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
5
+
6
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [generate text](#/en/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/en/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/en/generative-ai-inference/latest/EmbedTextResult/EmbedText).
7
+
8
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](/#/en/generative-ai/latest/) to [create a custom model](#/en/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](/#/en/generative-ai/latest/).
9
+
10
+ To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
11
+
12
+ * OpenAPI spec version: 20231130
13
+ *
14
+ *
15
+ * NOTE: This class is auto generated by OracleSDKGenerator.
16
+ * Do not edit the class manually.
17
+ *
18
+ * Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
19
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
20
+ */
21
+ var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
22
+ if (k2 === undefined) k2 = k;
23
+ Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
24
+ }) : (function(o, m, k, k2) {
25
+ if (k2 === undefined) k2 = k;
26
+ o[k2] = m[k];
27
+ }));
28
+ var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
29
+ Object.defineProperty(o, "default", { enumerable: true, value: v });
30
+ }) : function(o, v) {
31
+ o["default"] = v;
32
+ });
33
+ var __importStar = (this && this.__importStar) || function (mod) {
34
+ if (mod && mod.__esModule) return mod;
35
+ var result = {};
36
+ if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
37
+ __setModuleDefault(result, mod);
38
+ return result;
39
+ };
40
+ Object.defineProperty(exports, "__esModule", { value: true });
41
+ exports.GenerateTextResult = void 0;
42
+ const model = __importStar(require("../model"));
43
+ var GenerateTextResult;
44
+ (function (GenerateTextResult) {
45
+ function getJsonObj(obj) {
46
+ const jsonObj = Object.assign(Object.assign({}, obj), {
47
+ "inferenceResponse": obj.inferenceResponse
48
+ ? model.LlmInferenceResponse.getJsonObj(obj.inferenceResponse)
49
+ : undefined
50
+ });
51
+ return jsonObj;
52
+ }
53
+ GenerateTextResult.getJsonObj = getJsonObj;
54
+ function getDeserializedJsonObj(obj) {
55
+ const jsonObj = Object.assign(Object.assign({}, obj), {
56
+ "inferenceResponse": obj.inferenceResponse
57
+ ? model.LlmInferenceResponse.getDeserializedJsonObj(obj.inferenceResponse)
58
+ : undefined
59
+ });
60
+ return jsonObj;
61
+ }
62
+ GenerateTextResult.getDeserializedJsonObj = getDeserializedJsonObj;
63
+ })(GenerateTextResult = exports.GenerateTextResult || (exports.GenerateTextResult = {}));
64
+ //# sourceMappingURL=generate-text-result.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"generate-text-result.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/generate-text-result.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAkBlC,IAAiB,kBAAkB,CAyBlC;AAzBD,WAAiB,kBAAkB;IACjC,SAAgB,UAAU,CAAC,GAAuB;QAChD,MAAM,OAAO,mCACR,GAAG,GACH;YACD,mBAAmB,EAAE,GAAG,CAAC,iBAAiB;gBACxC,CAAC,CAAC,KAAK,CAAC,oBAAoB,CAAC,UAAU,CAAC,GAAG,CAAC,iBAAiB,CAAC;gBAC9D,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAXe,6BAAU,aAWzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAuB;QAC5D,MAAM,OAAO,mCACR,GAAG,GACH;YACD,mBAAmB,EAAE,GAAG,CAAC,iBAAiB;gBACxC,CAAC,CAAC,KAAK,CAAC,oBAAoB,CAAC,sBAAsB,CAAC,GAAG,CAAC,iBAAiB,CAAC;gBAC1E,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAXe,yCAAsB,yBAWrC,CAAA;AACH,CAAC,EAzBgB,kBAAkB,GAAlB,0BAAkB,KAAlB,0BAAkB,QAyBlC"}
@@ -0,0 +1,55 @@
1
+ /**
2
+ * Generative AI Service Inference API
3
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
4
+
5
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [generate text](#/en/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/en/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/en/generative-ai-inference/latest/EmbedTextResult/EmbedText).
6
+
7
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](/#/en/generative-ai/latest/) to [create a custom model](#/en/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](/#/en/generative-ai/latest/).
8
+
9
+ To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
10
+
11
+ * OpenAPI spec version: 20231130
12
+ *
13
+ *
14
+ * NOTE: This class is auto generated by OracleSDKGenerator.
15
+ * Do not edit the class manually.
16
+ *
17
+ * Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
18
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
19
+ */
20
+ import * as model from "../model";
21
+ /**
22
+ * The text generated during each run.
23
+ */
24
+ export interface GeneratedText {
25
+ /**
26
+ * A unique identifier for this text generation.
27
+ */
28
+ "id": string;
29
+ /**
30
+ * The generated text.
31
+ */
32
+ "text": string;
33
+ /**
34
+ * The overall likelihood of the generated text.
35
+ * <p>
36
+ When a large language model generates a new token for the output text, a likelihood is assigned to all tokens, where tokens with higher likelihoods are more likely to follow the current token. For example, it's more likely that the word favorite is followed by the word food or book rather than the word zebra. A lower likelihood means that it's less likely that token follows the current token.
37
+ * Note: Numbers greater than Number.MAX_SAFE_INTEGER will result in rounding issues.
38
+ */
39
+ "likelihood": number;
40
+ /**
41
+ * The reason why the model stopped generating tokens.
42
+ * <p>
43
+ A model stops generating tokens if the model hits a natural stop point or reaches a provided stop sequence.
44
+ *
45
+ */
46
+ "finishReason"?: string;
47
+ /**
48
+ * A collection of generated tokens and their corresponding likelihoods.
49
+ */
50
+ "tokenLikelihoods"?: Array<model.TokenLikelihood>;
51
+ }
52
+ export declare namespace GeneratedText {
53
+ function getJsonObj(obj: GeneratedText): object;
54
+ function getDeserializedJsonObj(obj: GeneratedText): object;
55
+ }
@@ -0,0 +1,68 @@
1
+ "use strict";
2
+ /**
3
+ * Generative AI Service Inference API
4
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
5
+
6
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [generate text](#/en/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/en/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/en/generative-ai-inference/latest/EmbedTextResult/EmbedText).
7
+
8
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](/#/en/generative-ai/latest/) to [create a custom model](#/en/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](/#/en/generative-ai/latest/).
9
+
10
+ To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
11
+
12
+ * OpenAPI spec version: 20231130
13
+ *
14
+ *
15
+ * NOTE: This class is auto generated by OracleSDKGenerator.
16
+ * Do not edit the class manually.
17
+ *
18
+ * Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
19
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
20
+ */
21
+ var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
22
+ if (k2 === undefined) k2 = k;
23
+ Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
24
+ }) : (function(o, m, k, k2) {
25
+ if (k2 === undefined) k2 = k;
26
+ o[k2] = m[k];
27
+ }));
28
+ var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
29
+ Object.defineProperty(o, "default", { enumerable: true, value: v });
30
+ }) : function(o, v) {
31
+ o["default"] = v;
32
+ });
33
+ var __importStar = (this && this.__importStar) || function (mod) {
34
+ if (mod && mod.__esModule) return mod;
35
+ var result = {};
36
+ if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
37
+ __setModuleDefault(result, mod);
38
+ return result;
39
+ };
40
+ Object.defineProperty(exports, "__esModule", { value: true });
41
+ exports.GeneratedText = void 0;
42
+ const model = __importStar(require("../model"));
43
+ var GeneratedText;
44
+ (function (GeneratedText) {
45
+ function getJsonObj(obj) {
46
+ const jsonObj = Object.assign(Object.assign({}, obj), {
47
+ "tokenLikelihoods": obj.tokenLikelihoods
48
+ ? obj.tokenLikelihoods.map(item => {
49
+ return model.TokenLikelihood.getJsonObj(item);
50
+ })
51
+ : undefined
52
+ });
53
+ return jsonObj;
54
+ }
55
+ GeneratedText.getJsonObj = getJsonObj;
56
+ function getDeserializedJsonObj(obj) {
57
+ const jsonObj = Object.assign(Object.assign({}, obj), {
58
+ "tokenLikelihoods": obj.tokenLikelihoods
59
+ ? obj.tokenLikelihoods.map(item => {
60
+ return model.TokenLikelihood.getDeserializedJsonObj(item);
61
+ })
62
+ : undefined
63
+ });
64
+ return jsonObj;
65
+ }
66
+ GeneratedText.getDeserializedJsonObj = getDeserializedJsonObj;
67
+ })(GeneratedText = exports.GeneratedText || (exports.GeneratedText = {}));
68
+ //# sourceMappingURL=generated-text.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"generated-text.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/generated-text.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAmClC,IAAiB,aAAa,CA6B7B;AA7BD,WAAiB,aAAa;IAC5B,SAAgB,UAAU,CAAC,GAAkB;QAC3C,MAAM,OAAO,mCACR,GAAG,GACH;YACD,kBAAkB,EAAE,GAAG,CAAC,gBAAgB;gBACtC,CAAC,CAAC,GAAG,CAAC,gBAAgB,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBAC9B,OAAO,KAAK,CAAC,eAAe,CAAC,UAAU,CAAC,IAAI,CAAC,CAAC;gBAChD,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAbe,wBAAU,aAazB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAkB;QACvD,MAAM,OAAO,mCACR,GAAG,GACH;YACD,kBAAkB,EAAE,GAAG,CAAC,gBAAgB;gBACtC,CAAC,CAAC,GAAG,CAAC,gBAAgB,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBAC9B,OAAO,KAAK,CAAC,eAAe,CAAC,sBAAsB,CAAC,IAAI,CAAC,CAAC;gBAC5D,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAbe,oCAAsB,yBAarC,CAAA;AACH,CAAC,EA7BgB,aAAa,GAAb,qBAAa,KAAb,qBAAa,QA6B7B"}
@@ -0,0 +1,57 @@
1
+ /**
2
+ * Generative AI Service Inference API
3
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
4
+
5
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [generate text](#/en/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/en/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/en/generative-ai-inference/latest/EmbedTextResult/EmbedText).
6
+
7
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](/#/en/generative-ai/latest/) to [create a custom model](#/en/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](/#/en/generative-ai/latest/).
8
+
9
+ To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
10
+
11
+ * OpenAPI spec version: 20231130
12
+ *
13
+ *
14
+ * NOTE: This class is auto generated by OracleSDKGenerator.
15
+ * Do not edit the class manually.
16
+ *
17
+ * Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
18
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
19
+ */
20
+ import * as Choice from "./choice";
21
+ export import Choice = Choice.Choice;
22
+ import * as EmbedTextDetails from "./embed-text-details";
23
+ export import EmbedTextDetails = EmbedTextDetails.EmbedTextDetails;
24
+ import * as EmbedTextResult from "./embed-text-result";
25
+ export import EmbedTextResult = EmbedTextResult.EmbedTextResult;
26
+ import * as GenerateTextDetails from "./generate-text-details";
27
+ export import GenerateTextDetails = GenerateTextDetails.GenerateTextDetails;
28
+ import * as GenerateTextResult from "./generate-text-result";
29
+ export import GenerateTextResult = GenerateTextResult.GenerateTextResult;
30
+ import * as GeneratedText from "./generated-text";
31
+ export import GeneratedText = GeneratedText.GeneratedText;
32
+ import * as LlmInferenceRequest from "./llm-inference-request";
33
+ export import LlmInferenceRequest = LlmInferenceRequest.LlmInferenceRequest;
34
+ import * as LlmInferenceResponse from "./llm-inference-response";
35
+ export import LlmInferenceResponse = LlmInferenceResponse.LlmInferenceResponse;
36
+ import * as Logprobs from "./logprobs";
37
+ export import Logprobs = Logprobs.Logprobs;
38
+ import * as ServingMode from "./serving-mode";
39
+ export import ServingMode = ServingMode.ServingMode;
40
+ import * as SummarizeTextDetails from "./summarize-text-details";
41
+ export import SummarizeTextDetails = SummarizeTextDetails.SummarizeTextDetails;
42
+ import * as SummarizeTextResult from "./summarize-text-result";
43
+ export import SummarizeTextResult = SummarizeTextResult.SummarizeTextResult;
44
+ import * as TokenLikelihood from "./token-likelihood";
45
+ export import TokenLikelihood = TokenLikelihood.TokenLikelihood;
46
+ import * as CohereLlmInferenceRequest from "./cohere-llm-inference-request";
47
+ export import CohereLlmInferenceRequest = CohereLlmInferenceRequest.CohereLlmInferenceRequest;
48
+ import * as CohereLlmInferenceResponse from "./cohere-llm-inference-response";
49
+ export import CohereLlmInferenceResponse = CohereLlmInferenceResponse.CohereLlmInferenceResponse;
50
+ import * as DedicatedServingMode from "./dedicated-serving-mode";
51
+ export import DedicatedServingMode = DedicatedServingMode.DedicatedServingMode;
52
+ import * as LlamaLlmInferenceRequest from "./llama-llm-inference-request";
53
+ export import LlamaLlmInferenceRequest = LlamaLlmInferenceRequest.LlamaLlmInferenceRequest;
54
+ import * as LlamaLlmInferenceResponse from "./llama-llm-inference-response";
55
+ export import LlamaLlmInferenceResponse = LlamaLlmInferenceResponse.LlamaLlmInferenceResponse;
56
+ import * as OnDemandServingMode from "./on-demand-serving-mode";
57
+ export import OnDemandServingMode = OnDemandServingMode.OnDemandServingMode;
@@ -0,0 +1,80 @@
1
+ "use strict";
2
+ /**
3
+ * Generative AI Service Inference API
4
+ * OCI Generative AI is a fully managed service that provides a set of state-of-the-art, customizable large language models (LLMs) that cover a wide range of use cases for text generation, summarization, and text embeddings.
5
+
6
+ Use the Generative AI service inference API to access your custom model endpoints, or to try the out-of-the-box models to [generate text](#/en/generative-ai-inference/latest/GenerateTextResult/GenerateText), [summarize](#/en/generative-ai-inference/latest/SummarizeTextResult/SummarizeText), and [create text embeddings](#/en/generative-ai-inference/latest/EmbedTextResult/EmbedText).
7
+
8
+ To use a Generative AI custom model for inference, you must first create an endpoint for that model. Use the [Generative AI service management API](/#/en/generative-ai/latest/) to [create a custom model](#/en/generative-ai/latest/Model/) by fine-tuning an out-of-the-box model, or a previous version of a custom model, using your own data. Fine-tune the custom model on a [fine-tuning dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/). Then, create a [hosting dedicated AI cluster](#/en/generative-ai/latest/DedicatedAiCluster/) with an [endpoint](#/en/generative-ai/latest/Endpoint/) to host your custom model. For resource management in the Generative AI service, use the [Generative AI service management API](/#/en/generative-ai/latest/).
9
+
10
+ To learn more about the service, see the [Generative AI documentation](/iaas/Content/generative-ai/home.htm).
11
+
12
+ * OpenAPI spec version: 20231130
13
+ *
14
+ *
15
+ * NOTE: This class is auto generated by OracleSDKGenerator.
16
+ * Do not edit the class manually.
17
+ *
18
+ * Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
19
+ * This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
20
+ */
21
+ var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
22
+ if (k2 === undefined) k2 = k;
23
+ Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
24
+ }) : (function(o, m, k, k2) {
25
+ if (k2 === undefined) k2 = k;
26
+ o[k2] = m[k];
27
+ }));
28
+ var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
29
+ Object.defineProperty(o, "default", { enumerable: true, value: v });
30
+ }) : function(o, v) {
31
+ o["default"] = v;
32
+ });
33
+ var __importStar = (this && this.__importStar) || function (mod) {
34
+ if (mod && mod.__esModule) return mod;
35
+ var result = {};
36
+ if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
37
+ __setModuleDefault(result, mod);
38
+ return result;
39
+ };
40
+ Object.defineProperty(exports, "__esModule", { value: true });
41
+ exports.OnDemandServingMode = exports.LlamaLlmInferenceResponse = exports.LlamaLlmInferenceRequest = exports.DedicatedServingMode = exports.CohereLlmInferenceResponse = exports.CohereLlmInferenceRequest = exports.TokenLikelihood = exports.SummarizeTextResult = exports.SummarizeTextDetails = exports.ServingMode = exports.Logprobs = exports.LlmInferenceResponse = exports.LlmInferenceRequest = exports.GeneratedText = exports.GenerateTextResult = exports.GenerateTextDetails = exports.EmbedTextResult = exports.EmbedTextDetails = exports.Choice = void 0;
42
+ const Choice = __importStar(require("./choice"));
43
+ exports.Choice = Choice.Choice;
44
+ const EmbedTextDetails = __importStar(require("./embed-text-details"));
45
+ exports.EmbedTextDetails = EmbedTextDetails.EmbedTextDetails;
46
+ const EmbedTextResult = __importStar(require("./embed-text-result"));
47
+ exports.EmbedTextResult = EmbedTextResult.EmbedTextResult;
48
+ const GenerateTextDetails = __importStar(require("./generate-text-details"));
49
+ exports.GenerateTextDetails = GenerateTextDetails.GenerateTextDetails;
50
+ const GenerateTextResult = __importStar(require("./generate-text-result"));
51
+ exports.GenerateTextResult = GenerateTextResult.GenerateTextResult;
52
+ const GeneratedText = __importStar(require("./generated-text"));
53
+ exports.GeneratedText = GeneratedText.GeneratedText;
54
+ const LlmInferenceRequest = __importStar(require("./llm-inference-request"));
55
+ exports.LlmInferenceRequest = LlmInferenceRequest.LlmInferenceRequest;
56
+ const LlmInferenceResponse = __importStar(require("./llm-inference-response"));
57
+ exports.LlmInferenceResponse = LlmInferenceResponse.LlmInferenceResponse;
58
+ const Logprobs = __importStar(require("./logprobs"));
59
+ exports.Logprobs = Logprobs.Logprobs;
60
+ const ServingMode = __importStar(require("./serving-mode"));
61
+ exports.ServingMode = ServingMode.ServingMode;
62
+ const SummarizeTextDetails = __importStar(require("./summarize-text-details"));
63
+ exports.SummarizeTextDetails = SummarizeTextDetails.SummarizeTextDetails;
64
+ const SummarizeTextResult = __importStar(require("./summarize-text-result"));
65
+ exports.SummarizeTextResult = SummarizeTextResult.SummarizeTextResult;
66
+ const TokenLikelihood = __importStar(require("./token-likelihood"));
67
+ exports.TokenLikelihood = TokenLikelihood.TokenLikelihood;
68
+ const CohereLlmInferenceRequest = __importStar(require("./cohere-llm-inference-request"));
69
+ exports.CohereLlmInferenceRequest = CohereLlmInferenceRequest.CohereLlmInferenceRequest;
70
+ const CohereLlmInferenceResponse = __importStar(require("./cohere-llm-inference-response"));
71
+ exports.CohereLlmInferenceResponse = CohereLlmInferenceResponse.CohereLlmInferenceResponse;
72
+ const DedicatedServingMode = __importStar(require("./dedicated-serving-mode"));
73
+ exports.DedicatedServingMode = DedicatedServingMode.DedicatedServingMode;
74
+ const LlamaLlmInferenceRequest = __importStar(require("./llama-llm-inference-request"));
75
+ exports.LlamaLlmInferenceRequest = LlamaLlmInferenceRequest.LlamaLlmInferenceRequest;
76
+ const LlamaLlmInferenceResponse = __importStar(require("./llama-llm-inference-response"));
77
+ exports.LlamaLlmInferenceResponse = LlamaLlmInferenceResponse.LlamaLlmInferenceResponse;
78
+ const OnDemandServingMode = __importStar(require("./on-demand-serving-mode"));
79
+ exports.OnDemandServingMode = OnDemandServingMode.OnDemandServingMode;
80
+ //# sourceMappingURL=index.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"index.js","sourceRoot":"","sources":["../../../../../lib/generativeaiinference/lib/model/index.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;GAkBG;;;;;;;;;;;;;;;;;;;;;;AAEH,iDAAmC;AACrB,QAAA,MAAM,GAAG,MAAM,CAAC,MAAM,CAAC;AACrC,uEAAyD;AAC3C,QAAA,gBAAgB,GAAG,gBAAgB,CAAC,gBAAgB,CAAC;AACnE,qEAAuD;AACzC,QAAA,eAAe,GAAG,eAAe,CAAC,eAAe,CAAC;AAChE,6EAA+D;AACjD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,2EAA6D;AAC/C,QAAA,kBAAkB,GAAG,kBAAkB,CAAC,kBAAkB,CAAC;AACzE,gEAAkD;AACpC,QAAA,aAAa,GAAG,aAAa,CAAC,aAAa,CAAC;AAC1D,6EAA+D;AACjD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,+EAAiE;AACnD,QAAA,oBAAoB,GAAG,oBAAoB,CAAC,oBAAoB,CAAC;AAC/E,qDAAuC;AACzB,QAAA,QAAQ,GAAG,QAAQ,CAAC,QAAQ,CAAC;AAC3C,4DAA8C;AAChC,QAAA,WAAW,GAAG,WAAW,CAAC,WAAW,CAAC;AACpD,+EAAiE;AACnD,QAAA,oBAAoB,GAAG,oBAAoB,CAAC,oBAAoB,CAAC;AAC/E,6EAA+D;AACjD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC;AAC5E,oEAAsD;AACxC,QAAA,eAAe,GAAG,eAAe,CAAC,eAAe,CAAC;AAEhE,0FAA4E;AAC9D,QAAA,yBAAyB,GAAG,yBAAyB,CAAC,yBAAyB,CAAC;AAC9F,4FAA8E;AAChE,QAAA,0BAA0B,GAAG,0BAA0B,CAAC,0BAA0B,CAAC;AACjG,+EAAiE;AACnD,QAAA,oBAAoB,GAAG,oBAAoB,CAAC,oBAAoB,CAAC;AAC/E,wFAA0E;AAC5D,QAAA,wBAAwB,GAAG,wBAAwB,CAAC,wBAAwB,CAAC;AAC3F,0FAA4E;AAC9D,QAAA,yBAAyB,GAAG,yBAAyB,CAAC,yBAAyB,CAAC;AAC9F,8EAAgE;AAClD,QAAA,mBAAmB,GAAG,mBAAmB,CAAC,mBAAmB,CAAC"}