vellum_ai 1.14.1 → 1.14.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/lib/requests.rb +2 -2
- data/lib/types_export.rb +1 -0
- data/lib/vellum_ai/types/integration_name.rb +2 -0
- data/lib/vellum_ai/types/ml_model_hosting_interface.rb +68 -0
- data/lib/vellum_ai/types/ml_model_read.rb +70 -2
- data/lib/vellum_ai/types/prompt_deployment_release_prompt_version.rb +9 -2
- data/lib/vellum_ai/workflow_deployments/client.rb +14 -2
- metadata +3 -2
checksums.yaml
CHANGED
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
---
|
|
2
2
|
SHA256:
|
|
3
|
-
metadata.gz:
|
|
4
|
-
data.tar.gz:
|
|
3
|
+
metadata.gz: 4b8a8f454fee1f291795592d7269917b93878d313195d21ef6f3c7ed218f5bd9
|
|
4
|
+
data.tar.gz: b7f45a55cd2e7ba92886f8adb40189f5bfec81fc18ed2e653e99a842b2120b0c
|
|
5
5
|
SHA512:
|
|
6
|
-
metadata.gz:
|
|
7
|
-
data.tar.gz:
|
|
6
|
+
metadata.gz: 99bef05a234be14be4ca0caef253f5445cd4e58dc1914be4a7d198ca7cdbaa50d9d9a093faefe310c226517602d2db01eaa46197523febc253dbaa6b9d152bba
|
|
7
|
+
data.tar.gz: 9abbeda556fc5ca31bb6e1685449d33c6bdf14aed143a30b84830490f00547d160d7693e305dab1194894fcb13f17401317e17a69feaab577694fa501a25661d
|
data/lib/requests.rb
CHANGED
|
@@ -56,7 +56,7 @@ end
|
|
|
56
56
|
end
|
|
57
57
|
# @return [Hash{String => String}]
|
|
58
58
|
def get_headers
|
|
59
|
-
headers = { "X-Fern-Language": 'Ruby', "X-Fern-SDK-Name": 'vellum_ai', "X-Fern-SDK-Version": '1.14.
|
|
59
|
+
headers = { "X-Fern-Language": 'Ruby', "X-Fern-SDK-Name": 'vellum_ai', "X-Fern-SDK-Version": '1.14.2' }
|
|
60
60
|
headers["X-API-KEY"] = ((@api_key.is_a? Method) ? @api_key.call : @api_key) unless @api_key.nil?
|
|
61
61
|
headers
|
|
62
62
|
end
|
|
@@ -107,7 +107,7 @@ end
|
|
|
107
107
|
end
|
|
108
108
|
# @return [Hash{String => String}]
|
|
109
109
|
def get_headers
|
|
110
|
-
headers = { "X-Fern-Language": 'Ruby', "X-Fern-SDK-Name": 'vellum_ai', "X-Fern-SDK-Version": '1.14.
|
|
110
|
+
headers = { "X-Fern-Language": 'Ruby', "X-Fern-SDK-Name": 'vellum_ai', "X-Fern-SDK-Version": '1.14.2' }
|
|
111
111
|
headers["X-API-KEY"] = ((@api_key.is_a? Method) ? @api_key.call : @api_key) unless @api_key.nil?
|
|
112
112
|
headers
|
|
113
113
|
end
|
data/lib/types_export.rb
CHANGED
|
@@ -294,6 +294,7 @@ require_relative "vellum_ai/types/json_vellum_value"
|
|
|
294
294
|
require_relative "vellum_ai/types/json_vellum_value_request"
|
|
295
295
|
require_relative "vellum_ai/types/logical_operator"
|
|
296
296
|
require_relative "vellum_ai/types/logprobs_enum"
|
|
297
|
+
require_relative "vellum_ai/types/ml_model_hosting_interface"
|
|
297
298
|
require_relative "vellum_ai/types/ml_model_read"
|
|
298
299
|
require_relative "vellum_ai/types/ml_model_usage"
|
|
299
300
|
require_relative "vellum_ai/types/ml_model_usage_wrapper"
|
|
@@ -22,6 +22,7 @@ module Vellum
|
|
|
22
22
|
# * `MONDAY` - Monday
|
|
23
23
|
# * `NEON` - Neon
|
|
24
24
|
# * `OUTLOOK` - Outlook
|
|
25
|
+
# * `MICROSOFT_TEAMS` - Microsoft Teams
|
|
25
26
|
# * `GITHUB` - Github
|
|
26
27
|
# * `GOOGLE_SHEETS` - Google Sheets
|
|
27
28
|
# * `GOOGLE_CALENDAR` - Google Calendar
|
|
@@ -131,6 +132,7 @@ module Vellum
|
|
|
131
132
|
MONDAY = "MONDAY"
|
|
132
133
|
NEON = "NEON"
|
|
133
134
|
OUTLOOK = "OUTLOOK"
|
|
135
|
+
MICROSOFT_TEAMS = "MICROSOFT_TEAMS"
|
|
134
136
|
GITHUB = "GITHUB"
|
|
135
137
|
GOOGLE_SHEETS = "GOOGLE_SHEETS"
|
|
136
138
|
GOOGLE_CALENDAR = "GOOGLE_CALENDAR"
|
|
@@ -0,0 +1,68 @@
|
|
|
1
|
+
# frozen_string_literal: true
|
|
2
|
+
|
|
3
|
+
module Vellum
|
|
4
|
+
# * `ANTHROPIC` - Anthropic
|
|
5
|
+
# * `AWS_BEDROCK` - AWS Bedrock
|
|
6
|
+
# * `AZURE_AI_FOUNDRY` - Azure AI Foundry
|
|
7
|
+
# * `AZURE_OPENAI` - Azure OpenAI
|
|
8
|
+
# * `BASETEN` - BaseTen
|
|
9
|
+
# * `CEREBRAS` - Cerebras
|
|
10
|
+
# * `COHERE` - Cohere
|
|
11
|
+
# * `CUSTOM` - Custom
|
|
12
|
+
# * `DEEP_SEEK` - DeepSeek
|
|
13
|
+
# * `FIREWORKS_AI` - Fireworks AI
|
|
14
|
+
# * `GOOGLE` - Google
|
|
15
|
+
# * `GOOGLE_VERTEX_AI` - Google Vertex AI
|
|
16
|
+
# * `GROQ` - Groq
|
|
17
|
+
# * `HUGGINGFACE` - HuggingFace
|
|
18
|
+
# * `IBM_WATSONX` - IBM WatsonX
|
|
19
|
+
# * `MISTRAL_AI` - Mistral AI
|
|
20
|
+
# * `MOSAICML` - MosaicML
|
|
21
|
+
# * `MYSTIC` - Mystic
|
|
22
|
+
# * `NVIDIA` - NVIDIA
|
|
23
|
+
# * `OPENAI` - OpenAI
|
|
24
|
+
# * `OPEN_ROUTER` - Open Router
|
|
25
|
+
# * `OPENPIPE` - OpenPipe
|
|
26
|
+
# * `PERPLEXITY` - Perplexity
|
|
27
|
+
# * `PYQ` - Pyq
|
|
28
|
+
# * `REPLICATE` - Replicate
|
|
29
|
+
# * `SAMBANOVA` - SambaNova
|
|
30
|
+
# * `TOGETHER_AI` - Together AI
|
|
31
|
+
# * `X_AI` - xAI
|
|
32
|
+
# * `FASTWEB` - Fastweb
|
|
33
|
+
# * `SWISSCOM` - Swisscom
|
|
34
|
+
class MlModelHostingInterface
|
|
35
|
+
|
|
36
|
+
ANTHROPIC = "ANTHROPIC"
|
|
37
|
+
AWS_BEDROCK = "AWS_BEDROCK"
|
|
38
|
+
AZURE_AI_FOUNDRY = "AZURE_AI_FOUNDRY"
|
|
39
|
+
AZURE_OPENAI = "AZURE_OPENAI"
|
|
40
|
+
BASETEN = "BASETEN"
|
|
41
|
+
CEREBRAS = "CEREBRAS"
|
|
42
|
+
COHERE = "COHERE"
|
|
43
|
+
CUSTOM = "CUSTOM"
|
|
44
|
+
DEEP_SEEK = "DEEP_SEEK"
|
|
45
|
+
FIREWORKS_AI = "FIREWORKS_AI"
|
|
46
|
+
GOOGLE = "GOOGLE"
|
|
47
|
+
GOOGLE_VERTEX_AI = "GOOGLE_VERTEX_AI"
|
|
48
|
+
GROQ = "GROQ"
|
|
49
|
+
HUGGINGFACE = "HUGGINGFACE"
|
|
50
|
+
IBM_WATSONX = "IBM_WATSONX"
|
|
51
|
+
MISTRAL_AI = "MISTRAL_AI"
|
|
52
|
+
MOSAICML = "MOSAICML"
|
|
53
|
+
MYSTIC = "MYSTIC"
|
|
54
|
+
NVIDIA = "NVIDIA"
|
|
55
|
+
OPENAI = "OPENAI"
|
|
56
|
+
OPEN_ROUTER = "OPEN_ROUTER"
|
|
57
|
+
OPENPIPE = "OPENPIPE"
|
|
58
|
+
PERPLEXITY = "PERPLEXITY"
|
|
59
|
+
PYQ = "PYQ"
|
|
60
|
+
REPLICATE = "REPLICATE"
|
|
61
|
+
SAMBANOVA = "SAMBANOVA"
|
|
62
|
+
TOGETHER_AI = "TOGETHER_AI"
|
|
63
|
+
X_AI = "X_AI"
|
|
64
|
+
FASTWEB = "FASTWEB"
|
|
65
|
+
SWISSCOM = "SWISSCOM"
|
|
66
|
+
|
|
67
|
+
end
|
|
68
|
+
end
|
|
@@ -1,4 +1,5 @@
|
|
|
1
1
|
# frozen_string_literal: true
|
|
2
|
+
require_relative "ml_model_hosting_interface"
|
|
2
3
|
require "date"
|
|
3
4
|
require "ostruct"
|
|
4
5
|
require "json"
|
|
@@ -10,6 +11,38 @@ module Vellum
|
|
|
10
11
|
attr_reader :id
|
|
11
12
|
# @return [String] The unique name of the ML Model.
|
|
12
13
|
attr_reader :name
|
|
14
|
+
# @return [Vellum::MlModelHostingInterface] The organization hosting the ML Model.
|
|
15
|
+
# * `ANTHROPIC` - Anthropic
|
|
16
|
+
# * `AWS_BEDROCK` - AWS Bedrock
|
|
17
|
+
# * `AZURE_AI_FOUNDRY` - Azure AI Foundry
|
|
18
|
+
# * `AZURE_OPENAI` - Azure OpenAI
|
|
19
|
+
# * `BASETEN` - BaseTen
|
|
20
|
+
# * `CEREBRAS` - Cerebras
|
|
21
|
+
# * `COHERE` - Cohere
|
|
22
|
+
# * `CUSTOM` - Custom
|
|
23
|
+
# * `DEEP_SEEK` - DeepSeek
|
|
24
|
+
# * `FIREWORKS_AI` - Fireworks AI
|
|
25
|
+
# * `GOOGLE` - Google
|
|
26
|
+
# * `GOOGLE_VERTEX_AI` - Google Vertex AI
|
|
27
|
+
# * `GROQ` - Groq
|
|
28
|
+
# * `HUGGINGFACE` - HuggingFace
|
|
29
|
+
# * `IBM_WATSONX` - IBM WatsonX
|
|
30
|
+
# * `MISTRAL_AI` - Mistral AI
|
|
31
|
+
# * `MOSAICML` - MosaicML
|
|
32
|
+
# * `MYSTIC` - Mystic
|
|
33
|
+
# * `NVIDIA` - NVIDIA
|
|
34
|
+
# * `OPENAI` - OpenAI
|
|
35
|
+
# * `OPEN_ROUTER` - Open Router
|
|
36
|
+
# * `OPENPIPE` - OpenPipe
|
|
37
|
+
# * `PERPLEXITY` - Perplexity
|
|
38
|
+
# * `PYQ` - Pyq
|
|
39
|
+
# * `REPLICATE` - Replicate
|
|
40
|
+
# * `SAMBANOVA` - SambaNova
|
|
41
|
+
# * `TOGETHER_AI` - Together AI
|
|
42
|
+
# * `X_AI` - xAI
|
|
43
|
+
# * `FASTWEB` - Fastweb
|
|
44
|
+
# * `SWISSCOM` - Swisscom
|
|
45
|
+
attr_reader :hosted_by
|
|
13
46
|
# @return [DateTime]
|
|
14
47
|
attr_reader :introduced_on
|
|
15
48
|
# @return [OpenStruct] Additional properties unmapped to the current class definition
|
|
@@ -22,15 +55,47 @@ module Vellum
|
|
|
22
55
|
|
|
23
56
|
# @param id [String]
|
|
24
57
|
# @param name [String] The unique name of the ML Model.
|
|
58
|
+
# @param hosted_by [Vellum::MlModelHostingInterface] The organization hosting the ML Model.
|
|
59
|
+
# * `ANTHROPIC` - Anthropic
|
|
60
|
+
# * `AWS_BEDROCK` - AWS Bedrock
|
|
61
|
+
# * `AZURE_AI_FOUNDRY` - Azure AI Foundry
|
|
62
|
+
# * `AZURE_OPENAI` - Azure OpenAI
|
|
63
|
+
# * `BASETEN` - BaseTen
|
|
64
|
+
# * `CEREBRAS` - Cerebras
|
|
65
|
+
# * `COHERE` - Cohere
|
|
66
|
+
# * `CUSTOM` - Custom
|
|
67
|
+
# * `DEEP_SEEK` - DeepSeek
|
|
68
|
+
# * `FIREWORKS_AI` - Fireworks AI
|
|
69
|
+
# * `GOOGLE` - Google
|
|
70
|
+
# * `GOOGLE_VERTEX_AI` - Google Vertex AI
|
|
71
|
+
# * `GROQ` - Groq
|
|
72
|
+
# * `HUGGINGFACE` - HuggingFace
|
|
73
|
+
# * `IBM_WATSONX` - IBM WatsonX
|
|
74
|
+
# * `MISTRAL_AI` - Mistral AI
|
|
75
|
+
# * `MOSAICML` - MosaicML
|
|
76
|
+
# * `MYSTIC` - Mystic
|
|
77
|
+
# * `NVIDIA` - NVIDIA
|
|
78
|
+
# * `OPENAI` - OpenAI
|
|
79
|
+
# * `OPEN_ROUTER` - Open Router
|
|
80
|
+
# * `OPENPIPE` - OpenPipe
|
|
81
|
+
# * `PERPLEXITY` - Perplexity
|
|
82
|
+
# * `PYQ` - Pyq
|
|
83
|
+
# * `REPLICATE` - Replicate
|
|
84
|
+
# * `SAMBANOVA` - SambaNova
|
|
85
|
+
# * `TOGETHER_AI` - Together AI
|
|
86
|
+
# * `X_AI` - xAI
|
|
87
|
+
# * `FASTWEB` - Fastweb
|
|
88
|
+
# * `SWISSCOM` - Swisscom
|
|
25
89
|
# @param introduced_on [DateTime]
|
|
26
90
|
# @param additional_properties [OpenStruct] Additional properties unmapped to the current class definition
|
|
27
91
|
# @return [Vellum::MlModelRead]
|
|
28
|
-
def initialize(id:, name:, introduced_on:, additional_properties: nil)
|
|
92
|
+
def initialize(id:, name:, hosted_by:, introduced_on:, additional_properties: nil)
|
|
29
93
|
@id = id
|
|
30
94
|
@name = name
|
|
95
|
+
@hosted_by = hosted_by
|
|
31
96
|
@introduced_on = introduced_on
|
|
32
97
|
@additional_properties = additional_properties
|
|
33
|
-
@_field_set = { "id": id, "name": name, "introduced_on": introduced_on }
|
|
98
|
+
@_field_set = { "id": id, "name": name, "hosted_by": hosted_by, "introduced_on": introduced_on }
|
|
34
99
|
end
|
|
35
100
|
# Deserialize a JSON object to an instance of MlModelRead
|
|
36
101
|
#
|
|
@@ -41,6 +106,7 @@ module Vellum
|
|
|
41
106
|
parsed_json = JSON.parse(json_object)
|
|
42
107
|
id = parsed_json["id"]
|
|
43
108
|
name = parsed_json["name"]
|
|
109
|
+
hosted_by = parsed_json["hosted_by"]
|
|
44
110
|
introduced_on = unless parsed_json["introduced_on"].nil?
|
|
45
111
|
DateTime.parse(parsed_json["introduced_on"])
|
|
46
112
|
else
|
|
@@ -49,6 +115,7 @@ end
|
|
|
49
115
|
new(
|
|
50
116
|
id: id,
|
|
51
117
|
name: name,
|
|
118
|
+
hosted_by: hosted_by,
|
|
52
119
|
introduced_on: introduced_on,
|
|
53
120
|
additional_properties: struct
|
|
54
121
|
)
|
|
@@ -68,6 +135,7 @@ end
|
|
|
68
135
|
def self.validate_raw(obj:)
|
|
69
136
|
obj.id.is_a?(String) != false || raise("Passed value for field obj.id is not the expected type, validation failed.")
|
|
70
137
|
obj.name.is_a?(String) != false || raise("Passed value for field obj.name is not the expected type, validation failed.")
|
|
138
|
+
obj.hosted_by.is_a?(Vellum::MlModelHostingInterface) != false || raise("Passed value for field obj.hosted_by is not the expected type, validation failed.")
|
|
71
139
|
obj.introduced_on.is_a?(DateTime) != false || raise("Passed value for field obj.introduced_on is not the expected type, validation failed.")
|
|
72
140
|
end
|
|
73
141
|
end
|
|
@@ -8,6 +8,8 @@ module Vellum
|
|
|
8
8
|
class PromptDeploymentReleasePromptVersion
|
|
9
9
|
# @return [String]
|
|
10
10
|
attr_reader :id
|
|
11
|
+
# @return [String]
|
|
12
|
+
attr_reader :ml_model_to_workspace_id
|
|
11
13
|
# @return [Vellum::COMPONENTS_SCHEMAS_PROMPT_VERSION_BUILD_CONFIG_SANDBOX] Configuration used to build this prompt version.
|
|
12
14
|
attr_reader :build_config
|
|
13
15
|
# @return [OpenStruct] Additional properties unmapped to the current class definition
|
|
@@ -19,14 +21,16 @@ module Vellum
|
|
|
19
21
|
OMIT = Object.new
|
|
20
22
|
|
|
21
23
|
# @param id [String]
|
|
24
|
+
# @param ml_model_to_workspace_id [String]
|
|
22
25
|
# @param build_config [Vellum::COMPONENTS_SCHEMAS_PROMPT_VERSION_BUILD_CONFIG_SANDBOX] Configuration used to build this prompt version.
|
|
23
26
|
# @param additional_properties [OpenStruct] Additional properties unmapped to the current class definition
|
|
24
27
|
# @return [Vellum::PromptDeploymentReleasePromptVersion]
|
|
25
|
-
def initialize(id:, build_config:, additional_properties: nil)
|
|
28
|
+
def initialize(id:, ml_model_to_workspace_id:, build_config:, additional_properties: nil)
|
|
26
29
|
@id = id
|
|
30
|
+
@ml_model_to_workspace_id = ml_model_to_workspace_id
|
|
27
31
|
@build_config = build_config
|
|
28
32
|
@additional_properties = additional_properties
|
|
29
|
-
@_field_set = { "id": id, "build_config": build_config }
|
|
33
|
+
@_field_set = { "id": id, "ml_model_to_workspace_id": ml_model_to_workspace_id, "build_config": build_config }
|
|
30
34
|
end
|
|
31
35
|
# Deserialize a JSON object to an instance of PromptDeploymentReleasePromptVersion
|
|
32
36
|
#
|
|
@@ -36,6 +40,7 @@ module Vellum
|
|
|
36
40
|
struct = JSON.parse(json_object, object_class: OpenStruct)
|
|
37
41
|
parsed_json = JSON.parse(json_object)
|
|
38
42
|
id = parsed_json["id"]
|
|
43
|
+
ml_model_to_workspace_id = parsed_json["ml_model_to_workspace_id"]
|
|
39
44
|
unless parsed_json["build_config"].nil?
|
|
40
45
|
build_config = parsed_json["build_config"].to_json
|
|
41
46
|
build_config = Vellum::PromptVersionBuildConfigSandbox.from_json(json_object: build_config)
|
|
@@ -44,6 +49,7 @@ module Vellum
|
|
|
44
49
|
end
|
|
45
50
|
new(
|
|
46
51
|
id: id,
|
|
52
|
+
ml_model_to_workspace_id: ml_model_to_workspace_id,
|
|
47
53
|
build_config: build_config,
|
|
48
54
|
additional_properties: struct
|
|
49
55
|
)
|
|
@@ -62,6 +68,7 @@ module Vellum
|
|
|
62
68
|
# @return [Void]
|
|
63
69
|
def self.validate_raw(obj:)
|
|
64
70
|
obj.id.is_a?(String) != false || raise("Passed value for field obj.id is not the expected type, validation failed.")
|
|
71
|
+
obj.ml_model_to_workspace_id.is_a?(String) != false || raise("Passed value for field obj.ml_model_to_workspace_id is not the expected type, validation failed.")
|
|
65
72
|
Vellum::PromptVersionBuildConfigSandbox.validate_raw(obj: obj.build_config)
|
|
66
73
|
end
|
|
67
74
|
end
|
|
@@ -40,7 +40,13 @@ module Vellum
|
|
|
40
40
|
# @param is_public [String] is_public
|
|
41
41
|
# @param limit [Integer] Number of results to return per page.
|
|
42
42
|
# @param offset [Integer] The initial index from which to return the results.
|
|
43
|
-
# @param ordering [String]
|
|
43
|
+
# @param ordering [String] Specifies the ordering of the returned Workflow Deployments. Valid values are:
|
|
44
|
+
# - `created` - Orders by creation time in ascending order.
|
|
45
|
+
# - `-created` - Orders by creation time in descending order. (default)
|
|
46
|
+
# - `name` - Orders by name in ascending order.
|
|
47
|
+
# - `-name` - Orders by name in descending order.
|
|
48
|
+
# Compound orderings are supported by separating fields with commas, e.g.,
|
|
49
|
+
# `-created,name`.
|
|
44
50
|
# @param owned_by [String] owned_by
|
|
45
51
|
# @param status [Vellum::WorkflowDeployments::WorkflowDeploymentsListRequestStatus] status
|
|
46
52
|
# @param request_options [Vellum::RequestOptions]
|
|
@@ -432,7 +438,13 @@ end
|
|
|
432
438
|
# @param is_public [String] is_public
|
|
433
439
|
# @param limit [Integer] Number of results to return per page.
|
|
434
440
|
# @param offset [Integer] The initial index from which to return the results.
|
|
435
|
-
# @param ordering [String]
|
|
441
|
+
# @param ordering [String] Specifies the ordering of the returned Workflow Deployments. Valid values are:
|
|
442
|
+
# - `created` - Orders by creation time in ascending order.
|
|
443
|
+
# - `-created` - Orders by creation time in descending order. (default)
|
|
444
|
+
# - `name` - Orders by name in ascending order.
|
|
445
|
+
# - `-name` - Orders by name in descending order.
|
|
446
|
+
# Compound orderings are supported by separating fields with commas, e.g.,
|
|
447
|
+
# `-created,name`.
|
|
436
448
|
# @param owned_by [String] owned_by
|
|
437
449
|
# @param status [Vellum::WorkflowDeployments::WorkflowDeploymentsListRequestStatus] status
|
|
438
450
|
# @param request_options [Vellum::RequestOptions]
|
metadata
CHANGED
|
@@ -1,14 +1,14 @@
|
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
|
2
2
|
name: vellum_ai
|
|
3
3
|
version: !ruby/object:Gem::Version
|
|
4
|
-
version: 1.14.
|
|
4
|
+
version: 1.14.2
|
|
5
5
|
platform: ruby
|
|
6
6
|
authors:
|
|
7
7
|
- Vellum
|
|
8
8
|
autorequire:
|
|
9
9
|
bindir: exe
|
|
10
10
|
cert_chain: []
|
|
11
|
-
date: 2026-
|
|
11
|
+
date: 2026-02-03 00:00:00.000000000 Z
|
|
12
12
|
dependencies:
|
|
13
13
|
- !ruby/object:Gem::Dependency
|
|
14
14
|
name: faraday
|
|
@@ -412,6 +412,7 @@ files:
|
|
|
412
412
|
- lib/vellum_ai/types/metric_definition_history_item.rb
|
|
413
413
|
- lib/vellum_ai/types/metric_definition_input.rb
|
|
414
414
|
- lib/vellum_ai/types/metric_node_result.rb
|
|
415
|
+
- lib/vellum_ai/types/ml_model_hosting_interface.rb
|
|
415
416
|
- lib/vellum_ai/types/ml_model_read.rb
|
|
416
417
|
- lib/vellum_ai/types/ml_model_usage.rb
|
|
417
418
|
- lib/vellum_ai/types/ml_model_usage_wrapper.rb
|