teradataml 20.0.0.3__py3-none-any.whl → 20.0.0.4__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of teradataml might be problematic. Click here for more details.
- teradataml/LICENSE-3RD-PARTY.pdf +0 -0
- teradataml/README.md +119 -0
- teradataml/_version.py +1 -1
- teradataml/analytics/analytic_function_executor.py +18 -6
- teradataml/analytics/byom/__init__.py +1 -1
- teradataml/analytics/sqle/__init__.py +4 -1
- teradataml/analytics/valib.py +18 -4
- teradataml/automl/__init__.py +51 -6
- teradataml/automl/data_preparation.py +56 -33
- teradataml/automl/data_transformation.py +58 -33
- teradataml/automl/feature_engineering.py +12 -5
- teradataml/automl/model_training.py +34 -13
- teradataml/common/__init__.py +1 -2
- teradataml/common/constants.py +64 -40
- teradataml/common/messagecodes.py +13 -3
- teradataml/common/messages.py +4 -1
- teradataml/common/sqlbundle.py +40 -10
- teradataml/common/utils.py +113 -39
- teradataml/common/warnings.py +11 -0
- teradataml/context/context.py +141 -17
- teradataml/data/amazon_reviews_25.csv +26 -0
- teradataml/data/byom_example.json +11 -0
- teradataml/data/docs/byom/docs/DataRobotPredict.py +2 -2
- teradataml/data/docs/byom/docs/DataikuPredict.py +40 -1
- teradataml/data/docs/byom/docs/H2OPredict.py +2 -2
- teradataml/data/docs/byom/docs/ONNXEmbeddings.py +242 -0
- teradataml/data/docs/byom/docs/ONNXPredict.py +2 -2
- teradataml/data/docs/byom/docs/PMMLPredict.py +2 -2
- teradataml/data/docs/sqle/docs_17_20/Shap.py +28 -6
- teradataml/data/docs/uaf/docs_17_20/DWT2D.py +4 -1
- teradataml/data/hnsw_alter_data.csv +5 -0
- teradataml/data/hnsw_data.csv +10 -0
- teradataml/data/jsons/byom/h2opredict.json +1 -1
- teradataml/data/jsons/byom/onnxembeddings.json +266 -0
- teradataml/data/jsons/sqle/17.20/TD_Shap.json +0 -1
- teradataml/data/jsons/sqle/20.00/TD_HNSW.json +296 -0
- teradataml/data/jsons/sqle/20.00/TD_HNSWPredict.json +206 -0
- teradataml/data/jsons/sqle/20.00/TD_HNSWSummary.json +32 -0
- teradataml/data/jsons/sqle/20.00/TD_KMeans.json +2 -2
- teradataml/data/jsons/sqle/20.00/TD_SMOTE.json +1 -1
- teradataml/data/jsons/sqle/20.00/TD_VectorDistance.json +5 -5
- teradataml/data/teradataml_example.json +8 -0
- teradataml/data/vectordistance_example.json +1 -1
- teradataml/dataframe/copy_to.py +8 -3
- teradataml/dataframe/data_transfer.py +11 -1
- teradataml/dataframe/dataframe.py +517 -121
- teradataml/dataframe/dataframe_utils.py +152 -20
- teradataml/dataframe/functions.py +26 -11
- teradataml/dataframe/setop.py +11 -6
- teradataml/dataframe/sql.py +2 -2
- teradataml/dbutils/dbutils.py +525 -129
- teradataml/hyperparameter_tuner/optimizer.py +12 -1
- teradataml/opensource/{sklearn/_sklearn_wrapper.py → _base.py} +317 -1011
- teradataml/opensource/_class.py +141 -17
- teradataml/opensource/{constants.py → _constants.py} +7 -3
- teradataml/opensource/_lightgbm.py +52 -53
- teradataml/opensource/_sklearn.py +1008 -0
- teradataml/opensource/_wrapper_utils.py +5 -5
- teradataml/options/__init__.py +47 -15
- teradataml/options/configure.py +103 -25
- teradataml/options/display.py +13 -2
- teradataml/plot/axis.py +47 -8
- teradataml/plot/figure.py +33 -0
- teradataml/plot/plot.py +63 -13
- teradataml/scriptmgmt/UserEnv.py +2 -2
- teradataml/scriptmgmt/lls_utils.py +63 -26
- teradataml/store/__init__.py +1 -2
- teradataml/store/feature_store/feature_store.py +102 -7
- teradataml/table_operators/Apply.py +32 -18
- teradataml/table_operators/Script.py +3 -1
- teradataml/table_operators/TableOperator.py +3 -1
- teradataml/utils/dtypes.py +47 -0
- teradataml/utils/internal_buffer.py +18 -0
- teradataml/utils/validators.py +68 -9
- {teradataml-20.0.0.3.dist-info → teradataml-20.0.0.4.dist-info}/METADATA +123 -2
- {teradataml-20.0.0.3.dist-info → teradataml-20.0.0.4.dist-info}/RECORD +79 -75
- teradataml/data/SQL_Fundamentals.pdf +0 -0
- teradataml/libaed_0_1.dylib +0 -0
- teradataml/libaed_0_1.so +0 -0
- teradataml/opensource/sklearn/__init__.py +0 -0
- teradataml/store/vector_store/__init__.py +0 -1586
- {teradataml-20.0.0.3.dist-info → teradataml-20.0.0.4.dist-info}/WHEEL +0 -0
- {teradataml-20.0.0.3.dist-info → teradataml-20.0.0.4.dist-info}/top_level.txt +0 -0
- {teradataml-20.0.0.3.dist-info → teradataml-20.0.0.4.dist-info}/zip-safe +0 -0
|
@@ -0,0 +1,266 @@
|
|
|
1
|
+
{
|
|
2
|
+
"json_schema_major_version": "1",
|
|
3
|
+
"json_schema_minor_version": "2",
|
|
4
|
+
"json_content_version": "1",
|
|
5
|
+
"function_name": "ONNXEmbeddings",
|
|
6
|
+
"function_version": "1.0",
|
|
7
|
+
"function_type": "byom",
|
|
8
|
+
"function_r_name": "aa.onnx.embeddings",
|
|
9
|
+
"function_alias_name": "ONNXEmbeddings",
|
|
10
|
+
"short_description": "This Function generates embeddings values using an ONNX model in Vantage",
|
|
11
|
+
"long_description": "This function is used to calculate embeddings values in Vantage with a HuggingFace model that has been created outside Vantage and exported to vantage using ONNX format",
|
|
12
|
+
"input_tables": [
|
|
13
|
+
{
|
|
14
|
+
"requiredInputKind": [
|
|
15
|
+
"PartitionByAny",
|
|
16
|
+
"PartitionByKey"
|
|
17
|
+
],
|
|
18
|
+
"isOrdered": false,
|
|
19
|
+
"partitionByOne": false,
|
|
20
|
+
"partitionByOneInclusive": false,
|
|
21
|
+
"name": "InputTable",
|
|
22
|
+
"alternateNames": [],
|
|
23
|
+
"isRequired": true,
|
|
24
|
+
"rDescription": "The input table that contains the text from which we generate embedding values ",
|
|
25
|
+
"description": "The input table that contains the text from which we generate embedding values ",
|
|
26
|
+
"datatype": "TABLE_ALIAS",
|
|
27
|
+
"allowsLists": false,
|
|
28
|
+
"rName": "newdata",
|
|
29
|
+
"useInR": true,
|
|
30
|
+
"rOrderNum": 1
|
|
31
|
+
},
|
|
32
|
+
{
|
|
33
|
+
"requiredInputKind": [
|
|
34
|
+
"Dimension"
|
|
35
|
+
],
|
|
36
|
+
"isOrdered": false,
|
|
37
|
+
"partitionByOne": false,
|
|
38
|
+
"partitionByOneInclusive": false,
|
|
39
|
+
"name": "ModelTable",
|
|
40
|
+
"alternateNames": [],
|
|
41
|
+
"isRequired": true,
|
|
42
|
+
"rDescription": "The model table to be used for calculating embedding values ",
|
|
43
|
+
"description": "The model table to be used for calculating embedding values ",
|
|
44
|
+
"datatype": "TABLE_ALIAS",
|
|
45
|
+
"allowsLists": false,
|
|
46
|
+
"rName": "modeldata",
|
|
47
|
+
"useInR": true,
|
|
48
|
+
"rOrderNum": 2
|
|
49
|
+
},
|
|
50
|
+
{
|
|
51
|
+
"requiredInputKind": [
|
|
52
|
+
"Dimension"
|
|
53
|
+
],
|
|
54
|
+
"isOrdered": false,
|
|
55
|
+
"partitionByOne": false,
|
|
56
|
+
"partitionByOneInclusive": false,
|
|
57
|
+
"name": "TokenizerTable",
|
|
58
|
+
"alternateNames": [],
|
|
59
|
+
"isRequired": true,
|
|
60
|
+
"rDescription": "The tokenizer table which contains the tokenizer json file ",
|
|
61
|
+
"description": "The tokenizer table which contains the tokenizer json file ",
|
|
62
|
+
"datatype": "TABLE_ALIAS",
|
|
63
|
+
"allowsLists": false,
|
|
64
|
+
"rName": "tokenizerdata",
|
|
65
|
+
"useInR": true,
|
|
66
|
+
"rOrderNum": 3
|
|
67
|
+
}
|
|
68
|
+
],
|
|
69
|
+
"argument_clauses": [
|
|
70
|
+
{
|
|
71
|
+
"targetTable": [
|
|
72
|
+
"InputTable"
|
|
73
|
+
],
|
|
74
|
+
"checkDuplicate": true,
|
|
75
|
+
"allowedTypes": [],
|
|
76
|
+
"allowedTypeGroups": [
|
|
77
|
+
"ALL"
|
|
78
|
+
],
|
|
79
|
+
"matchLengthOfArgument": "",
|
|
80
|
+
"allowPadding": true,
|
|
81
|
+
"name": "Accumulate",
|
|
82
|
+
"alternateNames": [],
|
|
83
|
+
"isRequired": true,
|
|
84
|
+
"rDescription": "Specifies the names of input_table columns to copy to the output table.",
|
|
85
|
+
"description": "Specify the names of the input columns to copy to the output table. ",
|
|
86
|
+
"datatype": "COLUMNS",
|
|
87
|
+
"allowsLists": true,
|
|
88
|
+
"rName": "accumulate",
|
|
89
|
+
"useInR": true,
|
|
90
|
+
"rOrderNum": 4
|
|
91
|
+
},
|
|
92
|
+
{
|
|
93
|
+
"checkDuplicate": true,
|
|
94
|
+
"allowedTypes": [],
|
|
95
|
+
"allowedTypeGroups": [
|
|
96
|
+
"ALL"
|
|
97
|
+
],
|
|
98
|
+
"matchLengthOfArgument": "",
|
|
99
|
+
"allowPadding": true,
|
|
100
|
+
"name": "ModelOutputTensor",
|
|
101
|
+
"alternateNames": [],
|
|
102
|
+
"isRequired": true,
|
|
103
|
+
"rDescription": "Specifies the column of the model's possible output fields that the user wants to calculate and output ",
|
|
104
|
+
"description": "Specifies the column of the model's possible output fields that the user wants to calculate and output ",
|
|
105
|
+
"datatype": "STRING",
|
|
106
|
+
"allowsLists": false,
|
|
107
|
+
"rName": "model.output.tensor",
|
|
108
|
+
"useInR": true,
|
|
109
|
+
"rOrderNum": 5
|
|
110
|
+
},
|
|
111
|
+
{
|
|
112
|
+
"defaultValue": 512,
|
|
113
|
+
"checkDuplicate": true,
|
|
114
|
+
"allowedTypes": [],
|
|
115
|
+
"allowedTypeGroups": [
|
|
116
|
+
"ALL"
|
|
117
|
+
],
|
|
118
|
+
"matchLengthOfArgument": "",
|
|
119
|
+
"allowPadding": true,
|
|
120
|
+
"name": "EncodeMaxLength",
|
|
121
|
+
"alternateNames": [],
|
|
122
|
+
"isRequired": false,
|
|
123
|
+
"rDescription": "Specifies the maximum length of the tokenizer output token encodings(only applies for models with symbolic dimensions) ",
|
|
124
|
+
"description": "Specifies the maximum length of the tokenizer output token encodings(only applies for models with symbolic dimensions) ",
|
|
125
|
+
"datatype": "INTEGER",
|
|
126
|
+
"allowsLists": false,
|
|
127
|
+
"rName": "encode.max.length",
|
|
128
|
+
"useInR": true,
|
|
129
|
+
"rOrderNum": 6
|
|
130
|
+
},
|
|
131
|
+
{
|
|
132
|
+
"defaultValue": false,
|
|
133
|
+
"checkDuplicate": true,
|
|
134
|
+
"allowedTypes": [],
|
|
135
|
+
"allowedTypeGroups": [
|
|
136
|
+
"ALL"
|
|
137
|
+
],
|
|
138
|
+
"matchLengthOfArgument": "",
|
|
139
|
+
"allowPadding": true,
|
|
140
|
+
"name": "ShowModelProperties",
|
|
141
|
+
"alternateNames": [],
|
|
142
|
+
"isRequired": false,
|
|
143
|
+
"rDescription": " Show default or expanded ModelInputFieldsMap based on input model for defaults or ModelInputFieldsMap for expansion. ",
|
|
144
|
+
"description": " Show default or expanded ModelInputFieldsMap based on input model for defaults or ModelInputFieldsMap for expansion. ",
|
|
145
|
+
"datatype": "BOOLEAN",
|
|
146
|
+
"allowsLists": false,
|
|
147
|
+
"rName": "show.model.properties",
|
|
148
|
+
"useInR": true,
|
|
149
|
+
"rOrderNum": 7
|
|
150
|
+
},
|
|
151
|
+
{
|
|
152
|
+
"defaultValue": "emb_",
|
|
153
|
+
"checkDuplicate": true,
|
|
154
|
+
"allowedTypes": [],
|
|
155
|
+
"allowedTypeGroups": [
|
|
156
|
+
"ALL"
|
|
157
|
+
],
|
|
158
|
+
"matchLengthOfArgument": "",
|
|
159
|
+
"allowPadding": true,
|
|
160
|
+
"name": "OutputColumnPrefix",
|
|
161
|
+
"alternateNames": [],
|
|
162
|
+
"isRequired": false,
|
|
163
|
+
"rDescription": "Specifies the column prefix for each of the output columns when using FLOAT32 OutputFormat ",
|
|
164
|
+
"description": "Specifies the column prefix for each of the output columns when using FLOAT32 OutputFormat ",
|
|
165
|
+
"datatype": "STRING",
|
|
166
|
+
"allowsLists": false,
|
|
167
|
+
"rName": "output.column.prefix",
|
|
168
|
+
"useInR": true,
|
|
169
|
+
"rOrderNum": 8
|
|
170
|
+
},
|
|
171
|
+
{
|
|
172
|
+
"defaultValue": "VARBYTE(3072)",
|
|
173
|
+
"checkDuplicate": true,
|
|
174
|
+
"allowedTypes": [],
|
|
175
|
+
"allowedTypeGroups": [
|
|
176
|
+
"ALL"
|
|
177
|
+
],
|
|
178
|
+
"matchLengthOfArgument": "",
|
|
179
|
+
"allowPadding": true,
|
|
180
|
+
"name": "OutputFormat",
|
|
181
|
+
"alternateNames": [],
|
|
182
|
+
"isRequired": false,
|
|
183
|
+
"rDescription": "Specifies the output format for the model embeddings output ",
|
|
184
|
+
"description": "Specifies the output format for the model embeddings output ",
|
|
185
|
+
"datatype": "STRING",
|
|
186
|
+
"allowsLists": false,
|
|
187
|
+
"rName": "output.format",
|
|
188
|
+
"useInR": true,
|
|
189
|
+
"rOrderNum": 9
|
|
190
|
+
},
|
|
191
|
+
{
|
|
192
|
+
"permittedValues": [
|
|
193
|
+
"true",
|
|
194
|
+
"t",
|
|
195
|
+
"yes",
|
|
196
|
+
"y",
|
|
197
|
+
"1",
|
|
198
|
+
"false",
|
|
199
|
+
"f",
|
|
200
|
+
"no",
|
|
201
|
+
"n",
|
|
202
|
+
"0",
|
|
203
|
+
"*",
|
|
204
|
+
"current_cached_model"
|
|
205
|
+
],
|
|
206
|
+
"defaultValue": "false",
|
|
207
|
+
"checkDuplicate": true,
|
|
208
|
+
"allowedTypes": [],
|
|
209
|
+
"allowedTypeGroups": [
|
|
210
|
+
"ALL"
|
|
211
|
+
],
|
|
212
|
+
"matchLengthOfArgument": "",
|
|
213
|
+
"allowPadding": true,
|
|
214
|
+
"name": "OverwriteCachedModel",
|
|
215
|
+
"alternateNames": [],
|
|
216
|
+
"isRequired": false,
|
|
217
|
+
"rDescription": "Specifies the model name that needs to be removed from the cache. * can also be used to remove the models ",
|
|
218
|
+
"description": " Specifies the model name that needs to be removed from the cache. * can also be used to remove the models ",
|
|
219
|
+
"datatype": "STRING",
|
|
220
|
+
"allowsLists": false,
|
|
221
|
+
"rName": "overwrite.cached.models",
|
|
222
|
+
"useInR": true,
|
|
223
|
+
"rOrderNum": 10
|
|
224
|
+
},
|
|
225
|
+
{
|
|
226
|
+
"defaultValue": false,
|
|
227
|
+
"checkDuplicate": true,
|
|
228
|
+
"allowedTypes": [],
|
|
229
|
+
"allowedTypeGroups": [
|
|
230
|
+
"ALL"
|
|
231
|
+
],
|
|
232
|
+
"matchLengthOfArgument": "",
|
|
233
|
+
"allowPadding": true,
|
|
234
|
+
"name": "IsDebug",
|
|
235
|
+
"alternateNames": [],
|
|
236
|
+
"isRequired": false,
|
|
237
|
+
"rDescription": "Print additional information in trace table regarding execution of ONNXPredict ",
|
|
238
|
+
"description": " Print additional information in trace table regarding execution of ONNXPredict ",
|
|
239
|
+
"datatype": "BOOLEAN",
|
|
240
|
+
"allowsLists": false,
|
|
241
|
+
"rName": "is.debug",
|
|
242
|
+
"useInR": true,
|
|
243
|
+
"rOrderNum": 11
|
|
244
|
+
},
|
|
245
|
+
{
|
|
246
|
+
"defaultValue": true,
|
|
247
|
+
"checkDuplicate": true,
|
|
248
|
+
"allowedTypes": [],
|
|
249
|
+
"allowedTypeGroups": [
|
|
250
|
+
"ALL"
|
|
251
|
+
],
|
|
252
|
+
"matchLengthOfArgument": "",
|
|
253
|
+
"allowPadding": true,
|
|
254
|
+
"name": "EnableMemoryCheck",
|
|
255
|
+
"alternateNames": [],
|
|
256
|
+
"isRequired": false,
|
|
257
|
+
"rDescription": "If true, verifies if there is enough native memory for large models ",
|
|
258
|
+
"description": "If true, verifies if there is enough native memory for large models ",
|
|
259
|
+
"datatype": "BOOLEAN",
|
|
260
|
+
"allowsLists": false,
|
|
261
|
+
"rName": "enable.memory.check",
|
|
262
|
+
"useInR": true,
|
|
263
|
+
"rOrderNum": 12
|
|
264
|
+
}
|
|
265
|
+
]
|
|
266
|
+
}
|
|
@@ -0,0 +1,296 @@
|
|
|
1
|
+
{
|
|
2
|
+
"json_schema_major_version": "1",
|
|
3
|
+
"json_schema_minor_version": "1",
|
|
4
|
+
"json_content_version": "1",
|
|
5
|
+
"function_name": "TD_HNSW",
|
|
6
|
+
"function_version": "1.0",
|
|
7
|
+
"function_type": "fastpath",
|
|
8
|
+
"function_category": "Model Training",
|
|
9
|
+
"function_alias_name": "TD_HNSW",
|
|
10
|
+
"function_r_name": "td.hnsw",
|
|
11
|
+
"short_description": "The TD_HNSW function generates the HNSW model using the input data points which is then used by TD_HNSWPredict function for determining the approximate nearest neighbors for any given input data point.",
|
|
12
|
+
"long_description": "The TD_HNSW function generates the HNSW model using the input data points which is then used by TD_HNSWPredict function for determining the approximate nearest neighbors for any given input data point.",
|
|
13
|
+
"input_tables": [
|
|
14
|
+
{
|
|
15
|
+
"requiredInputKind": [
|
|
16
|
+
"PartitionByAny",
|
|
17
|
+
"Dimension"
|
|
18
|
+
],
|
|
19
|
+
"isOrdered": false,
|
|
20
|
+
"partitionByOne": false,
|
|
21
|
+
"name": "InputTable",
|
|
22
|
+
"alternateNames": [],
|
|
23
|
+
"isRequired": true,
|
|
24
|
+
"rDescription": "The table that contains the input dataset for HNSW model training/update/delete operation.",
|
|
25
|
+
"description": "The table that contains the input dataset for HNSW model training/update/delete operation.",
|
|
26
|
+
"datatype": "TABLE_ALIAS",
|
|
27
|
+
"allowsLists": false,
|
|
28
|
+
"rName": "data",
|
|
29
|
+
"useInR": true,
|
|
30
|
+
"rOrderNum": 1
|
|
31
|
+
},
|
|
32
|
+
{
|
|
33
|
+
"requiredInputKind": [
|
|
34
|
+
"PartitionByAny"
|
|
35
|
+
],
|
|
36
|
+
"isOrdered": false,
|
|
37
|
+
"partitionByOne": false,
|
|
38
|
+
"name": "InputModelTable",
|
|
39
|
+
"alternateNames": [],
|
|
40
|
+
"isRequired": false,
|
|
41
|
+
"rDescription": "The table that contains the HNSW model for update/delete operation.",
|
|
42
|
+
"description": "The table that contains the HNSW model for update/delete operation.",
|
|
43
|
+
"datatype": "TABLE_ALIAS",
|
|
44
|
+
"allowsLists": false,
|
|
45
|
+
"rName": "object",
|
|
46
|
+
"useInR": true,
|
|
47
|
+
"rOrderNum": 2
|
|
48
|
+
}
|
|
49
|
+
],
|
|
50
|
+
"output_tables": [
|
|
51
|
+
{
|
|
52
|
+
"isOutputTable": true,
|
|
53
|
+
"omitPossible": false,
|
|
54
|
+
"name": "ModelTable",
|
|
55
|
+
"alternateNames": [],
|
|
56
|
+
"isRequired": true,
|
|
57
|
+
"rDescription": "Specifies the name of the table in which the generated HNSW model can be stored.",
|
|
58
|
+
"description": "Specifies the name of the table in which the generated HNSW model can be stored.",
|
|
59
|
+
"datatype": "TABLE_NAME",
|
|
60
|
+
"allowsLists": false,
|
|
61
|
+
"rName": "model.table",
|
|
62
|
+
"useInR": true,
|
|
63
|
+
"rOrderNum": 3
|
|
64
|
+
}
|
|
65
|
+
],
|
|
66
|
+
"argument_clauses": [
|
|
67
|
+
{
|
|
68
|
+
"targetTable": [
|
|
69
|
+
"InputTable"
|
|
70
|
+
],
|
|
71
|
+
"checkDuplicate": true,
|
|
72
|
+
"allowedTypes": [],
|
|
73
|
+
"allowedTypeGroups": [
|
|
74
|
+
"INTEGER"
|
|
75
|
+
],
|
|
76
|
+
"requiredLength": 1,
|
|
77
|
+
"matchLengthOfArgument": "",
|
|
78
|
+
"allowPadding": false,
|
|
79
|
+
"name": "IdColumn",
|
|
80
|
+
"alternateNames": [],
|
|
81
|
+
"isRequired": true,
|
|
82
|
+
"rDescription": "Specify the column name containing unique identifier of input rows.",
|
|
83
|
+
"description": "Specify the column name containing unique identifier of input rows.",
|
|
84
|
+
"datatype": "COLUMN_NAMES",
|
|
85
|
+
"allowsLists": false,
|
|
86
|
+
"rName": "id.column",
|
|
87
|
+
"useInR": true,
|
|
88
|
+
"rOrderNum": 4
|
|
89
|
+
},
|
|
90
|
+
{
|
|
91
|
+
"targetTable": [
|
|
92
|
+
"InputTable"
|
|
93
|
+
],
|
|
94
|
+
"checkDuplicate": true,
|
|
95
|
+
"allowedTypes": [],
|
|
96
|
+
"allowedTypeGroups": [
|
|
97
|
+
"VECTOR","BYTE","VARBYTE"
|
|
98
|
+
],
|
|
99
|
+
"requiredLength": 1,
|
|
100
|
+
"matchLengthOfArgument": "",
|
|
101
|
+
"allowPadding": false,
|
|
102
|
+
"name": "VectorColumn",
|
|
103
|
+
"alternateNames": [],
|
|
104
|
+
"isRequired": true,
|
|
105
|
+
"rDescription": "Specifies the column from the input table to be used for training the HNSW model.",
|
|
106
|
+
"description": "Specifies the column from the input table to be used for training the HNSW model.",
|
|
107
|
+
"datatype": "COLUMN_NAMES",
|
|
108
|
+
"allowsLists": false,
|
|
109
|
+
"rName": "vector.column",
|
|
110
|
+
"useInR": true,
|
|
111
|
+
"rOrderNum": 5
|
|
112
|
+
},
|
|
113
|
+
{
|
|
114
|
+
"lowerBound": 1,
|
|
115
|
+
"upperBound": 1024,
|
|
116
|
+
"lowerBoundType": "INCLUSIVE",
|
|
117
|
+
"upperBoundType": "INCLUSIVE",
|
|
118
|
+
"allowNaN": false,
|
|
119
|
+
"name": "NumLayer",
|
|
120
|
+
"alternateNames": [],
|
|
121
|
+
"isRequired": false,
|
|
122
|
+
"rDescription": "Specify the maximum number of layers for the HNSW model.",
|
|
123
|
+
"description": "Specify the maximum number of layers for the HNSW model.",
|
|
124
|
+
"datatype": "INTEGER",
|
|
125
|
+
"allowsLists": false,
|
|
126
|
+
"rName": "num.layer",
|
|
127
|
+
"useInR": true,
|
|
128
|
+
"rOrderNum": 6
|
|
129
|
+
},
|
|
130
|
+
{
|
|
131
|
+
"lowerBound": 0,
|
|
132
|
+
"upperBound": 2147483647,
|
|
133
|
+
"lowerBoundType": "INCLUSIVE",
|
|
134
|
+
"upperBoundType": "INCLUSIVE",
|
|
135
|
+
"allowNaN": false,
|
|
136
|
+
"name": "Seed",
|
|
137
|
+
"alternateNames": [],
|
|
138
|
+
"isRequired": false,
|
|
139
|
+
"rDescription": "Specify the random seed value for repeatable results.",
|
|
140
|
+
"description": "Specify the random seed value for repeatable results.",
|
|
141
|
+
"datatype": "INTEGER",
|
|
142
|
+
"allowsLists": false,
|
|
143
|
+
"rName": "seed",
|
|
144
|
+
"useInR": true,
|
|
145
|
+
"rOrderNum": 7
|
|
146
|
+
},
|
|
147
|
+
{
|
|
148
|
+
"defaultValue": 32,
|
|
149
|
+
"lowerBound": 1,
|
|
150
|
+
"upperBound": 1024,
|
|
151
|
+
"lowerBoundType": "INCLUSIVE",
|
|
152
|
+
"upperBoundType": "INCLUSIVE",
|
|
153
|
+
"allowNaN": false,
|
|
154
|
+
"name": "EfConstruction",
|
|
155
|
+
"alternateNames": [],
|
|
156
|
+
"isRequired": false,
|
|
157
|
+
"rDescription": "Specify the number of neighbors to search during training of HNSW model.",
|
|
158
|
+
"description": "Specify the number of neighbors to search during training of HNSW model.",
|
|
159
|
+
"datatype": "INTEGER",
|
|
160
|
+
"allowsLists": false,
|
|
161
|
+
"rName": "ef.construction",
|
|
162
|
+
"useInR": true,
|
|
163
|
+
"rOrderNum": 8
|
|
164
|
+
},
|
|
165
|
+
{
|
|
166
|
+
"defaultValue": 32,
|
|
167
|
+
"lowerBound": 1,
|
|
168
|
+
"upperBound": 1024,
|
|
169
|
+
"lowerBoundType": "INCLUSIVE",
|
|
170
|
+
"upperBoundType": "INCLUSIVE",
|
|
171
|
+
"allowNaN": false,
|
|
172
|
+
"name": "NumConnPerNode",
|
|
173
|
+
"alternateNames": [],
|
|
174
|
+
"isRequired": false,
|
|
175
|
+
"rDescription": "Specify number of connections for a node during training of HNSW model.",
|
|
176
|
+
"description": "Specify number of connections for a node during training of HNSW model.",
|
|
177
|
+
"datatype": "INTEGER",
|
|
178
|
+
"allowsLists": false,
|
|
179
|
+
"rName": "numconn.pernode",
|
|
180
|
+
"useInR": true,
|
|
181
|
+
"rOrderNum": 9
|
|
182
|
+
},
|
|
183
|
+
{
|
|
184
|
+
"defaultValue": 32,
|
|
185
|
+
"lowerBound": 1,
|
|
186
|
+
"upperBound": 1024,
|
|
187
|
+
"lowerBoundType": "INCLUSIVE",
|
|
188
|
+
"upperBoundType": "INCLUSIVE",
|
|
189
|
+
"allowNaN": false,
|
|
190
|
+
"name": "MaxNumConnPerNode",
|
|
191
|
+
"alternateNames": [],
|
|
192
|
+
"isRequired": false,
|
|
193
|
+
"rDescription": "Specify maximum number of connections allowed for a node during training of HNSW model.",
|
|
194
|
+
"description": "Specify maximum number of connections allowed for a node during training of HNSW model.",
|
|
195
|
+
"datatype": "INTEGER",
|
|
196
|
+
"allowsLists": false,
|
|
197
|
+
"rName": "maxnumconn.pernode",
|
|
198
|
+
"useInR": true,
|
|
199
|
+
"rOrderNum": 10
|
|
200
|
+
},
|
|
201
|
+
{
|
|
202
|
+
"permittedValues": [
|
|
203
|
+
"EUCLIDEAN",
|
|
204
|
+
"COSINE",
|
|
205
|
+
"DOTPRODUCT"
|
|
206
|
+
],
|
|
207
|
+
"defaultValue": "EUCLIDEAN",
|
|
208
|
+
"isOutputColumn": false,
|
|
209
|
+
"matchLengthOfArgument": "",
|
|
210
|
+
"allowPadding": false,
|
|
211
|
+
"name": "DistanceMeasure",
|
|
212
|
+
"alternateNames": [],
|
|
213
|
+
"isRequired": false,
|
|
214
|
+
"rDescription": "Specify the distance measure to be used for distance computation.",
|
|
215
|
+
"description": "Specify the distance measure to be used for distance computation.",
|
|
216
|
+
"datatype": "STRING",
|
|
217
|
+
"allowsLists": false,
|
|
218
|
+
"rName": "distance.measure",
|
|
219
|
+
"useInR": true,
|
|
220
|
+
"rOrderNum": 11
|
|
221
|
+
},
|
|
222
|
+
{
|
|
223
|
+
"lowerBound": 1,
|
|
224
|
+
"upperBound": 4096,
|
|
225
|
+
"lowerBoundType": "INCLUSIVE",
|
|
226
|
+
"upperBoundType": "INCLUSIVE",
|
|
227
|
+
"allowNaN": false,
|
|
228
|
+
"name": "EmbeddingSize",
|
|
229
|
+
"alternateNames": [],
|
|
230
|
+
"isRequired": false,
|
|
231
|
+
"rDescription": "Specify the embedding size of the vectors.",
|
|
232
|
+
"description": "Specify the embedding size of the vectors.",
|
|
233
|
+
"datatype": "INTEGER",
|
|
234
|
+
"allowsLists": false,
|
|
235
|
+
"rName": "embedding.size",
|
|
236
|
+
"useInR": true,
|
|
237
|
+
"rOrderNum": 12
|
|
238
|
+
},
|
|
239
|
+
{
|
|
240
|
+
"permittedValues": [],
|
|
241
|
+
"isOutputColumn": false,
|
|
242
|
+
"matchLengthOfArgument": "",
|
|
243
|
+
"allowPadding": false,
|
|
244
|
+
"defaultValue": false,
|
|
245
|
+
"name": "ApplyHeuristics",
|
|
246
|
+
"alternateNames": [],
|
|
247
|
+
"isRequired": false,
|
|
248
|
+
"rDescription": "Specify whether to apply heuristics optimizations during training of HNSW model.",
|
|
249
|
+
"description": "Specify whether to apply heuristics optimizations during training of HNSW model.",
|
|
250
|
+
"datatype": "BOOLEAN",
|
|
251
|
+
"allowsLists": false,
|
|
252
|
+
"rName": "apply.heuristics",
|
|
253
|
+
"useInR": true,
|
|
254
|
+
"rOrderNum": 13
|
|
255
|
+
},
|
|
256
|
+
{
|
|
257
|
+
"permittedValues": [
|
|
258
|
+
"UPDATE",
|
|
259
|
+
"DELETE"
|
|
260
|
+
],
|
|
261
|
+
"isOutputColumn": false,
|
|
262
|
+
"matchLengthOfArgument": "",
|
|
263
|
+
"allowPadding": false,
|
|
264
|
+
"name": "AlterOperation",
|
|
265
|
+
"alternateNames": [],
|
|
266
|
+
"isRequired": false,
|
|
267
|
+
"rDescription": "Specify the alter operation for HNSW model. This argument is required when InputModelTable is provided.",
|
|
268
|
+
"description": "Specify the alter operation for HNSW model. This argument is required when InputModelTable is provided.",
|
|
269
|
+
"datatype": "STRING",
|
|
270
|
+
"allowsLists": true,
|
|
271
|
+
"rName": "alter.operation",
|
|
272
|
+
"useInR": true,
|
|
273
|
+
"rOrderNum": 14
|
|
274
|
+
},
|
|
275
|
+
{
|
|
276
|
+
"permittedValues": [
|
|
277
|
+
"RECONSTRUCTION",
|
|
278
|
+
"DELETENODE"
|
|
279
|
+
],
|
|
280
|
+
"defaultValue": "RECONSTRUCTION",
|
|
281
|
+
"isOutputColumn": false,
|
|
282
|
+
"matchLengthOfArgument": "",
|
|
283
|
+
"allowPadding": false,
|
|
284
|
+
"name": "DeleteMethod",
|
|
285
|
+
"alternateNames": [],
|
|
286
|
+
"isRequired": false,
|
|
287
|
+
"rDescription": "Specify the method for delete operation.",
|
|
288
|
+
"description": "Specify the method for delete operation.",
|
|
289
|
+
"datatype": "STRING",
|
|
290
|
+
"allowsLists": true,
|
|
291
|
+
"rName": "delete.method",
|
|
292
|
+
"useInR": true,
|
|
293
|
+
"rOrderNum": 15
|
|
294
|
+
}
|
|
295
|
+
]
|
|
296
|
+
}
|