@maxim_mazurok/gapi.client.aiplatform-v1 0.0.20240501 → 0.0.20240510
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/index.d.ts +82 -839
- package/package.json +1 -1
package/index.d.ts
CHANGED
|
@@ -9,7 +9,7 @@
|
|
|
9
9
|
// This file was generated by https://github.com/Maxim-Mazurok/google-api-typings-generator. Please do not edit it manually.
|
|
10
10
|
// In case of any problems please post issue to https://github.com/Maxim-Mazurok/google-api-typings-generator
|
|
11
11
|
// Generated from: https://aiplatform.googleapis.com/$discovery/rest?version=v1
|
|
12
|
-
// Revision:
|
|
12
|
+
// Revision: 20240510
|
|
13
13
|
|
|
14
14
|
/// <reference types="gapi.client" />
|
|
15
15
|
|
|
@@ -24,10 +24,6 @@ declare namespace gapi.client {
|
|
|
24
24
|
function load(name: 'aiplatform', version: 'v1', callback: () => any): void;
|
|
25
25
|
|
|
26
26
|
namespace aiplatform {
|
|
27
|
-
interface CloudAiLargeModelsVisionEmbedVideoResponse {
|
|
28
|
-
/** The embedding vector for the video. */
|
|
29
|
-
videoEmbeddings?: any[];
|
|
30
|
-
}
|
|
31
27
|
interface CloudAiLargeModelsVisionFilteredText {
|
|
32
28
|
/** Confidence level */
|
|
33
29
|
category?: string;
|
|
@@ -76,10 +72,6 @@ declare namespace gapi.client {
|
|
|
76
72
|
/** Video */
|
|
77
73
|
video?: CloudAiLargeModelsVisionVideo;
|
|
78
74
|
}
|
|
79
|
-
interface CloudAiLargeModelsVisionMediaGenerateContentResponse {
|
|
80
|
-
/** Response to the user's request. */
|
|
81
|
-
response?: CloudAiNlLlmProtoServiceGenerateMultiModalResponse;
|
|
82
|
-
}
|
|
83
75
|
interface CloudAiLargeModelsVisionNamedBoundingBox {
|
|
84
76
|
classes?: string[];
|
|
85
77
|
entities?: string[];
|
|
@@ -95,22 +87,6 @@ declare namespace gapi.client {
|
|
|
95
87
|
/** List of rai scores mapping to the rai categories. Rounded to 1 decimal place. */
|
|
96
88
|
scores?: number[];
|
|
97
89
|
}
|
|
98
|
-
interface CloudAiLargeModelsVisionReasonVideoResponse {
|
|
99
|
-
/** Generated text responses. The generated responses for different segments within the same video. */
|
|
100
|
-
responses?: CloudAiLargeModelsVisionReasonVideoResponseTextResponse[];
|
|
101
|
-
}
|
|
102
|
-
interface CloudAiLargeModelsVisionReasonVideoResponseTextResponse {
|
|
103
|
-
/** Partition of the caption's video in time. This field is intended for video captioning. To represent the start time and end time of the caption's video. */
|
|
104
|
-
relativeTemporalPartition?: CloudAiLargeModelsVisionRelativeTemporalPartition;
|
|
105
|
-
/** Text information */
|
|
106
|
-
text?: string;
|
|
107
|
-
}
|
|
108
|
-
interface CloudAiLargeModelsVisionRelativeTemporalPartition {
|
|
109
|
-
/** End time offset of the partition. */
|
|
110
|
-
endOffset?: string;
|
|
111
|
-
/** Start time offset of the partition. */
|
|
112
|
-
startOffset?: string;
|
|
113
|
-
}
|
|
114
90
|
interface CloudAiLargeModelsVisionSemanticFilterResponse {
|
|
115
91
|
/** Class labels of the bounding boxes that failed the semantic filtering. Bounding box coordinates. */
|
|
116
92
|
namedBoundingBoxes?: CloudAiLargeModelsVisionNamedBoundingBox[];
|
|
@@ -123,225 +99,6 @@ declare namespace gapi.client {
|
|
|
123
99
|
/** Raw bytes. */
|
|
124
100
|
video?: string;
|
|
125
101
|
}
|
|
126
|
-
interface CloudAiNlLlmProtoServiceCandidate {
|
|
127
|
-
/** Source attribution of the generated content. */
|
|
128
|
-
citationMetadata?: CloudAiNlLlmProtoServiceCitationMetadata;
|
|
129
|
-
/** Content of the candidate. */
|
|
130
|
-
content?: CloudAiNlLlmProtoServiceContent;
|
|
131
|
-
/** A string that describes the filtering behavior in more detail. Only filled when reason is set. */
|
|
132
|
-
finishMessage?: string;
|
|
133
|
-
/** The reason why the model stopped generating tokens. */
|
|
134
|
-
finishReason?: string;
|
|
135
|
-
/** Grounding metadata. Combine with the facts list from response to generate grounding citations for this choice. */
|
|
136
|
-
groundingMetadata?: LearningGenaiRootGroundingMetadata;
|
|
137
|
-
/** Index of the candidate. */
|
|
138
|
-
index?: number;
|
|
139
|
-
/** Safety ratings of the generated content. */
|
|
140
|
-
safetyRatings?: CloudAiNlLlmProtoServiceSafetyRating[];
|
|
141
|
-
}
|
|
142
|
-
interface CloudAiNlLlmProtoServiceCitation {
|
|
143
|
-
/** End index into the content. */
|
|
144
|
-
endIndex?: number;
|
|
145
|
-
/** License of the attribution. */
|
|
146
|
-
license?: string;
|
|
147
|
-
/** Publication date of the attribution. */
|
|
148
|
-
publicationDate?: GoogleTypeDate;
|
|
149
|
-
/** Start index into the content. */
|
|
150
|
-
startIndex?: number;
|
|
151
|
-
/** Title of the attribution. */
|
|
152
|
-
title?: string;
|
|
153
|
-
/** Url reference of the attribution. */
|
|
154
|
-
uri?: string;
|
|
155
|
-
}
|
|
156
|
-
interface CloudAiNlLlmProtoServiceCitationMetadata {
|
|
157
|
-
/** List of citations. */
|
|
158
|
-
citations?: CloudAiNlLlmProtoServiceCitation[];
|
|
159
|
-
}
|
|
160
|
-
interface CloudAiNlLlmProtoServiceContent {
|
|
161
|
-
/** The parts of the message. */
|
|
162
|
-
parts?: CloudAiNlLlmProtoServicePart[];
|
|
163
|
-
/** The role of the current conversation participant. */
|
|
164
|
-
role?: string;
|
|
165
|
-
}
|
|
166
|
-
interface CloudAiNlLlmProtoServiceFact {
|
|
167
|
-
/** Query that is used to retrieve this fact. */
|
|
168
|
-
query?: string;
|
|
169
|
-
/** If present, the summary/snippet of the fact. */
|
|
170
|
-
summary?: string;
|
|
171
|
-
/** If present, it refers to the title of this fact. */
|
|
172
|
-
title?: string;
|
|
173
|
-
/** If present, this URL links to the webpage of the fact. */
|
|
174
|
-
url?: string;
|
|
175
|
-
}
|
|
176
|
-
interface CloudAiNlLlmProtoServiceFunctionCall {
|
|
177
|
-
/** The function parameters and values in JSON format. */
|
|
178
|
-
args?: {[P in string]: any};
|
|
179
|
-
/** Required. The name of the function to call. */
|
|
180
|
-
name?: string;
|
|
181
|
-
}
|
|
182
|
-
interface CloudAiNlLlmProtoServiceFunctionResponse {
|
|
183
|
-
/** Required. The name of the function to call. */
|
|
184
|
-
name?: string;
|
|
185
|
-
/** Required. The function response in JSON object format. */
|
|
186
|
-
response?: {[P in string]: any};
|
|
187
|
-
}
|
|
188
|
-
interface CloudAiNlLlmProtoServiceGenerateMultiModalResponse {
|
|
189
|
-
/** Possible candidate responses to the conversation up until this point. */
|
|
190
|
-
candidates?: CloudAiNlLlmProtoServiceCandidate[];
|
|
191
|
-
/** Debug information containing message metadata. Clients should not consume this field, and this is only populated for Flow Runner path. */
|
|
192
|
-
debugMetadata?: CloudAiNlLlmProtoServiceMessageMetadata;
|
|
193
|
-
/** External facts retrieved for factuality/grounding. */
|
|
194
|
-
facts?: CloudAiNlLlmProtoServiceFact[];
|
|
195
|
-
/** Content filter results for a prompt sent in the request. Note: Sent only in the first stream chunk. Only happens when no candidates were generated due to content violations. */
|
|
196
|
-
promptFeedback?: CloudAiNlLlmProtoServicePromptFeedback;
|
|
197
|
-
/** Billable prediction metrics. */
|
|
198
|
-
reportingMetrics?: IntelligenceCloudAutomlXpsReportingMetrics;
|
|
199
|
-
/** Usage metadata about the response(s). */
|
|
200
|
-
usageMetadata?: CloudAiNlLlmProtoServiceUsageMetadata;
|
|
201
|
-
}
|
|
202
|
-
interface CloudAiNlLlmProtoServiceMessageMetadata {
|
|
203
|
-
/** Factuality-related debug metadata. */
|
|
204
|
-
factualityDebugMetadata?: LearningGenaiRootPerRequestProcessorDebugMetadataFactualityDebugMetadata;
|
|
205
|
-
/** Filter metadata of the input messages. */
|
|
206
|
-
inputFilterInfo?: LearningServingLlmMessageMetadata;
|
|
207
|
-
/** This score is generated by the router model to decide which model to use */
|
|
208
|
-
modelRoutingDecision?: LearningGenaiRootRoutingDecision;
|
|
209
|
-
/** Filter metadata of the output messages. */
|
|
210
|
-
outputFilterInfo?: LearningServingLlmMessageMetadata[];
|
|
211
|
-
}
|
|
212
|
-
interface CloudAiNlLlmProtoServicePart {
|
|
213
|
-
/** Document metadata. The metadata should only be used by the Cloud LLM when supporting document mime types. It will only be populated when this image input part is converted from a document input part. */
|
|
214
|
-
documentMetadata?: CloudAiNlLlmProtoServicePartDocumentMetadata;
|
|
215
|
-
/** URI-based data. */
|
|
216
|
-
fileData?: CloudAiNlLlmProtoServicePartFileData;
|
|
217
|
-
/** Function call data. */
|
|
218
|
-
functionCall?: CloudAiNlLlmProtoServiceFunctionCall;
|
|
219
|
-
/** Function response data. */
|
|
220
|
-
functionResponse?: CloudAiNlLlmProtoServiceFunctionResponse;
|
|
221
|
-
/** Inline bytes data */
|
|
222
|
-
inlineData?: CloudAiNlLlmProtoServicePartBlob;
|
|
223
|
-
/** Metadata provides extra info for building the LM Root request. Note: High enough tag number for internal only fields. */
|
|
224
|
-
lmRootMetadata?: CloudAiNlLlmProtoServicePartLMRootMetadata;
|
|
225
|
-
/** Text input. */
|
|
226
|
-
text?: string;
|
|
227
|
-
/** Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. */
|
|
228
|
-
videoMetadata?: CloudAiNlLlmProtoServicePartVideoMetadata;
|
|
229
|
-
}
|
|
230
|
-
interface CloudAiNlLlmProtoServicePartBlob {
|
|
231
|
-
/** Inline data. */
|
|
232
|
-
data?: string;
|
|
233
|
-
/** The mime type corresponding to this input. */
|
|
234
|
-
mimeType?: string;
|
|
235
|
-
/** Original file data where the blob comes from. */
|
|
236
|
-
originalFileData?: CloudAiNlLlmProtoServicePartFileData;
|
|
237
|
-
}
|
|
238
|
-
interface CloudAiNlLlmProtoServicePartDocumentMetadata {
|
|
239
|
-
/** The original document blob. */
|
|
240
|
-
originalDocumentBlob?: CloudAiNlLlmProtoServicePartBlob;
|
|
241
|
-
/** The (1-indexed) page number of the image in the original document. The first page carries the original document content and mime type. */
|
|
242
|
-
pageNumber?: number;
|
|
243
|
-
}
|
|
244
|
-
interface CloudAiNlLlmProtoServicePartFileData {
|
|
245
|
-
/** Inline data. */
|
|
246
|
-
fileUri?: string;
|
|
247
|
-
/** The mime type corresponding to this input. */
|
|
248
|
-
mimeType?: string;
|
|
249
|
-
}
|
|
250
|
-
interface CloudAiNlLlmProtoServicePartLMRootMetadata {
|
|
251
|
-
/** Chunk id that will be used when mapping the part to the LM Root's chunk. */
|
|
252
|
-
chunkId?: string;
|
|
253
|
-
}
|
|
254
|
-
interface CloudAiNlLlmProtoServicePartVideoMetadata {
|
|
255
|
-
/** The end offset of the video. */
|
|
256
|
-
endOffset?: string;
|
|
257
|
-
/** The start offset of the video. */
|
|
258
|
-
startOffset?: string;
|
|
259
|
-
}
|
|
260
|
-
interface CloudAiNlLlmProtoServicePromptFeedback {
|
|
261
|
-
/** Blocked reason. */
|
|
262
|
-
blockReason?: string;
|
|
263
|
-
/** A readable block reason message. */
|
|
264
|
-
blockReasonMessage?: string;
|
|
265
|
-
/** Safety ratings. */
|
|
266
|
-
safetyRatings?: CloudAiNlLlmProtoServiceSafetyRating[];
|
|
267
|
-
}
|
|
268
|
-
interface CloudAiNlLlmProtoServiceRaiResult {
|
|
269
|
-
/** Recitation result from Aida recitation checker. */
|
|
270
|
-
aidaRecitationResult?: LanguageLabsAidaTrustRecitationProtoRecitationResult;
|
|
271
|
-
/** Use `triggered_blocklist`. */
|
|
272
|
-
blocked?: boolean;
|
|
273
|
-
/** The error codes indicate which RAI filters block the response. */
|
|
274
|
-
errorCodes?: number[];
|
|
275
|
-
/** Whether the text should be filtered and not shown to the end user. This is determined based on a combination of `triggered_recitation`, `triggered_blocklist`, `language_filter_result`, and `triggered_safety_filter`. */
|
|
276
|
-
filtered?: boolean;
|
|
277
|
-
/** Language filter result from SAFT LangId. */
|
|
278
|
-
languageFilterResult?: LearningGenaiRootLanguageFilterResult;
|
|
279
|
-
/** The RAI signals for the text. */
|
|
280
|
-
raiSignals?: CloudAiNlLlmProtoServiceRaiSignal[];
|
|
281
|
-
/** Translation request info during RAI for debugging purpose. Each TranslationRequestInfo corresponds to a request sent to the translation server. */
|
|
282
|
-
translationRequestInfos?: LearningGenaiRootTranslationRequestInfo[];
|
|
283
|
-
/** Whether the text triggered the blocklist. */
|
|
284
|
-
triggeredBlocklist?: boolean;
|
|
285
|
-
/** Whether the text should be blocked by the recitation result from Aida recitation checker. It is determined from aida_recitation_result. */
|
|
286
|
-
triggeredRecitation?: boolean;
|
|
287
|
-
/** Whether the text triggered the safety filter. Currently, this is due to CSAI triggering or one of four categories (derogatory, sexual, toxic, violent) having a score over the filter threshold. */
|
|
288
|
-
triggeredSafetyFilter?: boolean;
|
|
289
|
-
}
|
|
290
|
-
interface CloudAiNlLlmProtoServiceRaiSignal {
|
|
291
|
-
/** The confidence level for the RAI category. */
|
|
292
|
-
confidence?: string;
|
|
293
|
-
/** Whether the category is flagged as being present. Currently, this is set to true if score >= 0.5. */
|
|
294
|
-
flagged?: boolean;
|
|
295
|
-
/** The influential terms that could potentially block the response. */
|
|
296
|
-
influentialTerms?: CloudAiNlLlmProtoServiceRaiSignalInfluentialTerm[];
|
|
297
|
-
/** The RAI category. */
|
|
298
|
-
raiCategory?: string;
|
|
299
|
-
/** The score for the category, in the range [0.0, 1.0]. */
|
|
300
|
-
score?: number;
|
|
301
|
-
}
|
|
302
|
-
interface CloudAiNlLlmProtoServiceRaiSignalInfluentialTerm {
|
|
303
|
-
/** The beginning offset of the influential term. */
|
|
304
|
-
beginOffset?: number;
|
|
305
|
-
/** The confidence score of the influential term. */
|
|
306
|
-
confidence?: number;
|
|
307
|
-
/** The source of the influential term, prompt or response. */
|
|
308
|
-
source?: string;
|
|
309
|
-
/** The influential term. */
|
|
310
|
-
term?: string;
|
|
311
|
-
}
|
|
312
|
-
interface CloudAiNlLlmProtoServiceSafetyRating {
|
|
313
|
-
/** Indicates whether the content was filtered out because of this rating. */
|
|
314
|
-
blocked?: boolean;
|
|
315
|
-
/** Harm category. */
|
|
316
|
-
category?: string;
|
|
317
|
-
/** The influential terms that could potentially block the response. */
|
|
318
|
-
influentialTerms?: CloudAiNlLlmProtoServiceSafetyRatingInfluentialTerm[];
|
|
319
|
-
/** Harm probability levels in the content. */
|
|
320
|
-
probability?: string;
|
|
321
|
-
/** Harm probability score. */
|
|
322
|
-
probabilityScore?: number;
|
|
323
|
-
/** Harm severity levels in the content. */
|
|
324
|
-
severity?: string;
|
|
325
|
-
/** Harm severity score. */
|
|
326
|
-
severityScore?: number;
|
|
327
|
-
}
|
|
328
|
-
interface CloudAiNlLlmProtoServiceSafetyRatingInfluentialTerm {
|
|
329
|
-
/** The beginning offset of the influential term. */
|
|
330
|
-
beginOffset?: number;
|
|
331
|
-
/** The confidence score of the influential term. */
|
|
332
|
-
confidence?: number;
|
|
333
|
-
/** The source of the influential term, prompt or response. */
|
|
334
|
-
source?: string;
|
|
335
|
-
/** The influential term. */
|
|
336
|
-
term?: string;
|
|
337
|
-
}
|
|
338
|
-
interface CloudAiNlLlmProtoServiceUsageMetadata {
|
|
339
|
-
/** Number of tokens in the response(s). */
|
|
340
|
-
candidatesTokenCount?: number;
|
|
341
|
-
/** Number of tokens in the request. */
|
|
342
|
-
promptTokenCount?: number;
|
|
343
|
-
totalTokenCount?: number;
|
|
344
|
-
}
|
|
345
102
|
interface GoogleApiHttpBody {
|
|
346
103
|
/** The HTTP Content-Type header value specifying the content type of the body. */
|
|
347
104
|
contentType?: string;
|
|
@@ -1132,6 +889,8 @@ declare namespace gapi.client {
|
|
|
1132
889
|
metadataArtifact?: string;
|
|
1133
890
|
/** Required. Points to a YAML file stored on Google Cloud Storage describing additional information about the Dataset. The schema is defined as an OpenAPI 3.0.2 Schema Object. The schema files that can be used here are found in gs://google-cloud-aiplatform/schema/dataset/metadata/. */
|
|
1134
891
|
metadataSchemaUri?: string;
|
|
892
|
+
/** Optional. Reference to the public base model last used by the dataset. Only set for prompt datasets. */
|
|
893
|
+
modelReference?: string;
|
|
1135
894
|
/** Output only. The resource name of the Dataset. */
|
|
1136
895
|
name?: string;
|
|
1137
896
|
/** All SavedQueries belong to the Dataset will be returned in List/Get Dataset response. The annotation_specs field will not be populated except for UI cases which will only use annotation_spec_count. In CreateDataset request, a SavedQuery is created together if this field is set, up to one SavedQuery can be set in CreateDatasetRequest. The SavedQuery should not contain any AnnotationSpec. */
|
|
@@ -1150,6 +909,8 @@ declare namespace gapi.client {
|
|
|
1150
909
|
etag?: string;
|
|
1151
910
|
/** Required. Output only. Additional information about the DatasetVersion. */
|
|
1152
911
|
metadata?: any;
|
|
912
|
+
/** Output only. Reference to the public base model last used by the dataset version. Only set for prompt dataset versions. */
|
|
913
|
+
modelReference?: string;
|
|
1153
914
|
/** Output only. The resource name of the DatasetVersion. */
|
|
1154
915
|
name?: string;
|
|
1155
916
|
/** Output only. Timestamp when this DatasetVersion was last updated. */
|
|
@@ -1822,7 +1583,7 @@ declare namespace gapi.client {
|
|
|
1822
1583
|
versionColumnName?: string;
|
|
1823
1584
|
}
|
|
1824
1585
|
interface GoogleCloudAiplatformV1FeatureGroup {
|
|
1825
|
-
/** Indicates that features for this group come from BigQuery Table/View. By default treats the source as a sparse time series source
|
|
1586
|
+
/** Indicates that features for this group come from BigQuery Table/View. By default treats the source as a sparse time series source. The BigQuery source table or view must have at least one entity ID column and a column named `feature_timestamp`. */
|
|
1826
1587
|
bigQuery?: GoogleCloudAiplatformV1FeatureGroupBigQuery;
|
|
1827
1588
|
/** Output only. Timestamp when this FeatureGroup was created. */
|
|
1828
1589
|
createTime?: string;
|
|
@@ -2271,6 +2032,8 @@ declare namespace gapi.client {
|
|
|
2271
2032
|
presencePenalty?: number;
|
|
2272
2033
|
/** Optional. Output response mimetype of the generated candidate text. Supported mimetype: - `text/plain`: (default) Text output. - `application/json`: JSON response in the candidates. The model needs to be prompted to output the appropriate response type, otherwise the behavior is undefined. This is a preview feature. */
|
|
2273
2034
|
responseMimeType?: string;
|
|
2035
|
+
/** Optional. Control Three levels of creativity in the model output. Default: RESPONSE_STYLE_BALANCED */
|
|
2036
|
+
responseStyle?: string;
|
|
2274
2037
|
/** Optional. Stop sequences. */
|
|
2275
2038
|
stopSequences?: string[];
|
|
2276
2039
|
/** Optional. Controls the randomness of predictions. */
|
|
@@ -2477,9 +2240,9 @@ declare namespace gapi.client {
|
|
|
2477
2240
|
namespace?: string;
|
|
2478
2241
|
}
|
|
2479
2242
|
interface GoogleCloudAiplatformV1IndexDatapointSparseEmbedding {
|
|
2480
|
-
/**
|
|
2243
|
+
/** Required. The list of indexes for the embedding values of the sparse vector. */
|
|
2481
2244
|
dimensions?: string[];
|
|
2482
|
-
/**
|
|
2245
|
+
/** Required. The list of embedding values of the sparse vector. */
|
|
2483
2246
|
values?: number[];
|
|
2484
2247
|
}
|
|
2485
2248
|
interface GoogleCloudAiplatformV1IndexEndpoint {
|
|
@@ -3571,12 +3334,16 @@ declare namespace gapi.client {
|
|
|
3571
3334
|
interface GoogleCloudAiplatformV1NearestNeighborSearchOperationMetadataContentValidationStats {
|
|
3572
3335
|
/** Number of records in this file we skipped due to validate errors. */
|
|
3573
3336
|
invalidRecordCount?: string;
|
|
3337
|
+
/** Number of sparse records in this file we skipped due to validate errors. */
|
|
3338
|
+
invalidSparseRecordCount?: string;
|
|
3574
3339
|
/** The detail information of the partial failures encountered for those invalid records that couldn't be parsed. Up to 50 partial errors will be reported. */
|
|
3575
3340
|
partialErrors?: GoogleCloudAiplatformV1NearestNeighborSearchOperationMetadataRecordError[];
|
|
3576
3341
|
/** Cloud Storage URI pointing to the original file in user's bucket. */
|
|
3577
3342
|
sourceGcsUri?: string;
|
|
3578
3343
|
/** Number of records in this file that were successfully processed. */
|
|
3579
3344
|
validRecordCount?: string;
|
|
3345
|
+
/** Number of sparse records in this file that were successfully processed. */
|
|
3346
|
+
validSparseRecordCount?: string;
|
|
3580
3347
|
}
|
|
3581
3348
|
interface GoogleCloudAiplatformV1NearestNeighborSearchOperationMetadataRecordError {
|
|
3582
3349
|
/** Empty if the embedding id is failed to parse. */
|
|
@@ -3671,6 +3438,10 @@ declare namespace gapi.client {
|
|
|
3671
3438
|
runtimeState?: string;
|
|
3672
3439
|
/** Required. The user email of the NotebookRuntime. */
|
|
3673
3440
|
runtimeUser?: string;
|
|
3441
|
+
/** Output only. Reserved for future use. */
|
|
3442
|
+
satisfiesPzi?: boolean;
|
|
3443
|
+
/** Output only. Reserved for future use. */
|
|
3444
|
+
satisfiesPzs?: boolean;
|
|
3674
3445
|
/** Output only. The service account that the NotebookRuntime workload runs as. */
|
|
3675
3446
|
serviceAccount?: string;
|
|
3676
3447
|
/** Output only. Timestamp when this NotebookRuntime was most recently updated. */
|
|
@@ -6466,6 +6237,8 @@ declare namespace gapi.client {
|
|
|
6466
6237
|
createTime?: string;
|
|
6467
6238
|
/** Optional. The description of the TuningJob. */
|
|
6468
6239
|
description?: string;
|
|
6240
|
+
/** Customer-managed encryption key options for a TuningJob. If this is set, then all resources created by the TuningJob will be encrypted with the provided encryption key. */
|
|
6241
|
+
encryptionSpec?: GoogleCloudAiplatformV1EncryptionSpec;
|
|
6469
6242
|
/** Output only. Time when the TuningJob entered any of the following JobStates: `JOB_STATE_SUCCEEDED`, `JOB_STATE_FAILED`, `JOB_STATE_CANCELLED`, `JOB_STATE_EXPIRED`. */
|
|
6470
6243
|
endTime?: string;
|
|
6471
6244
|
/** Output only. Only populated when job's state is `JOB_STATE_FAILED` or `JOB_STATE_CANCELLED`. */
|
|
@@ -6794,598 +6567,6 @@ declare namespace gapi.client {
|
|
|
6794
6567
|
/** The whole units of the amount. For example if `currencyCode` is `"USD"`, then 1 unit is one US dollar. */
|
|
6795
6568
|
units?: string;
|
|
6796
6569
|
}
|
|
6797
|
-
interface IntelligenceCloudAutomlXpsMetricEntry {
|
|
6798
|
-
/** For billing metrics that are using legacy sku's, set the legacy billing metric id here. This will be sent to Chemist as the "cloudbilling.googleapis.com/argentum_metric_id" label. Otherwise leave empty. */
|
|
6799
|
-
argentumMetricId?: string;
|
|
6800
|
-
/** A double value. */
|
|
6801
|
-
doubleValue?: number;
|
|
6802
|
-
/** A signed 64-bit integer value. */
|
|
6803
|
-
int64Value?: string;
|
|
6804
|
-
/** The metric name defined in the service configuration. */
|
|
6805
|
-
metricName?: string;
|
|
6806
|
-
/** Billing system labels for this (metric, value) pair. */
|
|
6807
|
-
systemLabels?: IntelligenceCloudAutomlXpsMetricEntryLabel[];
|
|
6808
|
-
}
|
|
6809
|
-
interface IntelligenceCloudAutomlXpsMetricEntryLabel {
|
|
6810
|
-
/** The name of the label. */
|
|
6811
|
-
labelName?: string;
|
|
6812
|
-
/** The value of the label. */
|
|
6813
|
-
labelValue?: string;
|
|
6814
|
-
}
|
|
6815
|
-
interface IntelligenceCloudAutomlXpsReportingMetrics {
|
|
6816
|
-
/** The effective time training used. If set, this is used for quota management and billing. Deprecated. AutoML BE doesn't use this. Don't set. */
|
|
6817
|
-
effectiveTrainingDuration?: string;
|
|
6818
|
-
/** One entry per metric name. The values must be aggregated per metric name. */
|
|
6819
|
-
metricEntries?: IntelligenceCloudAutomlXpsMetricEntry[];
|
|
6820
|
-
}
|
|
6821
|
-
interface LanguageLabsAidaTrustRecitationProtoDocAttribution {
|
|
6822
|
-
amarnaId?: string;
|
|
6823
|
-
arxivId?: string;
|
|
6824
|
-
author?: string;
|
|
6825
|
-
bibkey?: string;
|
|
6826
|
-
/** ID of the paper in bioarxiv like ddoi.org/{biorxiv_id} eg: https://doi.org/10.1101/343517 */
|
|
6827
|
-
biorxivId?: string;
|
|
6828
|
-
bookTitle?: string;
|
|
6829
|
-
/** The Oceanographers full-view books dataset uses a 'volume id' as the unique ID of a book. There is a deterministic function from a volume id to a URL under the books.google.com domain. Marked as 'optional' since a volume ID of zero is potentially possible and we want to distinguish that from the volume ID not being set. */
|
|
6830
|
-
bookVolumeId?: string;
|
|
6831
|
-
category?: string;
|
|
6832
|
-
conversationId?: string;
|
|
6833
|
-
/** The dataset this document comes from. */
|
|
6834
|
-
dataset?: string;
|
|
6835
|
-
filepath?: string;
|
|
6836
|
-
geminiId?: string;
|
|
6837
|
-
gnewsArticleTitle?: string;
|
|
6838
|
-
goodallExampleId?: string;
|
|
6839
|
-
/** Whether the document is opted out. */
|
|
6840
|
-
isOptOut?: boolean;
|
|
6841
|
-
isPrompt?: boolean;
|
|
6842
|
-
lamdaExampleId?: string;
|
|
6843
|
-
license?: string;
|
|
6844
|
-
meenaConversationId?: string;
|
|
6845
|
-
/** Natural (not programming) language of the document. Language code as defined by http://www.unicode.org/reports/tr35/#Identifiers and https://tools.ietf.org/html/bcp47. Currently applicable to full-view books. Use docinfo-util.h to set & read language fields. See go/iii. */
|
|
6846
|
-
naturalLanguageCode?: string;
|
|
6847
|
-
/** True if this doc has no attribution information available. We use an explicit field for this instead of just implicitly leaving all the DocAttribution fields blank to distinguish a case where a bug/oversight has left the attribution information empty vs when we really have no attribution information available. */
|
|
6848
|
-
noAttribution?: boolean;
|
|
6849
|
-
podcastUtteranceId?: string;
|
|
6850
|
-
publicationDate?: GoogleTypeDate;
|
|
6851
|
-
/** This field is for opt-out experiment only, MUST never be used during actual production/serving. */
|
|
6852
|
-
qualityScoreExperimentOnly?: number;
|
|
6853
|
-
/** Github repository */
|
|
6854
|
-
repo?: string;
|
|
6855
|
-
/** URL of a webdoc */
|
|
6856
|
-
url?: string;
|
|
6857
|
-
volumeId?: string;
|
|
6858
|
-
/** Wikipedia article title. The Wikipedia TFDS dataset includes article titles but not URLs. While a URL is to the best of our knowledge a deterministic function of the title, we store the original title to reflect the information in the original dataset. */
|
|
6859
|
-
wikipediaArticleTitle?: string;
|
|
6860
|
-
/** The unique video id from Youtube. Example: AkoGsW52Ir0 */
|
|
6861
|
-
youtubeVideoId?: string;
|
|
6862
|
-
}
|
|
6863
|
-
interface LanguageLabsAidaTrustRecitationProtoRecitationResult {
|
|
6864
|
-
dynamicSegmentResults?: LanguageLabsAidaTrustRecitationProtoSegmentResult[];
|
|
6865
|
-
/** The recitation action for one given input. When its segments contain different actions, the overall action will be returned in the precedence of BLOCK > CITE > NO_ACTION. When the given input is not found in any source, the recitation action will not be specified. */
|
|
6866
|
-
recitationAction?: string;
|
|
6867
|
-
trainingSegmentResults?: LanguageLabsAidaTrustRecitationProtoSegmentResult[];
|
|
6868
|
-
}
|
|
6869
|
-
interface LanguageLabsAidaTrustRecitationProtoSegmentResult {
|
|
6870
|
-
/** The dataset the segment came from. Datasets change often as model evolves. Treat this field as informational only and avoid depending on it directly. */
|
|
6871
|
-
attributionDataset?: string;
|
|
6872
|
-
/** human-friendly string that contains information from doc_attribution which could be shown by clients */
|
|
6873
|
-
displayAttributionMessage?: string;
|
|
6874
|
-
docAttribution?: LanguageLabsAidaTrustRecitationProtoDocAttribution;
|
|
6875
|
-
/** number of documents that contained this segment */
|
|
6876
|
-
docOccurrences?: number;
|
|
6877
|
-
endIndex?: number;
|
|
6878
|
-
/** The raw text in the given input that is corresponding to the segment. It will be available only when 'return_segment_raw_text' is enabled in the request options. */
|
|
6879
|
-
rawText?: string;
|
|
6880
|
-
segmentRecitationAction?: string;
|
|
6881
|
-
/** The category of the source dataset where the segment came from. This is more stable than Dataset. */
|
|
6882
|
-
sourceCategory?: string;
|
|
6883
|
-
/** The segment boundary start (inclusive) and end index (exclusive) in the given text. In the streaming RPC, the indexes always start from the beginning of the first text in the entire stream. The indexes are measured in UTF-16 code units. */
|
|
6884
|
-
startIndex?: number;
|
|
6885
|
-
}
|
|
6886
|
-
interface LanguageLabsAidaTrustRecitationProtoStreamRecitationResult {
|
|
6887
|
-
/** The recitation result against the given dynamic data source. */
|
|
6888
|
-
dynamicSegmentResults?: LanguageLabsAidaTrustRecitationProtoSegmentResult[];
|
|
6889
|
-
/** Last index of input text fully checked for recitation in the entire streaming context. Would return `-1` if no Input was checked for recitation. */
|
|
6890
|
-
fullyCheckedTextIndex?: number;
|
|
6891
|
-
/** The recitation action for one given input. When its segments contain different actions, the overall action will be returned in the precedence of BLOCK > CITE > NO_ACTION. */
|
|
6892
|
-
recitationAction?: string;
|
|
6893
|
-
/** The recitation result against model training data. */
|
|
6894
|
-
trainingSegmentResults?: LanguageLabsAidaTrustRecitationProtoSegmentResult[];
|
|
6895
|
-
}
|
|
6896
|
-
interface LearningGenaiRecitationDocAttribution {
|
|
6897
|
-
amarnaId?: string;
|
|
6898
|
-
arxivId?: string;
|
|
6899
|
-
author?: string;
|
|
6900
|
-
bibkey?: string;
|
|
6901
|
-
/** ID of the paper in bioarxiv like ddoi.org/{biorxiv_id} eg: https://doi.org/10.1101/343517 */
|
|
6902
|
-
biorxivId?: string;
|
|
6903
|
-
bookTitle?: string;
|
|
6904
|
-
/** The Oceanographers full-view books dataset uses a 'volume id' as the unique ID of a book. There is a deterministic function from a volume id to a URL under the books.google.com domain. Marked as 'optional' since a volume ID of zero is potentially possible and we want to distinguish that from the volume ID not being set. */
|
|
6905
|
-
bookVolumeId?: string;
|
|
6906
|
-
conversationId?: string;
|
|
6907
|
-
/** The dataset this document comes from. */
|
|
6908
|
-
dataset?: string;
|
|
6909
|
-
filepath?: string;
|
|
6910
|
-
geminiId?: string;
|
|
6911
|
-
gnewsArticleTitle?: string;
|
|
6912
|
-
goodallExampleId?: string;
|
|
6913
|
-
/** Whether the document is opted out. */
|
|
6914
|
-
isOptOut?: boolean;
|
|
6915
|
-
/** When true, this attribution came from the user's prompt. */
|
|
6916
|
-
isPrompt?: boolean;
|
|
6917
|
-
lamdaExampleId?: string;
|
|
6918
|
-
license?: string;
|
|
6919
|
-
meenaConversationId?: string;
|
|
6920
|
-
/** Natural (not programming) language of the document. Language code as defined by http://www.unicode.org/reports/tr35/#Identifiers and https://tools.ietf.org/html/bcp47. Currently applicable to full-view books. Use docinfo-util.h to set & read language fields. See go/iii. */
|
|
6921
|
-
naturalLanguageCode?: string;
|
|
6922
|
-
/** True if this doc has no attribution information available. We use an explicit field for this instead of just implicitly leaving all the DocAttribution fields blank to distinguish a case where a bug/oversight has left the attribution information empty vs when we really have no attribution information available. */
|
|
6923
|
-
noAttribution?: boolean;
|
|
6924
|
-
podcastUtteranceId?: string;
|
|
6925
|
-
publicationDate?: GoogleTypeDate;
|
|
6926
|
-
/** This field is for opt-out experiment only, MUST never be used during actual production/serving. */
|
|
6927
|
-
qualityScoreExperimentOnly?: number;
|
|
6928
|
-
/** Github repository */
|
|
6929
|
-
repo?: string;
|
|
6930
|
-
/** URL of a webdoc */
|
|
6931
|
-
url?: string;
|
|
6932
|
-
volumeId?: string;
|
|
6933
|
-
/** Wikipedia article title. The Wikipedia TFDS dataset includes article titles but not URLs. While a URL is to the best of our knowledge a deterministic function of the title, we store the original title to reflect the information in the original dataset. */
|
|
6934
|
-
wikipediaArticleTitle?: string;
|
|
6935
|
-
youtubeVideoId?: string;
|
|
6936
|
-
}
|
|
6937
|
-
interface LearningGenaiRecitationRecitationResult {
|
|
6938
|
-
dynamicSegmentResults?: LearningGenaiRecitationSegmentResult[];
|
|
6939
|
-
/** The recitation action for one given input. When its segments contain different actions, the overall action will be returned in the precedence of BLOCK > CITE > NO_ACTION. When the given input is not found in any source, the recitation action will be NO_ACTION. */
|
|
6940
|
-
recitationAction?: string;
|
|
6941
|
-
trainingSegmentResults?: LearningGenaiRecitationSegmentResult[];
|
|
6942
|
-
}
|
|
6943
|
-
interface LearningGenaiRecitationSegmentResult {
|
|
6944
|
-
/** The dataset the segment came from. Datasets change often as model evolves. Treat this field as informational only and avoid depending on it directly. */
|
|
6945
|
-
attributionDataset?: string;
|
|
6946
|
-
/** human-friendly string that contains information from doc_attribution which could be shown by clients */
|
|
6947
|
-
displayAttributionMessage?: string;
|
|
6948
|
-
docAttribution?: LearningGenaiRecitationDocAttribution;
|
|
6949
|
-
/** number of documents that contained this segment */
|
|
6950
|
-
docOccurrences?: number;
|
|
6951
|
-
endIndex?: number;
|
|
6952
|
-
/** The raw text in the given input that is corresponding to the segment. It will be available only when 'return_segment_raw_text' is enabled in the request options. */
|
|
6953
|
-
rawText?: string;
|
|
6954
|
-
segmentRecitationAction?: string;
|
|
6955
|
-
/** The category of the source dataset where the segment came from. This is more stable than Dataset. */
|
|
6956
|
-
sourceCategory?: string;
|
|
6957
|
-
/** The segment boundary start (inclusive) and end index (exclusive) in the given text. In the streaming RPC, the indexes always start from the beginning of the first text in the entire stream. The indexes are measured in UTF-16 code units. */
|
|
6958
|
-
startIndex?: number;
|
|
6959
|
-
}
|
|
6960
|
-
interface LearningGenaiRootCalculationType {
|
|
6961
|
-
scoreType?: string;
|
|
6962
|
-
weights?: number;
|
|
6963
|
-
}
|
|
6964
|
-
interface LearningGenaiRootClassifierOutput {
|
|
6965
|
-
/** If set, this is the output of the first matching rule. */
|
|
6966
|
-
ruleOutput?: LearningGenaiRootRuleOutput;
|
|
6967
|
-
/** outputs of all matching rule. */
|
|
6968
|
-
ruleOutputs?: LearningGenaiRootRuleOutput[];
|
|
6969
|
-
/** The results of data_providers and metrics. */
|
|
6970
|
-
state?: LearningGenaiRootClassifierState;
|
|
6971
|
-
}
|
|
6972
|
-
interface LearningGenaiRootClassifierOutputSummary {
|
|
6973
|
-
metrics?: LearningGenaiRootMetricOutput[];
|
|
6974
|
-
/** Output of the first matching rule. */
|
|
6975
|
-
ruleOutput?: LearningGenaiRootRuleOutput;
|
|
6976
|
-
/** outputs of all matching rule. */
|
|
6977
|
-
ruleOutputs?: LearningGenaiRootRuleOutput[];
|
|
6978
|
-
}
|
|
6979
|
-
interface LearningGenaiRootClassifierState {
|
|
6980
|
-
dataProviderOutput?: LearningGenaiRootDataProviderOutput[];
|
|
6981
|
-
metricOutput?: LearningGenaiRootMetricOutput[];
|
|
6982
|
-
}
|
|
6983
|
-
interface LearningGenaiRootCodeyChatMetadata {
|
|
6984
|
-
/** Indicates the programming language of the code if the message is a code chunk. */
|
|
6985
|
-
codeLanguage?: string;
|
|
6986
|
-
}
|
|
6987
|
-
interface LearningGenaiRootCodeyCheckpoint {
|
|
6988
|
-
/** Metadata that describes what was truncated at this checkpoint. */
|
|
6989
|
-
codeyTruncatorMetadata?: LearningGenaiRootCodeyTruncatorMetadata;
|
|
6990
|
-
/** Current state of the sample after truncator. */
|
|
6991
|
-
currentSample?: string;
|
|
6992
|
-
/** Postprocessor run that yielded this checkpoint. */
|
|
6993
|
-
postInferenceStep?: string;
|
|
6994
|
-
}
|
|
6995
|
-
interface LearningGenaiRootCodeyCompletionMetadata {
|
|
6996
|
-
checkpoints?: LearningGenaiRootCodeyCheckpoint[];
|
|
6997
|
-
}
|
|
6998
|
-
interface LearningGenaiRootCodeyGenerationMetadata {
|
|
6999
|
-
/** Last state of the sample before getting dropped/returned. */
|
|
7000
|
-
output?: string;
|
|
7001
|
-
/** Last Codey postprocessing step for this sample before getting dropped/returned. */
|
|
7002
|
-
postInferenceStep?: string;
|
|
7003
|
-
}
|
|
7004
|
-
interface LearningGenaiRootCodeyOutput {
|
|
7005
|
-
codeyChatMetadata?: LearningGenaiRootCodeyChatMetadata;
|
|
7006
|
-
codeyCompletionMetadata?: LearningGenaiRootCodeyCompletionMetadata;
|
|
7007
|
-
codeyGenerationMetadata?: LearningGenaiRootCodeyGenerationMetadata;
|
|
7008
|
-
}
|
|
7009
|
-
interface LearningGenaiRootCodeyTruncatorMetadata {
|
|
7010
|
-
/** Index of the current sample that trims off truncated text. */
|
|
7011
|
-
cutoffIndex?: number;
|
|
7012
|
-
/** Text that was truncated at a specific checkpoint. */
|
|
7013
|
-
truncatedText?: string;
|
|
7014
|
-
}
|
|
7015
|
-
interface LearningGenaiRootControlDecodingConfigThreshold {
|
|
7016
|
-
policy?: string;
|
|
7017
|
-
scoreMax?: number;
|
|
7018
|
-
}
|
|
7019
|
-
interface LearningGenaiRootControlDecodingRecord {
|
|
7020
|
-
/** Prefixes feeded into scorer. */
|
|
7021
|
-
prefixes?: string;
|
|
7022
|
-
/** Per policy scores returned from Scorer. Expect to have the same number of scores as in `thresholds`. */
|
|
7023
|
-
scores?: LearningGenaiRootControlDecodingRecordPolicyScore[];
|
|
7024
|
-
/** Suffixes feeded into scorer. */
|
|
7025
|
-
suffiexes?: string;
|
|
7026
|
-
/** Per policy thresholds from user config. */
|
|
7027
|
-
thresholds?: LearningGenaiRootControlDecodingConfigThreshold[];
|
|
7028
|
-
}
|
|
7029
|
-
interface LearningGenaiRootControlDecodingRecordPolicyScore {
|
|
7030
|
-
policy?: string;
|
|
7031
|
-
score?: number;
|
|
7032
|
-
}
|
|
7033
|
-
interface LearningGenaiRootControlDecodingRecords {
|
|
7034
|
-
/** One ControlDecodingRecord record maps to one rewind. */
|
|
7035
|
-
records?: LearningGenaiRootControlDecodingRecord[];
|
|
7036
|
-
}
|
|
7037
|
-
interface LearningGenaiRootDataProviderOutput {
|
|
7038
|
-
name?: string;
|
|
7039
|
-
/** If set, this DataProvider failed and this is the error message. */
|
|
7040
|
-
status?: UtilStatusProto;
|
|
7041
|
-
}
|
|
7042
|
-
interface LearningGenaiRootFilterMetadata {
|
|
7043
|
-
/** Filter confidence. */
|
|
7044
|
-
confidence?: string;
|
|
7045
|
-
/** Debug info for the message. */
|
|
7046
|
-
debugInfo?: LearningGenaiRootFilterMetadataFilterDebugInfo;
|
|
7047
|
-
/** A fallback message chosen by the applied filter. */
|
|
7048
|
-
fallback?: string;
|
|
7049
|
-
/** Additional info for the filter. */
|
|
7050
|
-
info?: string;
|
|
7051
|
-
/** Name of the filter that triggered. */
|
|
7052
|
-
name?: string;
|
|
7053
|
-
/** Filter reason. */
|
|
7054
|
-
reason?: string;
|
|
7055
|
-
/** The input query or generated response that is getting filtered. */
|
|
7056
|
-
text?: string;
|
|
7057
|
-
}
|
|
7058
|
-
interface LearningGenaiRootFilterMetadataFilterDebugInfo {
|
|
7059
|
-
classifierOutput?: LearningGenaiRootClassifierOutput;
|
|
7060
|
-
defaultMetadata?: string;
|
|
7061
|
-
languageFilterResult?: LearningGenaiRootLanguageFilterResult;
|
|
7062
|
-
/** Safety filter output information for LLM Root RAI harm check. */
|
|
7063
|
-
raiOutput?: LearningGenaiRootRAIOutput;
|
|
7064
|
-
raiResult?: CloudAiNlLlmProtoServiceRaiResult;
|
|
7065
|
-
raiSignal?: CloudAiNlLlmProtoServiceRaiSignal;
|
|
7066
|
-
/** Number of rewinds by controlled decoding. */
|
|
7067
|
-
records?: LearningGenaiRootControlDecodingRecords;
|
|
7068
|
-
streamRecitationResult?: LanguageLabsAidaTrustRecitationProtoStreamRecitationResult;
|
|
7069
|
-
takedownResult?: LearningGenaiRootTakedownResult;
|
|
7070
|
-
toxicityResult?: LearningGenaiRootToxicityResult;
|
|
7071
|
-
}
|
|
7072
|
-
interface LearningGenaiRootGroundingMetadata {
|
|
7073
|
-
citations?: LearningGenaiRootGroundingMetadataCitation[];
|
|
7074
|
-
/** True if grounding is cancelled, for example, no facts being retrieved. */
|
|
7075
|
-
groundingCancelled?: boolean;
|
|
7076
|
-
searchQueries?: string[];
|
|
7077
|
-
}
|
|
7078
|
-
interface LearningGenaiRootGroundingMetadataCitation {
|
|
7079
|
-
/** Index in the prediction output where the citation ends (exclusive). Must be > start_index and <= len(output). */
|
|
7080
|
-
endIndex?: number;
|
|
7081
|
-
/** Index of the fact supporting this claim. Should be within the range of the `world_facts` in the GenerateResponse. */
|
|
7082
|
-
factIndex?: number;
|
|
7083
|
-
/** Confidence score of this entailment. Value is [0,1] with 1 is the most confidence. */
|
|
7084
|
-
score?: number;
|
|
7085
|
-
/** Index in the prediction output where the citation starts (inclusive). Must be >= 0 and < end_index. */
|
|
7086
|
-
startIndex?: number;
|
|
7087
|
-
}
|
|
7088
|
-
interface LearningGenaiRootHarm {
|
|
7089
|
-
/** Please do not use, this is still under development. */
|
|
7090
|
-
contextualDangerous?: boolean;
|
|
7091
|
-
csam?: boolean;
|
|
7092
|
-
fringe?: boolean;
|
|
7093
|
-
grailImageHarmType?: LearningGenaiRootHarmGrailImageHarmType;
|
|
7094
|
-
grailTextHarmType?: LearningGenaiRootHarmGrailTextHarmType;
|
|
7095
|
-
imageChild?: boolean;
|
|
7096
|
-
imageCsam?: boolean;
|
|
7097
|
-
imagePedo?: boolean;
|
|
7098
|
-
/** Image signals */
|
|
7099
|
-
imagePorn?: boolean;
|
|
7100
|
-
imageViolence?: boolean;
|
|
7101
|
-
pqc?: boolean;
|
|
7102
|
-
safetycat?: LearningGenaiRootHarmSafetyCatCategories;
|
|
7103
|
-
/** Spii Filter uses buckets http://google3/google/privacy/dlp/v2/storage.proto;l=77;rcl=584719820 to classify the input. LMRoot converts the bucket into double score. For example the score for "POSSIBLE" is 3 / 5 = 0.6 . */
|
|
7104
|
-
spii?: LearningGenaiRootHarmSpiiFilter;
|
|
7105
|
-
threshold?: number;
|
|
7106
|
-
videoFrameChild?: boolean;
|
|
7107
|
-
videoFrameCsam?: boolean;
|
|
7108
|
-
videoFramePedo?: boolean;
|
|
7109
|
-
/** Video frame signals */
|
|
7110
|
-
videoFramePorn?: boolean;
|
|
7111
|
-
videoFrameViolence?: boolean;
|
|
7112
|
-
}
|
|
7113
|
-
interface LearningGenaiRootHarmGrailImageHarmType {
|
|
7114
|
-
imageHarmType?: string[];
|
|
7115
|
-
}
|
|
7116
|
-
interface LearningGenaiRootHarmGrailTextHarmType {
|
|
7117
|
-
harmType?: string[];
|
|
7118
|
-
}
|
|
7119
|
-
interface LearningGenaiRootHarmSafetyCatCategories {
|
|
7120
|
-
categories?: string[];
|
|
7121
|
-
}
|
|
7122
|
-
interface LearningGenaiRootHarmSpiiFilter {
|
|
7123
|
-
usBankRoutingMicr?: boolean;
|
|
7124
|
-
usEmployerIdentificationNumber?: boolean;
|
|
7125
|
-
usSocialSecurityNumber?: boolean;
|
|
7126
|
-
}
|
|
7127
|
-
interface LearningGenaiRootInternalMetadata {
|
|
7128
|
-
scoredTokens?: LearningGenaiRootScoredToken[];
|
|
7129
|
-
}
|
|
7130
|
-
interface LearningGenaiRootLanguageFilterResult {
|
|
7131
|
-
/** False when query or response should be filtered out due to unsupported language. */
|
|
7132
|
-
allowed?: boolean;
|
|
7133
|
-
/** Language of the query or response. */
|
|
7134
|
-
detectedLanguage?: string;
|
|
7135
|
-
/** Probability of the language predicted as returned by LangID. */
|
|
7136
|
-
detectedLanguageProbability?: number;
|
|
7137
|
-
}
|
|
7138
|
-
interface LearningGenaiRootMetricOutput {
|
|
7139
|
-
debug?: string;
|
|
7140
|
-
/** Name of the metric. */
|
|
7141
|
-
name?: string;
|
|
7142
|
-
numericValue?: number;
|
|
7143
|
-
status?: UtilStatusProto;
|
|
7144
|
-
stringValue?: string;
|
|
7145
|
-
}
|
|
7146
|
-
interface LearningGenaiRootPerRequestProcessorDebugMetadataFactualityDebugMetadata {
|
|
7147
|
-
/** Latency spent on fact retrievals. There might be multiple retrievals from different fact providers. */
|
|
7148
|
-
factRetrievalMillisecondsByProvider?: {[P in string]: string};
|
|
7149
|
-
/** Latency spent on prompt2query. The procedure generates a search-friendly query given the original prompt. */
|
|
7150
|
-
prompt2queryMilliseconds?: string;
|
|
7151
|
-
}
|
|
7152
|
-
interface LearningGenaiRootRAIOutput {
|
|
7153
|
-
allowed?: boolean;
|
|
7154
|
-
harm?: LearningGenaiRootHarm;
|
|
7155
|
-
name?: string;
|
|
7156
|
-
score?: number;
|
|
7157
|
-
}
|
|
7158
|
-
interface LearningGenaiRootRegexTakedownResult {
|
|
7159
|
-
/** False when query or response should be taken down due to match with a blocked regex, true otherwise. */
|
|
7160
|
-
allowed?: boolean;
|
|
7161
|
-
/** Regex used to decide that query or response should be taken down. Empty when query or response is kept. */
|
|
7162
|
-
takedownRegex?: string;
|
|
7163
|
-
}
|
|
7164
|
-
interface LearningGenaiRootRequestResponseTakedownResult {
|
|
7165
|
-
/** False when response has to be taken down per above config. */
|
|
7166
|
-
allowed?: boolean;
|
|
7167
|
-
/** Regex used to match the request. */
|
|
7168
|
-
requestTakedownRegex?: string;
|
|
7169
|
-
/** Regex used to decide that response should be taken down. Empty when response is kept. */
|
|
7170
|
-
responseTakedownRegex?: string;
|
|
7171
|
-
}
|
|
7172
|
-
interface LearningGenaiRootRoutingDecision {
|
|
7173
|
-
metadata?: LearningGenaiRootRoutingDecisionMetadata;
|
|
7174
|
-
/** The selected model to route traffic to. */
|
|
7175
|
-
modelConfigId?: string;
|
|
7176
|
-
}
|
|
7177
|
-
interface LearningGenaiRootRoutingDecisionMetadata {
|
|
7178
|
-
scoreBasedRoutingMetadata?: LearningGenaiRootRoutingDecisionMetadataScoreBased;
|
|
7179
|
-
tokenLengthBasedRoutingMetadata?: LearningGenaiRootRoutingDecisionMetadataTokenLengthBased;
|
|
7180
|
-
}
|
|
7181
|
-
interface LearningGenaiRootRoutingDecisionMetadataScoreBased {
|
|
7182
|
-
/** The rule that was matched. */
|
|
7183
|
-
matchedRule?: LearningGenaiRootScoreBasedRoutingConfigRule;
|
|
7184
|
-
/** The score that was generated by the router i.e. the model. */
|
|
7185
|
-
score?: LearningGenaiRootScore;
|
|
7186
|
-
/** No rules were matched & therefore used the default fallback. */
|
|
7187
|
-
usedDefaultFallback?: boolean;
|
|
7188
|
-
}
|
|
7189
|
-
interface LearningGenaiRootRoutingDecisionMetadataTokenLengthBased {
|
|
7190
|
-
modelInputTokenMetadata?: LearningGenaiRootRoutingDecisionMetadataTokenLengthBasedModelInputTokenMetadata[];
|
|
7191
|
-
modelMaxTokenMetadata?: LearningGenaiRootRoutingDecisionMetadataTokenLengthBasedModelMaxTokenMetadata[];
|
|
7192
|
-
}
|
|
7193
|
-
interface LearningGenaiRootRoutingDecisionMetadataTokenLengthBasedModelInputTokenMetadata {
|
|
7194
|
-
/** The length computed by backends using the formatter & tokenizer specific to the model */
|
|
7195
|
-
computedInputTokenLength?: number;
|
|
7196
|
-
modelId?: string;
|
|
7197
|
-
/** If true, the model was selected as a fallback, since no model met requirements. */
|
|
7198
|
-
pickedAsFallback?: boolean;
|
|
7199
|
-
/** If true, the model was selected since it met the requriements. */
|
|
7200
|
-
selected?: boolean;
|
|
7201
|
-
}
|
|
7202
|
-
interface LearningGenaiRootRoutingDecisionMetadataTokenLengthBasedModelMaxTokenMetadata {
|
|
7203
|
-
maxNumInputTokens?: number;
|
|
7204
|
-
maxNumOutputTokens?: number;
|
|
7205
|
-
modelId?: string;
|
|
7206
|
-
}
|
|
7207
|
-
interface LearningGenaiRootRuleOutput {
|
|
7208
|
-
decision?: string;
|
|
7209
|
-
name?: string;
|
|
7210
|
-
}
|
|
7211
|
-
interface LearningGenaiRootScore {
|
|
7212
|
-
calculationType?: LearningGenaiRootCalculationType;
|
|
7213
|
-
/** The internal_metadata is intended to be used by internal processors and will be cleared before returns. */
|
|
7214
|
-
internalMetadata?: LearningGenaiRootInternalMetadata;
|
|
7215
|
-
thresholdType?: LearningGenaiRootThresholdType;
|
|
7216
|
-
/** Top candidate tokens and log probabilities at each decoding step. */
|
|
7217
|
-
tokensAndLogprobPerDecodingStep?: LearningGenaiRootTokensAndLogProbPerDecodingStep;
|
|
7218
|
-
value?: number;
|
|
7219
|
-
}
|
|
7220
|
-
interface LearningGenaiRootScoreBasedRoutingConfigRule {
|
|
7221
|
-
/** NOTE: Hardest examples have smaller values in their routing scores. */
|
|
7222
|
-
equalOrGreaterThan?: LearningGenaiRootScore;
|
|
7223
|
-
lessThan?: LearningGenaiRootScore;
|
|
7224
|
-
/** This model_config_id points to ModelConfig::id which allows us to find the ModelConfig to route to. This is part of the banks specified in the ModelBankConfig. */
|
|
7225
|
-
modelConfigId?: string;
|
|
7226
|
-
}
|
|
7227
|
-
interface LearningGenaiRootScoredSimilarityTakedownPhrase {
|
|
7228
|
-
phrase?: LearningGenaiRootSimilarityTakedownPhrase;
|
|
7229
|
-
similarityScore?: number;
|
|
7230
|
-
}
|
|
7231
|
-
interface LearningGenaiRootScoredToken {
|
|
7232
|
-
/** Each end_token_score is a logprob for how well the completion would end at a particular token. See http://google3/labs/language/aida/config/proto/model_config.proto;l=376;rcl=573039459 */
|
|
7233
|
-
endTokenScore?: number;
|
|
7234
|
-
/** Each score is the logprob for the token in model response. */
|
|
7235
|
-
score?: number;
|
|
7236
|
-
token?: string;
|
|
7237
|
-
}
|
|
7238
|
-
interface LearningGenaiRootSimilarityTakedownPhrase {
|
|
7239
|
-
blockedPhrase?: string;
|
|
7240
|
-
}
|
|
7241
|
-
interface LearningGenaiRootSimilarityTakedownResult {
|
|
7242
|
-
/** False when query or response should be taken down by any of the takedown rules, true otherwise. */
|
|
7243
|
-
allowed?: boolean;
|
|
7244
|
-
/** List of similar phrases with score. Set only if allowed=false. */
|
|
7245
|
-
scoredPhrases?: LearningGenaiRootScoredSimilarityTakedownPhrase[];
|
|
7246
|
-
}
|
|
7247
|
-
interface LearningGenaiRootTakedownResult {
|
|
7248
|
-
/** False when query or response should be taken down by any of the takedown rules, true otherwise. */
|
|
7249
|
-
allowed?: boolean;
|
|
7250
|
-
regexTakedownResult?: LearningGenaiRootRegexTakedownResult;
|
|
7251
|
-
requestResponseTakedownResult?: LearningGenaiRootRequestResponseTakedownResult;
|
|
7252
|
-
similarityTakedownResult?: LearningGenaiRootSimilarityTakedownResult;
|
|
7253
|
-
}
|
|
7254
|
-
interface LearningGenaiRootThresholdType {
|
|
7255
|
-
scoreType?: string;
|
|
7256
|
-
threshold?: number;
|
|
7257
|
-
}
|
|
7258
|
-
interface LearningGenaiRootTokensAndLogProbPerDecodingStep {
|
|
7259
|
-
/** Length = total number of decoding steps. The chosen candidates may or may not be in top_candidates. */
|
|
7260
|
-
chosenCandidates?: LearningGenaiRootTokensAndLogProbPerDecodingStepCandidate[];
|
|
7261
|
-
/** Length = total number of decoding steps. */
|
|
7262
|
-
topCandidates?: LearningGenaiRootTokensAndLogProbPerDecodingStepTopCandidates[];
|
|
7263
|
-
}
|
|
7264
|
-
interface LearningGenaiRootTokensAndLogProbPerDecodingStepCandidate {
|
|
7265
|
-
/** The candidate's log probability. */
|
|
7266
|
-
logProbability?: number;
|
|
7267
|
-
/** The candidate’s token value. */
|
|
7268
|
-
token?: string;
|
|
7269
|
-
}
|
|
7270
|
-
interface LearningGenaiRootTokensAndLogProbPerDecodingStepTopCandidates {
|
|
7271
|
-
/** Sorted by log probability in descending order. */
|
|
7272
|
-
candidates?: LearningGenaiRootTokensAndLogProbPerDecodingStepCandidate[];
|
|
7273
|
-
}
|
|
7274
|
-
interface LearningGenaiRootToxicityResult {
|
|
7275
|
-
signals?: LearningGenaiRootToxicitySignal[];
|
|
7276
|
-
}
|
|
7277
|
-
interface LearningGenaiRootToxicitySignal {
|
|
7278
|
-
allowed?: boolean;
|
|
7279
|
-
label?: string;
|
|
7280
|
-
score?: number;
|
|
7281
|
-
}
|
|
7282
|
-
interface LearningGenaiRootTranslationRequestInfo {
|
|
7283
|
-
/** The ISO-639 language code of source text in the initial request, detected automatically, if no source language was passed within the initial request. If the source language was passed, auto-detection of the language does not occur and this field is empty. */
|
|
7284
|
-
detectedLanguageCodes?: string[];
|
|
7285
|
-
/** The sum of the size of all the contents in the request. */
|
|
7286
|
-
totalContentSize?: string;
|
|
7287
|
-
}
|
|
7288
|
-
interface LearningServingLlmAtlasOutputMetadata {
|
|
7289
|
-
requestTopic?: string;
|
|
7290
|
-
source?: string;
|
|
7291
|
-
}
|
|
7292
|
-
interface LearningServingLlmMessageMetadata {
|
|
7293
|
-
atlasMetadata?: LearningServingLlmAtlasOutputMetadata;
|
|
7294
|
-
/** Summary of classifier output. We attach this to all messages regardless of whether classification rules triggered or not. */
|
|
7295
|
-
classifierSummary?: LearningGenaiRootClassifierOutputSummary;
|
|
7296
|
-
/** Contains metadata related to Codey Processors. */
|
|
7297
|
-
codeyOutput?: LearningGenaiRootCodeyOutput;
|
|
7298
|
-
currentStreamTextLength?: number;
|
|
7299
|
-
/** Whether the corresponding message has been deleted. */
|
|
7300
|
-
deleted?: boolean;
|
|
7301
|
-
/** Metadata for filters that triggered. */
|
|
7302
|
-
filterMeta?: LearningGenaiRootFilterMetadata[];
|
|
7303
|
-
/** This score is finally used for ranking the message. This will be same as the score present in `Message.score` field. */
|
|
7304
|
-
finalMessageScore?: LearningGenaiRootScore;
|
|
7305
|
-
/** NOT YET IMPLEMENTED. */
|
|
7306
|
-
finishReason?: string;
|
|
7307
|
-
groundingMetadata?: LearningGenaiRootGroundingMetadata;
|
|
7308
|
-
/** Applies to streaming response message only. Whether the message is a code. */
|
|
7309
|
-
isCode?: boolean;
|
|
7310
|
-
/** Applies to Response message only. Indicates whether the message is a fallback and the response would have otherwise been empty. */
|
|
7311
|
-
isFallback?: boolean;
|
|
7312
|
-
/** Result from nlp_saft DetectLanguage method. Currently the predicted language code and language probability is used. */
|
|
7313
|
-
langidResult?: NlpSaftLangIdResult;
|
|
7314
|
-
/** Detected language. */
|
|
7315
|
-
language?: string;
|
|
7316
|
-
/** The LM prefix used to generate this response. */
|
|
7317
|
-
lmPrefix?: string;
|
|
7318
|
-
/** The original text generated by LLM. This is the raw output for debugging purposes. */
|
|
7319
|
-
originalText?: string;
|
|
7320
|
-
/** Number of tokens decoded by the model as part of a stream. This count may be different from `per_stream_returned_token_count` which, is counted after any response rewriting or truncation. Applies to streaming response only. */
|
|
7321
|
-
perStreamDecodedTokenCount?: number;
|
|
7322
|
-
/** Number of tokens returned per stream in a response candidate after any response rewriting or truncation. Applies to streaming response only. Applies to Gemini models only. */
|
|
7323
|
-
perStreamReturnedTokenCount?: number;
|
|
7324
|
-
/** Results of running RAI on the query or this response candidate. One output per rai_config. It will be populated regardless of whether the threshold is exceeded or not. */
|
|
7325
|
-
raiOutputs?: LearningGenaiRootRAIOutput[];
|
|
7326
|
-
/** Recitation Results. It will be populated as long as Recitation processing is enabled, regardless of recitation outcome. */
|
|
7327
|
-
recitationResult?: LearningGenaiRecitationRecitationResult;
|
|
7328
|
-
/** NOT IMPLEMENTED TODO (b/334187574) Remove this field after Labs migrates to per_stream_returned_token_count and total_returned_token_count. */
|
|
7329
|
-
returnTokenCount?: number;
|
|
7330
|
-
/** All the different scores for a message are logged here. */
|
|
7331
|
-
scores?: LearningGenaiRootScore[];
|
|
7332
|
-
/** Whether the response is terminated during streaming return. Only used for streaming requests. */
|
|
7333
|
-
streamTerminated?: boolean;
|
|
7334
|
-
/** Total tokens decoded so far per response_candidate. For streaming: Count of all the tokens decoded so far (aggregated count). For unary: Count of all the tokens decoded per response_candidate. */
|
|
7335
|
-
totalDecodedTokenCount?: number;
|
|
7336
|
-
/** Total number of tokens returned in a response candidate. For streaming, it is the aggregated count (i.e. total so far) Applies to Gemini models only. */
|
|
7337
|
-
totalReturnedTokenCount?: number;
|
|
7338
|
-
/** Translated user-prompt used for RAI post processing. This is for internal processing only. We will translate in pre-processor and pass the translated text to the post processor using this field. It will be empty if non of the signals requested need translation. */
|
|
7339
|
-
translatedUserPrompts?: string[];
|
|
7340
|
-
/** The metadata from Vertex SafetyCat processors */
|
|
7341
|
-
vertexRaiResult?: CloudAiNlLlmProtoServiceRaiResult;
|
|
7342
|
-
}
|
|
7343
|
-
interface NlpSaftLangIdLocalesResult {
|
|
7344
|
-
/** List of locales in which the text would be considered acceptable. Sorted in descending order according to each locale's respective likelihood. For example, if a Portuguese text is acceptable in both Brazil and Portugal, but is more strongly associated with Brazil, then the predictions would be ["pt-BR", "pt-PT"], in that order. May be empty, indicating that the model did not predict any acceptable locales. */
|
|
7345
|
-
predictions?: NlpSaftLangIdLocalesResultLocale[];
|
|
7346
|
-
}
|
|
7347
|
-
interface NlpSaftLangIdLocalesResultLocale {
|
|
7348
|
-
/** A BCP 47 language code that includes region information. For example, "pt-BR" or "pt-PT". This field will always be populated. */
|
|
7349
|
-
languageCode?: string;
|
|
7350
|
-
}
|
|
7351
|
-
interface NlpSaftLangIdResult {
|
|
7352
|
-
/** The version of the model used to create these annotations. */
|
|
7353
|
-
modelVersion?: string;
|
|
7354
|
-
/** This field stores the n-best list of possible BCP 47 language code strings for a given input sorted in descending order according to each code's respective probability. */
|
|
7355
|
-
predictions?: NlpSaftLanguageSpan[];
|
|
7356
|
-
/** This field stores language predictions of subspans of the input, when available. Each LanguageSpanSequence is a sequence of LanguageSpans. A particular sequence of LanguageSpans has an associated probability, and need not necessarily cover the entire input. If no language could be predicted for any span, then this field may be empty. */
|
|
7357
|
-
spanPredictions?: NlpSaftLanguageSpanSequence[];
|
|
7358
|
-
}
|
|
7359
|
-
interface NlpSaftLanguageSpan {
|
|
7360
|
-
end?: number;
|
|
7361
|
-
/** A BCP 47 language code for this span. */
|
|
7362
|
-
languageCode?: string;
|
|
7363
|
-
/** Optional field containing any information that was predicted about the specific locale(s) of the span. */
|
|
7364
|
-
locales?: NlpSaftLangIdLocalesResult;
|
|
7365
|
-
/** A probability associated with this prediction. */
|
|
7366
|
-
probability?: number;
|
|
7367
|
-
/** Start and end byte offsets, inclusive, within the given input string. A value of -1 implies that this field is not set. Both fields must either be set with a nonnegative value or both are unset. If both are unset then this LanguageSpan applies to the entire input. */
|
|
7368
|
-
start?: number;
|
|
7369
|
-
}
|
|
7370
|
-
interface NlpSaftLanguageSpanSequence {
|
|
7371
|
-
/** A sequence of LanguageSpan objects, each assigning a language to a subspan of the input. */
|
|
7372
|
-
languageSpans?: NlpSaftLanguageSpan[];
|
|
7373
|
-
/** The probability of this sequence of LanguageSpans. */
|
|
7374
|
-
probability?: number;
|
|
7375
|
-
}
|
|
7376
|
-
interface Proto2BridgeMessageSet {}
|
|
7377
|
-
interface UtilStatusProto {
|
|
7378
|
-
/** The canonical error code (see codes.proto) that most closely corresponds to this status. This may be missing, and in the common case of the generic space, it definitely will be. */
|
|
7379
|
-
canonicalCode?: number;
|
|
7380
|
-
/** Numeric code drawn from the space specified below. Often, this is the canonical error space, and code is drawn from google3/util/task/codes.proto */
|
|
7381
|
-
code?: number;
|
|
7382
|
-
/** Detail message */
|
|
7383
|
-
message?: string;
|
|
7384
|
-
/** message_set associates an arbitrary proto message with the status. */
|
|
7385
|
-
messageSet?: any;
|
|
7386
|
-
/** The following are usually only present when code != 0 Space to which this status belongs */
|
|
7387
|
-
space?: string;
|
|
7388
|
-
}
|
|
7389
6570
|
interface BatchPredictionJobsResource {
|
|
7390
6571
|
/** Cancels a BatchPredictionJob. Starts asynchronous cancellation on the BatchPredictionJob. The server makes the best effort to cancel the job, but success is not guaranteed. Clients can use JobService.GetBatchPredictionJob or other methods to check whether the cancellation succeeded or whether the job completed despite cancellation. On a successful cancellation, the BatchPredictionJob is not deleted;instead its BatchPredictionJob.state is set to `CANCELLED`. Any files already outputted by the job are not deleted. */
|
|
7391
6572
|
cancel(request: {
|
|
@@ -9001,6 +8182,68 @@ declare namespace gapi.client {
|
|
|
9001
8182
|
/** Legacy upload protocol for media (e.g. "media", "multipart"). */
|
|
9002
8183
|
uploadType?: string;
|
|
9003
8184
|
}): Request<GoogleCloudAiplatformV1ListDatasetVersionsResponse>;
|
|
8185
|
+
/** Updates a DatasetVersion. */
|
|
8186
|
+
patch(request: {
|
|
8187
|
+
/** V1 error format. */
|
|
8188
|
+
'$.xgafv'?: string;
|
|
8189
|
+
/** OAuth access token. */
|
|
8190
|
+
access_token?: string;
|
|
8191
|
+
/** Data format for response. */
|
|
8192
|
+
alt?: string;
|
|
8193
|
+
/** JSONP */
|
|
8194
|
+
callback?: string;
|
|
8195
|
+
/** Selector specifying which fields to include in a partial response. */
|
|
8196
|
+
fields?: string;
|
|
8197
|
+
/** API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token. */
|
|
8198
|
+
key?: string;
|
|
8199
|
+
/** Output only. The resource name of the DatasetVersion. */
|
|
8200
|
+
name: string;
|
|
8201
|
+
/** OAuth 2.0 token for the current user. */
|
|
8202
|
+
oauth_token?: string;
|
|
8203
|
+
/** Returns response with indentations and line breaks. */
|
|
8204
|
+
prettyPrint?: boolean;
|
|
8205
|
+
/** Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters. */
|
|
8206
|
+
quotaUser?: string;
|
|
8207
|
+
/** Required. The update mask applies to the resource. For the `FieldMask` definition, see google.protobuf.FieldMask. Updatable fields: * `display_name` */
|
|
8208
|
+
updateMask?: string;
|
|
8209
|
+
/** Upload protocol for media (e.g. "raw", "multipart"). */
|
|
8210
|
+
upload_protocol?: string;
|
|
8211
|
+
/** Legacy upload protocol for media (e.g. "media", "multipart"). */
|
|
8212
|
+
uploadType?: string;
|
|
8213
|
+
/** Request body */
|
|
8214
|
+
resource: GoogleCloudAiplatformV1DatasetVersion;
|
|
8215
|
+
}): Request<GoogleCloudAiplatformV1DatasetVersion>;
|
|
8216
|
+
patch(
|
|
8217
|
+
request: {
|
|
8218
|
+
/** V1 error format. */
|
|
8219
|
+
'$.xgafv'?: string;
|
|
8220
|
+
/** OAuth access token. */
|
|
8221
|
+
access_token?: string;
|
|
8222
|
+
/** Data format for response. */
|
|
8223
|
+
alt?: string;
|
|
8224
|
+
/** JSONP */
|
|
8225
|
+
callback?: string;
|
|
8226
|
+
/** Selector specifying which fields to include in a partial response. */
|
|
8227
|
+
fields?: string;
|
|
8228
|
+
/** API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token. */
|
|
8229
|
+
key?: string;
|
|
8230
|
+
/** Output only. The resource name of the DatasetVersion. */
|
|
8231
|
+
name: string;
|
|
8232
|
+
/** OAuth 2.0 token for the current user. */
|
|
8233
|
+
oauth_token?: string;
|
|
8234
|
+
/** Returns response with indentations and line breaks. */
|
|
8235
|
+
prettyPrint?: boolean;
|
|
8236
|
+
/** Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters. */
|
|
8237
|
+
quotaUser?: string;
|
|
8238
|
+
/** Required. The update mask applies to the resource. For the `FieldMask` definition, see google.protobuf.FieldMask. Updatable fields: * `display_name` */
|
|
8239
|
+
updateMask?: string;
|
|
8240
|
+
/** Upload protocol for media (e.g. "raw", "multipart"). */
|
|
8241
|
+
upload_protocol?: string;
|
|
8242
|
+
/** Legacy upload protocol for media (e.g. "media", "multipart"). */
|
|
8243
|
+
uploadType?: string;
|
|
8244
|
+
},
|
|
8245
|
+
body: GoogleCloudAiplatformV1DatasetVersion
|
|
8246
|
+
): Request<GoogleCloudAiplatformV1DatasetVersion>;
|
|
9004
8247
|
/** Restores a dataset version. */
|
|
9005
8248
|
restore(request?: {
|
|
9006
8249
|
/** V1 error format. */
|