@maxim_mazurok/gapi.client.aiplatform-v1beta1 0.0.20250701 → 0.0.20250711
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/index.d.ts +21 -5
- package/package.json +1 -1
package/index.d.ts
CHANGED
|
@@ -9,7 +9,7 @@
|
|
|
9
9
|
// This file was generated by https://github.com/Maxim-Mazurok/google-api-typings-generator. Please do not edit it manually.
|
|
10
10
|
// In case of any problems please post issue to https://github.com/Maxim-Mazurok/google-api-typings-generator
|
|
11
11
|
// Generated from: https://aiplatform.googleapis.com/$discovery/rest?version=v1beta1
|
|
12
|
-
// Revision:
|
|
12
|
+
// Revision: 20250711
|
|
13
13
|
|
|
14
14
|
/// <reference types="gapi.client" />
|
|
15
15
|
|
|
@@ -1453,6 +1453,8 @@ declare namespace gapi.client {
|
|
|
1453
1453
|
interface GoogleCloudAiplatformV1beta1DedicatedResources {
|
|
1454
1454
|
/** Immutable. The metric specifications that overrides a resource utilization metric (CPU utilization, accelerator's duty cycle, and so on) target value (default to 60 if not set). At most one entry is allowed per metric. If machine_spec.accelerator_count is above 0, the autoscaling will be based on both CPU utilization and accelerator's duty cycle metrics and scale up when either metrics exceeds its target value while scale down if both metrics are under their target value. The default target value is 60 for both metrics. If machine_spec.accelerator_count is 0, the autoscaling will be based on CPU utilization metric only with default target value 60 if not explicitly set. For example, in the case of Online Prediction, if you want to override target CPU utilization to 80, you should set autoscaling_metric_specs.metric_name to `aiplatform.googleapis.com/prediction/online/cpu/utilization` and autoscaling_metric_specs.target to `80`. */
|
|
1455
1455
|
autoscalingMetricSpecs?: GoogleCloudAiplatformV1beta1AutoscalingMetricSpec[];
|
|
1456
|
+
/** Optional. Immutable. If set, use DWS resource to schedule the deployment workload. reference: (https://cloud.google.com/blog/products/compute/introducing-dynamic-workload-scheduler) */
|
|
1457
|
+
flexStart?: GoogleCloudAiplatformV1beta1FlexStart;
|
|
1456
1458
|
/** Required. Immutable. The specification of a single machine being used. */
|
|
1457
1459
|
machineSpec?: GoogleCloudAiplatformV1beta1MachineSpec;
|
|
1458
1460
|
/** Immutable. The maximum number of replicas that may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale to that many replicas is guaranteed (barring service outages). If traffic increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, will use min_replica_count as the default value. The value of this field impacts the charge against Vertex CPU and GPU quotas. Specifically, you will be charged for (max_replica_count * number of cores in the selected machine type) and (max_replica_count * number of GPUs per replica in the selected machine type). */
|
|
@@ -1529,6 +1531,8 @@ declare namespace gapi.client {
|
|
|
1529
1531
|
displayName?: string;
|
|
1530
1532
|
/** Optional. If true, private endpoint's access logs are sent to Cloud Logging. These logs are like standard server access logs, containing information like timestamp and latency for each MatchRequest. Note that logs may incur a cost, especially if the deployed index receives a high queries per second rate (QPS). Estimate your costs before enabling this option. */
|
|
1531
1533
|
enableAccessLogging?: boolean;
|
|
1534
|
+
/** Optional. If true, logs to Cloud Logging errors relating to datapoint upserts. Under normal operation conditions, these log entries should be very rare. However, if incompatible datapoint updates are being uploaded to an index, a high volume of log entries may be generated in a short period of time. Note that logs may incur a cost, especially if the deployed index receives a high volume of datapoint upserts. Estimate your costs before enabling this option. */
|
|
1535
|
+
enableDatapointUpsertLogging?: boolean;
|
|
1532
1536
|
/** Required. The user specified ID of the DeployedIndex. The ID can be up to 128 characters long and must start with a letter and only contain letters, numbers, and underscores. The ID must be unique within the project it is created in. */
|
|
1533
1537
|
id?: string;
|
|
1534
1538
|
/** Required. The name of the Index this is the deployment of. We may refer to this Index as the DeployedIndex's "original" Index. */
|
|
@@ -2103,6 +2107,8 @@ declare namespace gapi.client {
|
|
|
2103
2107
|
interface GoogleCloudAiplatformV1beta1EventMetadata {
|
|
2104
2108
|
/** Optional. The branch of the event. The format is like agent_1.agent_2.agent_3, where agent_1 is the parent of agent_2, and agent_2 is the parent of agent_3. Branch is used when multiple child agents shouldn't see their siblings' conversation history. */
|
|
2105
2109
|
branch?: string;
|
|
2110
|
+
/** The custom metadata of the LlmResponse. */
|
|
2111
|
+
customMetadata?: {[P in string]: any};
|
|
2106
2112
|
/** Optional. Metadata returned to client when grounding is enabled. */
|
|
2107
2113
|
groundingMetadata?: GoogleCloudAiplatformV1beta1GroundingMetadata;
|
|
2108
2114
|
/** Optional. Flag indicating that LLM was interrupted when generating the content. Usually it's due to user interruption during a bidi streaming. */
|
|
@@ -2940,7 +2946,7 @@ declare namespace gapi.client {
|
|
|
2940
2946
|
values?: GoogleCloudAiplatformV1beta1FeatureValue[];
|
|
2941
2947
|
}
|
|
2942
2948
|
interface GoogleCloudAiplatformV1beta1FeatureValueMetadata {
|
|
2943
|
-
/** Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. */
|
|
2949
|
+
/** Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. Legacy Feature Store: For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. */
|
|
2944
2950
|
generateTime?: string;
|
|
2945
2951
|
}
|
|
2946
2952
|
interface GoogleCloudAiplatformV1beta1FeatureView {
|
|
@@ -3235,6 +3241,10 @@ declare namespace gapi.client {
|
|
|
3235
3241
|
/** The distance between the neighbor and the query sparse_embedding. */
|
|
3236
3242
|
sparseDistance?: number;
|
|
3237
3243
|
}
|
|
3244
|
+
interface GoogleCloudAiplatformV1beta1FlexStart {
|
|
3245
|
+
/** The max duration of the deployment is max_runtime_duration. The deployment will be terminated after the duration. The max_runtime_duration can be set up to 7 days. */
|
|
3246
|
+
maxRuntimeDuration?: string;
|
|
3247
|
+
}
|
|
3238
3248
|
interface GoogleCloudAiplatformV1beta1FluencyInput {
|
|
3239
3249
|
/** Required. Fluency instance. */
|
|
3240
3250
|
instance?: GoogleCloudAiplatformV1beta1FluencyInstance;
|
|
@@ -6125,6 +6135,8 @@ declare namespace gapi.client {
|
|
|
6125
6135
|
taskId?: string;
|
|
6126
6136
|
/** Output only. The user specified name of the task that is defined in pipeline_spec. */
|
|
6127
6137
|
taskName?: string;
|
|
6138
|
+
/** Output only. The unique name of a task. This field is used by rerun pipeline job. Console UI and Vertex AI SDK will support triggering pipeline job reruns. The name is constructed by concatenating all the parent tasks name with the task name. For example, if a task named "child_task" has a parent task named "parent_task_1" and parent task 1 has a parent task named "parent_task_2", the task unique name will be "parent_task_2.parent_task_1.child_task". */
|
|
6139
|
+
taskUniqueName?: string;
|
|
6128
6140
|
}
|
|
6129
6141
|
interface GoogleCloudAiplatformV1beta1PipelineTaskDetailArtifactList {
|
|
6130
6142
|
/** Output only. A list of artifact metadata. */
|
|
@@ -7212,11 +7224,11 @@ declare namespace gapi.client {
|
|
|
7212
7224
|
similaritySearchConfig?: GoogleCloudAiplatformV1beta1ReasoningEngineContextSpecMemoryBankConfigSimilaritySearchConfig;
|
|
7213
7225
|
}
|
|
7214
7226
|
interface GoogleCloudAiplatformV1beta1ReasoningEngineContextSpecMemoryBankConfigGenerationConfig {
|
|
7215
|
-
/** Required. The model used to generate memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}
|
|
7227
|
+
/** Required. The model used to generate memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}`. */
|
|
7216
7228
|
model?: string;
|
|
7217
7229
|
}
|
|
7218
7230
|
interface GoogleCloudAiplatformV1beta1ReasoningEngineContextSpecMemoryBankConfigSimilaritySearchConfig {
|
|
7219
|
-
/** Required. The model used to generate embeddings to lookup similar memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}
|
|
7231
|
+
/** Required. The model used to generate embeddings to lookup similar memories. Format: `projects/{project}/locations/{location}/publishers/google/models/{model}`. */
|
|
7220
7232
|
embeddingModel?: string;
|
|
7221
7233
|
}
|
|
7222
7234
|
interface GoogleCloudAiplatformV1beta1ReasoningEngineSpec {
|
|
@@ -10490,7 +10502,7 @@ declare namespace gapi.client {
|
|
|
10490
10502
|
checkpoints?: GoogleCloudAiplatformV1beta1TunedModelCheckpoint[];
|
|
10491
10503
|
/** Output only. A resource name of an Endpoint. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}`. */
|
|
10492
10504
|
endpoint?: string;
|
|
10493
|
-
/** Output only. The resource name of the TunedModel. Format: `projects/{project}/locations/{location}/models/{model}
|
|
10505
|
+
/** Output only. The resource name of the TunedModel. Format: `projects/{project}/locations/{location}/models/{model}@{version_id}` When tuning from a base model, the version_id will be 1. For continuous tuning, the version id will be incremented by 1 from the last version id in the parent model. E.g., `projects/{project}/locations/{location}/models/{model}@{last_version_id + 1}` */
|
|
10494
10506
|
model?: string;
|
|
10495
10507
|
}
|
|
10496
10508
|
interface GoogleCloudAiplatformV1beta1TunedModelCheckpoint {
|
|
@@ -34481,6 +34493,8 @@ declare namespace gapi.client {
|
|
|
34481
34493
|
callback?: string;
|
|
34482
34494
|
/** Selector specifying which fields to include in a partial response. */
|
|
34483
34495
|
fields?: string;
|
|
34496
|
+
/** Optional. The standard list filter. Supported fields: * `timestamp` range (i.e. `timestamp>="2025-01-31T11:30:00-04:00"` where the timestamp is in RFC 3339 format) More detail in [AIP-160](https://google.aip.dev/160). */
|
|
34497
|
+
filter?: string;
|
|
34484
34498
|
/** API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token. */
|
|
34485
34499
|
key?: string;
|
|
34486
34500
|
/** OAuth 2.0 token for the current user. */
|
|
@@ -41297,6 +41311,8 @@ declare namespace gapi.client {
|
|
|
41297
41311
|
callback?: string;
|
|
41298
41312
|
/** Selector specifying which fields to include in a partial response. */
|
|
41299
41313
|
fields?: string;
|
|
41314
|
+
/** Optional. The standard list filter. Supported fields: * `timestamp` range (i.e. `timestamp>="2025-01-31T11:30:00-04:00"` where the timestamp is in RFC 3339 format) More detail in [AIP-160](https://google.aip.dev/160). */
|
|
41315
|
+
filter?: string;
|
|
41300
41316
|
/** API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token. */
|
|
41301
41317
|
key?: string;
|
|
41302
41318
|
/** OAuth 2.0 token for the current user. */
|