@maxim_mazurok/gapi.client.aiplatform-v1beta1 0.0.20250205 → 0.0.20250212

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (2) hide show
  1. package/index.d.ts +104 -18
  2. package/package.json +1 -1
package/index.d.ts CHANGED
@@ -9,7 +9,7 @@
9
9
  // This file was generated by https://github.com/Maxim-Mazurok/google-api-typings-generator. Please do not edit it manually.
10
10
  // In case of any problems please post issue to https://github.com/Maxim-Mazurok/google-api-typings-generator
11
11
  // Generated from: https://aiplatform.googleapis.com/$discovery/rest?version=v1beta1
12
- // Revision: 20250205
12
+ // Revision: 20250212
13
13
 
14
14
  /// <reference types="gapi.client" />
15
15
 
@@ -350,9 +350,9 @@ declare namespace gapi.client {
350
350
  serviceAccount?: string;
351
351
  }
352
352
  interface GoogleCloudAiplatformV1beta1AutomaticResources {
353
- /** Immutable. The maximum number of replicas this DeployedModel may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale the model to that many replicas is guaranteed (barring service outages). If traffic against the DeployedModel increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, a no upper bound for scaling under heavy traffic will be assume, though Vertex AI may be unable to scale beyond certain replica number. */
353
+ /** Immutable. The maximum number of replicas that may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale to that many replicas is guaranteed (barring service outages). If traffic increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, a no upper bound for scaling under heavy traffic will be assume, though Vertex AI may be unable to scale beyond certain replica number. */
354
354
  maxReplicaCount?: number;
355
- /** Immutable. The minimum number of replicas this DeployedModel will be always deployed on. If traffic against it increases, it may dynamically be deployed onto more replicas up to max_replica_count, and as traffic decreases, some of these extra replicas may be freed. If the requested value is too large, the deployment will error. */
355
+ /** Immutable. The minimum number of replicas that will be always deployed on. If traffic against it increases, it may dynamically be deployed onto more replicas up to max_replica_count, and as traffic decreases, some of these extra replicas may be freed. If the requested value is too large, the deployment will error. */
356
356
  minReplicaCount?: number;
357
357
  }
358
358
  interface GoogleCloudAiplatformV1beta1AutoraterConfig {
@@ -663,7 +663,7 @@ declare namespace gapi.client {
663
663
  displayName?: string;
664
664
  /** Timestamp of when this resource is considered expired. This is *always* provided on output, regardless of what was sent on input. */
665
665
  expireTime?: string;
666
- /** Immutable. The name of the publisher model to use for cached content. Format: projects/{project}/locations/{location}/publishers/{publisher}/models/{model} */
666
+ /** Immutable. The name of the `Model` to use for cached content. Currently, only the published Gemini base models are supported, in form of projects/{PROJECT}/locations/{LOCATION}/publishers/google/models/{MODEL} */
667
667
  model?: string;
668
668
  /** Immutable. Identifier. The server-generated resource name of the cached content Format: projects/{project}/locations/{location}/cachedContents/{cached_content} */
669
669
  name?: string;
@@ -1344,13 +1344,13 @@ declare namespace gapi.client {
1344
1344
  interface GoogleCloudAiplatformV1beta1DedicatedResources {
1345
1345
  /** Immutable. The metric specifications that overrides a resource utilization metric (CPU utilization, accelerator's duty cycle, and so on) target value (default to 60 if not set). At most one entry is allowed per metric. If machine_spec.accelerator_count is above 0, the autoscaling will be based on both CPU utilization and accelerator's duty cycle metrics and scale up when either metrics exceeds its target value while scale down if both metrics are under their target value. The default target value is 60 for both metrics. If machine_spec.accelerator_count is 0, the autoscaling will be based on CPU utilization metric only with default target value 60 if not explicitly set. For example, in the case of Online Prediction, if you want to override target CPU utilization to 80, you should set autoscaling_metric_specs.metric_name to `aiplatform.googleapis.com/prediction/online/cpu/utilization` and autoscaling_metric_specs.target to `80`. */
1346
1346
  autoscalingMetricSpecs?: GoogleCloudAiplatformV1beta1AutoscalingMetricSpec[];
1347
- /** Required. Immutable. The specification of a single machine used by the prediction. */
1347
+ /** Required. Immutable. The specification of a single machine being used. */
1348
1348
  machineSpec?: GoogleCloudAiplatformV1beta1MachineSpec;
1349
- /** Immutable. The maximum number of replicas this DeployedModel may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale the model to that many replicas is guaranteed (barring service outages). If traffic against the DeployedModel increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, will use min_replica_count as the default value. The value of this field impacts the charge against Vertex CPU and GPU quotas. Specifically, you will be charged for (max_replica_count * number of cores in the selected machine type) and (max_replica_count * number of GPUs per replica in the selected machine type). */
1349
+ /** Immutable. The maximum number of replicas that may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale to that many replicas is guaranteed (barring service outages). If traffic increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, will use min_replica_count as the default value. The value of this field impacts the charge against Vertex CPU and GPU quotas. Specifically, you will be charged for (max_replica_count * number of cores in the selected machine type) and (max_replica_count * number of GPUs per replica in the selected machine type). */
1350
1350
  maxReplicaCount?: number;
1351
- /** Required. Immutable. The minimum number of machine replicas this DeployedModel will be always deployed on. This value must be greater than or equal to 1. If traffic against the DeployedModel increases, it may dynamically be deployed onto more replicas, and as traffic decreases, some of these extra replicas may be freed. */
1351
+ /** Required. Immutable. The minimum number of machine replicas that will be always deployed on. This value must be greater than or equal to 1. If traffic increases, it may dynamically be deployed onto more replicas, and as traffic decreases, some of these extra replicas may be freed. */
1352
1352
  minReplicaCount?: number;
1353
- /** Optional. Number of required available replicas for the deployment to succeed. This field is only needed when partial model deployment/mutation is desired. If set, the model deploy/mutate operation will succeed once available_replica_count reaches required_replica_count, and the rest of the replicas will be retried. If not set, the default required_replica_count will be min_replica_count. */
1353
+ /** Optional. Number of required available replicas for the deployment to succeed. This field is only needed when partial deployment/mutation is desired. If set, the deploy/mutate operation will succeed once available_replica_count reaches required_replica_count, and the rest of the replicas will be retried. If not set, the default required_replica_count will be min_replica_count. */
1354
1354
  requiredReplicaCount?: number;
1355
1355
  /** Optional. If true, schedule the deployment workload on [spot VMs](https://cloud.google.com/kubernetes-engine/docs/concepts/spot-vms). */
1356
1356
  spot?: boolean;
@@ -1560,7 +1560,7 @@ declare namespace gapi.client {
1560
1560
  endpointDisplayName?: string;
1561
1561
  /** Optional. The Hugging Face read access token used to access the model artifacts of gated models. */
1562
1562
  huggingFaceAccessToken?: string;
1563
- /** Required. The name of the PublisherModel resource. Format: `publishers/{publisher}/models/{publisher_model}@{version_id}`, or `publishers/hf-{hugging-face-author}/models/{hugging-face-model-name}@001` or Hugging Face model ID like `google/gemma-2-2b-it`. */
1563
+ /** Required. The model to deploy. Format: 1. `publishers/{publisher}/models/{publisher_model}@{version_id}`, or `publishers/hf-{hugging-face-author}/models/{hugging-face-model-name}@001`. 2. Hugging Face model ID like `google/gemma-2-2b-it`. 3. Custom model Google Cloud Storage URI like `gs://bucket`. 4. Custom model zip file like `https://abc.com/a.zip`. */
1564
1564
  model?: string;
1565
1565
  /** Optional. The user-specified display name of the uploaded model. If not set, a default name will be used. */
1566
1566
  modelDisplayName?: string;
@@ -1762,6 +1762,16 @@ declare namespace gapi.client {
1762
1762
  /** Explanation type. For AutoML Image Classification models, possible values are: * `image-integrated-gradients` * `image-xrai` */
1763
1763
  explanationType?: string;
1764
1764
  }
1765
+ interface GoogleCloudAiplatformV1beta1EvaluateDatasetRequest {
1766
+ /** Optional. Autorater config used for evaluation. */
1767
+ autoraterConfig?: GoogleCloudAiplatformV1beta1AutoraterConfig;
1768
+ /** Required. The dataset used for evaluation. */
1769
+ dataset?: GoogleCloudAiplatformV1beta1EvaluationDataset;
1770
+ /** Required. The metrics used for evaluation. */
1771
+ metrics?: GoogleCloudAiplatformV1beta1Metric[];
1772
+ /** Required. Config for evaluation output. */
1773
+ outputConfig?: GoogleCloudAiplatformV1beta1OutputConfig;
1774
+ }
1765
1775
  interface GoogleCloudAiplatformV1beta1EvaluateInstancesRequest {
1766
1776
  /** Optional. Autorater config used for evaluation. */
1767
1777
  autoraterConfig?: GoogleCloudAiplatformV1beta1AutoraterConfig;
@@ -1892,6 +1902,12 @@ declare namespace gapi.client {
1892
1902
  /** Results for trajectory single tool use metric. */
1893
1903
  trajectorySingleToolUseResults?: GoogleCloudAiplatformV1beta1TrajectorySingleToolUseResults;
1894
1904
  }
1905
+ interface GoogleCloudAiplatformV1beta1EvaluationDataset {
1906
+ /** BigQuery source holds the dataset. */
1907
+ bigquerySource?: GoogleCloudAiplatformV1beta1BigQuerySource;
1908
+ /** Cloud storage source holds the dataset. */
1909
+ gcsSource?: GoogleCloudAiplatformV1beta1GcsSource;
1910
+ }
1895
1911
  interface GoogleCloudAiplatformV1beta1Event {
1896
1912
  /** Required. The relative resource name of the Artifact in the Event. */
1897
1913
  artifact?: string;
@@ -3121,8 +3137,6 @@ declare namespace gapi.client {
3121
3137
  stopSequences?: string[];
3122
3138
  /** Optional. Controls the randomness of predictions. */
3123
3139
  temperature?: number;
3124
- /** Optional. Config for thinking features. An error will be returned if this field is set for models that don't support thinking. */
3125
- thinkingConfig?: GoogleCloudAiplatformV1beta1GenerationConfigThinkingConfig;
3126
3140
  /** Optional. If specified, top-k sampling will be used. */
3127
3141
  topK?: number;
3128
3142
  /** Optional. If specified, nucleus sampling will be used. */
@@ -3142,10 +3156,6 @@ declare namespace gapi.client {
3142
3156
  /** The model name to use. Only the public LLM models are accepted. e.g. 'gemini-1.5-pro-001'. */
3143
3157
  modelName?: string;
3144
3158
  }
3145
- interface GoogleCloudAiplatformV1beta1GenerationConfigThinkingConfig {
3146
- /** Optional. Indicates whether to include thoughts in the response. If true, thoughts are returned only when available. */
3147
- includeThoughts?: boolean;
3148
- }
3149
3159
  interface GoogleCloudAiplatformV1beta1GenericOperationMetadata {
3150
3160
  /** Output only. Time when the operation was created. */
3151
3161
  createTime?: string;
@@ -3965,8 +3975,6 @@ declare namespace gapi.client {
3965
3975
  acceleratorType?: string;
3966
3976
  /** Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For DeployedModel this field is optional, and the default value is `n1-standard-2`. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. */
3967
3977
  machineType?: string;
3968
- /** Optional. Immutable. The number of nodes per replica for multihost GPU deployments. */
3969
- multihostGpuNodeCount?: number;
3970
3978
  /** Optional. Immutable. Configuration controlling how this resource pool consumes reservation. */
3971
3979
  reservationAffinity?: GoogleCloudAiplatformV1beta1ReservationAffinity;
3972
3980
  /** Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). */
@@ -4032,6 +4040,20 @@ declare namespace gapi.client {
4032
4040
  /** The disk utilization of the MetadataStore in bytes. */
4033
4041
  diskUtilizationBytes?: string;
4034
4042
  }
4043
+ interface GoogleCloudAiplatformV1beta1Metric {
4044
+ /** Optional. The aggregation metrics to use. */
4045
+ aggregationMetrics?: string[];
4046
+ /** Spec for bleu metric. */
4047
+ bleuSpec?: GoogleCloudAiplatformV1beta1BleuSpec;
4048
+ /** Spec for exact match metric. */
4049
+ exactMatchSpec?: any;
4050
+ /** Spec for pairwise metric. */
4051
+ pairwiseMetricSpec?: GoogleCloudAiplatformV1beta1PairwiseMetricSpec;
4052
+ /** Spec for pointwise metric. */
4053
+ pointwiseMetricSpec?: GoogleCloudAiplatformV1beta1PointwiseMetricSpec;
4054
+ /** Spec for rouge metric. */
4055
+ rougeSpec?: GoogleCloudAiplatformV1beta1RougeSpec;
4056
+ }
4035
4057
  interface GoogleCloudAiplatformV1beta1MetricxInput {
4036
4058
  /** Required. Metricx instance. */
4037
4059
  instance?: GoogleCloudAiplatformV1beta1MetricxInstance;
@@ -4847,6 +4869,8 @@ declare namespace gapi.client {
4847
4869
  checkpointId?: string;
4848
4870
  /** The epoch of the checkpoint. */
4849
4871
  epoch?: string;
4872
+ /** Identifier. The resource name of the ModelVersionCheckpoint. Format: `projects/{project}/locations/{location}/models/{model}/versions/{version}/checkpoints/{checkpoint}` */
4873
+ name?: string;
4850
4874
  /** The step of the checkpoint. */
4851
4875
  step?: string;
4852
4876
  }
@@ -5277,6 +5301,10 @@ declare namespace gapi.client {
5277
5301
  /** Optional. Post startup script config. */
5278
5302
  postStartupScriptConfig?: GoogleCloudAiplatformV1beta1PostStartupScriptConfig;
5279
5303
  }
5304
+ interface GoogleCloudAiplatformV1beta1OutputConfig {
5305
+ /** Cloud storage destination for evaluation output. */
5306
+ gcsDestination?: GoogleCloudAiplatformV1beta1GcsDestination;
5307
+ }
5280
5308
  interface GoogleCloudAiplatformV1beta1PairwiseMetricInput {
5281
5309
  /** Required. Pairwise metric instance. */
5282
5310
  instance?: GoogleCloudAiplatformV1beta1PairwiseMetricInstance;
@@ -7717,7 +7745,7 @@ declare namespace gapi.client {
7717
7745
  displayName?: string;
7718
7746
  }
7719
7747
  interface GoogleCloudAiplatformV1beta1SchemaTextDataItem {
7720
- /** Output only. Google Cloud Storage URI points to the original text in user's bucket. The text file is up to 10MB in size. */
7748
+ /** Output only. Google Cloud Storage URI points to a copy of the original text in the Vertex-managed bucket in the user's project. The text file is up to 10MB in size. */
7721
7749
  gcsUri?: string;
7722
7750
  }
7723
7751
  interface GoogleCloudAiplatformV1beta1SchemaTextDatasetMetadata {
@@ -36076,6 +36104,64 @@ declare namespace gapi.client {
36076
36104
  },
36077
36105
  body: GoogleCloudAiplatformV1beta1DeployPublisherModelRequest,
36078
36106
  ): Request<GoogleLongrunningOperation>;
36107
+ /** Evaluates a dataset based on a set of given metrics. */
36108
+ evaluateDataset(request: {
36109
+ /** V1 error format. */
36110
+ '$.xgafv'?: string;
36111
+ /** OAuth access token. */
36112
+ access_token?: string;
36113
+ /** Data format for response. */
36114
+ alt?: string;
36115
+ /** JSONP */
36116
+ callback?: string;
36117
+ /** Selector specifying which fields to include in a partial response. */
36118
+ fields?: string;
36119
+ /** API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token. */
36120
+ key?: string;
36121
+ /** Required. The resource name of the Location to evaluate the dataset. Format: `projects/{project}/locations/{location}` */
36122
+ location: string;
36123
+ /** OAuth 2.0 token for the current user. */
36124
+ oauth_token?: string;
36125
+ /** Returns response with indentations and line breaks. */
36126
+ prettyPrint?: boolean;
36127
+ /** Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters. */
36128
+ quotaUser?: string;
36129
+ /** Upload protocol for media (e.g. "raw", "multipart"). */
36130
+ upload_protocol?: string;
36131
+ /** Legacy upload protocol for media (e.g. "media", "multipart"). */
36132
+ uploadType?: string;
36133
+ /** Request body */
36134
+ resource: GoogleCloudAiplatformV1beta1EvaluateDatasetRequest;
36135
+ }): Request<GoogleLongrunningOperation>;
36136
+ evaluateDataset(
36137
+ request: {
36138
+ /** V1 error format. */
36139
+ '$.xgafv'?: string;
36140
+ /** OAuth access token. */
36141
+ access_token?: string;
36142
+ /** Data format for response. */
36143
+ alt?: string;
36144
+ /** JSONP */
36145
+ callback?: string;
36146
+ /** Selector specifying which fields to include in a partial response. */
36147
+ fields?: string;
36148
+ /** API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token. */
36149
+ key?: string;
36150
+ /** Required. The resource name of the Location to evaluate the dataset. Format: `projects/{project}/locations/{location}` */
36151
+ location: string;
36152
+ /** OAuth 2.0 token for the current user. */
36153
+ oauth_token?: string;
36154
+ /** Returns response with indentations and line breaks. */
36155
+ prettyPrint?: boolean;
36156
+ /** Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters. */
36157
+ quotaUser?: string;
36158
+ /** Upload protocol for media (e.g. "raw", "multipart"). */
36159
+ upload_protocol?: string;
36160
+ /** Legacy upload protocol for media (e.g. "media", "multipart"). */
36161
+ uploadType?: string;
36162
+ },
36163
+ body: GoogleCloudAiplatformV1beta1EvaluateDatasetRequest,
36164
+ ): Request<GoogleLongrunningOperation>;
36079
36165
  /** Evaluates instances based on a given metric. */
36080
36166
  evaluateInstances(request: {
36081
36167
  /** V1 error format. */
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@maxim_mazurok/gapi.client.aiplatform-v1beta1",
3
- "version": "0.0.20250205",
3
+ "version": "0.0.20250212",
4
4
  "description": "TypeScript typings for Vertex AI API v1beta1",
5
5
  "repository": {
6
6
  "type": "git",