@maxim_mazurok/gapi.client.bigquery-v2 0.0.20240602 → 0.0.20240623
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/index.d.ts +17 -15
- package/package.json +1 -1
package/index.d.ts
CHANGED
|
@@ -9,7 +9,7 @@
|
|
|
9
9
|
// This file was generated by https://github.com/Maxim-Mazurok/google-api-typings-generator. Please do not edit it manually.
|
|
10
10
|
// In case of any problems please post issue to https://github.com/Maxim-Mazurok/google-api-typings-generator
|
|
11
11
|
// Generated from: https://bigquery.googleapis.com/$discovery/rest?version=v2
|
|
12
|
-
// Revision:
|
|
12
|
+
// Revision: 20240623
|
|
13
13
|
|
|
14
14
|
/// <reference types="gapi.client" />
|
|
15
15
|
|
|
@@ -175,11 +175,11 @@ declare namespace gapi.client {
|
|
|
175
175
|
biEngineReasons?: BiEngineReason[];
|
|
176
176
|
}
|
|
177
177
|
interface BigLakeConfiguration {
|
|
178
|
-
/** Required. The connection specifying the credentials to be used to read and write to external storage, such as Cloud Storage. The connection_id can have the form
|
|
178
|
+
/** Required. The connection specifying the credentials to be used to read and write to external storage, such as Cloud Storage. The connection_id can have the form `{project}.{location}.{connection_id}` or `projects/{project}/locations/{location}/connections/{connection_id}". */
|
|
179
179
|
connectionId?: string;
|
|
180
180
|
/** Required. The file format the table data is stored in. */
|
|
181
181
|
fileFormat?: string;
|
|
182
|
-
/** Required. The fully qualified location prefix of the external folder where table data is stored. The '*' wildcard character is not allowed. The URI should be in the format
|
|
182
|
+
/** Required. The fully qualified location prefix of the external folder where table data is stored. The '*' wildcard character is not allowed. The URI should be in the format `gs://bucket/path_to_table/` */
|
|
183
183
|
storageUri?: string;
|
|
184
184
|
/** Required. The table format the metadata only snapshots are stored in. */
|
|
185
185
|
tableFormat?: string;
|
|
@@ -197,7 +197,7 @@ declare namespace gapi.client {
|
|
|
197
197
|
fieldName?: string;
|
|
198
198
|
/** Optional. If this is set, only the latest version of value in this column are exposed. 'onlyReadLatest' can also be set at the column family level. However, the setting at this level takes precedence if 'onlyReadLatest' is set at both levels. */
|
|
199
199
|
onlyReadLatest?: boolean;
|
|
200
|
-
/** [Required] Qualifier of the column. Columns in the parent column family that has this exact qualifier are exposed as
|
|
200
|
+
/** [Required] Qualifier of the column. Columns in the parent column family that has this exact qualifier are exposed as `.` field. If the qualifier is valid UTF-8 string, it can be specified in the qualifier_string field. Otherwise, a base-64 encoded value must be set to qualifier_encoded. The column field name is the same as the column qualifier. However, if the qualifier is not a valid BigQuery field identifier i.e. does not match a-zA-Z*, a valid identifier must be provided as field_name. */
|
|
201
201
|
qualifierEncoded?: string;
|
|
202
202
|
/** Qualifier string. */
|
|
203
203
|
qualifierString?: string;
|
|
@@ -205,7 +205,7 @@ declare namespace gapi.client {
|
|
|
205
205
|
type?: string;
|
|
206
206
|
}
|
|
207
207
|
interface BigtableColumnFamily {
|
|
208
|
-
/** Optional. Lists of columns that should be exposed as individual fields as opposed to a list of (column name, value) pairs. All columns whose qualifier matches a qualifier in this list can be accessed as
|
|
208
|
+
/** Optional. Lists of columns that should be exposed as individual fields as opposed to a list of (column name, value) pairs. All columns whose qualifier matches a qualifier in this list can be accessed as `.`. Other columns can be accessed as a list through `.Column` field. */
|
|
209
209
|
columns?: BigtableColumn[];
|
|
210
210
|
/** Optional. The encoding of the values when the type is not STRING. Acceptable encoding values are: TEXT - indicates values are alphanumeric text strings. BINARY - indicates values are encoded using HBase Bytes.toBytes family of functions. This can be overridden for a specific column by listing that column in 'columns' and specifying an encoding for it. */
|
|
211
211
|
encoding?: string;
|
|
@@ -389,7 +389,7 @@ declare namespace gapi.client {
|
|
|
389
389
|
groupByEmail?: string;
|
|
390
390
|
/** [Pick one] Some other type of member that appears in the IAM Policy but isn't a user, group, domain, or special group. */
|
|
391
391
|
iamMember?: string;
|
|
392
|
-
/** An IAM role ID that should be granted to the user, group, or domain specified in this access entry. The following legacy mappings will be applied: OWNER <=> roles/bigquery.dataOwner WRITER <=> roles/bigquery.dataEditor READER <=> roles/bigquery.dataViewer This field will accept any of the above formats, but will return only the legacy format. For example, if you set this field to "roles/bigquery.dataOwner", it will be returned back as "OWNER". */
|
|
392
|
+
/** An IAM role ID that should be granted to the user, group, or domain specified in this access entry. The following legacy mappings will be applied: * OWNER <=> roles/bigquery.dataOwner * WRITER <=> roles/bigquery.dataEditor * READER <=> roles/bigquery.dataViewer This field will accept any of the above formats, but will return only the legacy format. For example, if you set this field to "roles/bigquery.dataOwner", it will be returned back as "OWNER". */
|
|
393
393
|
role?: string;
|
|
394
394
|
/** [Pick one] A routine from a different dataset to grant access to. Queries executed against that routine will have read access to views/tables/routines in this dataset. Only UDF is supported for now. The role field is not required when this field is set. If that routine is updated by any user, access to the routine needs to be granted again via an update operation. */
|
|
395
395
|
routine?: RoutineReference;
|
|
@@ -430,7 +430,7 @@ declare namespace gapi.client {
|
|
|
430
430
|
isCaseInsensitive?: boolean;
|
|
431
431
|
/** Output only. The resource type. */
|
|
432
432
|
kind?: string;
|
|
433
|
-
/** The labels associated with this dataset. You can use these to organize and group your datasets. You can set this property when inserting or updating a dataset. See Creating and Updating Dataset Labels for more information. */
|
|
433
|
+
/** The labels associated with this dataset. You can use these to organize and group your datasets. You can set this property when inserting or updating a dataset. See [Creating and Updating Dataset Labels](https://cloud.google.com/bigquery/docs/creating-managing-labels#creating_and_updating_dataset_labels) for more information. */
|
|
434
434
|
labels?: {[P in string]: string};
|
|
435
435
|
/** Output only. The date when this dataset was last modified, in milliseconds since the epoch. */
|
|
436
436
|
lastModifiedTime?: string;
|
|
@@ -716,7 +716,7 @@ declare namespace gapi.client {
|
|
|
716
716
|
bigtableOptions?: BigtableOptions;
|
|
717
717
|
/** Optional. The compression type of the data source. Possible values include GZIP and NONE. The default value is NONE. This setting is ignored for Google Cloud Bigtable, Google Cloud Datastore backups, Avro, ORC and Parquet formats. An empty string is an invalid value. */
|
|
718
718
|
compression?: string;
|
|
719
|
-
/** Optional. The connection specifying the credentials to be used to read external storage, such as Azure Blob, Cloud Storage, or S3. The connection_id can have the form
|
|
719
|
+
/** Optional. The connection specifying the credentials to be used to read external storage, such as Azure Blob, Cloud Storage, or S3. The connection_id can have the form `{project_id}.{location_id};{connection_id}` or `projects/{project_id}/locations/{location_id}/connections/{connection_id}`. */
|
|
720
720
|
connectionId?: string;
|
|
721
721
|
/** Optional. Additional properties to set if sourceFormat is set to CSV. */
|
|
722
722
|
csvOptions?: CsvOptions;
|
|
@@ -1617,11 +1617,11 @@ declare namespace gapi.client {
|
|
|
1617
1617
|
mapTargetType?: string;
|
|
1618
1618
|
}
|
|
1619
1619
|
interface PartitionedColumn {
|
|
1620
|
-
/**
|
|
1620
|
+
/** Required. The name of the partition column. */
|
|
1621
1621
|
field?: string;
|
|
1622
1622
|
}
|
|
1623
1623
|
interface PartitioningDefinition {
|
|
1624
|
-
/**
|
|
1624
|
+
/** Optional. Details about each partitioning column. This field is output only for all partitioning types other than metastore partitioned tables. BigQuery native tables only support 1 partitioning column. Other table types may support 0, 1 or more partitioning columns. For metastore partitioned tables, the order must match the definition order in the Hive Metastore, where it must match the physical layout of the table. For example, CREATE TABLE a_table(id BIGINT, name STRING) PARTITIONED BY (city STRING, state STRING). In this case the values must be ['city', 'state'] in that order. */
|
|
1625
1625
|
partitionedColumn?: PartitionedColumn[];
|
|
1626
1626
|
}
|
|
1627
1627
|
interface PartitionSkew {
|
|
@@ -2212,7 +2212,7 @@ declare namespace gapi.client {
|
|
|
2212
2212
|
numTotalLogicalBytes?: string;
|
|
2213
2213
|
/** Output only. The physical size of this table in bytes. This also includes storage used for time travel. This data is not kept in real time, and might be delayed by a few seconds to a few minutes. */
|
|
2214
2214
|
numTotalPhysicalBytes?: string;
|
|
2215
|
-
/**
|
|
2215
|
+
/** Optional. The partition information for all table formats, including managed partitioned tables, hive partitioned tables, iceberg partitioned, and metastore partitioned tables. This field is only populated for metastore partitioned tables. For other table formats, this is an output only field. */
|
|
2216
2216
|
partitionDefinition?: PartitioningDefinition;
|
|
2217
2217
|
/** If specified, configures range partitioning for this table. */
|
|
2218
2218
|
rangePartitioning?: RangePartitioning;
|
|
@@ -2401,6 +2401,8 @@ declare namespace gapi.client {
|
|
|
2401
2401
|
interface TableMetadataCacheUsage {
|
|
2402
2402
|
/** Free form human-readable reason metadata caching was unused for the job. */
|
|
2403
2403
|
explanation?: string;
|
|
2404
|
+
/** Duration since last refresh as of this job for managed tables (indicates metadata cache staleness as seen by this job). */
|
|
2405
|
+
staleness?: string;
|
|
2404
2406
|
/** Metadata caching eligible table referenced in the query. */
|
|
2405
2407
|
tableReference?: TableReference;
|
|
2406
2408
|
/** [Table type](/bigquery/docs/reference/rest/v2/tables#Table.FIELDS.type). */
|
|
@@ -2829,7 +2831,7 @@ declare namespace gapi.client {
|
|
|
2829
2831
|
callback?: string;
|
|
2830
2832
|
/** Selector specifying which fields to include in a partial response. */
|
|
2831
2833
|
fields?: string;
|
|
2832
|
-
/** An expression for filtering the results of the request by label. The syntax is
|
|
2834
|
+
/** An expression for filtering the results of the request by label. The syntax is `labels.[:]`. Multiple filters can be ANDed together by connecting with a space. Example: `labels.department:receiving labels.active`. See [Filtering datasets using labels](https://cloud.google.com/bigquery/docs/filtering-labels#filtering_datasets_using_labels) for details. */
|
|
2833
2835
|
filter?: string;
|
|
2834
2836
|
/** API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token. */
|
|
2835
2837
|
key?: string;
|
|
@@ -3054,7 +3056,7 @@ declare namespace gapi.client {
|
|
|
3054
3056
|
jobId: string;
|
|
3055
3057
|
/** API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token. */
|
|
3056
3058
|
key?: string;
|
|
3057
|
-
/** The geographic location of the job. You must specify the location to run the job for the following scenarios:
|
|
3059
|
+
/** The geographic location of the job. You must specify the location to run the job for the following scenarios: * If the location to run a job is not in the `us` or the `eu` multi-regional location * If the job's location is in a single region (for example, `us-central1`) For more information, see https://cloud.google.com/bigquery/docs/locations#specifying_your_location. */
|
|
3058
3060
|
location?: string;
|
|
3059
3061
|
/** OAuth 2.0 token for the current user. */
|
|
3060
3062
|
oauth_token?: string;
|
|
@@ -3116,7 +3118,7 @@ declare namespace gapi.client {
|
|
|
3116
3118
|
jobId: string;
|
|
3117
3119
|
/** API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token. */
|
|
3118
3120
|
key?: string;
|
|
3119
|
-
/** The geographic location of the job. You must specify the location to run the job for the following scenarios:
|
|
3121
|
+
/** The geographic location of the job. You must specify the location to run the job for the following scenarios: * If the location to run a job is not in the `us` or the `eu` multi-regional location * If the job's location is in a single region (for example, `us-central1`) For more information, see https://cloud.google.com/bigquery/docs/locations#specifying_your_location. */
|
|
3120
3122
|
location?: string;
|
|
3121
3123
|
/** OAuth 2.0 token for the current user. */
|
|
3122
3124
|
oauth_token?: string;
|
|
@@ -3149,7 +3151,7 @@ declare namespace gapi.client {
|
|
|
3149
3151
|
jobId: string;
|
|
3150
3152
|
/** API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token. */
|
|
3151
3153
|
key?: string;
|
|
3152
|
-
/** The geographic location of the job. You must specify the location to run the job for the following scenarios:
|
|
3154
|
+
/** The geographic location of the job. You must specify the location to run the job for the following scenarios: * If the location to run a job is not in the `us` or the `eu` multi-regional location * If the job's location is in a single region (for example, `us-central1`) For more information, see https://cloud.google.com/bigquery/docs/locations#specifying_your_location. */
|
|
3153
3155
|
location?: string;
|
|
3154
3156
|
/** Maximum number of results to read. */
|
|
3155
3157
|
maxResults?: number;
|