diff --git a/bigquery/v2/bigquery-api.json b/bigquery/v2/bigquery-api.json index 286f3cbe44..88d45fcbc3 100644 --- a/bigquery/v2/bigquery-api.json +++ b/bigquery/v2/bigquery-api.json @@ -529,7 +529,7 @@ "type": "string" }, "location": { - "description": "The geographic location of the job. You must specify the location to run the job for the following scenarios: * If the location to run a job is not in the `us` or the `eu` multi-regional location * If the job's location is in a single region (for example, `us-central1`) For more information, see https://cloud.google.com/bigquery/docs/locations#specifying_your_location.", + "description": "The geographic location of the job. You must [specify the location](https://cloud.google.com/bigquery/docs/locations#specify_locations) to run the job for the following scenarios: * If the location to run a job is not in the `us` or the `eu` multi-regional location * If the job's location is in a single region (for example, `us-central1`)", "location": "query", "type": "string" }, @@ -568,7 +568,7 @@ "type": "string" }, "location": { - "description": "The geographic location of the job. Required. See details at: https://cloud.google.com/bigquery/docs/locations#specifying_your_location.", + "description": "The geographic location of the job. Required. For more information, see how to [specify locations](https://cloud.google.com/bigquery/docs/locations#specify_locations).", "location": "query", "type": "string" }, @@ -604,7 +604,7 @@ "type": "string" }, "location": { - "description": "The geographic location of the job. You must specify the location to run the job for the following scenarios: * If the location to run a job is not in the `us` or the `eu` multi-regional location * If the job's location is in a single region (for example, `us-central1`) For more information, see https://cloud.google.com/bigquery/docs/locations#specifying_your_location.", + "description": "The geographic location of the job. You must specify the location to run the job for the following scenarios: * If the location to run a job is not in the `us` or the `eu` multi-regional location * If the job's location is in a single region (for example, `us-central1`) For more information, see how to [specify locations](https://cloud.google.com/bigquery/docs/locations#specify_locations).", "location": "query", "type": "string" }, @@ -649,7 +649,7 @@ "type": "string" }, "location": { - "description": "The geographic location of the job. You must specify the location to run the job for the following scenarios: * If the location to run a job is not in the `us` or the `eu` multi-regional location * If the job's location is in a single region (for example, `us-central1`) For more information, see https://cloud.google.com/bigquery/docs/locations#specifying_your_location.", + "description": "The geographic location of the job. You must specify the location to run the job for the following scenarios: * If the location to run a job is not in the `us` or the `eu` multi-regional location * If the job's location is in a single region (for example, `us-central1`) For more information, see how to [specify locations](https://cloud.google.com/bigquery/docs/locations#specify_locations).", "location": "query", "type": "string" }, @@ -2009,7 +2009,7 @@ } } }, - "revision": "20241115", + "revision": "20241222", "rootUrl": "https://bigquery.googleapis.com/", "schemas": { "AggregateClassificationMetrics": { @@ -3019,7 +3019,7 @@ "id": "Clustering", "properties": { "fields": { - "description": "One or more fields on which data should be clustered. Only top-level, non-repeated, simple-type fields are supported. The ordering of the clustering fields should be prioritized from most to least important for filtering purposes. Additional information on limitations can be found here: https://cloud.google.com/bigquery/docs/creating-clustered-tables#limitations", + "description": "One or more fields on which data should be clustered. Only top-level, non-repeated, simple-type fields are supported. The ordering of the clustering fields should be prioritized from most to least important for filtering purposes. For additional information, see [Introduction to clustered tables](https://cloud.google.com/bigquery/docs/clustered-tables#limitations).", "items": { "type": "string" }, @@ -4015,7 +4015,7 @@ "type": "object" }, "ExternalCatalogDatasetOptions": { - "description": "Options defining open source compatible datasets living in the BigQuery catalog. Contains metadata of open source database, schema or namespace represented by the current dataset.", + "description": "Options defining open source compatible datasets living in the BigQuery catalog. Contains metadata of open source database, schema, or namespace represented by the current dataset.", "id": "ExternalCatalogDatasetOptions", "properties": { "defaultStorageLocationUri": { @@ -4026,25 +4026,25 @@ "additionalProperties": { "type": "string" }, - "description": "Optional. A map of key value pairs defining the parameters and properties of the open source schema. Maximum size of 2Mib.", + "description": "Optional. A map of key value pairs defining the parameters and properties of the open source schema. Maximum size of 2MiB.", "type": "object" } }, "type": "object" }, "ExternalCatalogTableOptions": { - "description": "Metadata about open source compatible table. The fields contained in these options correspond to hive metastore's table level properties.", + "description": "Metadata about open source compatible table. The fields contained in these options correspond to Hive metastore's table-level properties.", "id": "ExternalCatalogTableOptions", "properties": { "connectionId": { - "description": "Optional. The connection specifying the credentials to be used to read external storage, such as Azure Blob, Cloud Storage, or S3. The connection is needed to read the open source table from BigQuery Engine. The connection_id can have the form `..` or `projects//locations//connections/`.", + "description": "Optional. A connection ID that specifies the credentials to be used to read external storage, such as Azure Blob, Cloud Storage, or Amazon S3. This connection is needed to read the open source table from BigQuery. The connection_id format must be either `..` or `projects//locations//connections/`.", "type": "string" }, "parameters": { "additionalProperties": { "type": "string" }, - "description": "Optional. A map of key value pairs defining the parameters and properties of the open source table. Corresponds with hive meta store table parameters. Maximum size of 4Mib.", + "description": "Optional. A map of the key-value pairs defining the parameters and properties of the open source table. Corresponds with Hive metastore table parameters. Maximum size of 4MiB.", "type": "object" }, "storageDescriptor": { @@ -5934,7 +5934,7 @@ "readOnly": true }, "statementType": { - "description": "Output only. The type of query statement, if valid. Possible values: * `SELECT`: [`SELECT`](https://cloud.google.com/bigquery/docs/reference/standard-sql/query-syntax#select_list) statement. * `ASSERT`: [`ASSERT`](https://cloud.google.com/bigquery/docs/reference/standard-sql/debugging-statements#assert) statement. * `INSERT`: [`INSERT`](https://cloud.google.com/bigquery/docs/reference/standard-sql/dml-syntax#insert_statement) statement. * `UPDATE`: [`UPDATE`](https://cloud.google.com/bigquery/docs/reference/standard-sql/query-syntax#update_statement) statement. * `DELETE`: [`DELETE`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-manipulation-language) statement. * `MERGE`: [`MERGE`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-manipulation-language) statement. * `CREATE_TABLE`: [`CREATE TABLE`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_table_statement) statement, without `AS SELECT`. * `CREATE_TABLE_AS_SELECT`: [`CREATE TABLE AS SELECT`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#query_statement) statement. * `CREATE_VIEW`: [`CREATE VIEW`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_view_statement) statement. * `CREATE_MODEL`: [`CREATE MODEL`](https://cloud.google.com/bigquery-ml/docs/reference/standard-sql/bigqueryml-syntax-create#create_model_statement) statement. * `CREATE_MATERIALIZED_VIEW`: [`CREATE MATERIALIZED VIEW`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_materialized_view_statement) statement. * `CREATE_FUNCTION`: [`CREATE FUNCTION`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_function_statement) statement. * `CREATE_TABLE_FUNCTION`: [`CREATE TABLE FUNCTION`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_table_function_statement) statement. * `CREATE_PROCEDURE`: [`CREATE PROCEDURE`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_procedure) statement. * `CREATE_ROW_ACCESS_POLICY`: [`CREATE ROW ACCESS POLICY`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_row_access_policy_statement) statement. * `CREATE_SCHEMA`: [`CREATE SCHEMA`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_schema_statement) statement. * `CREATE_SNAPSHOT_TABLE`: [`CREATE SNAPSHOT TABLE`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_snapshot_table_statement) statement. * `CREATE_SEARCH_INDEX`: [`CREATE SEARCH INDEX`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_search_index_statement) statement. * `DROP_TABLE`: [`DROP TABLE`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_table_statement) statement. * `DROP_EXTERNAL_TABLE`: [`DROP EXTERNAL TABLE`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_external_table_statement) statement. * `DROP_VIEW`: [`DROP VIEW`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_view_statement) statement. * `DROP_MODEL`: [`DROP MODEL`](https://cloud.google.com/bigquery-ml/docs/reference/standard-sql/bigqueryml-syntax-drop-model) statement. * `DROP_MATERIALIZED_VIEW`: [`DROP MATERIALIZED VIEW`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_materialized_view_statement) statement. * `DROP_FUNCTION` : [`DROP FUNCTION`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_function_statement) statement. * `DROP_TABLE_FUNCTION` : [`DROP TABLE FUNCTION`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_table_function) statement. * `DROP_PROCEDURE`: [`DROP PROCEDURE`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_procedure_statement) statement. * `DROP_SEARCH_INDEX`: [`DROP SEARCH INDEX`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_search_index) statement. * `DROP_SCHEMA`: [`DROP SCHEMA`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_schema_statement) statement. * `DROP_SNAPSHOT_TABLE`: [`DROP SNAPSHOT TABLE`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_snapshot_table_statement) statement. * `DROP_ROW_ACCESS_POLICY`: [`DROP [ALL] ROW ACCESS POLICY|POLICIES`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_row_access_policy_statement) statement. * `ALTER_TABLE`: [`ALTER TABLE`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#alter_table_set_options_statement) statement. * `ALTER_VIEW`: [`ALTER VIEW`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#alter_view_set_options_statement) statement. * `ALTER_MATERIALIZED_VIEW`: [`ALTER MATERIALIZED VIEW`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#alter_materialized_view_set_options_statement) statement. * `ALTER_SCHEMA`: [`ALTER SCHEMA`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#aalter_schema_set_options_statement) statement. * `SCRIPT`: [`SCRIPT`](https://cloud.google.com/bigquery/docs/reference/standard-sql/procedural-language). * `TRUNCATE_TABLE`: [`TRUNCATE TABLE`](https://cloud.google.com/bigquery/docs/reference/standard-sql/dml-syntax#truncate_table_statement) statement. * `CREATE_EXTERNAL_TABLE`: [`CREATE EXTERNAL TABLE`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_external_table_statement) statement. * `EXPORT_DATA`: [`EXPORT DATA`](https://cloud.google.com/bigquery/docs/reference/standard-sql/other-statements#export_data_statement) statement. * `EXPORT_MODEL`: [`EXPORT MODEL`](https://cloud.google.com/bigquery-ml/docs/reference/standard-sql/bigqueryml-syntax-export-model) statement. * `LOAD_DATA`: [`LOAD DATA`](https://cloud.google.com/bigquery/docs/reference/standard-sql/other-statements#load_data_statement) statement. * `CALL`: [`CALL`](https://cloud.google.com/bigquery/docs/reference/standard-sql/procedural-language#call) statement.", + "description": "Output only. The type of query statement, if valid. Possible values: * `SELECT`: [`SELECT`](https://cloud.google.com/bigquery/docs/reference/standard-sql/query-syntax#select_list) statement. * `ASSERT`: [`ASSERT`](https://cloud.google.com/bigquery/docs/reference/standard-sql/debugging-statements#assert) statement. * `INSERT`: [`INSERT`](https://cloud.google.com/bigquery/docs/reference/standard-sql/dml-syntax#insert_statement) statement. * `UPDATE`: [`UPDATE`](https://cloud.google.com/bigquery/docs/reference/standard-sql/dml-syntax#update_statement) statement. * `DELETE`: [`DELETE`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-manipulation-language) statement. * `MERGE`: [`MERGE`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-manipulation-language) statement. * `CREATE_TABLE`: [`CREATE TABLE`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_table_statement) statement, without `AS SELECT`. * `CREATE_TABLE_AS_SELECT`: [`CREATE TABLE AS SELECT`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_table_statement) statement. * `CREATE_VIEW`: [`CREATE VIEW`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_view_statement) statement. * `CREATE_MODEL`: [`CREATE MODEL`](https://cloud.google.com/bigquery-ml/docs/reference/standard-sql/bigqueryml-syntax-create#create_model_statement) statement. * `CREATE_MATERIALIZED_VIEW`: [`CREATE MATERIALIZED VIEW`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_materialized_view_statement) statement. * `CREATE_FUNCTION`: [`CREATE FUNCTION`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_function_statement) statement. * `CREATE_TABLE_FUNCTION`: [`CREATE TABLE FUNCTION`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_table_function_statement) statement. * `CREATE_PROCEDURE`: [`CREATE PROCEDURE`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_procedure) statement. * `CREATE_ROW_ACCESS_POLICY`: [`CREATE ROW ACCESS POLICY`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_row_access_policy_statement) statement. * `CREATE_SCHEMA`: [`CREATE SCHEMA`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_schema_statement) statement. * `CREATE_SNAPSHOT_TABLE`: [`CREATE SNAPSHOT TABLE`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_snapshot_table_statement) statement. * `CREATE_SEARCH_INDEX`: [`CREATE SEARCH INDEX`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_search_index_statement) statement. * `DROP_TABLE`: [`DROP TABLE`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_table_statement) statement. * `DROP_EXTERNAL_TABLE`: [`DROP EXTERNAL TABLE`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_external_table_statement) statement. * `DROP_VIEW`: [`DROP VIEW`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_view_statement) statement. * `DROP_MODEL`: [`DROP MODEL`](https://cloud.google.com/bigquery-ml/docs/reference/standard-sql/bigqueryml-syntax-drop-model) statement. * `DROP_MATERIALIZED_VIEW`: [`DROP MATERIALIZED VIEW`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_materialized_view_statement) statement. * `DROP_FUNCTION` : [`DROP FUNCTION`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_function_statement) statement. * `DROP_TABLE_FUNCTION` : [`DROP TABLE FUNCTION`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_table_function) statement. * `DROP_PROCEDURE`: [`DROP PROCEDURE`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_procedure_statement) statement. * `DROP_SEARCH_INDEX`: [`DROP SEARCH INDEX`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_search_index) statement. * `DROP_SCHEMA`: [`DROP SCHEMA`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_schema_statement) statement. * `DROP_SNAPSHOT_TABLE`: [`DROP SNAPSHOT TABLE`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_snapshot_table_statement) statement. * `DROP_ROW_ACCESS_POLICY`: [`DROP [ALL] ROW ACCESS POLICY|POLICIES`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_row_access_policy_statement) statement. * `ALTER_TABLE`: [`ALTER TABLE`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#alter_table_set_options_statement) statement. * `ALTER_VIEW`: [`ALTER VIEW`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#alter_view_set_options_statement) statement. * `ALTER_MATERIALIZED_VIEW`: [`ALTER MATERIALIZED VIEW`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#alter_materialized_view_set_options_statement) statement. * `ALTER_SCHEMA`: [`ALTER SCHEMA`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#alter_schema_set_options_statement) statement. * `SCRIPT`: [`SCRIPT`](https://cloud.google.com/bigquery/docs/reference/standard-sql/procedural-language). * `TRUNCATE_TABLE`: [`TRUNCATE TABLE`](https://cloud.google.com/bigquery/docs/reference/standard-sql/dml-syntax#truncate_table_statement) statement. * `CREATE_EXTERNAL_TABLE`: [`CREATE EXTERNAL TABLE`](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_external_table_statement) statement. * `EXPORT_DATA`: [`EXPORT DATA`](https://cloud.google.com/bigquery/docs/reference/standard-sql/other-statements#export_data_statement) statement. * `EXPORT_MODEL`: [`EXPORT MODEL`](https://cloud.google.com/bigquery-ml/docs/reference/standard-sql/bigqueryml-syntax-export-model) statement. * `LOAD_DATA`: [`LOAD DATA`](https://cloud.google.com/bigquery/docs/reference/standard-sql/other-statements#load_data_statement) statement. * `CALL`: [`CALL`](https://cloud.google.com/bigquery/docs/reference/standard-sql/procedural-language#call) statement.", "readOnly": true, "type": "string" }, @@ -6346,7 +6346,7 @@ "View has no cached data because it has not refreshed yet.", "The estimated cost of the view is more expensive than another view or the base table. Note: The estimate cost might not match the billed cost.", "View has no cached data because a base table is truncated.", - "View is invalidated because of a data change in one or more base tables. It could be any recent change if the [`max_staleness`](https://cloud.google.com/bigquery/docs/materialized-views-create#max_staleness) option is not set for the view, or otherwise any change outside of the staleness window.", + "View is invalidated because of a data change in one or more base tables. It could be any recent change if the [`maxStaleness`](https://cloud.google.com/bigquery/docs/reference/rest/v2/tables#Table.FIELDS.max_staleness) option is not set for the view, or otherwise any change outside of the staleness window.", "View is invalidated because a base table's partition expiration has changed.", "View is invalidated because a base table's partition has expired.", "View is invalidated because a base table has an incompatible metadata change.", @@ -7219,7 +7219,7 @@ "type": "object" }, "location": { - "description": "The geographic location where the job should run. See details at https://cloud.google.com/bigquery/docs/locations#specifying_your_location.", + "description": "The geographic location where the job should run. For more information, see how to [specify locations](https://cloud.google.com/bigquery/docs/locations#specify_locations).", "type": "string" }, "maxResults": { @@ -9465,7 +9465,7 @@ "type": "string" }, "dataSplitColumn": { - "description": "The column to split data with. This column won't be used as a feature. 1. When data_split_method is CUSTOM, the corresponding column should be boolean. The rows with true value tag are eval data, and the false are training data. 2. When data_split_method is SEQ, the first DATA_SPLIT_EVAL_FRACTION rows (from smallest to largest) in the corresponding column are used as training data, and the rest are eval data. It respects the order in Orderable data types: https://cloud.google.com/bigquery/docs/reference/standard-sql/data-types#data-type-properties", + "description": "The column to split data with. This column won't be used as a feature. 1. When data_split_method is CUSTOM, the corresponding column should be boolean. The rows with true value tag are eval data, and the false are training data. 2. When data_split_method is SEQ, the first DATA_SPLIT_EVAL_FRACTION rows (from smallest to largest) in the corresponding column are used as training data, and the rest are eval data. It respects the order in Orderable data types: https://cloud.google.com/bigquery/docs/reference/standard-sql/data-types#data_type_properties", "type": "string" }, "dataSplitEvalFraction": { diff --git a/bigquery/v2/bigquery-gen.go b/bigquery/v2/bigquery-gen.go index cfdfe03558..25a53a0ebf 100644 --- a/bigquery/v2/bigquery-gen.go +++ b/bigquery/v2/bigquery-gen.go @@ -1627,8 +1627,9 @@ type Clustering struct { // Fields: One or more fields on which data should be clustered. Only // top-level, non-repeated, simple-type fields are supported. The ordering of // the clustering fields should be prioritized from most to least important for - // filtering purposes. Additional information on limitations can be found here: - // https://cloud.google.com/bigquery/docs/creating-clustered-tables#limitations + // filtering purposes. For additional information, see Introduction to + // clustered tables + // (https://cloud.google.com/bigquery/docs/clustered-tables#limitations). Fields []string `json:"fields,omitempty"` // ForceSendFields is a list of field names (e.g. "Fields") to unconditionally // include in API requests. By default, fields with empty or default values are @@ -3070,14 +3071,14 @@ func (s Expr) MarshalJSON() ([]byte, error) { // ExternalCatalogDatasetOptions: Options defining open source compatible // datasets living in the BigQuery catalog. Contains metadata of open source -// database, schema or namespace represented by the current dataset. +// database, schema, or namespace represented by the current dataset. type ExternalCatalogDatasetOptions struct { // DefaultStorageLocationUri: Optional. The storage location URI for all tables // in the dataset. Equivalent to hive metastore's database locationUri. Maximum // length of 1024 characters. DefaultStorageLocationUri string `json:"defaultStorageLocationUri,omitempty"` // Parameters: Optional. A map of key value pairs defining the parameters and - // properties of the open source schema. Maximum size of 2Mib. + // properties of the open source schema. Maximum size of 2MiB. Parameters map[string]string `json:"parameters,omitempty"` // ForceSendFields is a list of field names (e.g. "DefaultStorageLocationUri") // to unconditionally include in API requests. By default, fields with empty or @@ -3098,17 +3099,18 @@ func (s ExternalCatalogDatasetOptions) MarshalJSON() ([]byte, error) { } // ExternalCatalogTableOptions: Metadata about open source compatible table. -// The fields contained in these options correspond to hive metastore's table -// level properties. +// The fields contained in these options correspond to Hive metastore's +// table-level properties. type ExternalCatalogTableOptions struct { - // ConnectionId: Optional. The connection specifying the credentials to be used - // to read external storage, such as Azure Blob, Cloud Storage, or S3. The - // connection is needed to read the open source table from BigQuery Engine. The - // connection_id can have the form `..` or `projects//locations//connections/`. + // ConnectionId: Optional. A connection ID that specifies the credentials to be + // used to read external storage, such as Azure Blob, Cloud Storage, or Amazon + // S3. This connection is needed to read the open source table from BigQuery. + // The connection_id format must be either `..` or + // `projects//locations//connections/`. ConnectionId string `json:"connectionId,omitempty"` - // Parameters: Optional. A map of key value pairs defining the parameters and - // properties of the open source table. Corresponds with hive meta store table - // parameters. Maximum size of 4Mib. + // Parameters: Optional. A map of the key-value pairs defining the parameters + // and properties of the open source table. Corresponds with Hive metastore + // table parameters. Maximum size of 4MiB. Parameters map[string]string `json:"parameters,omitempty"` // StorageDescriptor: Optional. A storage descriptor containing information // about the physical storage of this table. @@ -5284,7 +5286,7 @@ type JobStatistics2 struct { // statement. * `INSERT`: `INSERT` // (https://cloud.google.com/bigquery/docs/reference/standard-sql/dml-syntax#insert_statement) // statement. * `UPDATE`: `UPDATE` - // (https://cloud.google.com/bigquery/docs/reference/standard-sql/query-syntax#update_statement) + // (https://cloud.google.com/bigquery/docs/reference/standard-sql/dml-syntax#update_statement) // statement. * `DELETE`: `DELETE` // (https://cloud.google.com/bigquery/docs/reference/standard-sql/data-manipulation-language) // statement. * `MERGE`: `MERGE` @@ -5293,7 +5295,7 @@ type JobStatistics2 struct { // (https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_table_statement) // statement, without `AS SELECT`. * `CREATE_TABLE_AS_SELECT`: `CREATE TABLE AS // SELECT` - // (https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#query_statement) + // (https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_table_statement) // statement. * `CREATE_VIEW`: `CREATE VIEW` // (https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_view_statement) // statement. * `CREATE_MODEL`: `CREATE MODEL` @@ -5346,7 +5348,7 @@ type JobStatistics2 struct { // statement. * `ALTER_MATERIALIZED_VIEW`: `ALTER MATERIALIZED VIEW` // (https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#alter_materialized_view_set_options_statement) // statement. * `ALTER_SCHEMA`: `ALTER SCHEMA` - // (https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#aalter_schema_set_options_statement) + // (https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#alter_schema_set_options_statement) // statement. * `SCRIPT`: `SCRIPT` // (https://cloud.google.com/bigquery/docs/reference/standard-sql/procedural-language). // * `TRUNCATE_TABLE`: `TRUNCATE TABLE` @@ -5859,9 +5861,9 @@ type MaterializedView struct { // truncated. // "BASE_TABLE_DATA_CHANGE" - View is invalidated because of a data change in // one or more base tables. It could be any recent change if the - // [`max_staleness`](https://cloud.google.com/bigquery/docs/materialized-views-c - // reate#max_staleness) option is not set for the view, or otherwise any change - // outside of the staleness window. + // [`maxStaleness`](https://cloud.google.com/bigquery/docs/reference/rest/v2/tab + // les#Table.FIELDS.max_staleness) option is not set for the view, or otherwise + // any change outside of the staleness window. // "BASE_TABLE_PARTITION_EXPIRATION_CHANGE" - View is invalidated because a // base table's partition expiration has changed. // "BASE_TABLE_EXPIRED_PARTITION" - View is invalidated because a base @@ -6929,8 +6931,9 @@ type QueryRequest struct { // must start with a letter and each label in the list must have a different // key. Labels map[string]string `json:"labels,omitempty"` - // Location: The geographic location where the job should run. See details at - // https://cloud.google.com/bigquery/docs/locations#specifying_your_location. + // Location: The geographic location where the job should run. For more + // information, see how to specify locations + // (https://cloud.google.com/bigquery/docs/locations#specify_locations). Location string `json:"location,omitempty"` // MaxResults: Optional. The maximum number of rows of data to return per page // of results. Setting this flag to a small value such as 1000 and then paging @@ -9625,7 +9628,7 @@ type TrainingOptions struct { // DATA_SPLIT_EVAL_FRACTION rows (from smallest to largest) in the // corresponding column are used as training data, and the rest are eval data. // It respects the order in Orderable data types: - // https://cloud.google.com/bigquery/docs/reference/standard-sql/data-types#data-type-properties + // https://cloud.google.com/bigquery/docs/reference/standard-sql/data-types#data_type_properties DataSplitColumn string `json:"dataSplitColumn,omitempty"` // DataSplitEvalFraction: The fraction of evaluation data over the whole input // data. The rest of data will be used as training data. The format should be @@ -11291,11 +11294,11 @@ func (r *JobsService) Cancel(projectId string, jobId string) *JobsCancelCall { } // Location sets the optional parameter "location": The geographic location of -// the job. You must specify the location to run the job for the following -// scenarios: * If the location to run a job is not in the `us` or the `eu` -// multi-regional location * If the job's location is in a single region (for -// example, `us-central1`) For more information, see -// https://cloud.google.com/bigquery/docs/locations#specifying_your_location. +// the job. You must specify the location +// (https://cloud.google.com/bigquery/docs/locations#specify_locations) to run +// the job for the following scenarios: * If the location to run a job is not +// in the `us` or the `eu` multi-regional location * If the job's location is +// in a single region (for example, `us-central1`) func (c *JobsCancelCall) Location(location string) *JobsCancelCall { c.urlParams_.Set("location", location) return c @@ -11408,8 +11411,8 @@ func (r *JobsService) Delete(projectId string, jobId string) *JobsDeleteCall { } // Location sets the optional parameter "location": The geographic location of -// the job. Required. See details at: -// https://cloud.google.com/bigquery/docs/locations#specifying_your_location. +// the job. Required. For more information, see how to specify locations +// (https://cloud.google.com/bigquery/docs/locations#specify_locations). func (c *JobsDeleteCall) Location(location string) *JobsDeleteCall { c.urlParams_.Set("location", location) return c @@ -11499,8 +11502,8 @@ func (r *JobsService) Get(projectId string, jobId string) *JobsGetCall { // the job. You must specify the location to run the job for the following // scenarios: * If the location to run a job is not in the `us` or the `eu` // multi-regional location * If the job's location is in a single region (for -// example, `us-central1`) For more information, see -// https://cloud.google.com/bigquery/docs/locations#specifying_your_location. +// example, `us-central1`) For more information, see how to specify locations +// (https://cloud.google.com/bigquery/docs/locations#specify_locations). func (c *JobsGetCall) Location(location string) *JobsGetCall { c.urlParams_.Set("location", location) return c @@ -11631,8 +11634,8 @@ func (c *JobsGetQueryResultsCall) FormatOptionsUseInt64Timestamp(formatOptionsUs // the job. You must specify the location to run the job for the following // scenarios: * If the location to run a job is not in the `us` or the `eu` // multi-regional location * If the job's location is in a single region (for -// example, `us-central1`) For more information, see -// https://cloud.google.com/bigquery/docs/locations#specifying_your_location. +// example, `us-central1`) For more information, see how to specify locations +// (https://cloud.google.com/bigquery/docs/locations#specify_locations). func (c *JobsGetQueryResultsCall) Location(location string) *JobsGetQueryResultsCall { c.urlParams_.Set("location", location) return c diff --git a/serviceusage/v1/serviceusage-api.json b/serviceusage/v1/serviceusage-api.json index 097ca7f58d..de2e334e63 100644 --- a/serviceusage/v1/serviceusage-api.json +++ b/serviceusage/v1/serviceusage-api.json @@ -426,7 +426,7 @@ } } }, - "revision": "20241205", + "revision": "20241219", "rootUrl": "https://serviceusage.googleapis.com/", "schemas": { "AddEnableRulesMetadata": { @@ -502,12 +502,14 @@ "enum": [ "ANALYSIS_TYPE_UNSPECIFIED", "ANALYSIS_TYPE_DEPENDENCY", - "ANALYSIS_TYPE_RESOURCE_USAGE" + "ANALYSIS_TYPE_RESOURCE_USAGE", + "ANALYSIS_TYPE_RESOURCE_EXISTENCE" ], "enumDescriptions": [ "Unspecified analysis type. Do not use.", "The analysis of service dependencies.", - "The analysis of service resource usage." + "The analysis of service resource usage.", + "The analysis of service resource existence." ], "readOnly": true, "type": "string" @@ -2305,14 +2307,20 @@ "description": "Output only. The type of impact.", "enum": [ "IMPACT_TYPE_UNSPECIFIED", - "DEPENDENCY_MISSING_DEPENDENCIES" + "DEPENDENCY_MISSING_DEPENDENCIES", + "RESOURCE_EXISTENCE_PROJECT" ], "enumDescriptions": [ - "Reserved Blocks (Block n contains codes from 100n to 100(n+1) -1 Block 0 - Special/Admin codes Block 1 - Impact Type of ANALYSIS_TYPE_DEPENDENCY Block 2 - Impact Type of ANALYSIS_TYPE_RESOURCE_USAGE ...", - "Block 1 - Impact Type of ANALYSIS_TYPE_DEPENDENCY" + "Reserved Blocks (Block n contains codes from 100n to 100(n+1) -1 Block 0 - Special/Admin codes Block 1 - Impact Type of ANALYSIS_TYPE_DEPENDENCY Block 2 - Impact Type of ANALYSIS_TYPE_RESOURCE_USAGE Block 3 - Impact Type of ANALYSIS_TYPE_RESOURCE_EXISTENCE ...", + "Block 1 - Impact Type of ANALYSIS_TYPE_DEPENDENCY", + "Block 3 - Impact Type of ANALYSIS_TYPE_RESOURCE_EXISTENCE" ], "readOnly": true, "type": "string" + }, + "parent": { + "description": "The parent resource that the analysis is based on and the service name that the analysis is for. Example: `projects/100/services/compute.googleapis.com`, folders/101/services/compute.googleapis.com` and `organizations/102/services/compute.googleapis.com`. Usually, the parent resource here is same as the parent resource of the analyzed policy. However, for some analysis types, the parent can be different. For example, for resource existence analysis, if the parent resource of the analyzed policy is a folder or an organization, the parent resource here can still be the project that contains the resources.", + "type": "string" } }, "type": "object" diff --git a/serviceusage/v1/serviceusage-gen.go b/serviceusage/v1/serviceusage-gen.go index 5934e7095e..0178941563 100644 --- a/serviceusage/v1/serviceusage-gen.go +++ b/serviceusage/v1/serviceusage-gen.go @@ -282,6 +282,8 @@ type Analysis struct { // "ANALYSIS_TYPE_UNSPECIFIED" - Unspecified analysis type. Do not use. // "ANALYSIS_TYPE_DEPENDENCY" - The analysis of service dependencies. // "ANALYSIS_TYPE_RESOURCE_USAGE" - The analysis of service resource usage. + // "ANALYSIS_TYPE_RESOURCE_EXISTENCE" - The analysis of service resource + // existence. AnalysisType string `json:"analysisType,omitempty"` // DisplayName: Output only. The user friendly display name of the analysis // type. E.g. service dependency analysis, service resource usage analysis, @@ -2744,10 +2746,24 @@ type Impact struct { // "IMPACT_TYPE_UNSPECIFIED" - Reserved Blocks (Block n contains codes from // 100n to 100(n+1) -1 Block 0 - Special/Admin codes Block 1 - Impact Type of // ANALYSIS_TYPE_DEPENDENCY Block 2 - Impact Type of - // ANALYSIS_TYPE_RESOURCE_USAGE ... + // ANALYSIS_TYPE_RESOURCE_USAGE Block 3 - Impact Type of + // ANALYSIS_TYPE_RESOURCE_EXISTENCE ... // "DEPENDENCY_MISSING_DEPENDENCIES" - Block 1 - Impact Type of // ANALYSIS_TYPE_DEPENDENCY + // "RESOURCE_EXISTENCE_PROJECT" - Block 3 - Impact Type of + // ANALYSIS_TYPE_RESOURCE_EXISTENCE ImpactType string `json:"impactType,omitempty"` + // Parent: The parent resource that the analysis is based on and the service + // name that the analysis is for. Example: + // `projects/100/services/compute.googleapis.com`, + // folders/101/services/compute.googleapis.com` and + // `organizations/102/services/compute.googleapis.com`. Usually, the parent + // resource here is same as the parent resource of the analyzed policy. + // However, for some analysis types, the parent can be different. For example, + // for resource existence analysis, if the parent resource of the analyzed + // policy is a folder or an organization, the parent resource here can still be + // the project that contains the resources. + Parent string `json:"parent,omitempty"` // ForceSendFields is a list of field names (e.g. "Detail") to unconditionally // include in API requests. By default, fields with empty or default values are // omitted from API requests. See diff --git a/sqladmin/v1beta4/sqladmin-api.json b/sqladmin/v1beta4/sqladmin-api.json index e3bad66ab3..3c04df7a72 100644 --- a/sqladmin/v1beta4/sqladmin-api.json +++ b/sqladmin/v1beta4/sqladmin-api.json @@ -2366,7 +2366,7 @@ } } }, - "revision": "20241204", + "revision": "20241210", "rootUrl": "https://sqladmin.googleapis.com/", "schemas": { "AclEntry": { @@ -5217,7 +5217,6 @@ "description": "Optional. The sub operation based on the operation type." }, "targetId": { - "description": "Name of the database instance related to this operation.", "type": "string" }, "targetLink": { diff --git a/sqladmin/v1beta4/sqladmin-gen.go b/sqladmin/v1beta4/sqladmin-gen.go index f92dfc6c1f..f057f2d83d 100644 --- a/sqladmin/v1beta4/sqladmin-gen.go +++ b/sqladmin/v1beta4/sqladmin-gen.go @@ -3290,9 +3290,8 @@ type Operation struct { Status string `json:"status,omitempty"` // SubOperationType: Optional. The sub operation based on the operation type. SubOperationType *SqlSubOperationType `json:"subOperationType,omitempty"` - // TargetId: Name of the database instance related to this operation. - TargetId string `json:"targetId,omitempty"` - TargetLink string `json:"targetLink,omitempty"` + TargetId string `json:"targetId,omitempty"` + TargetLink string `json:"targetLink,omitempty"` // TargetProject: The project ID of the target instance related to this // operation. TargetProject string `json:"targetProject,omitempty"`