From f252225da4b56f1d195de418657853c59af91c7f Mon Sep 17 00:00:00 2001 From: Yoshi Automation Date: Fri, 15 Aug 2025 01:37:18 +0000 Subject: [PATCH 01/11] feat(alloydb): update the API #### alloydb:v1alpha The following keys were added: - schemas.StorageDatabasecenterPartnerapiV1mainDatabaseResourceFeed.properties.skipIngestion.description - schemas.StorageDatabasecenterPartnerapiV1mainDatabaseResourceFeed.properties.skipIngestion.type The following keys were changed: - schemas.StorageDatabasecenterPartnerapiV1mainDatabaseResourceFeed.description #### alloydb:v1beta The following keys were added: - schemas.StorageDatabasecenterPartnerapiV1mainDatabaseResourceFeed.properties.skipIngestion.description - schemas.StorageDatabasecenterPartnerapiV1mainDatabaseResourceFeed.properties.skipIngestion.type The following keys were changed: - schemas.StorageDatabasecenterPartnerapiV1mainDatabaseResourceFeed.description #### alloydb:v1 The following keys were added: - schemas.StorageDatabasecenterPartnerapiV1mainDatabaseResourceFeed.properties.skipIngestion.description - schemas.StorageDatabasecenterPartnerapiV1mainDatabaseResourceFeed.properties.skipIngestion.type The following keys were changed: - schemas.StorageDatabasecenterPartnerapiV1mainDatabaseResourceFeed.description --- discovery/alloydb-v1.json | 8 ++++++-- discovery/alloydb-v1alpha.json | 8 ++++++-- discovery/alloydb-v1beta.json | 8 ++++++-- src/apis/alloydb/v1.ts | 6 +++++- src/apis/alloydb/v1alpha.ts | 6 +++++- src/apis/alloydb/v1beta.ts | 6 +++++- 6 files changed, 33 insertions(+), 9 deletions(-) diff --git a/discovery/alloydb-v1.json b/discovery/alloydb-v1.json index b77bde8cd55..b2c4d1fe036 100644 --- a/discovery/alloydb-v1.json +++ b/discovery/alloydb-v1.json @@ -1622,7 +1622,7 @@ } } }, - "revision": "20250731", + "revision": "20250807", "rootUrl": "https://alloydb.googleapis.com/", "schemas": { "AuthorizedNetwork": { @@ -4243,7 +4243,7 @@ "type": "object" }, "StorageDatabasecenterPartnerapiV1mainDatabaseResourceFeed": { - "description": "DatabaseResourceFeed is the top level proto to be used to ingest different database resource level events into Condor platform. Next ID: 10", + "description": "DatabaseResourceFeed is the top level proto to be used to ingest different database resource level events into Condor platform. Next ID: 11", "id": "StorageDatabasecenterPartnerapiV1mainDatabaseResourceFeed", "properties": { "backupdrMetadata": { @@ -4297,6 +4297,10 @@ }, "resourceMetadata": { "$ref": "StorageDatabasecenterPartnerapiV1mainDatabaseResourceMetadata" + }, + "skipIngestion": { + "description": "Optional. If true, the feed won't be ingested by DB Center. This indicates that the feed is intentionally skipped. For example, BackupDR feeds are only needed for resources integrated with DB Center (e.g., CloudSQL, AlloyDB). Feeds for non-integrated resources (e.g., Compute Engine, Persistent Disk) can be skipped.", + "type": "boolean" } }, "type": "object" diff --git a/discovery/alloydb-v1alpha.json b/discovery/alloydb-v1alpha.json index 20a19339da1..f569c008651 100644 --- a/discovery/alloydb-v1alpha.json +++ b/discovery/alloydb-v1alpha.json @@ -1622,7 +1622,7 @@ } } }, - "revision": "20250731", + "revision": "20250807", "rootUrl": "https://alloydb.googleapis.com/", "schemas": { "AuthorizedNetwork": { @@ -4420,7 +4420,7 @@ "type": "object" }, "StorageDatabasecenterPartnerapiV1mainDatabaseResourceFeed": { - "description": "DatabaseResourceFeed is the top level proto to be used to ingest different database resource level events into Condor platform. Next ID: 10", + "description": "DatabaseResourceFeed is the top level proto to be used to ingest different database resource level events into Condor platform. Next ID: 11", "id": "StorageDatabasecenterPartnerapiV1mainDatabaseResourceFeed", "properties": { "backupdrMetadata": { @@ -4474,6 +4474,10 @@ }, "resourceMetadata": { "$ref": "StorageDatabasecenterPartnerapiV1mainDatabaseResourceMetadata" + }, + "skipIngestion": { + "description": "Optional. If true, the feed won't be ingested by DB Center. This indicates that the feed is intentionally skipped. For example, BackupDR feeds are only needed for resources integrated with DB Center (e.g., CloudSQL, AlloyDB). Feeds for non-integrated resources (e.g., Compute Engine, Persistent Disk) can be skipped.", + "type": "boolean" } }, "type": "object" diff --git a/discovery/alloydb-v1beta.json b/discovery/alloydb-v1beta.json index d97a619f7d8..85eb558d66e 100644 --- a/discovery/alloydb-v1beta.json +++ b/discovery/alloydb-v1beta.json @@ -1619,7 +1619,7 @@ } } }, - "revision": "20250731", + "revision": "20250807", "rootUrl": "https://alloydb.googleapis.com/", "schemas": { "AuthorizedNetwork": { @@ -4396,7 +4396,7 @@ "type": "object" }, "StorageDatabasecenterPartnerapiV1mainDatabaseResourceFeed": { - "description": "DatabaseResourceFeed is the top level proto to be used to ingest different database resource level events into Condor platform. Next ID: 10", + "description": "DatabaseResourceFeed is the top level proto to be used to ingest different database resource level events into Condor platform. Next ID: 11", "id": "StorageDatabasecenterPartnerapiV1mainDatabaseResourceFeed", "properties": { "backupdrMetadata": { @@ -4450,6 +4450,10 @@ }, "resourceMetadata": { "$ref": "StorageDatabasecenterPartnerapiV1mainDatabaseResourceMetadata" + }, + "skipIngestion": { + "description": "Optional. If true, the feed won't be ingested by DB Center. This indicates that the feed is intentionally skipped. For example, BackupDR feeds are only needed for resources integrated with DB Center (e.g., CloudSQL, AlloyDB). Feeds for non-integrated resources (e.g., Compute Engine, Persistent Disk) can be skipped.", + "type": "boolean" } }, "type": "object" diff --git a/src/apis/alloydb/v1.ts b/src/apis/alloydb/v1.ts index 5e59389b3a0..8695bc874d1 100644 --- a/src/apis/alloydb/v1.ts +++ b/src/apis/alloydb/v1.ts @@ -1812,7 +1812,7 @@ export namespace alloydb_v1 { internalResourceMetadata?: Schema$StorageDatabasecenterPartnerapiV1mainInternalResourceMetadata[]; } /** - * DatabaseResourceFeed is the top level proto to be used to ingest different database resource level events into Condor platform. Next ID: 10 + * DatabaseResourceFeed is the top level proto to be used to ingest different database resource level events into Condor platform. Next ID: 11 */ export interface Schema$StorageDatabasecenterPartnerapiV1mainDatabaseResourceFeed { /** @@ -1839,6 +1839,10 @@ export namespace alloydb_v1 { */ resourceId?: Schema$StorageDatabasecenterPartnerapiV1mainDatabaseResourceId; resourceMetadata?: Schema$StorageDatabasecenterPartnerapiV1mainDatabaseResourceMetadata; + /** + * Optional. If true, the feed won't be ingested by DB Center. This indicates that the feed is intentionally skipped. For example, BackupDR feeds are only needed for resources integrated with DB Center (e.g., CloudSQL, AlloyDB). Feeds for non-integrated resources (e.g., Compute Engine, Persistent Disk) can be skipped. + */ + skipIngestion?: boolean | null; } /** * Common model for database resource health signal data. diff --git a/src/apis/alloydb/v1alpha.ts b/src/apis/alloydb/v1alpha.ts index 3ee5af1a604..57251843959 100644 --- a/src/apis/alloydb/v1alpha.ts +++ b/src/apis/alloydb/v1alpha.ts @@ -1930,7 +1930,7 @@ export namespace alloydb_v1alpha { internalResourceMetadata?: Schema$StorageDatabasecenterPartnerapiV1mainInternalResourceMetadata[]; } /** - * DatabaseResourceFeed is the top level proto to be used to ingest different database resource level events into Condor platform. Next ID: 10 + * DatabaseResourceFeed is the top level proto to be used to ingest different database resource level events into Condor platform. Next ID: 11 */ export interface Schema$StorageDatabasecenterPartnerapiV1mainDatabaseResourceFeed { /** @@ -1957,6 +1957,10 @@ export namespace alloydb_v1alpha { */ resourceId?: Schema$StorageDatabasecenterPartnerapiV1mainDatabaseResourceId; resourceMetadata?: Schema$StorageDatabasecenterPartnerapiV1mainDatabaseResourceMetadata; + /** + * Optional. If true, the feed won't be ingested by DB Center. This indicates that the feed is intentionally skipped. For example, BackupDR feeds are only needed for resources integrated with DB Center (e.g., CloudSQL, AlloyDB). Feeds for non-integrated resources (e.g., Compute Engine, Persistent Disk) can be skipped. + */ + skipIngestion?: boolean | null; } /** * Common model for database resource health signal data. diff --git a/src/apis/alloydb/v1beta.ts b/src/apis/alloydb/v1beta.ts index 5cbd148e1ef..c0b5848e940 100644 --- a/src/apis/alloydb/v1beta.ts +++ b/src/apis/alloydb/v1beta.ts @@ -1914,7 +1914,7 @@ export namespace alloydb_v1beta { internalResourceMetadata?: Schema$StorageDatabasecenterPartnerapiV1mainInternalResourceMetadata[]; } /** - * DatabaseResourceFeed is the top level proto to be used to ingest different database resource level events into Condor platform. Next ID: 10 + * DatabaseResourceFeed is the top level proto to be used to ingest different database resource level events into Condor platform. Next ID: 11 */ export interface Schema$StorageDatabasecenterPartnerapiV1mainDatabaseResourceFeed { /** @@ -1941,6 +1941,10 @@ export namespace alloydb_v1beta { */ resourceId?: Schema$StorageDatabasecenterPartnerapiV1mainDatabaseResourceId; resourceMetadata?: Schema$StorageDatabasecenterPartnerapiV1mainDatabaseResourceMetadata; + /** + * Optional. If true, the feed won't be ingested by DB Center. This indicates that the feed is intentionally skipped. For example, BackupDR feeds are only needed for resources integrated with DB Center (e.g., CloudSQL, AlloyDB). Feeds for non-integrated resources (e.g., Compute Engine, Persistent Disk) can be skipped. + */ + skipIngestion?: boolean | null; } /** * Common model for database resource health signal data. From d09fe0c91d9c728c49c81615ff7e906e96537db1 Mon Sep 17 00:00:00 2001 From: Yoshi Automation Date: Fri, 15 Aug 2025 01:37:18 +0000 Subject: [PATCH 02/11] fix(compute): update the API #### compute:v1 The following keys were changed: - schemas.BackendService.properties.localityLbPolicy.description - schemas.BackendService.properties.localityLbPolicy.enumDescriptions - schemas.BackendService.properties.serviceLbPolicy.description - schemas.BackendServiceLocalityLoadBalancingPolicyConfigPolicy.properties.name.enumDescriptions - schemas.TargetHttpsProxy.description - schemas.UpcomingMaintenance.properties.maintenanceReasons.items.enum - schemas.UpcomingMaintenance.properties.maintenanceReasons.items.enumDescriptions --- discovery/compute-v1.json | 14 ++++++++------ src/apis/compute/v1.ts | 6 +++--- 2 files changed, 11 insertions(+), 9 deletions(-) diff --git a/discovery/compute-v1.json b/discovery/compute-v1.json index 8a69d7c86d3..ed527bc2bf5 100644 --- a/discovery/compute-v1.json +++ b/discovery/compute-v1.json @@ -39996,7 +39996,7 @@ } } }, - "revision": "20250729", + "revision": "20250807", "rootUrl": "https://compute.googleapis.com/", "schemas": { "AWSV4Signature": { @@ -43341,7 +43341,7 @@ "type": "array" }, "localityLbPolicy": { - "description": "The load balancing algorithm used within the scope of the locality. The possible values are: - ROUND_ROBIN: This is a simple policy in which each healthy backend is selected in round robin order. This is the default. - LEAST_REQUEST: An O(1) algorithm which selects two random healthy hosts and picks the host which has fewer active requests. - RING_HASH: The ring/modulo hash load balancer implements consistent hashing to backends. The algorithm has the property that the addition/removal of a host from a set of N hosts only affects 1/N of the requests. - RANDOM: The load balancer selects a random healthy host. - ORIGINAL_DESTINATION: Backend host is selected based on the client connection metadata, i.e., connections are opened to the same address as the destination address of the incoming connection before the connection was redirected to the load balancer. - MAGLEV: used as a drop in replacement for the ring hash load balancer. Maglev is not as stable as ring hash but has faster table lookup build times and host selection times. For more information about Maglev, see https://ai.google/research/pubs/pub44824 - WEIGHTED_ROUND_ROBIN: Per-endpoint Weighted Round Robin Load Balancing using weights computed from Backend reported Custom Metrics. If set, the Backend Service responses are expected to contain non-standard HTTP response header field Endpoint-Load-Metrics. The reported metrics to use for computing the weights are specified via the customMetrics field. This field is applicable to either: - A regional backend service with the service_protocol set to HTTP, HTTPS, HTTP2 or H2C, and load_balancing_scheme set to INTERNAL_MANAGED. - A global backend service with the load_balancing_scheme set to INTERNAL_SELF_MANAGED, INTERNAL_MANAGED, or EXTERNAL_MANAGED. If sessionAffinity is not configured—that is, if session affinity remains at the default value of NONE—then the default value for localityLbPolicy is ROUND_ROBIN. If session affinity is set to a value other than NONE, then the default value for localityLbPolicy is MAGLEV. Only ROUND_ROBIN and RING_HASH are supported when the backend service is referenced by a URL map that is bound to target gRPC proxy that has validateForProxyless field set to true. localityLbPolicy cannot be specified with haPolicy.", + "description": "The load balancing algorithm used within the scope of the locality. The possible values are: - ROUND_ROBIN: This is a simple policy in which each healthy backend is selected in round robin order. This is the default. - LEAST_REQUEST: An O(1) algorithm which selects two random healthy hosts and picks the host which has fewer active requests. - RING_HASH: The ring/modulo hash load balancer implements consistent hashing to backends. The algorithm has the property that the addition/removal of a host from a set of N hosts only affects 1/N of the requests. - RANDOM: The load balancer selects a random healthy host. - ORIGINAL_DESTINATION: Backend host is selected based on the client connection metadata, i.e., connections are opened to the same address as the destination address of the incoming connection before the connection was redirected to the load balancer. - MAGLEV: used as a drop in replacement for the ring hash load balancer. Maglev is not as stable as ring hash but has faster table lookup build times and host selection times. For more information about Maglev, see Maglev: A Fast and Reliable Software Network Load Balancer. - WEIGHTED_ROUND_ROBIN: Per-endpoint Weighted Round Robin Load Balancing using weights computed from Backend reported Custom Metrics. If set, the Backend Service responses are expected to contain non-standard HTTP response header field Endpoint-Load-Metrics. The reported metrics to use for computing the weights are specified via the customMetrics field. This field is applicable to either: - A regional backend service with the service_protocol set to HTTP, HTTPS, HTTP2 or H2C, and load_balancing_scheme set to INTERNAL_MANAGED. - A global backend service with the load_balancing_scheme set to INTERNAL_SELF_MANAGED, INTERNAL_MANAGED, or EXTERNAL_MANAGED. If sessionAffinity is not configured—that is, if session affinity remains at the default value of NONE—then the default value for localityLbPolicy is ROUND_ROBIN. If session affinity is set to a value other than NONE, then the default value for localityLbPolicy is MAGLEV. Only ROUND_ROBIN and RING_HASH are supported when the backend service is referenced by a URL map that is bound to target gRPC proxy that has validateForProxyless field set to true. localityLbPolicy cannot be specified with haPolicy.", "enum": [ "INVALID_LB_POLICY", "LEAST_REQUEST", @@ -43356,7 +43356,7 @@ "enumDescriptions": [ "", "An O(1) algorithm which selects two random healthy hosts and picks the host which has fewer active requests.", - "This algorithm implements consistent hashing to backends. Maglev can be used as a drop in replacement for the ring hash load balancer. Maglev is not as stable as ring hash but has faster table lookup build times and host selection times. For more information about Maglev, see https://ai.google/research/pubs/pub44824", + "This algorithm implements consistent hashing to backends. Maglev can be used as a drop in replacement for the ring hash load balancer. Maglev is not as stable as ring hash but has faster table lookup build times and host selection times. For more information about Maglev, see Maglev: A Fast and Reliable Software Network Load Balancer.", "Backend host is selected based on the client connection metadata, i.e., connections are opened to the same address as the destination address of the incoming connection before the connection was redirected to the load balancer.", "The load balancer selects a random healthy host.", "The ring/modulo hash load balancer implements consistent hashing to backends. The algorithm has the property that the addition/removal of a host from a set of N hosts only affects 1/N of the requests.", @@ -43458,7 +43458,7 @@ "type": "array" }, "serviceLbPolicy": { - "description": "URL to networkservices.ServiceLbPolicy resource. Can only be set if load balancing scheme is EXTERNAL, EXTERNAL_MANAGED, INTERNAL_MANAGED or INTERNAL_SELF_MANAGED and the scope is global.", + "description": "URL to networkservices.ServiceLbPolicy resource. Can only be set if load balancing scheme is EXTERNAL_MANAGED, INTERNAL_MANAGED or INTERNAL_SELF_MANAGED and the scope is global.", "type": "string" }, "sessionAffinity": { @@ -44336,7 +44336,7 @@ "enumDescriptions": [ "", "An O(1) algorithm which selects two random healthy hosts and picks the host which has fewer active requests.", - "This algorithm implements consistent hashing to backends. Maglev can be used as a drop in replacement for the ring hash load balancer. Maglev is not as stable as ring hash but has faster table lookup build times and host selection times. For more information about Maglev, see https://ai.google/research/pubs/pub44824", + "This algorithm implements consistent hashing to backends. Maglev can be used as a drop in replacement for the ring hash load balancer. Maglev is not as stable as ring hash but has faster table lookup build times and host selection times. For more information about Maglev, see Maglev: A Fast and Reliable Software Network Load Balancer.", "Backend host is selected based on the client connection metadata, i.e., connections are opened to the same address as the destination address of the incoming connection before the connection was redirected to the load balancer.", "The load balancer selects a random healthy host.", "The ring/modulo hash load balancer implements consistent hashing to backends. The algorithm has the property that the addition/removal of a host from a set of N hosts only affects 1/N of the requests.", @@ -86210,7 +86210,7 @@ "type": "object" }, "TargetHttpsProxy": { - "description": "Represents a Target HTTPS Proxy resource. Google Compute Engine has two Target HTTPS Proxy resources: * [Global](/compute/docs/reference/rest/v1/targetHttpsProxies) * [Regional](/compute/docs/reference/rest/v1/regionTargetHttpsProxies) A target HTTPS proxy is a component of Google Cloud HTTPS load balancers. * targetHttpProxies are used by global external Application Load Balancers, classic Application Load Balancers, cross-region internal Application Load Balancers, and Traffic Director. * regionTargetHttpProxies are used by regional internal Application Load Balancers and regional external Application Load Balancers. Forwarding rules reference a target HTTPS proxy, and the target proxy then references a URL map. For more information, read Using Target Proxies and Forwarding rule concepts.", + "description": "Represents a Target HTTPS Proxy resource. Google Compute Engine has two Target HTTPS Proxy resources: * [Global](/compute/docs/reference/rest/v1/targetHttpsProxies) * [Regional](/compute/docs/reference/rest/v1/regionTargetHttpsProxies) A target HTTPS proxy is a component of Google Cloud HTTPS load balancers. * targetHttpsProxies are used by global external Application Load Balancers, classic Application Load Balancers, cross-region internal Application Load Balancers, and Traffic Director. * regionTargetHttpsProxies are used by regional internal Application Load Balancers and regional external Application Load Balancers. Forwarding rules reference a target HTTPS proxy, and the target proxy then references a URL map. For more information, read Using Target Proxies and Forwarding rule concepts.", "id": "TargetHttpsProxy", "properties": { "authorizationPolicy": { @@ -89247,6 +89247,7 @@ "FAILURE_MEMORY", "FAILURE_NETWORK", "FAILURE_NVLINK", + "FAILURE_REDUNDANT_HARDWARE_FAULT", "INFRASTRUCTURE_RELOCATION", "MAINTENANCE_REASON_UNKNOWN", "PLANNED_NETWORK_UPDATE", @@ -89262,6 +89263,7 @@ "Maintenance due to memory errors.", "Maintenance due to network errors.", "Maintenance due to NVLink failure.", + "Maintenance due to redundant hardware fault.", "Maintenance due to infrastructure relocation.", "Unknown maintenance reason. Do not use this value.", "Maintenance due to planned network update.", diff --git a/src/apis/compute/v1.ts b/src/apis/compute/v1.ts index f517fa58506..c974a10c9ed 100644 --- a/src/apis/compute/v1.ts +++ b/src/apis/compute/v1.ts @@ -1804,7 +1804,7 @@ export namespace compute_v1 { */ localityLbPolicies?: Schema$BackendServiceLocalityLoadBalancingPolicyConfig[]; /** - * The load balancing algorithm used within the scope of the locality. The possible values are: - ROUND_ROBIN: This is a simple policy in which each healthy backend is selected in round robin order. This is the default. - LEAST_REQUEST: An O(1) algorithm which selects two random healthy hosts and picks the host which has fewer active requests. - RING_HASH: The ring/modulo hash load balancer implements consistent hashing to backends. The algorithm has the property that the addition/removal of a host from a set of N hosts only affects 1/N of the requests. - RANDOM: The load balancer selects a random healthy host. - ORIGINAL_DESTINATION: Backend host is selected based on the client connection metadata, i.e., connections are opened to the same address as the destination address of the incoming connection before the connection was redirected to the load balancer. - MAGLEV: used as a drop in replacement for the ring hash load balancer. Maglev is not as stable as ring hash but has faster table lookup build times and host selection times. For more information about Maglev, see https://ai.google/research/pubs/pub44824 - WEIGHTED_ROUND_ROBIN: Per-endpoint Weighted Round Robin Load Balancing using weights computed from Backend reported Custom Metrics. If set, the Backend Service responses are expected to contain non-standard HTTP response header field Endpoint-Load-Metrics. The reported metrics to use for computing the weights are specified via the customMetrics field. This field is applicable to either: - A regional backend service with the service_protocol set to HTTP, HTTPS, HTTP2 or H2C, and load_balancing_scheme set to INTERNAL_MANAGED. - A global backend service with the load_balancing_scheme set to INTERNAL_SELF_MANAGED, INTERNAL_MANAGED, or EXTERNAL_MANAGED. If sessionAffinity is not configured—that is, if session affinity remains at the default value of NONE—then the default value for localityLbPolicy is ROUND_ROBIN. If session affinity is set to a value other than NONE, then the default value for localityLbPolicy is MAGLEV. Only ROUND_ROBIN and RING_HASH are supported when the backend service is referenced by a URL map that is bound to target gRPC proxy that has validateForProxyless field set to true. localityLbPolicy cannot be specified with haPolicy. + * The load balancing algorithm used within the scope of the locality. The possible values are: - ROUND_ROBIN: This is a simple policy in which each healthy backend is selected in round robin order. This is the default. - LEAST_REQUEST: An O(1) algorithm which selects two random healthy hosts and picks the host which has fewer active requests. - RING_HASH: The ring/modulo hash load balancer implements consistent hashing to backends. The algorithm has the property that the addition/removal of a host from a set of N hosts only affects 1/N of the requests. - RANDOM: The load balancer selects a random healthy host. - ORIGINAL_DESTINATION: Backend host is selected based on the client connection metadata, i.e., connections are opened to the same address as the destination address of the incoming connection before the connection was redirected to the load balancer. - MAGLEV: used as a drop in replacement for the ring hash load balancer. Maglev is not as stable as ring hash but has faster table lookup build times and host selection times. For more information about Maglev, see Maglev: A Fast and Reliable Software Network Load Balancer. - WEIGHTED_ROUND_ROBIN: Per-endpoint Weighted Round Robin Load Balancing using weights computed from Backend reported Custom Metrics. If set, the Backend Service responses are expected to contain non-standard HTTP response header field Endpoint-Load-Metrics. The reported metrics to use for computing the weights are specified via the customMetrics field. This field is applicable to either: - A regional backend service with the service_protocol set to HTTP, HTTPS, HTTP2 or H2C, and load_balancing_scheme set to INTERNAL_MANAGED. - A global backend service with the load_balancing_scheme set to INTERNAL_SELF_MANAGED, INTERNAL_MANAGED, or EXTERNAL_MANAGED. If sessionAffinity is not configured—that is, if session affinity remains at the default value of NONE—then the default value for localityLbPolicy is ROUND_ROBIN. If session affinity is set to a value other than NONE, then the default value for localityLbPolicy is MAGLEV. Only ROUND_ROBIN and RING_HASH are supported when the backend service is referenced by a URL map that is bound to target gRPC proxy that has validateForProxyless field set to true. localityLbPolicy cannot be specified with haPolicy. */ localityLbPolicy?: string | null; /** @@ -1868,7 +1868,7 @@ export namespace compute_v1 { */ serviceBindings?: string[] | null; /** - * URL to networkservices.ServiceLbPolicy resource. Can only be set if load balancing scheme is EXTERNAL, EXTERNAL_MANAGED, INTERNAL_MANAGED or INTERNAL_SELF_MANAGED and the scope is global. + * URL to networkservices.ServiceLbPolicy resource. Can only be set if load balancing scheme is EXTERNAL_MANAGED, INTERNAL_MANAGED or INTERNAL_SELF_MANAGED and the scope is global. */ serviceLbPolicy?: string | null; /** @@ -18144,7 +18144,7 @@ export namespace compute_v1 { sslCertificates?: string[] | null; } /** - * Represents a Target HTTPS Proxy resource. Google Compute Engine has two Target HTTPS Proxy resources: * [Global](/compute/docs/reference/rest/v1/targetHttpsProxies) * [Regional](/compute/docs/reference/rest/v1/regionTargetHttpsProxies) A target HTTPS proxy is a component of Google Cloud HTTPS load balancers. * targetHttpProxies are used by global external Application Load Balancers, classic Application Load Balancers, cross-region internal Application Load Balancers, and Traffic Director. * regionTargetHttpProxies are used by regional internal Application Load Balancers and regional external Application Load Balancers. Forwarding rules reference a target HTTPS proxy, and the target proxy then references a URL map. For more information, read Using Target Proxies and Forwarding rule concepts. + * Represents a Target HTTPS Proxy resource. Google Compute Engine has two Target HTTPS Proxy resources: * [Global](/compute/docs/reference/rest/v1/targetHttpsProxies) * [Regional](/compute/docs/reference/rest/v1/regionTargetHttpsProxies) A target HTTPS proxy is a component of Google Cloud HTTPS load balancers. * targetHttpsProxies are used by global external Application Load Balancers, classic Application Load Balancers, cross-region internal Application Load Balancers, and Traffic Director. * regionTargetHttpsProxies are used by regional internal Application Load Balancers and regional external Application Load Balancers. Forwarding rules reference a target HTTPS proxy, and the target proxy then references a URL map. For more information, read Using Target Proxies and Forwarding rule concepts. */ export interface Schema$TargetHttpsProxy { /** From 36318083c236c1a65fa8dba975c421c80620e69c Mon Sep 17 00:00:00 2001 From: Yoshi Automation Date: Fri, 15 Aug 2025 01:37:18 +0000 Subject: [PATCH 03/11] feat(dataplex): update the API #### dataplex:v1 The following keys were added: - schemas.GoogleCloudDataplexV1DataProfileResult.properties.profile.readOnly - schemas.GoogleCloudDataplexV1DataProfileResult.properties.rowCount.readOnly - schemas.GoogleCloudDataplexV1DataProfileResult.properties.scannedData.readOnly - schemas.GoogleCloudDataplexV1DataProfileResultProfile.properties.fields.readOnly - schemas.GoogleCloudDataplexV1DataProfileResultProfileField.properties.mode.readOnly - schemas.GoogleCloudDataplexV1DataProfileResultProfileField.properties.name.readOnly - schemas.GoogleCloudDataplexV1DataProfileResultProfileField.properties.profile.readOnly - schemas.GoogleCloudDataplexV1DataProfileResultProfileField.properties.type.readOnly - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfo.properties.distinctRatio.readOnly - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfo.properties.nullRatio.readOnly - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfo.properties.topNValues.readOnly - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoDoubleFieldInfo.properties.average.readOnly - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoDoubleFieldInfo.properties.max.readOnly - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoDoubleFieldInfo.properties.min.readOnly - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoDoubleFieldInfo.properties.quartiles.readOnly - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoDoubleFieldInfo.properties.standardDeviation.readOnly - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoIntegerFieldInfo.properties.average.readOnly - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoIntegerFieldInfo.properties.max.readOnly - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoIntegerFieldInfo.properties.min.readOnly - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoIntegerFieldInfo.properties.quartiles.readOnly - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoIntegerFieldInfo.properties.standardDeviation.readOnly - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoStringFieldInfo.properties.averageLength.readOnly - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoStringFieldInfo.properties.maxLength.readOnly - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoStringFieldInfo.properties.minLength.readOnly - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoTopNValue.properties.count.readOnly - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoTopNValue.properties.ratio.readOnly - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoTopNValue.properties.value.readOnly - schemas.GoogleCloudDataplexV1DataQualityDimension.properties.name.readOnly - schemas.GoogleCloudDataplexV1ScannedDataIncrementalField.properties.end.readOnly - schemas.GoogleCloudDataplexV1ScannedDataIncrementalField.properties.field.readOnly - schemas.GoogleCloudDataplexV1ScannedDataIncrementalField.properties.start.readOnly The following keys were changed: - schemas.GoogleCloudDataplexV1DataProfileResult.properties.profile.description - schemas.GoogleCloudDataplexV1DataProfileResult.properties.rowCount.description - schemas.GoogleCloudDataplexV1DataProfileResult.properties.scannedData.description - schemas.GoogleCloudDataplexV1DataProfileResultProfile.properties.fields.description - schemas.GoogleCloudDataplexV1DataProfileResultProfileField.properties.mode.description - schemas.GoogleCloudDataplexV1DataProfileResultProfileField.properties.name.description - schemas.GoogleCloudDataplexV1DataProfileResultProfileField.properties.profile.description - schemas.GoogleCloudDataplexV1DataProfileResultProfileField.properties.type.description - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfo.properties.distinctRatio.description - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfo.properties.nullRatio.description - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfo.properties.topNValues.description - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoDoubleFieldInfo.properties.average.description - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoDoubleFieldInfo.properties.max.description - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoDoubleFieldInfo.properties.min.description - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoDoubleFieldInfo.properties.quartiles.description - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoDoubleFieldInfo.properties.standardDeviation.description - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoIntegerFieldInfo.properties.average.description - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoIntegerFieldInfo.properties.max.description - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoIntegerFieldInfo.properties.min.description - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoIntegerFieldInfo.properties.quartiles.description - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoIntegerFieldInfo.properties.standardDeviation.description - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoStringFieldInfo.properties.averageLength.description - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoStringFieldInfo.properties.maxLength.description - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoStringFieldInfo.properties.minLength.description - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoTopNValue.properties.count.description - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoTopNValue.properties.ratio.description - schemas.GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoTopNValue.properties.value.description - schemas.GoogleCloudDataplexV1DataProfileSpec.properties.rowFilter.description - schemas.GoogleCloudDataplexV1DataProfileSpecPostScanActionsBigQueryExport.properties.resultsTable.description - schemas.GoogleCloudDataplexV1DataQualityDimension.properties.name.description - schemas.GoogleCloudDataplexV1ScannedDataIncrementalField.properties.end.description - schemas.GoogleCloudDataplexV1ScannedDataIncrementalField.properties.field.description - schemas.GoogleCloudDataplexV1ScannedDataIncrementalField.properties.start.description --- discovery/dataplex-v1.json | 99 +++++++++++++++++++++++++------------- src/apis/dataplex/v1.ts | 66 ++++++++++++------------- 2 files changed, 98 insertions(+), 67 deletions(-) diff --git a/discovery/dataplex-v1.json b/discovery/dataplex-v1.json index f1c96fc1a78..e8a42d3b1f0 100644 --- a/discovery/dataplex-v1.json +++ b/discovery/dataplex-v1.json @@ -7211,7 +7211,7 @@ } } }, - "revision": "20250805", + "revision": "20250808", "rootUrl": "https://dataplex.googleapis.com/", "schemas": { "Empty": { @@ -8594,16 +8594,19 @@ }, "profile": { "$ref": "GoogleCloudDataplexV1DataProfileResultProfile", - "description": "The profile information per field." + "description": "Output only. The profile information per field.", + "readOnly": true }, "rowCount": { - "description": "The count of rows scanned.", + "description": "Output only. The count of rows scanned.", "format": "int64", + "readOnly": true, "type": "string" }, "scannedData": { "$ref": "GoogleCloudDataplexV1ScannedData", - "description": "The data scanned for this result." + "description": "Output only. The data scanned for this result.", + "readOnly": true } }, "type": "object" @@ -8654,10 +8657,11 @@ "id": "GoogleCloudDataplexV1DataProfileResultProfile", "properties": { "fields": { - "description": "List of fields with structural and profile information for each field.", + "description": "Output only. List of fields with structural and profile information for each field.", "items": { "$ref": "GoogleCloudDataplexV1DataProfileResultProfileField" }, + "readOnly": true, "type": "array" } }, @@ -8668,19 +8672,23 @@ "id": "GoogleCloudDataplexV1DataProfileResultProfileField", "properties": { "mode": { - "description": "The mode of the field. Possible values include: REQUIRED, if it is a required field. NULLABLE, if it is an optional field. REPEATED, if it is a repeated field.", + "description": "Output only. The mode of the field. Possible values include: REQUIRED, if it is a required field. NULLABLE, if it is an optional field. REPEATED, if it is a repeated field.", + "readOnly": true, "type": "string" }, "name": { - "description": "The name of the field.", + "description": "Output only. The name of the field.", + "readOnly": true, "type": "string" }, "profile": { "$ref": "GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfo", - "description": "Profile information for the corresponding field." + "description": "Output only. Profile information for the corresponding field.", + "readOnly": true }, "type": { - "description": "The data type retrieved from the schema of the data source. For instance, for a BigQuery native table, it is the BigQuery Table Schema (https://cloud.google.com/bigquery/docs/reference/rest/v2/tables#tablefieldschema). For a Dataplex Universal Catalog Entity, it is the Entity Schema (https://cloud.google.com/dataplex/docs/reference/rpc/google.cloud.dataplex.v1#type_3).", + "description": "Output only. The data type retrieved from the schema of the data source. For instance, for a BigQuery native table, it is the BigQuery Table Schema (https://cloud.google.com/bigquery/docs/reference/rest/v2/tables#tablefieldschema). For a Dataplex Universal Catalog Entity, it is the Entity Schema (https://cloud.google.com/dataplex/docs/reference/rpc/google.cloud.dataplex.v1#type_3).", + "readOnly": true, "type": "string" } }, @@ -8691,8 +8699,9 @@ "id": "GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfo", "properties": { "distinctRatio": { - "description": "Ratio of rows with distinct values against total scanned rows. Not available for complex non-groupable field type, including RECORD, ARRAY, GEOGRAPHY, and JSON, as well as fields with REPEATABLE mode.", + "description": "Output only. Ratio of rows with distinct values against total scanned rows. Not available for complex non-groupable field type, including RECORD, ARRAY, GEOGRAPHY, and JSON, as well as fields with REPEATABLE mode.", "format": "double", + "readOnly": true, "type": "number" }, "doubleProfile": { @@ -8704,8 +8713,9 @@ "description": "Integer type field information." }, "nullRatio": { - "description": "Ratio of rows with null value against total scanned rows.", + "description": "Output only. Ratio of rows with null value against total scanned rows.", "format": "double", + "readOnly": true, "type": "number" }, "stringProfile": { @@ -8713,10 +8723,11 @@ "description": "String type field information." }, "topNValues": { - "description": "The list of top N non-null values, frequency and ratio with which they occur in the scanned data. N is 10 or equal to the number of distinct values in the field, whichever is smaller. Not available for complex non-groupable field type, including RECORD, ARRAY, GEOGRAPHY, and JSON, as well as fields with REPEATABLE mode.", + "description": "Output only. The list of top N non-null values, frequency and ratio with which they occur in the scanned data. N is 10 or equal to the number of distinct values in the field, whichever is smaller. Not available for complex non-groupable field type, including RECORD, ARRAY, GEOGRAPHY, and JSON, as well as fields with REPEATABLE mode.", "items": { "$ref": "GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoTopNValue" }, + "readOnly": true, "type": "array" } }, @@ -8727,31 +8738,36 @@ "id": "GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoDoubleFieldInfo", "properties": { "average": { - "description": "Average of non-null values in the scanned data. NaN, if the field has a NaN.", + "description": "Output only. Average of non-null values in the scanned data. NaN, if the field has a NaN.", "format": "double", + "readOnly": true, "type": "number" }, "max": { - "description": "Maximum of non-null values in the scanned data. NaN, if the field has a NaN.", + "description": "Output only. Maximum of non-null values in the scanned data. NaN, if the field has a NaN.", "format": "double", + "readOnly": true, "type": "number" }, "min": { - "description": "Minimum of non-null values in the scanned data. NaN, if the field has a NaN.", + "description": "Output only. Minimum of non-null values in the scanned data. NaN, if the field has a NaN.", "format": "double", + "readOnly": true, "type": "number" }, "quartiles": { - "description": "A quartile divides the number of data points into four parts, or quarters, of more-or-less equal size. Three main quartiles used are: The first quartile (Q1) splits off the lowest 25% of data from the highest 75%. It is also known as the lower or 25th empirical quartile, as 25% of the data is below this point. The second quartile (Q2) is the median of a data set. So, 50% of the data lies below this point. The third quartile (Q3) splits off the highest 25% of data from the lowest 75%. It is known as the upper or 75th empirical quartile, as 75% of the data lies below this point. Here, the quartiles is provided as an ordered list of quartile values for the scanned data, occurring in order Q1, median, Q3.", + "description": "Output only. A quartile divides the number of data points into four parts, or quarters, of more-or-less equal size. Three main quartiles used are: The first quartile (Q1) splits off the lowest 25% of data from the highest 75%. It is also known as the lower or 25th empirical quartile, as 25% of the data is below this point. The second quartile (Q2) is the median of a data set. So, 50% of the data lies below this point. The third quartile (Q3) splits off the highest 25% of data from the lowest 75%. It is known as the upper or 75th empirical quartile, as 75% of the data lies below this point. Here, the quartiles is provided as an ordered list of quartile values for the scanned data, occurring in order Q1, median, Q3.", "items": { "format": "double", "type": "number" }, + "readOnly": true, "type": "array" }, "standardDeviation": { - "description": "Standard deviation of non-null values in the scanned data. NaN, if the field has a NaN.", + "description": "Output only. Standard deviation of non-null values in the scanned data. NaN, if the field has a NaN.", "format": "double", + "readOnly": true, "type": "number" } }, @@ -8762,31 +8778,36 @@ "id": "GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoIntegerFieldInfo", "properties": { "average": { - "description": "Average of non-null values in the scanned data. NaN, if the field has a NaN.", + "description": "Output only. Average of non-null values in the scanned data. NaN, if the field has a NaN.", "format": "double", + "readOnly": true, "type": "number" }, "max": { - "description": "Maximum of non-null values in the scanned data. NaN, if the field has a NaN.", + "description": "Output only. Maximum of non-null values in the scanned data. NaN, if the field has a NaN.", "format": "int64", + "readOnly": true, "type": "string" }, "min": { - "description": "Minimum of non-null values in the scanned data. NaN, if the field has a NaN.", + "description": "Output only. Minimum of non-null values in the scanned data. NaN, if the field has a NaN.", "format": "int64", + "readOnly": true, "type": "string" }, "quartiles": { - "description": "A quartile divides the number of data points into four parts, or quarters, of more-or-less equal size. Three main quartiles used are: The first quartile (Q1) splits off the lowest 25% of data from the highest 75%. It is also known as the lower or 25th empirical quartile, as 25% of the data is below this point. The second quartile (Q2) is the median of a data set. So, 50% of the data lies below this point. The third quartile (Q3) splits off the highest 25% of data from the lowest 75%. It is known as the upper or 75th empirical quartile, as 75% of the data lies below this point. Here, the quartiles is provided as an ordered list of approximate quartile values for the scanned data, occurring in order Q1, median, Q3.", + "description": "Output only. A quartile divides the number of data points into four parts, or quarters, of more-or-less equal size. Three main quartiles used are: The first quartile (Q1) splits off the lowest 25% of data from the highest 75%. It is also known as the lower or 25th empirical quartile, as 25% of the data is below this point. The second quartile (Q2) is the median of a data set. So, 50% of the data lies below this point. The third quartile (Q3) splits off the highest 25% of data from the lowest 75%. It is known as the upper or 75th empirical quartile, as 75% of the data lies below this point. Here, the quartiles is provided as an ordered list of approximate quartile values for the scanned data, occurring in order Q1, median, Q3.", "items": { "format": "int64", "type": "string" }, + "readOnly": true, "type": "array" }, "standardDeviation": { - "description": "Standard deviation of non-null values in the scanned data. NaN, if the field has a NaN.", + "description": "Output only. Standard deviation of non-null values in the scanned data. NaN, if the field has a NaN.", "format": "double", + "readOnly": true, "type": "number" } }, @@ -8797,18 +8818,21 @@ "id": "GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoStringFieldInfo", "properties": { "averageLength": { - "description": "Average length of non-null values in the scanned data.", + "description": "Output only. Average length of non-null values in the scanned data.", "format": "double", + "readOnly": true, "type": "number" }, "maxLength": { - "description": "Maximum length of non-null values in the scanned data.", + "description": "Output only. Maximum length of non-null values in the scanned data.", "format": "int64", + "readOnly": true, "type": "string" }, "minLength": { - "description": "Minimum length of non-null values in the scanned data.", + "description": "Output only. Minimum length of non-null values in the scanned data.", "format": "int64", + "readOnly": true, "type": "string" } }, @@ -8819,17 +8843,20 @@ "id": "GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoTopNValue", "properties": { "count": { - "description": "Count of the corresponding value in the scanned data.", + "description": "Output only. Count of the corresponding value in the scanned data.", "format": "int64", + "readOnly": true, "type": "string" }, "ratio": { - "description": "Ratio of the corresponding value in the field against the total number of rows in the scanned data.", + "description": "Output only. Ratio of the corresponding value in the field against the total number of rows in the scanned data.", "format": "double", + "readOnly": true, "type": "number" }, "value": { - "description": "String value of a top N non-null value.", + "description": "Output only. String value of a top N non-null value.", + "readOnly": true, "type": "string" } }, @@ -8852,7 +8879,7 @@ "description": "Optional. Actions to take upon job completion.." }, "rowFilter": { - "description": "Optional. A filter applied to all rows in a single DataScan job. The filter needs to be a valid SQL expression for a WHERE clause in GoogleSQL syntax (https://cloud.google.com/bigquery/docs/reference/standard-sql/query-syntax#where_clause).Example: col1 >= 0 AND col2 < 10", + "description": "Optional. A filter applied to all rows in a single DataScan job. The filter needs to be a valid SQL expression for a WHERE clause in BigQuery standard SQL syntax. Example: col1 >= 0 AND col2 < 10", "type": "string" }, "samplingPercent": { @@ -8879,7 +8906,7 @@ "id": "GoogleCloudDataplexV1DataProfileSpecPostScanActionsBigQueryExport", "properties": { "resultsTable": { - "description": "Optional. The BigQuery table to export DataProfileScan results to. Format: //bigquery.googleapis.com/projects/PROJECT_ID/datasets/DATASET_ID/tables/TABLE_ID or projects/PROJECT_ID/datasets/DATASET_ID/tables/TABLE_ID", + "description": "Optional. The BigQuery table to export DataProfileScan results to. Format: //bigquery.googleapis.com/projects/PROJECT_ID/datasets/DATASET_ID/tables/TABLE_ID", "type": "string" } }, @@ -8935,7 +8962,8 @@ "id": "GoogleCloudDataplexV1DataQualityDimension", "properties": { "name": { - "description": "Optional. The dimension name a rule belongs to. Custom dimension name is supported with all uppercase letters and maximum length of 30 characters.", + "description": "Output only. The dimension name a rule belongs to. Custom dimension name is supported with all uppercase letters and maximum length of 30 characters.", + "readOnly": true, "type": "string" } }, @@ -12854,15 +12882,18 @@ "id": "GoogleCloudDataplexV1ScannedDataIncrementalField", "properties": { "end": { - "description": "Value that marks the end of the range.", + "description": "Output only. Value that marks the end of the range.", + "readOnly": true, "type": "string" }, "field": { - "description": "The field that contains values which monotonically increases over time (e.g. a timestamp column).", + "description": "Output only. The field that contains values which monotonically increases over time (e.g. a timestamp column).", + "readOnly": true, "type": "string" }, "start": { - "description": "Value that marks the start of the range.", + "description": "Output only. Value that marks the start of the range.", + "readOnly": true, "type": "string" } }, diff --git a/src/apis/dataplex/v1.ts b/src/apis/dataplex/v1.ts index 88764e8e8ec..8dcfc68d279 100644 --- a/src/apis/dataplex/v1.ts +++ b/src/apis/dataplex/v1.ts @@ -1106,15 +1106,15 @@ export namespace dataplex_v1 { */ postScanActionsResult?: Schema$GoogleCloudDataplexV1DataProfileResultPostScanActionsResult; /** - * The profile information per field. + * Output only. The profile information per field. */ profile?: Schema$GoogleCloudDataplexV1DataProfileResultProfile; /** - * The count of rows scanned. + * Output only. The count of rows scanned. */ rowCount?: string | null; /** - * The data scanned for this result. + * Output only. The data scanned for this result. */ scannedData?: Schema$GoogleCloudDataplexV1ScannedData; } @@ -1145,7 +1145,7 @@ export namespace dataplex_v1 { */ export interface Schema$GoogleCloudDataplexV1DataProfileResultProfile { /** - * List of fields with structural and profile information for each field. + * Output only. List of fields with structural and profile information for each field. */ fields?: Schema$GoogleCloudDataplexV1DataProfileResultProfileField[]; } @@ -1154,19 +1154,19 @@ export namespace dataplex_v1 { */ export interface Schema$GoogleCloudDataplexV1DataProfileResultProfileField { /** - * The mode of the field. Possible values include: REQUIRED, if it is a required field. NULLABLE, if it is an optional field. REPEATED, if it is a repeated field. + * Output only. The mode of the field. Possible values include: REQUIRED, if it is a required field. NULLABLE, if it is an optional field. REPEATED, if it is a repeated field. */ mode?: string | null; /** - * The name of the field. + * Output only. The name of the field. */ name?: string | null; /** - * Profile information for the corresponding field. + * Output only. Profile information for the corresponding field. */ profile?: Schema$GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfo; /** - * The data type retrieved from the schema of the data source. For instance, for a BigQuery native table, it is the BigQuery Table Schema (https://cloud.google.com/bigquery/docs/reference/rest/v2/tables#tablefieldschema). For a Dataplex Universal Catalog Entity, it is the Entity Schema (https://cloud.google.com/dataplex/docs/reference/rpc/google.cloud.dataplex.v1#type_3). + * Output only. The data type retrieved from the schema of the data source. For instance, for a BigQuery native table, it is the BigQuery Table Schema (https://cloud.google.com/bigquery/docs/reference/rest/v2/tables#tablefieldschema). For a Dataplex Universal Catalog Entity, it is the Entity Schema (https://cloud.google.com/dataplex/docs/reference/rpc/google.cloud.dataplex.v1#type_3). */ type?: string | null; } @@ -1175,7 +1175,7 @@ export namespace dataplex_v1 { */ export interface Schema$GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfo { /** - * Ratio of rows with distinct values against total scanned rows. Not available for complex non-groupable field type, including RECORD, ARRAY, GEOGRAPHY, and JSON, as well as fields with REPEATABLE mode. + * Output only. Ratio of rows with distinct values against total scanned rows. Not available for complex non-groupable field type, including RECORD, ARRAY, GEOGRAPHY, and JSON, as well as fields with REPEATABLE mode. */ distinctRatio?: number | null; /** @@ -1187,7 +1187,7 @@ export namespace dataplex_v1 { */ integerProfile?: Schema$GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoIntegerFieldInfo; /** - * Ratio of rows with null value against total scanned rows. + * Output only. Ratio of rows with null value against total scanned rows. */ nullRatio?: number | null; /** @@ -1195,7 +1195,7 @@ export namespace dataplex_v1 { */ stringProfile?: Schema$GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoStringFieldInfo; /** - * The list of top N non-null values, frequency and ratio with which they occur in the scanned data. N is 10 or equal to the number of distinct values in the field, whichever is smaller. Not available for complex non-groupable field type, including RECORD, ARRAY, GEOGRAPHY, and JSON, as well as fields with REPEATABLE mode. + * Output only. The list of top N non-null values, frequency and ratio with which they occur in the scanned data. N is 10 or equal to the number of distinct values in the field, whichever is smaller. Not available for complex non-groupable field type, including RECORD, ARRAY, GEOGRAPHY, and JSON, as well as fields with REPEATABLE mode. */ topNValues?: Schema$GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoTopNValue[]; } @@ -1204,23 +1204,23 @@ export namespace dataplex_v1 { */ export interface Schema$GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoDoubleFieldInfo { /** - * Average of non-null values in the scanned data. NaN, if the field has a NaN. + * Output only. Average of non-null values in the scanned data. NaN, if the field has a NaN. */ average?: number | null; /** - * Maximum of non-null values in the scanned data. NaN, if the field has a NaN. + * Output only. Maximum of non-null values in the scanned data. NaN, if the field has a NaN. */ max?: number | null; /** - * Minimum of non-null values in the scanned data. NaN, if the field has a NaN. + * Output only. Minimum of non-null values in the scanned data. NaN, if the field has a NaN. */ min?: number | null; /** - * A quartile divides the number of data points into four parts, or quarters, of more-or-less equal size. Three main quartiles used are: The first quartile (Q1) splits off the lowest 25% of data from the highest 75%. It is also known as the lower or 25th empirical quartile, as 25% of the data is below this point. The second quartile (Q2) is the median of a data set. So, 50% of the data lies below this point. The third quartile (Q3) splits off the highest 25% of data from the lowest 75%. It is known as the upper or 75th empirical quartile, as 75% of the data lies below this point. Here, the quartiles is provided as an ordered list of quartile values for the scanned data, occurring in order Q1, median, Q3. + * Output only. A quartile divides the number of data points into four parts, or quarters, of more-or-less equal size. Three main quartiles used are: The first quartile (Q1) splits off the lowest 25% of data from the highest 75%. It is also known as the lower or 25th empirical quartile, as 25% of the data is below this point. The second quartile (Q2) is the median of a data set. So, 50% of the data lies below this point. The third quartile (Q3) splits off the highest 25% of data from the lowest 75%. It is known as the upper or 75th empirical quartile, as 75% of the data lies below this point. Here, the quartiles is provided as an ordered list of quartile values for the scanned data, occurring in order Q1, median, Q3. */ quartiles?: number[] | null; /** - * Standard deviation of non-null values in the scanned data. NaN, if the field has a NaN. + * Output only. Standard deviation of non-null values in the scanned data. NaN, if the field has a NaN. */ standardDeviation?: number | null; } @@ -1229,23 +1229,23 @@ export namespace dataplex_v1 { */ export interface Schema$GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoIntegerFieldInfo { /** - * Average of non-null values in the scanned data. NaN, if the field has a NaN. + * Output only. Average of non-null values in the scanned data. NaN, if the field has a NaN. */ average?: number | null; /** - * Maximum of non-null values in the scanned data. NaN, if the field has a NaN. + * Output only. Maximum of non-null values in the scanned data. NaN, if the field has a NaN. */ max?: string | null; /** - * Minimum of non-null values in the scanned data. NaN, if the field has a NaN. + * Output only. Minimum of non-null values in the scanned data. NaN, if the field has a NaN. */ min?: string | null; /** - * A quartile divides the number of data points into four parts, or quarters, of more-or-less equal size. Three main quartiles used are: The first quartile (Q1) splits off the lowest 25% of data from the highest 75%. It is also known as the lower or 25th empirical quartile, as 25% of the data is below this point. The second quartile (Q2) is the median of a data set. So, 50% of the data lies below this point. The third quartile (Q3) splits off the highest 25% of data from the lowest 75%. It is known as the upper or 75th empirical quartile, as 75% of the data lies below this point. Here, the quartiles is provided as an ordered list of approximate quartile values for the scanned data, occurring in order Q1, median, Q3. + * Output only. A quartile divides the number of data points into four parts, or quarters, of more-or-less equal size. Three main quartiles used are: The first quartile (Q1) splits off the lowest 25% of data from the highest 75%. It is also known as the lower or 25th empirical quartile, as 25% of the data is below this point. The second quartile (Q2) is the median of a data set. So, 50% of the data lies below this point. The third quartile (Q3) splits off the highest 25% of data from the lowest 75%. It is known as the upper or 75th empirical quartile, as 75% of the data lies below this point. Here, the quartiles is provided as an ordered list of approximate quartile values for the scanned data, occurring in order Q1, median, Q3. */ quartiles?: string[] | null; /** - * Standard deviation of non-null values in the scanned data. NaN, if the field has a NaN. + * Output only. Standard deviation of non-null values in the scanned data. NaN, if the field has a NaN. */ standardDeviation?: number | null; } @@ -1254,15 +1254,15 @@ export namespace dataplex_v1 { */ export interface Schema$GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoStringFieldInfo { /** - * Average length of non-null values in the scanned data. + * Output only. Average length of non-null values in the scanned data. */ averageLength?: number | null; /** - * Maximum length of non-null values in the scanned data. + * Output only. Maximum length of non-null values in the scanned data. */ maxLength?: string | null; /** - * Minimum length of non-null values in the scanned data. + * Output only. Minimum length of non-null values in the scanned data. */ minLength?: string | null; } @@ -1271,15 +1271,15 @@ export namespace dataplex_v1 { */ export interface Schema$GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoTopNValue { /** - * Count of the corresponding value in the scanned data. + * Output only. Count of the corresponding value in the scanned data. */ count?: string | null; /** - * Ratio of the corresponding value in the field against the total number of rows in the scanned data. + * Output only. Ratio of the corresponding value in the field against the total number of rows in the scanned data. */ ratio?: number | null; /** - * String value of a top N non-null value. + * Output only. String value of a top N non-null value. */ value?: string | null; } @@ -1300,7 +1300,7 @@ export namespace dataplex_v1 { */ postScanActions?: Schema$GoogleCloudDataplexV1DataProfileSpecPostScanActions; /** - * Optional. A filter applied to all rows in a single DataScan job. The filter needs to be a valid SQL expression for a WHERE clause in GoogleSQL syntax (https://cloud.google.com/bigquery/docs/reference/standard-sql/query-syntax#where_clause).Example: col1 \>= 0 AND col2 < 10 + * Optional. A filter applied to all rows in a single DataScan job. The filter needs to be a valid SQL expression for a WHERE clause in BigQuery standard SQL syntax. Example: col1 \>= 0 AND col2 < 10 */ rowFilter?: string | null; /** @@ -1322,7 +1322,7 @@ export namespace dataplex_v1 { */ export interface Schema$GoogleCloudDataplexV1DataProfileSpecPostScanActionsBigQueryExport { /** - * Optional. The BigQuery table to export DataProfileScan results to. Format: //bigquery.googleapis.com/projects/PROJECT_ID/datasets/DATASET_ID/tables/TABLE_ID or projects/PROJECT_ID/datasets/DATASET_ID/tables/TABLE_ID + * Optional. The BigQuery table to export DataProfileScan results to. Format: //bigquery.googleapis.com/projects/PROJECT_ID/datasets/DATASET_ID/tables/TABLE_ID */ resultsTable?: string | null; } @@ -1361,7 +1361,7 @@ export namespace dataplex_v1 { */ export interface Schema$GoogleCloudDataplexV1DataQualityDimension { /** - * Optional. The dimension name a rule belongs to. Custom dimension name is supported with all uppercase letters and maximum length of 30 characters. + * Output only. The dimension name a rule belongs to. Custom dimension name is supported with all uppercase letters and maximum length of 30 characters. */ name?: string | null; } @@ -3955,15 +3955,15 @@ export namespace dataplex_v1 { */ export interface Schema$GoogleCloudDataplexV1ScannedDataIncrementalField { /** - * Value that marks the end of the range. + * Output only. Value that marks the end of the range. */ end?: string | null; /** - * The field that contains values which monotonically increases over time (e.g. a timestamp column). + * Output only. The field that contains values which monotonically increases over time (e.g. a timestamp column). */ field?: string | null; /** - * Value that marks the start of the range. + * Output only. Value that marks the start of the range. */ start?: string | null; } From f4af5d4bc2b2ca7c0fca919a7c139eaf31aa6fbf Mon Sep 17 00:00:00 2001 From: Yoshi Automation Date: Fri, 15 Aug 2025 01:37:18 +0000 Subject: [PATCH 04/11] fix(dialogflow): update the API #### dialogflow:v3 The following keys were changed: - schemas.GoogleCloudDialogflowCxV3SecuritySettingsAudioExportSettings.properties.audioExportPattern.description --- discovery/dialogflow-v3.json | 4 ++-- src/apis/dialogflow/v3.ts | 2 +- 2 files changed, 3 insertions(+), 3 deletions(-) diff --git a/discovery/dialogflow-v3.json b/discovery/dialogflow-v3.json index e1a819856df..0a3129976ac 100644 --- a/discovery/dialogflow-v3.json +++ b/discovery/dialogflow-v3.json @@ -5309,7 +5309,7 @@ } } }, - "revision": "20250804", + "revision": "20250812", "rootUrl": "https://dialogflow.googleapis.com/", "schemas": { "GoogleCloudDialogflowCxV3Action": { @@ -10800,7 +10800,7 @@ "id": "GoogleCloudDialogflowCxV3SecuritySettingsAudioExportSettings", "properties": { "audioExportPattern": { - "description": "Filename pattern for exported audio.", + "description": "Filename pattern for exported audio. {conversation} and {timestamp} are placeholders that will be replaced with the conversation ID and epoch micros of the conversation. For example, \"{conversation}/recording_{timestamp}.mulaw\".", "type": "string" }, "audioFormat": { diff --git a/src/apis/dialogflow/v3.ts b/src/apis/dialogflow/v3.ts index 606d1fadd4c..39846a18039 100644 --- a/src/apis/dialogflow/v3.ts +++ b/src/apis/dialogflow/v3.ts @@ -5837,7 +5837,7 @@ export namespace dialogflow_v3 { */ export interface Schema$GoogleCloudDialogflowCxV3SecuritySettingsAudioExportSettings { /** - * Filename pattern for exported audio. + * Filename pattern for exported audio. {conversation\} and {timestamp\} are placeholders that will be replaced with the conversation ID and epoch micros of the conversation. For example, "{conversation\}/recording_{timestamp\}.mulaw". */ audioExportPattern?: string | null; /** From dc726ed5e6a29434a39f653d43b387f5b9de536e Mon Sep 17 00:00:00 2001 From: Yoshi Automation Date: Fri, 15 Aug 2025 01:37:18 +0000 Subject: [PATCH 05/11] feat(documentai): update the API #### documentai:v1beta3 The following keys were added: - schemas.GoogleCloudDocumentaiV1beta3Document.properties.validationOutputs.description - schemas.GoogleCloudDocumentaiV1beta3Document.properties.validationOutputs.items.$ref - schemas.GoogleCloudDocumentaiV1beta3Document.properties.validationOutputs.type - schemas.GoogleCloudDocumentaiV1beta3DocumentEntity.properties.method.description - schemas.GoogleCloudDocumentaiV1beta3DocumentEntity.properties.method.enum - schemas.GoogleCloudDocumentaiV1beta3DocumentEntity.properties.method.enumDescriptions - schemas.GoogleCloudDocumentaiV1beta3DocumentEntity.properties.method.type - schemas.GoogleCloudDocumentaiV1beta3DocumentValidationOutput.description - schemas.GoogleCloudDocumentaiV1beta3DocumentValidationOutput.id - schemas.GoogleCloudDocumentaiV1beta3DocumentValidationOutput.properties.passAllRules.description - schemas.GoogleCloudDocumentaiV1beta3DocumentValidationOutput.properties.passAllRules.type - schemas.GoogleCloudDocumentaiV1beta3DocumentValidationOutput.properties.validationResults.description - schemas.GoogleCloudDocumentaiV1beta3DocumentValidationOutput.properties.validationResults.items.$ref - schemas.GoogleCloudDocumentaiV1beta3DocumentValidationOutput.properties.validationResults.type - schemas.GoogleCloudDocumentaiV1beta3DocumentValidationOutput.type - schemas.GoogleCloudDocumentaiV1beta3DocumentValidationOutputValidationResult.description - schemas.GoogleCloudDocumentaiV1beta3DocumentValidationOutputValidationResult.id - schemas.GoogleCloudDocumentaiV1beta3DocumentValidationOutputValidationResult.properties.ruleDescription.description - schemas.GoogleCloudDocumentaiV1beta3DocumentValidationOutputValidationResult.properties.ruleDescription.type - schemas.GoogleCloudDocumentaiV1beta3DocumentValidationOutputValidationResult.properties.ruleName.description - schemas.GoogleCloudDocumentaiV1beta3DocumentValidationOutputValidationResult.properties.ruleName.type - schemas.GoogleCloudDocumentaiV1beta3DocumentValidationOutputValidationResult.properties.validationDetails.description - schemas.GoogleCloudDocumentaiV1beta3DocumentValidationOutputValidationResult.properties.validationDetails.type - schemas.GoogleCloudDocumentaiV1beta3DocumentValidationOutputValidationResult.properties.validationResultType.description - schemas.GoogleCloudDocumentaiV1beta3DocumentValidationOutputValidationResult.properties.validationResultType.enum - schemas.GoogleCloudDocumentaiV1beta3DocumentValidationOutputValidationResult.properties.validationResultType.enumDescriptions - schemas.GoogleCloudDocumentaiV1beta3DocumentValidationOutputValidationResult.properties.validationResultType.type - schemas.GoogleCloudDocumentaiV1beta3DocumentValidationOutputValidationResult.type #### documentai:v1 The following keys were added: - schemas.GoogleCloudDocumentaiV1Document.properties.validationOutputs.description - schemas.GoogleCloudDocumentaiV1Document.properties.validationOutputs.items.$ref - schemas.GoogleCloudDocumentaiV1Document.properties.validationOutputs.type - schemas.GoogleCloudDocumentaiV1DocumentEntity.properties.method.description - schemas.GoogleCloudDocumentaiV1DocumentEntity.properties.method.enum - schemas.GoogleCloudDocumentaiV1DocumentEntity.properties.method.enumDescriptions - schemas.GoogleCloudDocumentaiV1DocumentEntity.properties.method.type - schemas.GoogleCloudDocumentaiV1DocumentValidationOutput.description - schemas.GoogleCloudDocumentaiV1DocumentValidationOutput.id - schemas.GoogleCloudDocumentaiV1DocumentValidationOutput.properties.passAllRules.description - schemas.GoogleCloudDocumentaiV1DocumentValidationOutput.properties.passAllRules.type - schemas.GoogleCloudDocumentaiV1DocumentValidationOutput.properties.validationResults.description - schemas.GoogleCloudDocumentaiV1DocumentValidationOutput.properties.validationResults.items.$ref - schemas.GoogleCloudDocumentaiV1DocumentValidationOutput.properties.validationResults.type - schemas.GoogleCloudDocumentaiV1DocumentValidationOutput.type - schemas.GoogleCloudDocumentaiV1DocumentValidationOutputValidationResult.description - schemas.GoogleCloudDocumentaiV1DocumentValidationOutputValidationResult.id - schemas.GoogleCloudDocumentaiV1DocumentValidationOutputValidationResult.properties.ruleDescription.description - schemas.GoogleCloudDocumentaiV1DocumentValidationOutputValidationResult.properties.ruleDescription.type - schemas.GoogleCloudDocumentaiV1DocumentValidationOutputValidationResult.properties.ruleName.description - schemas.GoogleCloudDocumentaiV1DocumentValidationOutputValidationResult.properties.ruleName.type - schemas.GoogleCloudDocumentaiV1DocumentValidationOutputValidationResult.properties.validationDetails.description - schemas.GoogleCloudDocumentaiV1DocumentValidationOutputValidationResult.properties.validationDetails.type - schemas.GoogleCloudDocumentaiV1DocumentValidationOutputValidationResult.properties.validationResultType.description - schemas.GoogleCloudDocumentaiV1DocumentValidationOutputValidationResult.properties.validationResultType.enum - schemas.GoogleCloudDocumentaiV1DocumentValidationOutputValidationResult.properties.validationResultType.enumDescriptions - schemas.GoogleCloudDocumentaiV1DocumentValidationOutputValidationResult.properties.validationResultType.type - schemas.GoogleCloudDocumentaiV1DocumentValidationOutputValidationResult.type --- discovery/documentai-v1.json | 78 ++++++++++++++++++++++++++++++- discovery/documentai-v1beta3.json | 78 ++++++++++++++++++++++++++++++- src/apis/documentai/v1.ts | 42 +++++++++++++++++ src/apis/documentai/v1beta3.ts | 42 +++++++++++++++++ 4 files changed, 238 insertions(+), 2 deletions(-) diff --git a/discovery/documentai-v1.json b/discovery/documentai-v1.json index 6ca26e37823..6e619e91802 100644 --- a/discovery/documentai-v1.json +++ b/discovery/documentai-v1.json @@ -1048,7 +1048,7 @@ } } }, - "revision": "20250716", + "revision": "20250811", "rootUrl": "https://documentai.googleapis.com/", "schemas": { "GoogleCloudDocumentaiUiv1beta3AutoLabelDocumentsMetadata": { @@ -2302,6 +2302,13 @@ "uri": { "description": "Optional. Currently supports Google Cloud Storage URI of the form `gs://bucket_name/object_name`. Object versioning is not supported. For more information, refer to [Google Cloud Storage Request URIs](https://cloud.google.com/storage/docs/reference-uris).", "type": "string" + }, + "validationOutputs": { + "description": "The output of the validation given the document and the validation rules. The output is appended to the document in the processing order.", + "items": { + "$ref": "GoogleCloudDocumentaiV1DocumentValidationOutput" + }, + "type": "array" } }, "type": "object" @@ -2607,6 +2614,20 @@ "description": "Optional. Text value of the entity e.g. `1600 Amphitheatre Pkwy`.", "type": "string" }, + "method": { + "description": "Optional. Specifies how the entity's value is obtained.", + "enum": [ + "METHOD_UNSPECIFIED", + "EXTRACT", + "DERIVE" + ], + "enumDescriptions": [ + "When the method is not specified, it should be treated as `EXTRACT`.", + "The entity's value is directly extracted as-is from the document text.", + "The entity's value is derived through inference and is not necessarily an exact text extraction from the document." + ], + "type": "string" + }, "normalizedValue": { "$ref": "GoogleCloudDocumentaiV1DocumentEntityNormalizedValue", "description": "Optional. Normalized entity value. Absent if the extracted value could not be converted or the type (e.g. address) is not supported for certain parsers. This field is also only populated for certain supported document types." @@ -3886,6 +3907,61 @@ }, "type": "object" }, + "GoogleCloudDocumentaiV1DocumentValidationOutput": { + "description": "The output of the validation given the document and the validation rules.", + "id": "GoogleCloudDocumentaiV1DocumentValidationOutput", + "properties": { + "passAllRules": { + "description": "The overall result of the validation, true if all applicable rules are valid.", + "type": "boolean" + }, + "validationResults": { + "description": "The result of each validation rule.", + "items": { + "$ref": "GoogleCloudDocumentaiV1DocumentValidationOutputValidationResult" + }, + "type": "array" + } + }, + "type": "object" + }, + "GoogleCloudDocumentaiV1DocumentValidationOutputValidationResult": { + "description": "Validation result for a single validation rule.", + "id": "GoogleCloudDocumentaiV1DocumentValidationOutputValidationResult", + "properties": { + "ruleDescription": { + "description": "The description of the validation rule.", + "type": "string" + }, + "ruleName": { + "description": "The name of the validation rule.", + "type": "string" + }, + "validationDetails": { + "description": "The detailed information of the running the validation process using the entity from the document based on the validation rule.", + "type": "string" + }, + "validationResultType": { + "description": "The result of the validation rule.", + "enum": [ + "VALIDATION_RESULT_TYPE_UNSPECIFIED", + "VALIDATION_RESULT_TYPE_VALID", + "VALIDATION_RESULT_TYPE_INVALID", + "VALIDATION_RESULT_TYPE_SKIPPED", + "VALIDATION_RESULT_TYPE_NOT_APPLICABLE" + ], + "enumDescriptions": [ + "The validation result type is unspecified.", + "The validation is valid.", + "The validation is invalid.", + "The validation is skipped.", + "The validation is not applicable." + ], + "type": "string" + } + }, + "type": "object" + }, "GoogleCloudDocumentaiV1EnableProcessorMetadata": { "description": "The long-running operation metadata for the EnableProcessor method.", "id": "GoogleCloudDocumentaiV1EnableProcessorMetadata", diff --git a/discovery/documentai-v1beta3.json b/discovery/documentai-v1beta3.json index be108955617..b1a62c3f91c 100644 --- a/discovery/documentai-v1beta3.json +++ b/discovery/documentai-v1beta3.json @@ -1290,7 +1290,7 @@ } } }, - "revision": "20250716", + "revision": "20250811", "rootUrl": "https://documentai.googleapis.com/", "schemas": { "GoogleCloudDocumentaiUiv1beta3AutoLabelDocumentsMetadata": { @@ -3194,6 +3194,13 @@ "uri": { "description": "Optional. Currently supports Google Cloud Storage URI of the form `gs://bucket_name/object_name`. Object versioning is not supported. For more information, refer to [Google Cloud Storage Request URIs](https://cloud.google.com/storage/docs/reference-uris).", "type": "string" + }, + "validationOutputs": { + "description": "The output of the validation given the document and the validation rules. The output is appended to the document in the processing order.", + "items": { + "$ref": "GoogleCloudDocumentaiV1beta3DocumentValidationOutput" + }, + "type": "array" } }, "type": "object" @@ -3625,6 +3632,20 @@ "description": "Optional. Text value of the entity e.g. `1600 Amphitheatre Pkwy`.", "type": "string" }, + "method": { + "description": "Optional. Specifies how the entity's value is obtained.", + "enum": [ + "METHOD_UNSPECIFIED", + "EXTRACT", + "DERIVE" + ], + "enumDescriptions": [ + "When the method is not specified, it should be treated as `EXTRACT`.", + "The entity's value is directly extracted as-is from the document text.", + "The entity's value is derived through inference and is not necessarily an exact text extraction from the document." + ], + "type": "string" + }, "normalizedValue": { "$ref": "GoogleCloudDocumentaiV1beta3DocumentEntityNormalizedValue", "description": "Optional. Normalized entity value. Absent if the extracted value could not be converted or the type (e.g. address) is not supported for certain parsers. This field is also only populated for certain supported document types." @@ -5018,6 +5039,61 @@ }, "type": "object" }, + "GoogleCloudDocumentaiV1beta3DocumentValidationOutput": { + "description": "The output of the validation given the document and the validation rules.", + "id": "GoogleCloudDocumentaiV1beta3DocumentValidationOutput", + "properties": { + "passAllRules": { + "description": "The overall result of the validation, true if all applicable rules are valid.", + "type": "boolean" + }, + "validationResults": { + "description": "The result of each validation rule.", + "items": { + "$ref": "GoogleCloudDocumentaiV1beta3DocumentValidationOutputValidationResult" + }, + "type": "array" + } + }, + "type": "object" + }, + "GoogleCloudDocumentaiV1beta3DocumentValidationOutputValidationResult": { + "description": "Validation result for a single validation rule.", + "id": "GoogleCloudDocumentaiV1beta3DocumentValidationOutputValidationResult", + "properties": { + "ruleDescription": { + "description": "The description of the validation rule.", + "type": "string" + }, + "ruleName": { + "description": "The name of the validation rule.", + "type": "string" + }, + "validationDetails": { + "description": "The detailed information of the running the validation process using the entity from the document based on the validation rule.", + "type": "string" + }, + "validationResultType": { + "description": "The result of the validation rule.", + "enum": [ + "VALIDATION_RESULT_TYPE_UNSPECIFIED", + "VALIDATION_RESULT_TYPE_VALID", + "VALIDATION_RESULT_TYPE_INVALID", + "VALIDATION_RESULT_TYPE_SKIPPED", + "VALIDATION_RESULT_TYPE_NOT_APPLICABLE" + ], + "enumDescriptions": [ + "The validation result type is unspecified.", + "The validation is valid.", + "The validation is invalid.", + "The validation is skipped.", + "The validation is not applicable." + ], + "type": "string" + } + }, + "type": "object" + }, "GoogleCloudDocumentaiV1beta3EnableProcessorMetadata": { "description": "The long-running operation metadata for the EnableProcessor method.", "id": "GoogleCloudDocumentaiV1beta3EnableProcessorMetadata", diff --git a/src/apis/documentai/v1.ts b/src/apis/documentai/v1.ts index 4b81e73549e..ba287507365 100644 --- a/src/apis/documentai/v1.ts +++ b/src/apis/documentai/v1.ts @@ -1552,6 +1552,10 @@ export namespace documentai_v1 { * Optional. Currently supports Google Cloud Storage URI of the form `gs://bucket_name/object_name`. Object versioning is not supported. For more information, refer to [Google Cloud Storage Request URIs](https://cloud.google.com/storage/docs/reference-uris). */ uri?: string | null; + /** + * The output of the validation given the document and the validation rules. The output is appended to the document in the processing order. + */ + validationOutputs?: Schema$GoogleCloudDocumentaiV1DocumentValidationOutput[]; } /** * Represents the chunks that the document is divided into. @@ -1783,6 +1787,10 @@ export namespace documentai_v1 { * Optional. Text value of the entity e.g. `1600 Amphitheatre Pkwy`. */ mentionText?: string | null; + /** + * Optional. Specifies how the entity's value is obtained. + */ + method?: string | null; /** * Optional. Normalized entity value. Absent if the extracted value could not be converted or the type (e.g. address) is not supported for certain parsers. This field is also only populated for certain supported document types. */ @@ -2712,6 +2720,40 @@ export namespace documentai_v1 { */ textAnchor?: Schema$GoogleCloudDocumentaiV1DocumentTextAnchor; } + /** + * The output of the validation given the document and the validation rules. + */ + export interface Schema$GoogleCloudDocumentaiV1DocumentValidationOutput { + /** + * The overall result of the validation, true if all applicable rules are valid. + */ + passAllRules?: boolean | null; + /** + * The result of each validation rule. + */ + validationResults?: Schema$GoogleCloudDocumentaiV1DocumentValidationOutputValidationResult[]; + } + /** + * Validation result for a single validation rule. + */ + export interface Schema$GoogleCloudDocumentaiV1DocumentValidationOutputValidationResult { + /** + * The description of the validation rule. + */ + ruleDescription?: string | null; + /** + * The name of the validation rule. + */ + ruleName?: string | null; + /** + * The detailed information of the running the validation process using the entity from the document based on the validation rule. + */ + validationDetails?: string | null; + /** + * The result of the validation rule. + */ + validationResultType?: string | null; + } /** * The long-running operation metadata for the EnableProcessor method. */ diff --git a/src/apis/documentai/v1beta3.ts b/src/apis/documentai/v1beta3.ts index d5d1da0a8f6..32906860d18 100644 --- a/src/apis/documentai/v1beta3.ts +++ b/src/apis/documentai/v1beta3.ts @@ -1253,6 +1253,10 @@ export namespace documentai_v1beta3 { * Optional. Currently supports Google Cloud Storage URI of the form `gs://bucket_name/object_name`. Object versioning is not supported. For more information, refer to [Google Cloud Storage Request URIs](https://cloud.google.com/storage/docs/reference-uris). */ uri?: string | null; + /** + * The output of the validation given the document and the validation rules. The output is appended to the document in the processing order. + */ + validationOutputs?: Schema$GoogleCloudDocumentaiV1beta3DocumentValidationOutput[]; } /** * Represents the annotation of a block or a chunk. @@ -1594,6 +1598,10 @@ export namespace documentai_v1beta3 { * Optional. Text value of the entity e.g. `1600 Amphitheatre Pkwy`. */ mentionText?: string | null; + /** + * Optional. Specifies how the entity's value is obtained. + */ + method?: string | null; /** * Optional. Normalized entity value. Absent if the extracted value could not be converted or the type (e.g. address) is not supported for certain parsers. This field is also only populated for certain supported document types. */ @@ -2603,6 +2611,40 @@ export namespace documentai_v1beta3 { */ textAnchor?: Schema$GoogleCloudDocumentaiV1beta3DocumentTextAnchor; } + /** + * The output of the validation given the document and the validation rules. + */ + export interface Schema$GoogleCloudDocumentaiV1beta3DocumentValidationOutput { + /** + * The overall result of the validation, true if all applicable rules are valid. + */ + passAllRules?: boolean | null; + /** + * The result of each validation rule. + */ + validationResults?: Schema$GoogleCloudDocumentaiV1beta3DocumentValidationOutputValidationResult[]; + } + /** + * Validation result for a single validation rule. + */ + export interface Schema$GoogleCloudDocumentaiV1beta3DocumentValidationOutputValidationResult { + /** + * The description of the validation rule. + */ + ruleDescription?: string | null; + /** + * The name of the validation rule. + */ + ruleName?: string | null; + /** + * The detailed information of the running the validation process using the entity from the document based on the validation rule. + */ + validationDetails?: string | null; + /** + * The result of the validation rule. + */ + validationResultType?: string | null; + } /** * The long-running operation metadata for the EnableProcessor method. */ From 2c5a3b422ff5fa2aefc19ce6deb42776abc265e6 Mon Sep 17 00:00:00 2001 From: Yoshi Automation Date: Fri, 15 Aug 2025 01:37:19 +0000 Subject: [PATCH 06/11] fix(firebasedataconnect): update the API #### firebasedataconnect:v1beta The following keys were changed: - schemas.PostgreSql.properties.ephemeral.description #### firebasedataconnect:v1 The following keys were changed: - schemas.PostgreSql.properties.ephemeral.description --- discovery/firebasedataconnect-v1.json | 4 ++-- discovery/firebasedataconnect-v1beta.json | 4 ++-- src/apis/firebasedataconnect/v1.ts | 2 +- src/apis/firebasedataconnect/v1beta.ts | 2 +- 4 files changed, 6 insertions(+), 6 deletions(-) diff --git a/discovery/firebasedataconnect-v1.json b/discovery/firebasedataconnect-v1.json index 50fed91a981..c029310656e 100644 --- a/discovery/firebasedataconnect-v1.json +++ b/discovery/firebasedataconnect-v1.json @@ -1076,7 +1076,7 @@ } } }, - "revision": "20250803", + "revision": "20250810", "rootUrl": "https://firebasedataconnect.googleapis.com/", "schemas": { "CancelOperationRequest": { @@ -1680,7 +1680,7 @@ "type": "string" }, "ephemeral": { - "description": "Output only. Ephemeral is true if this data connect service is served from in-memory emulation of Postgres. Data Connect service will transfer the data on a best-effort basis to the Cloud SQL instance once it's provisioned and linked. WARNING: When `ephemeral=true`, mutations to the database are not guaranteed to be durably persisted, even if an OK status code is returned. All or parts of the data may be lost or reverted to earlier versions.", + "description": "Output only. Ephemeral is true if this data connect service is served from temporary in-memory emulation of Postgres. While Cloud SQL is being provisioned, the data connect service provides the ephemeral service to help developers get started. Once the Cloud SQL is provisioned, Data Connect service will transfer its data on a best-effort basis to the Cloud SQL instance. WARNING: Ephemeral data sources will expire after 24 hour. The data will be lost if they aren't transferred to the Cloud SQL instance. WARNING: When `ephemeral=true`, mutations to the database are not guaranteed to be durably persisted, even if an OK status code is returned. All or parts of the data may be lost or reverted to earlier versions.", "readOnly": true, "type": "boolean" }, diff --git a/discovery/firebasedataconnect-v1beta.json b/discovery/firebasedataconnect-v1beta.json index 3ac2b9bc1ab..4c9f6d5ad77 100644 --- a/discovery/firebasedataconnect-v1beta.json +++ b/discovery/firebasedataconnect-v1beta.json @@ -1076,7 +1076,7 @@ } } }, - "revision": "20250803", + "revision": "20250810", "rootUrl": "https://firebasedataconnect.googleapis.com/", "schemas": { "CancelOperationRequest": { @@ -1680,7 +1680,7 @@ "type": "string" }, "ephemeral": { - "description": "Output only. Ephemeral is true if this data connect service is served from in-memory emulation of Postgres. Data Connect service will transfer the data on a best-effort basis to the Cloud SQL instance once it's provisioned and linked. WARNING: When `ephemeral=true`, mutations to the database are not guaranteed to be durably persisted, even if an OK status code is returned. All or parts of the data may be lost or reverted to earlier versions.", + "description": "Output only. Ephemeral is true if this data connect service is served from temporary in-memory emulation of Postgres. While Cloud SQL is being provisioned, the data connect service provides the ephemeral service to help developers get started. Once the Cloud SQL is provisioned, Data Connect service will transfer its data on a best-effort basis to the Cloud SQL instance. WARNING: Ephemeral data sources will expire after 24 hour. The data will be lost if they aren't transferred to the Cloud SQL instance. WARNING: When `ephemeral=true`, mutations to the database are not guaranteed to be durably persisted, even if an OK status code is returned. All or parts of the data may be lost or reverted to earlier versions.", "readOnly": true, "type": "boolean" }, diff --git a/src/apis/firebasedataconnect/v1.ts b/src/apis/firebasedataconnect/v1.ts index 5a473923a62..65676fe8164 100644 --- a/src/apis/firebasedataconnect/v1.ts +++ b/src/apis/firebasedataconnect/v1.ts @@ -535,7 +535,7 @@ export namespace firebasedataconnect_v1 { */ database?: string | null; /** - * Output only. Ephemeral is true if this data connect service is served from in-memory emulation of Postgres. Data Connect service will transfer the data on a best-effort basis to the Cloud SQL instance once it's provisioned and linked. WARNING: When `ephemeral=true`, mutations to the database are not guaranteed to be durably persisted, even if an OK status code is returned. All or parts of the data may be lost or reverted to earlier versions. + * Output only. Ephemeral is true if this data connect service is served from temporary in-memory emulation of Postgres. While Cloud SQL is being provisioned, the data connect service provides the ephemeral service to help developers get started. Once the Cloud SQL is provisioned, Data Connect service will transfer its data on a best-effort basis to the Cloud SQL instance. WARNING: Ephemeral data sources will expire after 24 hour. The data will be lost if they aren't transferred to the Cloud SQL instance. WARNING: When `ephemeral=true`, mutations to the database are not guaranteed to be durably persisted, even if an OK status code is returned. All or parts of the data may be lost or reverted to earlier versions. */ ephemeral?: boolean | null; /** diff --git a/src/apis/firebasedataconnect/v1beta.ts b/src/apis/firebasedataconnect/v1beta.ts index e80d72ac2bb..6f0762d0f80 100644 --- a/src/apis/firebasedataconnect/v1beta.ts +++ b/src/apis/firebasedataconnect/v1beta.ts @@ -535,7 +535,7 @@ export namespace firebasedataconnect_v1beta { */ database?: string | null; /** - * Output only. Ephemeral is true if this data connect service is served from in-memory emulation of Postgres. Data Connect service will transfer the data on a best-effort basis to the Cloud SQL instance once it's provisioned and linked. WARNING: When `ephemeral=true`, mutations to the database are not guaranteed to be durably persisted, even if an OK status code is returned. All or parts of the data may be lost or reverted to earlier versions. + * Output only. Ephemeral is true if this data connect service is served from temporary in-memory emulation of Postgres. While Cloud SQL is being provisioned, the data connect service provides the ephemeral service to help developers get started. Once the Cloud SQL is provisioned, Data Connect service will transfer its data on a best-effort basis to the Cloud SQL instance. WARNING: Ephemeral data sources will expire after 24 hour. The data will be lost if they aren't transferred to the Cloud SQL instance. WARNING: When `ephemeral=true`, mutations to the database are not guaranteed to be durably persisted, even if an OK status code is returned. All or parts of the data may be lost or reverted to earlier versions. */ ephemeral?: boolean | null; /** From 63a107abdc58a4328fd23e4f14563b889c4bc654 Mon Sep 17 00:00:00 2001 From: Yoshi Automation Date: Fri, 15 Aug 2025 01:37:19 +0000 Subject: [PATCH 07/11] feat(firebaseml)!: update the API BREAKING CHANGE: This release has breaking changes. #### firebaseml:v2beta The following keys were deleted: - schemas.GoogleCloudAiplatformV1beta1GoogleMaps.properties.apiAuth.$ref - schemas.GoogleCloudAiplatformV1beta1GoogleMaps.properties.apiAuth.deprecated - schemas.GoogleCloudAiplatformV1beta1GoogleMaps.properties.apiAuth.description - schemas.GoogleCloudAiplatformV1beta1GoogleMaps.properties.authConfig.$ref - schemas.GoogleCloudAiplatformV1beta1GoogleMaps.properties.authConfig.description The following keys were added: - schemas.GoogleCloudAiplatformV1beta1GroundingChunkRetrievedContext.properties.documentName.description - schemas.GoogleCloudAiplatformV1beta1GroundingChunkRetrievedContext.properties.documentName.readOnly - schemas.GoogleCloudAiplatformV1beta1GroundingChunkRetrievedContext.properties.documentName.type --- discovery/firebaseml-v2beta.json | 19 +++++++------------ src/apis/firebaseml/v2beta.ts | 15 +++++---------- 2 files changed, 12 insertions(+), 22 deletions(-) diff --git a/discovery/firebaseml-v2beta.json b/discovery/firebaseml-v2beta.json index 73145a8391e..e7932b64cc6 100644 --- a/discovery/firebaseml-v2beta.json +++ b/discovery/firebaseml-v2beta.json @@ -206,7 +206,7 @@ } } }, - "revision": "20250731", + "revision": "20250812", "rootUrl": "https://firebaseml.googleapis.com/", "schemas": { "Date": { @@ -1384,17 +1384,7 @@ "GoogleCloudAiplatformV1beta1GoogleMaps": { "description": "Tool to retrieve public maps data for grounding, powered by Google.", "id": "GoogleCloudAiplatformV1beta1GoogleMaps", - "properties": { - "apiAuth": { - "$ref": "GoogleCloudAiplatformV1beta1ApiAuth", - "deprecated": true, - "description": "The authentication config to access the API. Deprecated. Please use auth_config instead." - }, - "authConfig": { - "$ref": "GoogleCloudAiplatformV1beta1AuthConfig", - "description": "The authentication config to access the API. Only API key is supported." - } - }, + "properties": {}, "type": "object" }, "GoogleCloudAiplatformV1beta1GoogleSearchRetrieval": { @@ -1522,6 +1512,11 @@ "description": "Chunk from context retrieved by the retrieval tools.", "id": "GoogleCloudAiplatformV1beta1GroundingChunkRetrievedContext", "properties": { + "documentName": { + "description": "Output only. The full document name for the referenced Vertex AI Search document.", + "readOnly": true, + "type": "string" + }, "ragChunk": { "$ref": "GoogleCloudAiplatformV1beta1RagChunk", "description": "Additional context for the RAG retrieval result. This is only populated when using the RAG retrieval tool." diff --git a/src/apis/firebaseml/v2beta.ts b/src/apis/firebaseml/v2beta.ts index fec81c97586..83a4aa74d48 100644 --- a/src/apis/firebaseml/v2beta.ts +++ b/src/apis/firebaseml/v2beta.ts @@ -892,16 +892,7 @@ export namespace firebaseml_v2beta { /** * Tool to retrieve public maps data for grounding, powered by Google. */ - export interface Schema$GoogleCloudAiplatformV1beta1GoogleMaps { - /** - * The authentication config to access the API. Deprecated. Please use auth_config instead. - */ - apiAuth?: Schema$GoogleCloudAiplatformV1beta1ApiAuth; - /** - * The authentication config to access the API. Only API key is supported. - */ - authConfig?: Schema$GoogleCloudAiplatformV1beta1AuthConfig; - } + export interface Schema$GoogleCloudAiplatformV1beta1GoogleMaps {} /** * Tool to retrieve public web data for grounding, powered by Google. */ @@ -1012,6 +1003,10 @@ export namespace firebaseml_v2beta { * Chunk from context retrieved by the retrieval tools. */ export interface Schema$GoogleCloudAiplatformV1beta1GroundingChunkRetrievedContext { + /** + * Output only. The full document name for the referenced Vertex AI Search document. + */ + documentName?: string | null; /** * Additional context for the RAG retrieval result. This is only populated when using the RAG retrieval tool. */ From 6120524ec2af7cd56150e59ea3ddfdae091b5505 Mon Sep 17 00:00:00 2001 From: Yoshi Automation Date: Fri, 15 Aug 2025 01:37:19 +0000 Subject: [PATCH 08/11] feat(run): update the API #### run:v2 The following keys were added: - schemas.GoogleCloudRunV2VolumeMount.properties.subPath.description - schemas.GoogleCloudRunV2VolumeMount.properties.subPath.type --- discovery/run-v2.json | 6 +++++- src/apis/run/v2.ts | 4 ++++ 2 files changed, 9 insertions(+), 1 deletion(-) diff --git a/discovery/run-v2.json b/discovery/run-v2.json index 4621bfd9765..8b72bacfc35 100644 --- a/discovery/run-v2.json +++ b/discovery/run-v2.json @@ -1922,7 +1922,7 @@ } } }, - "revision": "20250801", + "revision": "20250808", "rootUrl": "https://run.googleapis.com/", "schemas": { "GoogleCloudRunV2BinaryAuthorization": { @@ -4581,6 +4581,10 @@ "name": { "description": "Required. This must match the Name of a Volume.", "type": "string" + }, + "subPath": { + "description": "Optional. Path within the volume from which the container's volume should be mounted. Defaults to \"\" (volume's root).", + "type": "string" } }, "type": "object" diff --git a/src/apis/run/v2.ts b/src/apis/run/v2.ts index 5d6f0dba7ff..73d475a4422 100644 --- a/src/apis/run/v2.ts +++ b/src/apis/run/v2.ts @@ -1922,6 +1922,10 @@ export namespace run_v2 { * Required. This must match the Name of a Volume. */ name?: string | null; + /** + * Optional. Path within the volume from which the container's volume should be mounted. Defaults to "" (volume's root). + */ + subPath?: string | null; } /** * VPC Access settings. For more information on sending traffic to a VPC network, visit https://cloud.google.com/run/docs/configuring/connecting-vpc. From 222a3b895e82805762ad6058dab14063ab022279 Mon Sep 17 00:00:00 2001 From: Yoshi Automation Date: Fri, 15 Aug 2025 01:37:19 +0000 Subject: [PATCH 09/11] fix(saasservicemgmt): update the API #### saasservicemgmt:v1beta1 The following keys were changed: - description - schemas.Saas.properties.locations.description - schemas.Tenant.properties.consumerResource.description - schemas.Tenant.properties.saas.description - schemas.ToMapping.properties.ignoreForLookup.description - schemas.UnitKind.properties.saas.description - schemas.UnitOperation.description --- discovery/saasservicemgmt-v1beta1.json | 16 ++++++++-------- src/apis/saasservicemgmt/v1beta1.ts | 14 +++++++------- 2 files changed, 15 insertions(+), 15 deletions(-) diff --git a/discovery/saasservicemgmt-v1beta1.json b/discovery/saasservicemgmt-v1beta1.json index 3081b652b7c..395641a8a66 100644 --- a/discovery/saasservicemgmt-v1beta1.json +++ b/discovery/saasservicemgmt-v1beta1.json @@ -12,7 +12,7 @@ "baseUrl": "https://saasservicemgmt.googleapis.com/", "batchPath": "batch", "canonicalName": "SaaS Service Management", - "description": "", + "description": "Model, deploy, and operate your SaaS at scale. ", "discoveryVersion": "v1", "documentationLink": "https://cloud.google.com/saas-runtime/docs", "fullyEncodeReservedExpansion": true, @@ -1804,7 +1804,7 @@ } } }, - "revision": "20250528", + "revision": "20250806", "rootUrl": "https://saasservicemgmt.googleapis.com/", "schemas": { "Aggregate": { @@ -2586,7 +2586,7 @@ "type": "object" }, "locations": { - "description": "Optional. Immutable. List of locations that the service is available in. Rollout refers to the list to generate a rollout plan.", + "description": "Optional. List of locations that the service is available in. Rollout refers to the list to generate a rollout plan.", "items": { "$ref": "Location" }, @@ -2634,7 +2634,7 @@ "type": "object" }, "consumerResource": { - "description": "Optional. Immutable. A reference to the consumer resource this SaaS Tenant is representing. The relationship with a consumer resource can be used by EasySaaS for retrieving consumer-defined settings and policies such as maintenance policies (using Unified Maintenance Policy API).", + "description": "Optional. Immutable. A reference to the consumer resource this SaaS Tenant is representing. The relationship with a consumer resource can be used by SaaS Runtime for retrieving consumer-defined settings and policies such as maintenance policies (using Unified Maintenance Policy API).", "type": "string" }, "createTime": { @@ -2660,7 +2660,7 @@ "type": "string" }, "saas": { - "description": "Required. Immutable. A reference to the Saas that defines the product (managed service) that the producer wants to manage with EasySaaS. Part of the EasySaaS common data model.", + "description": "Required. Immutable. A reference to the Saas that defines the product (managed service) that the producer wants to manage with SaaS Runtime. Part of the SaaS Runtime common data model.", "type": "string" }, "uid": { @@ -2686,7 +2686,7 @@ "type": "string" }, "ignoreForLookup": { - "description": "Optional. Tells EasySaaS if this mapping should be used during lookup or not", + "description": "Optional. Tells SaaS Runtime if this mapping should be used during lookup or not", "type": "boolean" }, "inputVariable": { @@ -3014,7 +3014,7 @@ "type": "array" }, "saas": { - "description": "Required. Immutable. A reference to the Saas that defines the product (managed service) that the producer wants to manage with EasySaaS. Part of the EasySaaS common data model. Immutable once set.", + "description": "Required. Immutable. A reference to the Saas that defines the product (managed service) that the producer wants to manage with SaaS Runtime. Part of the SaaS Runtime common data model. Immutable once set.", "type": "string" }, "uid": { @@ -3032,7 +3032,7 @@ "type": "object" }, "UnitOperation": { - "description": "UnitOperation encapsulates the intent of changing/interacting with the service component represented by the specific Unit. Multiple UnitOperations can be created (requested) and scheduled in the future, however only one will be allowed to execute at a time (that can change in the future for non-mutating operations). UnitOperations allow different actors interacting with the same unit to focus only on the change they have requested. This is a base object that contains the common fields in all unit operations.", + "description": "UnitOperation encapsulates the intent of changing/interacting with the service component represented by the specific Unit. Multiple UnitOperations can be created (requested) and scheduled in the future, however only one will be allowed to execute at a time (that can change in the future for non-mutating operations). UnitOperations allow different actors interacting with the same unit to focus only on the change they have requested. This is a base object that contains the common fields in all unit operations. Next: 19", "id": "UnitOperation", "properties": { "annotations": { diff --git a/src/apis/saasservicemgmt/v1beta1.ts b/src/apis/saasservicemgmt/v1beta1.ts index 55a64ad814b..cddb5141afd 100644 --- a/src/apis/saasservicemgmt/v1beta1.ts +++ b/src/apis/saasservicemgmt/v1beta1.ts @@ -102,7 +102,7 @@ export namespace saasservicemgmt_v1beta1 { /** * SaaS Runtime API * - * + * Model, deploy, and operate your SaaS at scale. * * @example * ```js @@ -662,7 +662,7 @@ export namespace saasservicemgmt_v1beta1 { */ labels?: {[key: string]: string} | null; /** - * Optional. Immutable. List of locations that the service is available in. Rollout refers to the list to generate a rollout plan. + * Optional. List of locations that the service is available in. Rollout refers to the list to generate a rollout plan. */ locations?: Schema$Location[]; /** @@ -696,7 +696,7 @@ export namespace saasservicemgmt_v1beta1 { */ annotations?: {[key: string]: string} | null; /** - * Optional. Immutable. A reference to the consumer resource this SaaS Tenant is representing. The relationship with a consumer resource can be used by EasySaaS for retrieving consumer-defined settings and policies such as maintenance policies (using Unified Maintenance Policy API). + * Optional. Immutable. A reference to the consumer resource this SaaS Tenant is representing. The relationship with a consumer resource can be used by SaaS Runtime for retrieving consumer-defined settings and policies such as maintenance policies (using Unified Maintenance Policy API). */ consumerResource?: string | null; /** @@ -716,7 +716,7 @@ export namespace saasservicemgmt_v1beta1 { */ name?: string | null; /** - * Required. Immutable. A reference to the Saas that defines the product (managed service) that the producer wants to manage with EasySaaS. Part of the EasySaaS common data model. + * Required. Immutable. A reference to the Saas that defines the product (managed service) that the producer wants to manage with SaaS Runtime. Part of the SaaS Runtime common data model. */ saas?: string | null; /** @@ -737,7 +737,7 @@ export namespace saasservicemgmt_v1beta1 { */ dependency?: string | null; /** - * Optional. Tells EasySaaS if this mapping should be used during lookup or not + * Optional. Tells SaaS Runtime if this mapping should be used during lookup or not */ ignoreForLookup?: boolean | null; /** @@ -921,7 +921,7 @@ export namespace saasservicemgmt_v1beta1 { */ outputVariableMappings?: Schema$VariableMapping[]; /** - * Required. Immutable. A reference to the Saas that defines the product (managed service) that the producer wants to manage with EasySaaS. Part of the EasySaaS common data model. Immutable once set. + * Required. Immutable. A reference to the Saas that defines the product (managed service) that the producer wants to manage with SaaS Runtime. Part of the SaaS Runtime common data model. Immutable once set. */ saas?: string | null; /** @@ -934,7 +934,7 @@ export namespace saasservicemgmt_v1beta1 { updateTime?: string | null; } /** - * UnitOperation encapsulates the intent of changing/interacting with the service component represented by the specific Unit. Multiple UnitOperations can be created (requested) and scheduled in the future, however only one will be allowed to execute at a time (that can change in the future for non-mutating operations). UnitOperations allow different actors interacting with the same unit to focus only on the change they have requested. This is a base object that contains the common fields in all unit operations. + * UnitOperation encapsulates the intent of changing/interacting with the service component represented by the specific Unit. Multiple UnitOperations can be created (requested) and scheduled in the future, however only one will be allowed to execute at a time (that can change in the future for non-mutating operations). UnitOperations allow different actors interacting with the same unit to focus only on the change they have requested. This is a base object that contains the common fields in all unit operations. Next: 19 */ export interface Schema$UnitOperation { /** From cc134ef0c60c3e8da281b4c130bf1dc3a736c3c5 Mon Sep 17 00:00:00 2001 From: Yoshi Automation Date: Fri, 15 Aug 2025 01:37:19 +0000 Subject: [PATCH 10/11] feat(workloadmanager): update the API #### workloadmanager:v1 The following keys were added: - schemas.AgentStates.properties.hanaMonitoring.$ref - schemas.AgentStates.properties.hanaMonitoring.description - schemas.AgentStates.properties.isFullyEnabled.description - schemas.AgentStates.properties.isFullyEnabled.type - schemas.AgentStates.properties.processMetrics.$ref - schemas.AgentStates.properties.processMetrics.description - schemas.AgentStates.properties.systemDiscovery.$ref - schemas.AgentStates.properties.systemDiscovery.description - schemas.Execution.properties.engine.description - schemas.Execution.properties.engine.enum - schemas.Execution.properties.engine.enumDescriptions - schemas.Execution.properties.engine.type - schemas.IAMPermission.description - schemas.IAMPermission.id - schemas.IAMPermission.properties.granted.description - schemas.IAMPermission.properties.granted.readOnly - schemas.IAMPermission.properties.granted.type - schemas.IAMPermission.properties.name.description - schemas.IAMPermission.properties.name.readOnly - schemas.IAMPermission.properties.name.type - schemas.IAMPermission.type - schemas.ServiceStates.description - schemas.ServiceStates.id - schemas.ServiceStates.properties.iamPermissions.description - schemas.ServiceStates.properties.iamPermissions.items.$ref - schemas.ServiceStates.properties.iamPermissions.readOnly - schemas.ServiceStates.properties.iamPermissions.type - schemas.ServiceStates.properties.state.description - schemas.ServiceStates.properties.state.enum - schemas.ServiceStates.properties.state.enumDescriptions - schemas.ServiceStates.properties.state.readOnly - schemas.ServiceStates.properties.state.type - schemas.ServiceStates.type --- discovery/workloadmanager-v1.json | 85 ++++++++++++++++++++++++++++++- src/apis/workloadmanager/v1.ts | 47 +++++++++++++++++ 2 files changed, 131 insertions(+), 1 deletion(-) diff --git a/discovery/workloadmanager-v1.json b/discovery/workloadmanager-v1.json index cce3b743284..3706ff04c05 100644 --- a/discovery/workloadmanager-v1.json +++ b/discovery/workloadmanager-v1.json @@ -879,7 +879,7 @@ } } }, - "revision": "20250727", + "revision": "20250806", "rootUrl": "https://workloadmanager.googleapis.com/", "schemas": { "AgentCommand": { @@ -908,9 +908,25 @@ "description": "Optional. The available version of the agent in artifact registry.", "type": "string" }, + "hanaMonitoring": { + "$ref": "ServiceStates", + "description": "Optional. HANA monitoring metrics of the agent." + }, "installedVersion": { "description": "Optional. The installed version of the agent on the host.", "type": "string" + }, + "isFullyEnabled": { + "description": "Optional. Whether the agent is fully enabled. If false, the agent is has some issues.", + "type": "boolean" + }, + "processMetrics": { + "$ref": "ServiceStates", + "description": "Optional. The Process metrics of the agent." + }, + "systemDiscovery": { + "$ref": "ServiceStates", + "description": "Optional. The System discovery metrics of the agent." } }, "type": "object" @@ -1444,6 +1460,20 @@ "readOnly": true, "type": "string" }, + "engine": { + "description": "Optional. Engine", + "enum": [ + "ENGINE_UNSPECIFIED", + "ENGINE_SCANNER", + "V2" + ], + "enumDescriptions": [ + "The original CG", + "SlimCG / Scanner", + "Evaluation Engine V2" + ], + "type": "string" + }, "evaluationId": { "description": "Output only. [Output only] Evaluation ID", "readOnly": true, @@ -1631,6 +1661,23 @@ }, "type": "object" }, + "IAMPermission": { + "description": "The IAM permission status.", + "id": "IAMPermission", + "properties": { + "granted": { + "description": "Output only. Whether the permission is granted.", + "readOnly": true, + "type": "boolean" + }, + "name": { + "description": "Output only. The name of the permission.", + "readOnly": true, + "type": "string" + } + }, + "type": "object" + }, "Insight": { "description": "A presentation of host resource usage where the workload runs.", "id": "Insight", @@ -2975,6 +3022,42 @@ }, "type": "object" }, + "ServiceStates": { + "description": "The state of the service.", + "id": "ServiceStates", + "properties": { + "iamPermissions": { + "description": "Optional. Output only. The IAM permissions for the service.", + "items": { + "$ref": "IAMPermission" + }, + "readOnly": true, + "type": "array" + }, + "state": { + "description": "Output only. The overall state of the service.", + "enum": [ + "STATE_UNSPECIFIED", + "CONFIG_FAILURE", + "IAM_FAILURE", + "FUNCTIONAILITY_FAILURE", + "ENABLED", + "DISABLED" + ], + "enumDescriptions": [ + "The state is unspecified.", + "The state means the service has config errors.", + "The state means the service has IAM permission errors.", + "The state means the service has functionality errors.", + "The state means the service has no error.", + "The state means the service disabled." + ], + "readOnly": true, + "type": "string" + } + }, + "type": "object" + }, "ShellCommand": { "description": "* A ShellCommand is invoked via the agent's command line executor", "id": "ShellCommand", diff --git a/src/apis/workloadmanager/v1.ts b/src/apis/workloadmanager/v1.ts index 372a35d9198..cc0b31c0a89 100644 --- a/src/apis/workloadmanager/v1.ts +++ b/src/apis/workloadmanager/v1.ts @@ -145,10 +145,26 @@ export namespace workloadmanager_v1 { * Optional. The available version of the agent in artifact registry. */ availableVersion?: string | null; + /** + * Optional. HANA monitoring metrics of the agent. + */ + hanaMonitoring?: Schema$ServiceStates; /** * Optional. The installed version of the agent on the host. */ installedVersion?: string | null; + /** + * Optional. Whether the agent is fully enabled. If false, the agent is has some issues. + */ + isFullyEnabled?: boolean | null; + /** + * Optional. The Process metrics of the agent. + */ + processMetrics?: Schema$ServiceStates; + /** + * Optional. The System discovery metrics of the agent. + */ + systemDiscovery?: Schema$ServiceStates; } /** * The schema of agent status data. @@ -429,6 +445,10 @@ export namespace workloadmanager_v1 { * Output only. [Output only] End time stamp */ endTime?: string | null; + /** + * Optional. Engine + */ + engine?: string | null; /** * Output only. [Output only] Evaluation ID */ @@ -541,6 +561,19 @@ export namespace workloadmanager_v1 { */ serviceAccounts?: string[] | null; } + /** + * The IAM permission status. + */ + export interface Schema$IAMPermission { + /** + * Output only. Whether the permission is granted. + */ + granted?: boolean | null; + /** + * Output only. The name of the permission. + */ + name?: string | null; + } /** * A presentation of host resource usage where the workload runs. */ @@ -1443,6 +1476,19 @@ export namespace workloadmanager_v1 { */ type?: string | null; } + /** + * The state of the service. + */ + export interface Schema$ServiceStates { + /** + * Optional. Output only. The IAM permissions for the service. + */ + iamPermissions?: Schema$IAMPermission[]; + /** + * Output only. The overall state of the service. + */ + state?: string | null; + } /** * * A ShellCommand is invoked via the agent's command line executor */ @@ -3065,6 +3111,7 @@ export namespace workloadmanager_v1 { * // Example response * // { * // "endTime": "my_endTime", + * // "engine": "my_engine", * // "evaluationId": "my_evaluationId", * // "externalDataSources": [], * // "inventoryTime": "my_inventoryTime", From e75ef0e3aabe60ba002dafb5d338d5ba7b2d7d64 Mon Sep 17 00:00:00 2001 From: Yoshi Automation Date: Fri, 15 Aug 2025 01:37:19 +0000 Subject: [PATCH 11/11] feat: regenerate index files --- discovery/healthcare-v1.json | 4 +- discovery/healthcare-v1beta1.json | 98 +------------------------------ 2 files changed, 4 insertions(+), 98 deletions(-) diff --git a/discovery/healthcare-v1.json b/discovery/healthcare-v1.json index 7433bf6a451..e41b9f481f2 100644 --- a/discovery/healthcare-v1.json +++ b/discovery/healthcare-v1.json @@ -5132,7 +5132,7 @@ } } }, - "revision": "20250710", + "revision": "20250724", "rootUrl": "https://healthcare.googleapis.com/", "schemas": { "AccessDeterminationLogConfig": { @@ -5151,7 +5151,7 @@ "No log level specified. This value is unused.", "No additional consent-related logging is added to audit logs.", "The following information is included: * One of the following [`consentMode`](https://cloud.google.com/healthcare-api/docs/fhir-consent#audit_logs) fields: (`off`|`emptyScope`|`enforced`|`btg`|`bypass`). * The accessor's request headers * The `log_level` of the AccessDeterminationLogConfig * The final consent evaluation (`PERMIT`, `DENY`, or `NO_CONSENT`) * A human-readable summary of the evaluation", - "Includes `MINIMUM` and, for each resource owner, returns: * The resource owner's name * Most specific part of the `X-Consent-Scope` resulting in consensual determination * Timestamp of the applied enforcement leading to the decision * Enforcement version at the time the applicable consents were applied * The Consent resource name * The timestamp of the Consent resource used for enforcement * Policy type (`PATIENT` or `ADMIN`) Note that this mode adds some overhead to CRUD operations." + "Includes `MINIMUM` and, for each resource owner, returns: * The resource owner's name * Most specific part of the `X-Consent-Scope` resulting in consensual determination * Timestamp of the applied enforcement leading to the decision * Enforcement version at the time the applicable consents were applied * The Consent resource name * The timestamp of the Consent resource used for enforcement * Policy type (`PATIENT` or `ADMIN`) Due to the limited space for logging, this mode is the same as `MINIMUM` for methods that return multiple resources (such as FHIR Search)." ], "type": "string" } diff --git a/discovery/healthcare-v1beta1.json b/discovery/healthcare-v1beta1.json index e86687ad68d..a667fec0f7c 100644 --- a/discovery/healthcare-v1beta1.json +++ b/discovery/healthcare-v1beta1.json @@ -469,100 +469,6 @@ } }, "resources": { - "annotationStores": { - "methods": { - "getIamPolicy": { - "description": "Gets the access control policy for a resource. Returns an empty policy if the resource exists and does not have a policy set.", - "flatPath": "v1beta1/projects/{projectsId}/locations/{locationsId}/datasets/{datasetsId}/annotationStores/{annotationStoresId}:getIamPolicy", - "httpMethod": "GET", - "id": "healthcare.projects.locations.datasets.annotationStores.getIamPolicy", - "parameterOrder": [ - "resource" - ], - "parameters": { - "options.requestedPolicyVersion": { - "description": "Optional. The maximum policy version that will be used to format the policy. Valid values are 0, 1, and 3. Requests specifying an invalid value will be rejected. Requests for policies with any conditional role bindings must specify version 3. Policies with no conditional role bindings may specify any valid value or leave the field unset. The policy in the response might use the policy version that you specified, or it might use a lower policy version. For example, if you specify version 3, but the policy has no conditional role bindings, the response uses version 1. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).", - "format": "int32", - "location": "query", - "type": "integer" - }, - "resource": { - "description": "REQUIRED: The resource for which the policy is being requested. See [Resource names](https://cloud.google.com/apis/design/resource_names) for the appropriate value for this field.", - "location": "path", - "pattern": "^projects/[^/]+/locations/[^/]+/datasets/[^/]+/annotationStores/[^/]+$", - "required": true, - "type": "string" - } - }, - "path": "v1beta1/{+resource}:getIamPolicy", - "response": { - "$ref": "Policy" - }, - "scopes": [ - "https://www.googleapis.com/auth/cloud-healthcare", - "https://www.googleapis.com/auth/cloud-platform" - ] - }, - "setIamPolicy": { - "description": "Sets the access control policy on the specified resource. Replaces any existing policy. Can return `NOT_FOUND`, `INVALID_ARGUMENT`, and `PERMISSION_DENIED` errors.", - "flatPath": "v1beta1/projects/{projectsId}/locations/{locationsId}/datasets/{datasetsId}/annotationStores/{annotationStoresId}:setIamPolicy", - "httpMethod": "POST", - "id": "healthcare.projects.locations.datasets.annotationStores.setIamPolicy", - "parameterOrder": [ - "resource" - ], - "parameters": { - "resource": { - "description": "REQUIRED: The resource for which the policy is being specified. See [Resource names](https://cloud.google.com/apis/design/resource_names) for the appropriate value for this field.", - "location": "path", - "pattern": "^projects/[^/]+/locations/[^/]+/datasets/[^/]+/annotationStores/[^/]+$", - "required": true, - "type": "string" - } - }, - "path": "v1beta1/{+resource}:setIamPolicy", - "request": { - "$ref": "SetIamPolicyRequest" - }, - "response": { - "$ref": "Policy" - }, - "scopes": [ - "https://www.googleapis.com/auth/cloud-healthcare", - "https://www.googleapis.com/auth/cloud-platform" - ] - }, - "testIamPermissions": { - "description": "Returns permissions that a caller has on the specified resource. If the resource does not exist, this will return an empty set of permissions, not a `NOT_FOUND` error. Note: This operation is designed to be used for building permission-aware UIs and command-line tools, not for authorization checking. This operation may \"fail open\" without warning.", - "flatPath": "v1beta1/projects/{projectsId}/locations/{locationsId}/datasets/{datasetsId}/annotationStores/{annotationStoresId}:testIamPermissions", - "httpMethod": "POST", - "id": "healthcare.projects.locations.datasets.annotationStores.testIamPermissions", - "parameterOrder": [ - "resource" - ], - "parameters": { - "resource": { - "description": "REQUIRED: The resource for which the policy detail is being requested. See [Resource names](https://cloud.google.com/apis/design/resource_names) for the appropriate value for this field.", - "location": "path", - "pattern": "^projects/[^/]+/locations/[^/]+/datasets/[^/]+/annotationStores/[^/]+$", - "required": true, - "type": "string" - } - }, - "path": "v1beta1/{+resource}:testIamPermissions", - "request": { - "$ref": "TestIamPermissionsRequest" - }, - "response": { - "$ref": "TestIamPermissionsResponse" - }, - "scopes": [ - "https://www.googleapis.com/auth/cloud-healthcare", - "https://www.googleapis.com/auth/cloud-platform" - ] - } - } - }, "consentStores": { "methods": { "checkDataAccess": { @@ -5588,7 +5494,7 @@ } } }, - "revision": "20250710", + "revision": "20250724", "rootUrl": "https://healthcare.googleapis.com/", "schemas": { "AccessDeterminationLogConfig": { @@ -5607,7 +5513,7 @@ "No log level specified. This value is unused.", "No additional consent-related logging is added to audit logs.", "The following information is included: * One of the following [`consentMode`](https://cloud.google.com/healthcare-api/docs/fhir-consent#audit_logs) fields: (`off`|`emptyScope`|`enforced`|`btg`|`bypass`). * The accessor's request headers * The `log_level` of the AccessDeterminationLogConfig * The final consent evaluation (`PERMIT`, `DENY`, or `NO_CONSENT`) * A human-readable summary of the evaluation", - "Includes `MINIMUM` and, for each resource owner, returns: * The resource owner's name * Most specific part of the `X-Consent-Scope` resulting in consensual determination * Timestamp of the applied enforcement leading to the decision * Enforcement version at the time the applicable consents were applied * The Consent resource name * The timestamp of the Consent resource used for enforcement * Policy type (`PATIENT` or `ADMIN`) Note that this mode adds some overhead to CRUD operations." + "Includes `MINIMUM` and, for each resource owner, returns: * The resource owner's name * Most specific part of the `X-Consent-Scope` resulting in consensual determination * Timestamp of the applied enforcement leading to the decision * Enforcement version at the time the applicable consents were applied * The Consent resource name * The timestamp of the Consent resource used for enforcement * Policy type (`PATIENT` or `ADMIN`) Due to the limited space for logging, this mode is the same as `MINIMUM` for methods that return multiple resources (such as FHIR Search)." ], "type": "string" }