docs: update docs/dyn (#1096)
This PR was generated using Autosynth. :rainbow:
Synth log will be available here:
https://source.cloud.google.com/results/invocations/6f0f288a-a1e8-4b2d-a85f-00b1c6150185/targets
- [ ] To automatically regenerate this PR, check this box.
Source-Link: https://github.com/googleapis/synthtool/commit/39b7149da4026765385403632db3c6f63db96b2c
Source-Link: https://github.com/googleapis/synthtool/commit/9a7d9fbb7045c34c9d3d22c1ff766eeae51f04c9
Source-Link: https://github.com/googleapis/synthtool/commit/dc9903a8c30c3662b6098f0e4a97f221d67268b2
Source-Link: https://github.com/googleapis/synthtool/commit/7fcc405a579d5d53a726ff3da1b7c8c08f0f2d58
Source-Link: https://github.com/googleapis/synthtool/commit/d5fc0bcf9ea9789c5b0e3154a9e3b29e5cea6116
Source-Link: https://github.com/googleapis/synthtool/commit/e89175cf074dccc4babb4eca66ae913696e47a71
Source-Link: https://github.com/googleapis/synthtool/commit/7d652819519dfa24da9e14548232e4aaba71a11c
Source-Link: https://github.com/googleapis/synthtool/commit/7db8a6c5ffb12a6e4c2f799c18f00f7f3d60e279
Source-Link: https://github.com/googleapis/synthtool/commit/1f1148d3c7a7a52f0c98077f976bd9b3c948ee2b
Source-Link: https://github.com/googleapis/synthtool/commit/2c8aecedd55b0480fb4e123b6e07fa5b12953862
Source-Link: https://github.com/googleapis/synthtool/commit/3d3e94c4e02370f307a9a200b0c743c3d8d19f29
Source-Link: https://github.com/googleapis/synthtool/commit/c7824ea48ff6d4d42dfae0849aec8a85acd90bd9
Source-Link: https://github.com/googleapis/synthtool/commit/ba9918cd22874245b55734f57470c719b577e591
Source-Link: https://github.com/googleapis/synthtool/commit/b19b401571e77192f8dd38eab5fb2300a0de9324
Source-Link: https://github.com/googleapis/synthtool/commit/6542bd723403513626f61642fc02ddca528409aa
diff --git a/docs/dyn/sqladmin_v1beta4.operations.html b/docs/dyn/sqladmin_v1beta4.operations.html
index 6876ac0..19739ed 100644
--- a/docs/dyn/sqladmin_v1beta4.operations.html
+++ b/docs/dyn/sqladmin_v1beta4.operations.html
@@ -81,7 +81,7 @@
<code><a href="#get">get(project, operation, x__xgafv=None)</a></code></p>
<p class="firstline">Retrieves an instance operation that has been performed on an instance.</p>
<p class="toc_element">
- <code><a href="#list">list(project, instance=None, pageToken=None, maxResults=None, x__xgafv=None)</a></code></p>
+ <code><a href="#list">list(project, instance=None, maxResults=None, pageToken=None, x__xgafv=None)</a></code></p>
<p class="firstline">Lists all instance operations that have been performed on the given Cloud SQL instance in the reverse chronological order of the start time.</p>
<p class="toc_element">
<code><a href="#list_next">list_next(previous_request, previous_response)</a></code></p>
@@ -108,56 +108,7 @@
An object of the form:
{ # An Operation resource. For successful operations that return an Operation resource, only the fields relevant to the operation are populated in the resource.
- "importContext": { # Database instance import context. # The context for import operation, if applicable.
- "database": "A String", # The target database for the import. If *fileType* is *SQL*, this field is required only if the import file does not specify a database, and is overridden by any database specification in the import file. If *fileType* is *CSV*, one database must be specified.
- "csvImportOptions": { # Options for importing data as CSV.
- "columns": [ # The columns to which CSV data is imported. If not specified, all columns of the database table are loaded with CSV data.
- "A String",
- ],
- "table": "A String", # The table to which CSV data is imported.
- },
- "kind": "A String", # This is always *sql#importContext*.
- "importUser": "A String", # The PostgreSQL user for this import operation. PostgreSQL instances only.
- "fileType": "A String", # The file type for the specified uri. *SQL*: The file contains SQL statements. *CSV*: The file contains CSV data.
- "uri": "A String", # Path to the import file in Cloud Storage, in the form *gs: //bucketName/fileName*. Compressed gzip files (.gz) are supported // when *fileType* is *SQL*. The instance must have // write permissions to the bucket and read access to the file.
- "bakImportOptions": { # Import parameters specific to SQL Server .BAK files
- "encryptionOptions": {
- "pvkPassword": "A String", # Password that encrypts the private key
- "certPath": "A String", # Path to the Certificate (.cer) in Cloud Storage, in the form *gs://bucketName/fileName*. The instance must have write permissions to the bucket and read access to the file.
- "pvkPath": "A String", # Path to the Certificate Private Key (.pvk) in Cloud Storage, in the form *gs://bucketName/fileName*. The instance must have write permissions to the bucket and read access to the file.
- },
- },
- },
- "exportContext": { # Database instance export context. # The context for export operation, if applicable.
- "uri": "A String", # The path to the file in Google Cloud Storage where the export will be stored. The URI is in the form *gs: //bucketName/fileName*. If the file already exists, the requests // succeeds, but the operation fails. If *fileType* is // *SQL* and the filename ends with .gz, the contents are // compressed.
- "sqlExportOptions": { # Options for exporting data as SQL statements.
- "mysqlExportOptions": { # Options for exporting from MySQL.
- "masterData": 42, # Option to include SQL statement required to set up replication. If set to *1*, the dump file includes a CHANGE MASTER TO statement with the binary log coordinates. If set to *2*, the CHANGE MASTER TO statement is written as a SQL comment, and has no effect. All other values are ignored.
- },
- "schemaOnly": True or False, # Export only schemas.
- "tables": [ # Tables to export, or that were exported, from the specified database. If you specify tables, specify one and only one database. For PostgreSQL instances, you can specify only one table.
- "A String",
- ],
- },
- "kind": "A String", # This is always *sql#exportContext*.
- "csvExportOptions": { # Options for exporting data as CSV.
- "selectQuery": "A String", # The select query used to extract the data.
- },
- "fileType": "A String", # The file type for the specified uri. *SQL*: The file contains SQL statements. *CSV*: The file contains CSV data.
- "offload": True or False, # Option for export offload.
- "databases": [ # Databases to be exported. *MySQL instances:* If *fileType* is *SQL* and no database is specified, all databases are exported, except for the *mysql* system database. If *fileType* is *CSV*, you can specify one database, either by using this property or by using the *csvExportOptions.selectQuery* property, which takes precedence over this property. *PostgreSQL instances:* You must specify one database to be exported. If *fileType* is *CSV*, this database must match the one specified in the *csvExportOptions.selectQuery* property.
- "A String",
- ],
- },
- "endTime": "A String", # The time this operation finished in UTC timezone in RFC 3339 format, for example *2012-11-15T16:19:00.094Z*.
- "user": "A String", # The email address of the user who initiated this operation.
- "selfLink": "A String", # The URI of this resource.
"status": "A String", # The status of an operation. Valid values are: *PENDING* *RUNNING* *DONE* *SQL_OPERATION_STATUS_UNSPECIFIED*
- "insertTime": "A String", # The time this operation was enqueued in UTC timezone in RFC 3339 format, for example *2012-11-15T16:19:00.094Z*.
- "targetId": "A String", # Name of the database instance related to this operation.
- "targetProject": "A String", # The project ID of the target instance related to this operation.
- "operationType": "A String", # The type of the operation. Valid values are: *CREATE* *DELETE* *UPDATE* *RESTART* *IMPORT* *EXPORT* *BACKUP_VOLUME* *RESTORE_VOLUME* *CREATE_USER* *DELETE_USER* *CREATE_DATABASE* *DELETE_DATABASE*
- "kind": "A String", # This is always *sql#operation*.
"error": { # Database instance operation errors list wrapper. # If errors occurred during processing of this operation, this field will be populated.
"kind": "A String", # This is always *sql#operationErrors*.
"errors": [ # The list of errors encountered while processing this operation.
@@ -168,21 +119,70 @@
},
],
},
+ "endTime": "A String", # The time this operation finished in UTC timezone in RFC 3339 format, for example *2012-11-15T16:19:00.094Z*.
"targetLink": "A String",
"name": "A String", # An identifier that uniquely identifies the operation. You can use this identifier to retrieve the Operations resource that has information about the operation.
+ "exportContext": { # Database instance export context. # The context for export operation, if applicable.
+ "offload": True or False, # Option for export offload.
+ "csvExportOptions": { # Options for exporting data as CSV.
+ "selectQuery": "A String", # The select query used to extract the data.
+ },
+ "databases": [ # Databases to be exported. *MySQL instances:* If *fileType* is *SQL* and no database is specified, all databases are exported, except for the *mysql* system database. If *fileType* is *CSV*, you can specify one database, either by using this property or by using the *csvExportOptions.selectQuery* property, which takes precedence over this property. *PostgreSQL instances:* You must specify one database to be exported. If *fileType* is *CSV*, this database must match the one specified in the *csvExportOptions.selectQuery* property.
+ "A String",
+ ],
+ "kind": "A String", # This is always *sql#exportContext*.
+ "sqlExportOptions": { # Options for exporting data as SQL statements.
+ "schemaOnly": True or False, # Export only schemas.
+ "tables": [ # Tables to export, or that were exported, from the specified database. If you specify tables, specify one and only one database. For PostgreSQL instances, you can specify only one table.
+ "A String",
+ ],
+ "mysqlExportOptions": { # Options for exporting from MySQL.
+ "masterData": 42, # Option to include SQL statement required to set up replication. If set to *1*, the dump file includes a CHANGE MASTER TO statement with the binary log coordinates. If set to *2*, the CHANGE MASTER TO statement is written as a SQL comment, and has no effect. All other values are ignored.
+ },
+ },
+ "fileType": "A String", # The file type for the specified uri. *SQL*: The file contains SQL statements. *CSV*: The file contains CSV data.
+ "uri": "A String", # The path to the file in Google Cloud Storage where the export will be stored. The URI is in the form *gs: //bucketName/fileName*. If the file already exists, the requests // succeeds, but the operation fails. If *fileType* is // *SQL* and the filename ends with .gz, the contents are // compressed.
+ },
+ "operationType": "A String", # The type of the operation. Valid values are: *CREATE* *DELETE* *UPDATE* *RESTART* *IMPORT* *EXPORT* *BACKUP_VOLUME* *RESTORE_VOLUME* *CREATE_USER* *DELETE_USER* *CREATE_DATABASE* *DELETE_DATABASE*
"startTime": "A String", # The time this operation actually started in UTC timezone in RFC 3339 format, for example *2012-11-15T16:19:00.094Z*.
+ "selfLink": "A String", # The URI of this resource.
+ "kind": "A String", # This is always *sql#operation*.
+ "insertTime": "A String", # The time this operation was enqueued in UTC timezone in RFC 3339 format, for example *2012-11-15T16:19:00.094Z*.
+ "targetProject": "A String", # The project ID of the target instance related to this operation.
+ "importContext": { # Database instance import context. # The context for import operation, if applicable.
+ "uri": "A String", # Path to the import file in Cloud Storage, in the form *gs: //bucketName/fileName*. Compressed gzip files (.gz) are supported // when *fileType* is *SQL*. The instance must have // write permissions to the bucket and read access to the file.
+ "fileType": "A String", # The file type for the specified uri. *SQL*: The file contains SQL statements. *CSV*: The file contains CSV data.
+ "csvImportOptions": { # Options for importing data as CSV.
+ "columns": [ # The columns to which CSV data is imported. If not specified, all columns of the database table are loaded with CSV data.
+ "A String",
+ ],
+ "table": "A String", # The table to which CSV data is imported.
+ },
+ "database": "A String", # The target database for the import. If *fileType* is *SQL*, this field is required only if the import file does not specify a database, and is overridden by any database specification in the import file. If *fileType* is *CSV*, one database must be specified.
+ "importUser": "A String", # The PostgreSQL user for this import operation. PostgreSQL instances only.
+ "kind": "A String", # This is always *sql#importContext*.
+ "bakImportOptions": { # Import parameters specific to SQL Server .BAK files
+ "encryptionOptions": {
+ "certPath": "A String", # Path to the Certificate (.cer) in Cloud Storage, in the form *gs://bucketName/fileName*. The instance must have write permissions to the bucket and read access to the file.
+ "pvkPassword": "A String", # Password that encrypts the private key
+ "pvkPath": "A String", # Path to the Certificate Private Key (.pvk) in Cloud Storage, in the form *gs://bucketName/fileName*. The instance must have write permissions to the bucket and read access to the file.
+ },
+ },
+ },
+ "user": "A String", # The email address of the user who initiated this operation.
+ "targetId": "A String", # Name of the database instance related to this operation.
}</pre>
</div>
<div class="method">
- <code class="details" id="list">list(project, instance=None, pageToken=None, maxResults=None, x__xgafv=None)</code>
+ <code class="details" id="list">list(project, instance=None, maxResults=None, pageToken=None, x__xgafv=None)</code>
<pre>Lists all instance operations that have been performed on the given Cloud SQL instance in the reverse chronological order of the start time.
Args:
project: string, Project ID of the project that contains the instance. (required)
instance: string, Cloud SQL instance ID. This does not include the project ID.
- pageToken: string, A previously-returned page token representing part of the larger set of results to view.
maxResults: integer, Maximum number of operations per response.
+ pageToken: string, A previously-returned page token representing part of the larger set of results to view.
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -192,59 +192,11 @@
An object of the form:
{ # Database instance list operations response.
+ "kind": "A String", # This is always *sql#operationsList*.
"nextPageToken": "A String", # The continuation token, used to page through large result sets. Provide this value in a subsequent request to return the next page of results.
"items": [ # List of operation resources.
{ # An Operation resource. For successful operations that return an Operation resource, only the fields relevant to the operation are populated in the resource.
- "importContext": { # Database instance import context. # The context for import operation, if applicable.
- "database": "A String", # The target database for the import. If *fileType* is *SQL*, this field is required only if the import file does not specify a database, and is overridden by any database specification in the import file. If *fileType* is *CSV*, one database must be specified.
- "csvImportOptions": { # Options for importing data as CSV.
- "columns": [ # The columns to which CSV data is imported. If not specified, all columns of the database table are loaded with CSV data.
- "A String",
- ],
- "table": "A String", # The table to which CSV data is imported.
- },
- "kind": "A String", # This is always *sql#importContext*.
- "importUser": "A String", # The PostgreSQL user for this import operation. PostgreSQL instances only.
- "fileType": "A String", # The file type for the specified uri. *SQL*: The file contains SQL statements. *CSV*: The file contains CSV data.
- "uri": "A String", # Path to the import file in Cloud Storage, in the form *gs: //bucketName/fileName*. Compressed gzip files (.gz) are supported // when *fileType* is *SQL*. The instance must have // write permissions to the bucket and read access to the file.
- "bakImportOptions": { # Import parameters specific to SQL Server .BAK files
- "encryptionOptions": {
- "pvkPassword": "A String", # Password that encrypts the private key
- "certPath": "A String", # Path to the Certificate (.cer) in Cloud Storage, in the form *gs://bucketName/fileName*. The instance must have write permissions to the bucket and read access to the file.
- "pvkPath": "A String", # Path to the Certificate Private Key (.pvk) in Cloud Storage, in the form *gs://bucketName/fileName*. The instance must have write permissions to the bucket and read access to the file.
- },
- },
- },
- "exportContext": { # Database instance export context. # The context for export operation, if applicable.
- "uri": "A String", # The path to the file in Google Cloud Storage where the export will be stored. The URI is in the form *gs: //bucketName/fileName*. If the file already exists, the requests // succeeds, but the operation fails. If *fileType* is // *SQL* and the filename ends with .gz, the contents are // compressed.
- "sqlExportOptions": { # Options for exporting data as SQL statements.
- "mysqlExportOptions": { # Options for exporting from MySQL.
- "masterData": 42, # Option to include SQL statement required to set up replication. If set to *1*, the dump file includes a CHANGE MASTER TO statement with the binary log coordinates. If set to *2*, the CHANGE MASTER TO statement is written as a SQL comment, and has no effect. All other values are ignored.
- },
- "schemaOnly": True or False, # Export only schemas.
- "tables": [ # Tables to export, or that were exported, from the specified database. If you specify tables, specify one and only one database. For PostgreSQL instances, you can specify only one table.
- "A String",
- ],
- },
- "kind": "A String", # This is always *sql#exportContext*.
- "csvExportOptions": { # Options for exporting data as CSV.
- "selectQuery": "A String", # The select query used to extract the data.
- },
- "fileType": "A String", # The file type for the specified uri. *SQL*: The file contains SQL statements. *CSV*: The file contains CSV data.
- "offload": True or False, # Option for export offload.
- "databases": [ # Databases to be exported. *MySQL instances:* If *fileType* is *SQL* and no database is specified, all databases are exported, except for the *mysql* system database. If *fileType* is *CSV*, you can specify one database, either by using this property or by using the *csvExportOptions.selectQuery* property, which takes precedence over this property. *PostgreSQL instances:* You must specify one database to be exported. If *fileType* is *CSV*, this database must match the one specified in the *csvExportOptions.selectQuery* property.
- "A String",
- ],
- },
- "endTime": "A String", # The time this operation finished in UTC timezone in RFC 3339 format, for example *2012-11-15T16:19:00.094Z*.
- "user": "A String", # The email address of the user who initiated this operation.
- "selfLink": "A String", # The URI of this resource.
"status": "A String", # The status of an operation. Valid values are: *PENDING* *RUNNING* *DONE* *SQL_OPERATION_STATUS_UNSPECIFIED*
- "insertTime": "A String", # The time this operation was enqueued in UTC timezone in RFC 3339 format, for example *2012-11-15T16:19:00.094Z*.
- "targetId": "A String", # Name of the database instance related to this operation.
- "targetProject": "A String", # The project ID of the target instance related to this operation.
- "operationType": "A String", # The type of the operation. Valid values are: *CREATE* *DELETE* *UPDATE* *RESTART* *IMPORT* *EXPORT* *BACKUP_VOLUME* *RESTORE_VOLUME* *CREATE_USER* *DELETE_USER* *CREATE_DATABASE* *DELETE_DATABASE*
- "kind": "A String", # This is always *sql#operation*.
"error": { # Database instance operation errors list wrapper. # If errors occurred during processing of this operation, this field will be populated.
"kind": "A String", # This is always *sql#operationErrors*.
"errors": [ # The list of errors encountered while processing this operation.
@@ -255,12 +207,60 @@
},
],
},
+ "endTime": "A String", # The time this operation finished in UTC timezone in RFC 3339 format, for example *2012-11-15T16:19:00.094Z*.
"targetLink": "A String",
"name": "A String", # An identifier that uniquely identifies the operation. You can use this identifier to retrieve the Operations resource that has information about the operation.
+ "exportContext": { # Database instance export context. # The context for export operation, if applicable.
+ "offload": True or False, # Option for export offload.
+ "csvExportOptions": { # Options for exporting data as CSV.
+ "selectQuery": "A String", # The select query used to extract the data.
+ },
+ "databases": [ # Databases to be exported. *MySQL instances:* If *fileType* is *SQL* and no database is specified, all databases are exported, except for the *mysql* system database. If *fileType* is *CSV*, you can specify one database, either by using this property or by using the *csvExportOptions.selectQuery* property, which takes precedence over this property. *PostgreSQL instances:* You must specify one database to be exported. If *fileType* is *CSV*, this database must match the one specified in the *csvExportOptions.selectQuery* property.
+ "A String",
+ ],
+ "kind": "A String", # This is always *sql#exportContext*.
+ "sqlExportOptions": { # Options for exporting data as SQL statements.
+ "schemaOnly": True or False, # Export only schemas.
+ "tables": [ # Tables to export, or that were exported, from the specified database. If you specify tables, specify one and only one database. For PostgreSQL instances, you can specify only one table.
+ "A String",
+ ],
+ "mysqlExportOptions": { # Options for exporting from MySQL.
+ "masterData": 42, # Option to include SQL statement required to set up replication. If set to *1*, the dump file includes a CHANGE MASTER TO statement with the binary log coordinates. If set to *2*, the CHANGE MASTER TO statement is written as a SQL comment, and has no effect. All other values are ignored.
+ },
+ },
+ "fileType": "A String", # The file type for the specified uri. *SQL*: The file contains SQL statements. *CSV*: The file contains CSV data.
+ "uri": "A String", # The path to the file in Google Cloud Storage where the export will be stored. The URI is in the form *gs: //bucketName/fileName*. If the file already exists, the requests // succeeds, but the operation fails. If *fileType* is // *SQL* and the filename ends with .gz, the contents are // compressed.
+ },
+ "operationType": "A String", # The type of the operation. Valid values are: *CREATE* *DELETE* *UPDATE* *RESTART* *IMPORT* *EXPORT* *BACKUP_VOLUME* *RESTORE_VOLUME* *CREATE_USER* *DELETE_USER* *CREATE_DATABASE* *DELETE_DATABASE*
"startTime": "A String", # The time this operation actually started in UTC timezone in RFC 3339 format, for example *2012-11-15T16:19:00.094Z*.
+ "selfLink": "A String", # The URI of this resource.
+ "kind": "A String", # This is always *sql#operation*.
+ "insertTime": "A String", # The time this operation was enqueued in UTC timezone in RFC 3339 format, for example *2012-11-15T16:19:00.094Z*.
+ "targetProject": "A String", # The project ID of the target instance related to this operation.
+ "importContext": { # Database instance import context. # The context for import operation, if applicable.
+ "uri": "A String", # Path to the import file in Cloud Storage, in the form *gs: //bucketName/fileName*. Compressed gzip files (.gz) are supported // when *fileType* is *SQL*. The instance must have // write permissions to the bucket and read access to the file.
+ "fileType": "A String", # The file type for the specified uri. *SQL*: The file contains SQL statements. *CSV*: The file contains CSV data.
+ "csvImportOptions": { # Options for importing data as CSV.
+ "columns": [ # The columns to which CSV data is imported. If not specified, all columns of the database table are loaded with CSV data.
+ "A String",
+ ],
+ "table": "A String", # The table to which CSV data is imported.
+ },
+ "database": "A String", # The target database for the import. If *fileType* is *SQL*, this field is required only if the import file does not specify a database, and is overridden by any database specification in the import file. If *fileType* is *CSV*, one database must be specified.
+ "importUser": "A String", # The PostgreSQL user for this import operation. PostgreSQL instances only.
+ "kind": "A String", # This is always *sql#importContext*.
+ "bakImportOptions": { # Import parameters specific to SQL Server .BAK files
+ "encryptionOptions": {
+ "certPath": "A String", # Path to the Certificate (.cer) in Cloud Storage, in the form *gs://bucketName/fileName*. The instance must have write permissions to the bucket and read access to the file.
+ "pvkPassword": "A String", # Password that encrypts the private key
+ "pvkPath": "A String", # Path to the Certificate Private Key (.pvk) in Cloud Storage, in the form *gs://bucketName/fileName*. The instance must have write permissions to the bucket and read access to the file.
+ },
+ },
+ },
+ "user": "A String", # The email address of the user who initiated this operation.
+ "targetId": "A String", # Name of the database instance related to this operation.
},
],
- "kind": "A String", # This is always *sql#operationsList*.
}</pre>
</div>