docs: update docs/dyn (#1096)
This PR was generated using Autosynth. :rainbow:
Synth log will be available here:
https://source.cloud.google.com/results/invocations/6f0f288a-a1e8-4b2d-a85f-00b1c6150185/targets
- [ ] To automatically regenerate this PR, check this box.
Source-Link: https://github.com/googleapis/synthtool/commit/39b7149da4026765385403632db3c6f63db96b2c
Source-Link: https://github.com/googleapis/synthtool/commit/9a7d9fbb7045c34c9d3d22c1ff766eeae51f04c9
Source-Link: https://github.com/googleapis/synthtool/commit/dc9903a8c30c3662b6098f0e4a97f221d67268b2
Source-Link: https://github.com/googleapis/synthtool/commit/7fcc405a579d5d53a726ff3da1b7c8c08f0f2d58
Source-Link: https://github.com/googleapis/synthtool/commit/d5fc0bcf9ea9789c5b0e3154a9e3b29e5cea6116
Source-Link: https://github.com/googleapis/synthtool/commit/e89175cf074dccc4babb4eca66ae913696e47a71
Source-Link: https://github.com/googleapis/synthtool/commit/7d652819519dfa24da9e14548232e4aaba71a11c
Source-Link: https://github.com/googleapis/synthtool/commit/7db8a6c5ffb12a6e4c2f799c18f00f7f3d60e279
Source-Link: https://github.com/googleapis/synthtool/commit/1f1148d3c7a7a52f0c98077f976bd9b3c948ee2b
Source-Link: https://github.com/googleapis/synthtool/commit/2c8aecedd55b0480fb4e123b6e07fa5b12953862
Source-Link: https://github.com/googleapis/synthtool/commit/3d3e94c4e02370f307a9a200b0c743c3d8d19f29
Source-Link: https://github.com/googleapis/synthtool/commit/c7824ea48ff6d4d42dfae0849aec8a85acd90bd9
Source-Link: https://github.com/googleapis/synthtool/commit/ba9918cd22874245b55734f57470c719b577e591
Source-Link: https://github.com/googleapis/synthtool/commit/b19b401571e77192f8dd38eab5fb2300a0de9324
Source-Link: https://github.com/googleapis/synthtool/commit/6542bd723403513626f61642fc02ddca528409aa
diff --git a/docs/dyn/sqladmin_v1beta4.users.html b/docs/dyn/sqladmin_v1beta4.users.html
index a5ce15a..d6e8b29 100644
--- a/docs/dyn/sqladmin_v1beta4.users.html
+++ b/docs/dyn/sqladmin_v1beta4.users.html
@@ -87,7 +87,7 @@
<code><a href="#list">list(project, instance, x__xgafv=None)</a></code></p>
<p class="firstline">Lists users in the specified Cloud SQL instance.</p>
<p class="toc_element">
- <code><a href="#update">update(project, instance, body=None, name=None, host=None, x__xgafv=None)</a></code></p>
+ <code><a href="#update">update(project, instance, body=None, host=None, name=None, x__xgafv=None)</a></code></p>
<p class="firstline">Updates an existing user in a Cloud SQL instance.</p>
<h3>Method Details</h3>
<div class="method">
@@ -113,56 +113,7 @@
An object of the form:
{ # An Operation resource. For successful operations that return an Operation resource, only the fields relevant to the operation are populated in the resource.
- "importContext": { # Database instance import context. # The context for import operation, if applicable.
- "database": "A String", # The target database for the import. If *fileType* is *SQL*, this field is required only if the import file does not specify a database, and is overridden by any database specification in the import file. If *fileType* is *CSV*, one database must be specified.
- "csvImportOptions": { # Options for importing data as CSV.
- "columns": [ # The columns to which CSV data is imported. If not specified, all columns of the database table are loaded with CSV data.
- "A String",
- ],
- "table": "A String", # The table to which CSV data is imported.
- },
- "kind": "A String", # This is always *sql#importContext*.
- "importUser": "A String", # The PostgreSQL user for this import operation. PostgreSQL instances only.
- "fileType": "A String", # The file type for the specified uri. *SQL*: The file contains SQL statements. *CSV*: The file contains CSV data.
- "uri": "A String", # Path to the import file in Cloud Storage, in the form *gs: //bucketName/fileName*. Compressed gzip files (.gz) are supported // when *fileType* is *SQL*. The instance must have // write permissions to the bucket and read access to the file.
- "bakImportOptions": { # Import parameters specific to SQL Server .BAK files
- "encryptionOptions": {
- "pvkPassword": "A String", # Password that encrypts the private key
- "certPath": "A String", # Path to the Certificate (.cer) in Cloud Storage, in the form *gs://bucketName/fileName*. The instance must have write permissions to the bucket and read access to the file.
- "pvkPath": "A String", # Path to the Certificate Private Key (.pvk) in Cloud Storage, in the form *gs://bucketName/fileName*. The instance must have write permissions to the bucket and read access to the file.
- },
- },
- },
- "exportContext": { # Database instance export context. # The context for export operation, if applicable.
- "uri": "A String", # The path to the file in Google Cloud Storage where the export will be stored. The URI is in the form *gs: //bucketName/fileName*. If the file already exists, the requests // succeeds, but the operation fails. If *fileType* is // *SQL* and the filename ends with .gz, the contents are // compressed.
- "sqlExportOptions": { # Options for exporting data as SQL statements.
- "mysqlExportOptions": { # Options for exporting from MySQL.
- "masterData": 42, # Option to include SQL statement required to set up replication. If set to *1*, the dump file includes a CHANGE MASTER TO statement with the binary log coordinates. If set to *2*, the CHANGE MASTER TO statement is written as a SQL comment, and has no effect. All other values are ignored.
- },
- "schemaOnly": True or False, # Export only schemas.
- "tables": [ # Tables to export, or that were exported, from the specified database. If you specify tables, specify one and only one database. For PostgreSQL instances, you can specify only one table.
- "A String",
- ],
- },
- "kind": "A String", # This is always *sql#exportContext*.
- "csvExportOptions": { # Options for exporting data as CSV.
- "selectQuery": "A String", # The select query used to extract the data.
- },
- "fileType": "A String", # The file type for the specified uri. *SQL*: The file contains SQL statements. *CSV*: The file contains CSV data.
- "offload": True or False, # Option for export offload.
- "databases": [ # Databases to be exported. *MySQL instances:* If *fileType* is *SQL* and no database is specified, all databases are exported, except for the *mysql* system database. If *fileType* is *CSV*, you can specify one database, either by using this property or by using the *csvExportOptions.selectQuery* property, which takes precedence over this property. *PostgreSQL instances:* You must specify one database to be exported. If *fileType* is *CSV*, this database must match the one specified in the *csvExportOptions.selectQuery* property.
- "A String",
- ],
- },
- "endTime": "A String", # The time this operation finished in UTC timezone in RFC 3339 format, for example *2012-11-15T16:19:00.094Z*.
- "user": "A String", # The email address of the user who initiated this operation.
- "selfLink": "A String", # The URI of this resource.
"status": "A String", # The status of an operation. Valid values are: *PENDING* *RUNNING* *DONE* *SQL_OPERATION_STATUS_UNSPECIFIED*
- "insertTime": "A String", # The time this operation was enqueued in UTC timezone in RFC 3339 format, for example *2012-11-15T16:19:00.094Z*.
- "targetId": "A String", # Name of the database instance related to this operation.
- "targetProject": "A String", # The project ID of the target instance related to this operation.
- "operationType": "A String", # The type of the operation. Valid values are: *CREATE* *DELETE* *UPDATE* *RESTART* *IMPORT* *EXPORT* *BACKUP_VOLUME* *RESTORE_VOLUME* *CREATE_USER* *DELETE_USER* *CREATE_DATABASE* *DELETE_DATABASE*
- "kind": "A String", # This is always *sql#operation*.
"error": { # Database instance operation errors list wrapper. # If errors occurred during processing of this operation, this field will be populated.
"kind": "A String", # This is always *sql#operationErrors*.
"errors": [ # The list of errors encountered while processing this operation.
@@ -173,9 +124,58 @@
},
],
},
+ "endTime": "A String", # The time this operation finished in UTC timezone in RFC 3339 format, for example *2012-11-15T16:19:00.094Z*.
"targetLink": "A String",
"name": "A String", # An identifier that uniquely identifies the operation. You can use this identifier to retrieve the Operations resource that has information about the operation.
+ "exportContext": { # Database instance export context. # The context for export operation, if applicable.
+ "offload": True or False, # Option for export offload.
+ "csvExportOptions": { # Options for exporting data as CSV.
+ "selectQuery": "A String", # The select query used to extract the data.
+ },
+ "databases": [ # Databases to be exported. *MySQL instances:* If *fileType* is *SQL* and no database is specified, all databases are exported, except for the *mysql* system database. If *fileType* is *CSV*, you can specify one database, either by using this property or by using the *csvExportOptions.selectQuery* property, which takes precedence over this property. *PostgreSQL instances:* You must specify one database to be exported. If *fileType* is *CSV*, this database must match the one specified in the *csvExportOptions.selectQuery* property.
+ "A String",
+ ],
+ "kind": "A String", # This is always *sql#exportContext*.
+ "sqlExportOptions": { # Options for exporting data as SQL statements.
+ "schemaOnly": True or False, # Export only schemas.
+ "tables": [ # Tables to export, or that were exported, from the specified database. If you specify tables, specify one and only one database. For PostgreSQL instances, you can specify only one table.
+ "A String",
+ ],
+ "mysqlExportOptions": { # Options for exporting from MySQL.
+ "masterData": 42, # Option to include SQL statement required to set up replication. If set to *1*, the dump file includes a CHANGE MASTER TO statement with the binary log coordinates. If set to *2*, the CHANGE MASTER TO statement is written as a SQL comment, and has no effect. All other values are ignored.
+ },
+ },
+ "fileType": "A String", # The file type for the specified uri. *SQL*: The file contains SQL statements. *CSV*: The file contains CSV data.
+ "uri": "A String", # The path to the file in Google Cloud Storage where the export will be stored. The URI is in the form *gs: //bucketName/fileName*. If the file already exists, the requests // succeeds, but the operation fails. If *fileType* is // *SQL* and the filename ends with .gz, the contents are // compressed.
+ },
+ "operationType": "A String", # The type of the operation. Valid values are: *CREATE* *DELETE* *UPDATE* *RESTART* *IMPORT* *EXPORT* *BACKUP_VOLUME* *RESTORE_VOLUME* *CREATE_USER* *DELETE_USER* *CREATE_DATABASE* *DELETE_DATABASE*
"startTime": "A String", # The time this operation actually started in UTC timezone in RFC 3339 format, for example *2012-11-15T16:19:00.094Z*.
+ "selfLink": "A String", # The URI of this resource.
+ "kind": "A String", # This is always *sql#operation*.
+ "insertTime": "A String", # The time this operation was enqueued in UTC timezone in RFC 3339 format, for example *2012-11-15T16:19:00.094Z*.
+ "targetProject": "A String", # The project ID of the target instance related to this operation.
+ "importContext": { # Database instance import context. # The context for import operation, if applicable.
+ "uri": "A String", # Path to the import file in Cloud Storage, in the form *gs: //bucketName/fileName*. Compressed gzip files (.gz) are supported // when *fileType* is *SQL*. The instance must have // write permissions to the bucket and read access to the file.
+ "fileType": "A String", # The file type for the specified uri. *SQL*: The file contains SQL statements. *CSV*: The file contains CSV data.
+ "csvImportOptions": { # Options for importing data as CSV.
+ "columns": [ # The columns to which CSV data is imported. If not specified, all columns of the database table are loaded with CSV data.
+ "A String",
+ ],
+ "table": "A String", # The table to which CSV data is imported.
+ },
+ "database": "A String", # The target database for the import. If *fileType* is *SQL*, this field is required only if the import file does not specify a database, and is overridden by any database specification in the import file. If *fileType* is *CSV*, one database must be specified.
+ "importUser": "A String", # The PostgreSQL user for this import operation. PostgreSQL instances only.
+ "kind": "A String", # This is always *sql#importContext*.
+ "bakImportOptions": { # Import parameters specific to SQL Server .BAK files
+ "encryptionOptions": {
+ "certPath": "A String", # Path to the Certificate (.cer) in Cloud Storage, in the form *gs://bucketName/fileName*. The instance must have write permissions to the bucket and read access to the file.
+ "pvkPassword": "A String", # Password that encrypts the private key
+ "pvkPath": "A String", # Path to the Certificate Private Key (.pvk) in Cloud Storage, in the form *gs://bucketName/fileName*. The instance must have write permissions to the bucket and read access to the file.
+ },
+ },
+ },
+ "user": "A String", # The email address of the user who initiated this operation.
+ "targetId": "A String", # Name of the database instance related to this operation.
}</pre>
</div>
@@ -190,21 +190,21 @@
The object takes the form of:
{ # A Cloud SQL user resource.
- "etag": "A String", # This field is deprecated and will be removed from a future version of the API.
- "host": "A String", # The host name from which the user can connect. For *insert* operations, host defaults to an empty string. For *update* operations, host is specified as part of the request URL. The host name cannot be updated after insertion.
- "kind": "A String", # This is always *sql#user*.
- "project": "A String", # The project ID of the project containing the Cloud SQL database. The Google apps domain is prefixed if applicable. Can be omitted for *update* since it is already specified on the URL.
- "type": "A String", # The user type. It determines the method to authenticate the user during login. The default is the database's built-in user type.
- "name": "A String", # The name of the user in the Cloud SQL instance. Can be omitted for *update* since it is already specified in the URL.
- "instance": "A String", # The name of the Cloud SQL instance. This does not include the project ID. Can be omitted for *update* since it is already specified on the URL.
- "password": "A String", # The password for the user.
- "sqlserverUserDetails": { # Represents a Sql Server user on the Cloud SQL instance.
- "disabled": True or False, # If the user has been disabled
- "serverRoles": [ # The server roles for this user
- "A String",
- ],
- },
-}
+ "password": "A String", # The password for the user.
+ "sqlserverUserDetails": { # Represents a Sql Server user on the Cloud SQL instance.
+ "serverRoles": [ # The server roles for this user
+ "A String",
+ ],
+ "disabled": True or False, # If the user has been disabled
+ },
+ "kind": "A String", # This is always *sql#user*.
+ "name": "A String", # The name of the user in the Cloud SQL instance. Can be omitted for *update* since it is already specified in the URL.
+ "project": "A String", # The project ID of the project containing the Cloud SQL database. The Google apps domain is prefixed if applicable. Can be omitted for *update* since it is already specified on the URL.
+ "instance": "A String", # The name of the Cloud SQL instance. This does not include the project ID. Can be omitted for *update* since it is already specified on the URL.
+ "type": "A String", # The user type. It determines the method to authenticate the user during login. The default is the database's built-in user type.
+ "etag": "A String", # This field is deprecated and will be removed from a future version of the API.
+ "host": "A String", # The host name from which the user can connect. For *insert* operations, host defaults to an empty string. For *update* operations, host is specified as part of the request URL. The host name cannot be updated after insertion.
+ }
x__xgafv: string, V1 error format.
Allowed values
@@ -215,56 +215,7 @@
An object of the form:
{ # An Operation resource. For successful operations that return an Operation resource, only the fields relevant to the operation are populated in the resource.
- "importContext": { # Database instance import context. # The context for import operation, if applicable.
- "database": "A String", # The target database for the import. If *fileType* is *SQL*, this field is required only if the import file does not specify a database, and is overridden by any database specification in the import file. If *fileType* is *CSV*, one database must be specified.
- "csvImportOptions": { # Options for importing data as CSV.
- "columns": [ # The columns to which CSV data is imported. If not specified, all columns of the database table are loaded with CSV data.
- "A String",
- ],
- "table": "A String", # The table to which CSV data is imported.
- },
- "kind": "A String", # This is always *sql#importContext*.
- "importUser": "A String", # The PostgreSQL user for this import operation. PostgreSQL instances only.
- "fileType": "A String", # The file type for the specified uri. *SQL*: The file contains SQL statements. *CSV*: The file contains CSV data.
- "uri": "A String", # Path to the import file in Cloud Storage, in the form *gs: //bucketName/fileName*. Compressed gzip files (.gz) are supported // when *fileType* is *SQL*. The instance must have // write permissions to the bucket and read access to the file.
- "bakImportOptions": { # Import parameters specific to SQL Server .BAK files
- "encryptionOptions": {
- "pvkPassword": "A String", # Password that encrypts the private key
- "certPath": "A String", # Path to the Certificate (.cer) in Cloud Storage, in the form *gs://bucketName/fileName*. The instance must have write permissions to the bucket and read access to the file.
- "pvkPath": "A String", # Path to the Certificate Private Key (.pvk) in Cloud Storage, in the form *gs://bucketName/fileName*. The instance must have write permissions to the bucket and read access to the file.
- },
- },
- },
- "exportContext": { # Database instance export context. # The context for export operation, if applicable.
- "uri": "A String", # The path to the file in Google Cloud Storage where the export will be stored. The URI is in the form *gs: //bucketName/fileName*. If the file already exists, the requests // succeeds, but the operation fails. If *fileType* is // *SQL* and the filename ends with .gz, the contents are // compressed.
- "sqlExportOptions": { # Options for exporting data as SQL statements.
- "mysqlExportOptions": { # Options for exporting from MySQL.
- "masterData": 42, # Option to include SQL statement required to set up replication. If set to *1*, the dump file includes a CHANGE MASTER TO statement with the binary log coordinates. If set to *2*, the CHANGE MASTER TO statement is written as a SQL comment, and has no effect. All other values are ignored.
- },
- "schemaOnly": True or False, # Export only schemas.
- "tables": [ # Tables to export, or that were exported, from the specified database. If you specify tables, specify one and only one database. For PostgreSQL instances, you can specify only one table.
- "A String",
- ],
- },
- "kind": "A String", # This is always *sql#exportContext*.
- "csvExportOptions": { # Options for exporting data as CSV.
- "selectQuery": "A String", # The select query used to extract the data.
- },
- "fileType": "A String", # The file type for the specified uri. *SQL*: The file contains SQL statements. *CSV*: The file contains CSV data.
- "offload": True or False, # Option for export offload.
- "databases": [ # Databases to be exported. *MySQL instances:* If *fileType* is *SQL* and no database is specified, all databases are exported, except for the *mysql* system database. If *fileType* is *CSV*, you can specify one database, either by using this property or by using the *csvExportOptions.selectQuery* property, which takes precedence over this property. *PostgreSQL instances:* You must specify one database to be exported. If *fileType* is *CSV*, this database must match the one specified in the *csvExportOptions.selectQuery* property.
- "A String",
- ],
- },
- "endTime": "A String", # The time this operation finished in UTC timezone in RFC 3339 format, for example *2012-11-15T16:19:00.094Z*.
- "user": "A String", # The email address of the user who initiated this operation.
- "selfLink": "A String", # The URI of this resource.
"status": "A String", # The status of an operation. Valid values are: *PENDING* *RUNNING* *DONE* *SQL_OPERATION_STATUS_UNSPECIFIED*
- "insertTime": "A String", # The time this operation was enqueued in UTC timezone in RFC 3339 format, for example *2012-11-15T16:19:00.094Z*.
- "targetId": "A String", # Name of the database instance related to this operation.
- "targetProject": "A String", # The project ID of the target instance related to this operation.
- "operationType": "A String", # The type of the operation. Valid values are: *CREATE* *DELETE* *UPDATE* *RESTART* *IMPORT* *EXPORT* *BACKUP_VOLUME* *RESTORE_VOLUME* *CREATE_USER* *DELETE_USER* *CREATE_DATABASE* *DELETE_DATABASE*
- "kind": "A String", # This is always *sql#operation*.
"error": { # Database instance operation errors list wrapper. # If errors occurred during processing of this operation, this field will be populated.
"kind": "A String", # This is always *sql#operationErrors*.
"errors": [ # The list of errors encountered while processing this operation.
@@ -275,9 +226,58 @@
},
],
},
+ "endTime": "A String", # The time this operation finished in UTC timezone in RFC 3339 format, for example *2012-11-15T16:19:00.094Z*.
"targetLink": "A String",
"name": "A String", # An identifier that uniquely identifies the operation. You can use this identifier to retrieve the Operations resource that has information about the operation.
+ "exportContext": { # Database instance export context. # The context for export operation, if applicable.
+ "offload": True or False, # Option for export offload.
+ "csvExportOptions": { # Options for exporting data as CSV.
+ "selectQuery": "A String", # The select query used to extract the data.
+ },
+ "databases": [ # Databases to be exported. *MySQL instances:* If *fileType* is *SQL* and no database is specified, all databases are exported, except for the *mysql* system database. If *fileType* is *CSV*, you can specify one database, either by using this property or by using the *csvExportOptions.selectQuery* property, which takes precedence over this property. *PostgreSQL instances:* You must specify one database to be exported. If *fileType* is *CSV*, this database must match the one specified in the *csvExportOptions.selectQuery* property.
+ "A String",
+ ],
+ "kind": "A String", # This is always *sql#exportContext*.
+ "sqlExportOptions": { # Options for exporting data as SQL statements.
+ "schemaOnly": True or False, # Export only schemas.
+ "tables": [ # Tables to export, or that were exported, from the specified database. If you specify tables, specify one and only one database. For PostgreSQL instances, you can specify only one table.
+ "A String",
+ ],
+ "mysqlExportOptions": { # Options for exporting from MySQL.
+ "masterData": 42, # Option to include SQL statement required to set up replication. If set to *1*, the dump file includes a CHANGE MASTER TO statement with the binary log coordinates. If set to *2*, the CHANGE MASTER TO statement is written as a SQL comment, and has no effect. All other values are ignored.
+ },
+ },
+ "fileType": "A String", # The file type for the specified uri. *SQL*: The file contains SQL statements. *CSV*: The file contains CSV data.
+ "uri": "A String", # The path to the file in Google Cloud Storage where the export will be stored. The URI is in the form *gs: //bucketName/fileName*. If the file already exists, the requests // succeeds, but the operation fails. If *fileType* is // *SQL* and the filename ends with .gz, the contents are // compressed.
+ },
+ "operationType": "A String", # The type of the operation. Valid values are: *CREATE* *DELETE* *UPDATE* *RESTART* *IMPORT* *EXPORT* *BACKUP_VOLUME* *RESTORE_VOLUME* *CREATE_USER* *DELETE_USER* *CREATE_DATABASE* *DELETE_DATABASE*
"startTime": "A String", # The time this operation actually started in UTC timezone in RFC 3339 format, for example *2012-11-15T16:19:00.094Z*.
+ "selfLink": "A String", # The URI of this resource.
+ "kind": "A String", # This is always *sql#operation*.
+ "insertTime": "A String", # The time this operation was enqueued in UTC timezone in RFC 3339 format, for example *2012-11-15T16:19:00.094Z*.
+ "targetProject": "A String", # The project ID of the target instance related to this operation.
+ "importContext": { # Database instance import context. # The context for import operation, if applicable.
+ "uri": "A String", # Path to the import file in Cloud Storage, in the form *gs: //bucketName/fileName*. Compressed gzip files (.gz) are supported // when *fileType* is *SQL*. The instance must have // write permissions to the bucket and read access to the file.
+ "fileType": "A String", # The file type for the specified uri. *SQL*: The file contains SQL statements. *CSV*: The file contains CSV data.
+ "csvImportOptions": { # Options for importing data as CSV.
+ "columns": [ # The columns to which CSV data is imported. If not specified, all columns of the database table are loaded with CSV data.
+ "A String",
+ ],
+ "table": "A String", # The table to which CSV data is imported.
+ },
+ "database": "A String", # The target database for the import. If *fileType* is *SQL*, this field is required only if the import file does not specify a database, and is overridden by any database specification in the import file. If *fileType* is *CSV*, one database must be specified.
+ "importUser": "A String", # The PostgreSQL user for this import operation. PostgreSQL instances only.
+ "kind": "A String", # This is always *sql#importContext*.
+ "bakImportOptions": { # Import parameters specific to SQL Server .BAK files
+ "encryptionOptions": {
+ "certPath": "A String", # Path to the Certificate (.cer) in Cloud Storage, in the form *gs://bucketName/fileName*. The instance must have write permissions to the bucket and read access to the file.
+ "pvkPassword": "A String", # Password that encrypts the private key
+ "pvkPath": "A String", # Path to the Certificate Private Key (.pvk) in Cloud Storage, in the form *gs://bucketName/fileName*. The instance must have write permissions to the bucket and read access to the file.
+ },
+ },
+ },
+ "user": "A String", # The email address of the user who initiated this operation.
+ "targetId": "A String", # Name of the database instance related to this operation.
}</pre>
</div>
@@ -297,31 +297,31 @@
An object of the form:
{ # User list response.
- "kind": "A String", # This is always *sql#usersList*.
- "nextPageToken": "A String", # An identifier that uniquely identifies the operation. You can use this identifier to retrieve the Operations resource that has information about the operation.
"items": [ # List of user resources in the instance.
{ # A Cloud SQL user resource.
- "etag": "A String", # This field is deprecated and will be removed from a future version of the API.
- "host": "A String", # The host name from which the user can connect. For *insert* operations, host defaults to an empty string. For *update* operations, host is specified as part of the request URL. The host name cannot be updated after insertion.
- "kind": "A String", # This is always *sql#user*.
- "project": "A String", # The project ID of the project containing the Cloud SQL database. The Google apps domain is prefixed if applicable. Can be omitted for *update* since it is already specified on the URL.
- "type": "A String", # The user type. It determines the method to authenticate the user during login. The default is the database's built-in user type.
- "name": "A String", # The name of the user in the Cloud SQL instance. Can be omitted for *update* since it is already specified in the URL.
- "instance": "A String", # The name of the Cloud SQL instance. This does not include the project ID. Can be omitted for *update* since it is already specified on the URL.
- "password": "A String", # The password for the user.
- "sqlserverUserDetails": { # Represents a Sql Server user on the Cloud SQL instance.
- "disabled": True or False, # If the user has been disabled
- "serverRoles": [ # The server roles for this user
- "A String",
- ],
+ "password": "A String", # The password for the user.
+ "sqlserverUserDetails": { # Represents a Sql Server user on the Cloud SQL instance.
+ "serverRoles": [ # The server roles for this user
+ "A String",
+ ],
+ "disabled": True or False, # If the user has been disabled
+ },
+ "kind": "A String", # This is always *sql#user*.
+ "name": "A String", # The name of the user in the Cloud SQL instance. Can be omitted for *update* since it is already specified in the URL.
+ "project": "A String", # The project ID of the project containing the Cloud SQL database. The Google apps domain is prefixed if applicable. Can be omitted for *update* since it is already specified on the URL.
+ "instance": "A String", # The name of the Cloud SQL instance. This does not include the project ID. Can be omitted for *update* since it is already specified on the URL.
+ "type": "A String", # The user type. It determines the method to authenticate the user during login. The default is the database's built-in user type.
+ "etag": "A String", # This field is deprecated and will be removed from a future version of the API.
+ "host": "A String", # The host name from which the user can connect. For *insert* operations, host defaults to an empty string. For *update* operations, host is specified as part of the request URL. The host name cannot be updated after insertion.
},
- },
],
+ "kind": "A String", # This is always *sql#usersList*.
+ "nextPageToken": "A String", # An identifier that uniquely identifies the operation. You can use this identifier to retrieve the Operations resource that has information about the operation.
}</pre>
</div>
<div class="method">
- <code class="details" id="update">update(project, instance, body=None, name=None, host=None, x__xgafv=None)</code>
+ <code class="details" id="update">update(project, instance, body=None, host=None, name=None, x__xgafv=None)</code>
<pre>Updates an existing user in a Cloud SQL instance.
Args:
@@ -331,24 +331,24 @@
The object takes the form of:
{ # A Cloud SQL user resource.
- "etag": "A String", # This field is deprecated and will be removed from a future version of the API.
- "host": "A String", # The host name from which the user can connect. For *insert* operations, host defaults to an empty string. For *update* operations, host is specified as part of the request URL. The host name cannot be updated after insertion.
- "kind": "A String", # This is always *sql#user*.
- "project": "A String", # The project ID of the project containing the Cloud SQL database. The Google apps domain is prefixed if applicable. Can be omitted for *update* since it is already specified on the URL.
- "type": "A String", # The user type. It determines the method to authenticate the user during login. The default is the database's built-in user type.
- "name": "A String", # The name of the user in the Cloud SQL instance. Can be omitted for *update* since it is already specified in the URL.
- "instance": "A String", # The name of the Cloud SQL instance. This does not include the project ID. Can be omitted for *update* since it is already specified on the URL.
- "password": "A String", # The password for the user.
- "sqlserverUserDetails": { # Represents a Sql Server user on the Cloud SQL instance.
- "disabled": True or False, # If the user has been disabled
- "serverRoles": [ # The server roles for this user
- "A String",
- ],
- },
-}
+ "password": "A String", # The password for the user.
+ "sqlserverUserDetails": { # Represents a Sql Server user on the Cloud SQL instance.
+ "serverRoles": [ # The server roles for this user
+ "A String",
+ ],
+ "disabled": True or False, # If the user has been disabled
+ },
+ "kind": "A String", # This is always *sql#user*.
+ "name": "A String", # The name of the user in the Cloud SQL instance. Can be omitted for *update* since it is already specified in the URL.
+ "project": "A String", # The project ID of the project containing the Cloud SQL database. The Google apps domain is prefixed if applicable. Can be omitted for *update* since it is already specified on the URL.
+ "instance": "A String", # The name of the Cloud SQL instance. This does not include the project ID. Can be omitted for *update* since it is already specified on the URL.
+ "type": "A String", # The user type. It determines the method to authenticate the user during login. The default is the database's built-in user type.
+ "etag": "A String", # This field is deprecated and will be removed from a future version of the API.
+ "host": "A String", # The host name from which the user can connect. For *insert* operations, host defaults to an empty string. For *update* operations, host is specified as part of the request URL. The host name cannot be updated after insertion.
+ }
- name: string, Name of the user in the instance.
host: string, Optional. Host of the user in the instance.
+ name: string, Name of the user in the instance.
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -358,56 +358,7 @@
An object of the form:
{ # An Operation resource. For successful operations that return an Operation resource, only the fields relevant to the operation are populated in the resource.
- "importContext": { # Database instance import context. # The context for import operation, if applicable.
- "database": "A String", # The target database for the import. If *fileType* is *SQL*, this field is required only if the import file does not specify a database, and is overridden by any database specification in the import file. If *fileType* is *CSV*, one database must be specified.
- "csvImportOptions": { # Options for importing data as CSV.
- "columns": [ # The columns to which CSV data is imported. If not specified, all columns of the database table are loaded with CSV data.
- "A String",
- ],
- "table": "A String", # The table to which CSV data is imported.
- },
- "kind": "A String", # This is always *sql#importContext*.
- "importUser": "A String", # The PostgreSQL user for this import operation. PostgreSQL instances only.
- "fileType": "A String", # The file type for the specified uri. *SQL*: The file contains SQL statements. *CSV*: The file contains CSV data.
- "uri": "A String", # Path to the import file in Cloud Storage, in the form *gs: //bucketName/fileName*. Compressed gzip files (.gz) are supported // when *fileType* is *SQL*. The instance must have // write permissions to the bucket and read access to the file.
- "bakImportOptions": { # Import parameters specific to SQL Server .BAK files
- "encryptionOptions": {
- "pvkPassword": "A String", # Password that encrypts the private key
- "certPath": "A String", # Path to the Certificate (.cer) in Cloud Storage, in the form *gs://bucketName/fileName*. The instance must have write permissions to the bucket and read access to the file.
- "pvkPath": "A String", # Path to the Certificate Private Key (.pvk) in Cloud Storage, in the form *gs://bucketName/fileName*. The instance must have write permissions to the bucket and read access to the file.
- },
- },
- },
- "exportContext": { # Database instance export context. # The context for export operation, if applicable.
- "uri": "A String", # The path to the file in Google Cloud Storage where the export will be stored. The URI is in the form *gs: //bucketName/fileName*. If the file already exists, the requests // succeeds, but the operation fails. If *fileType* is // *SQL* and the filename ends with .gz, the contents are // compressed.
- "sqlExportOptions": { # Options for exporting data as SQL statements.
- "mysqlExportOptions": { # Options for exporting from MySQL.
- "masterData": 42, # Option to include SQL statement required to set up replication. If set to *1*, the dump file includes a CHANGE MASTER TO statement with the binary log coordinates. If set to *2*, the CHANGE MASTER TO statement is written as a SQL comment, and has no effect. All other values are ignored.
- },
- "schemaOnly": True or False, # Export only schemas.
- "tables": [ # Tables to export, or that were exported, from the specified database. If you specify tables, specify one and only one database. For PostgreSQL instances, you can specify only one table.
- "A String",
- ],
- },
- "kind": "A String", # This is always *sql#exportContext*.
- "csvExportOptions": { # Options for exporting data as CSV.
- "selectQuery": "A String", # The select query used to extract the data.
- },
- "fileType": "A String", # The file type for the specified uri. *SQL*: The file contains SQL statements. *CSV*: The file contains CSV data.
- "offload": True or False, # Option for export offload.
- "databases": [ # Databases to be exported. *MySQL instances:* If *fileType* is *SQL* and no database is specified, all databases are exported, except for the *mysql* system database. If *fileType* is *CSV*, you can specify one database, either by using this property or by using the *csvExportOptions.selectQuery* property, which takes precedence over this property. *PostgreSQL instances:* You must specify one database to be exported. If *fileType* is *CSV*, this database must match the one specified in the *csvExportOptions.selectQuery* property.
- "A String",
- ],
- },
- "endTime": "A String", # The time this operation finished in UTC timezone in RFC 3339 format, for example *2012-11-15T16:19:00.094Z*.
- "user": "A String", # The email address of the user who initiated this operation.
- "selfLink": "A String", # The URI of this resource.
"status": "A String", # The status of an operation. Valid values are: *PENDING* *RUNNING* *DONE* *SQL_OPERATION_STATUS_UNSPECIFIED*
- "insertTime": "A String", # The time this operation was enqueued in UTC timezone in RFC 3339 format, for example *2012-11-15T16:19:00.094Z*.
- "targetId": "A String", # Name of the database instance related to this operation.
- "targetProject": "A String", # The project ID of the target instance related to this operation.
- "operationType": "A String", # The type of the operation. Valid values are: *CREATE* *DELETE* *UPDATE* *RESTART* *IMPORT* *EXPORT* *BACKUP_VOLUME* *RESTORE_VOLUME* *CREATE_USER* *DELETE_USER* *CREATE_DATABASE* *DELETE_DATABASE*
- "kind": "A String", # This is always *sql#operation*.
"error": { # Database instance operation errors list wrapper. # If errors occurred during processing of this operation, this field will be populated.
"kind": "A String", # This is always *sql#operationErrors*.
"errors": [ # The list of errors encountered while processing this operation.
@@ -418,9 +369,58 @@
},
],
},
+ "endTime": "A String", # The time this operation finished in UTC timezone in RFC 3339 format, for example *2012-11-15T16:19:00.094Z*.
"targetLink": "A String",
"name": "A String", # An identifier that uniquely identifies the operation. You can use this identifier to retrieve the Operations resource that has information about the operation.
+ "exportContext": { # Database instance export context. # The context for export operation, if applicable.
+ "offload": True or False, # Option for export offload.
+ "csvExportOptions": { # Options for exporting data as CSV.
+ "selectQuery": "A String", # The select query used to extract the data.
+ },
+ "databases": [ # Databases to be exported. *MySQL instances:* If *fileType* is *SQL* and no database is specified, all databases are exported, except for the *mysql* system database. If *fileType* is *CSV*, you can specify one database, either by using this property or by using the *csvExportOptions.selectQuery* property, which takes precedence over this property. *PostgreSQL instances:* You must specify one database to be exported. If *fileType* is *CSV*, this database must match the one specified in the *csvExportOptions.selectQuery* property.
+ "A String",
+ ],
+ "kind": "A String", # This is always *sql#exportContext*.
+ "sqlExportOptions": { # Options for exporting data as SQL statements.
+ "schemaOnly": True or False, # Export only schemas.
+ "tables": [ # Tables to export, or that were exported, from the specified database. If you specify tables, specify one and only one database. For PostgreSQL instances, you can specify only one table.
+ "A String",
+ ],
+ "mysqlExportOptions": { # Options for exporting from MySQL.
+ "masterData": 42, # Option to include SQL statement required to set up replication. If set to *1*, the dump file includes a CHANGE MASTER TO statement with the binary log coordinates. If set to *2*, the CHANGE MASTER TO statement is written as a SQL comment, and has no effect. All other values are ignored.
+ },
+ },
+ "fileType": "A String", # The file type for the specified uri. *SQL*: The file contains SQL statements. *CSV*: The file contains CSV data.
+ "uri": "A String", # The path to the file in Google Cloud Storage where the export will be stored. The URI is in the form *gs: //bucketName/fileName*. If the file already exists, the requests // succeeds, but the operation fails. If *fileType* is // *SQL* and the filename ends with .gz, the contents are // compressed.
+ },
+ "operationType": "A String", # The type of the operation. Valid values are: *CREATE* *DELETE* *UPDATE* *RESTART* *IMPORT* *EXPORT* *BACKUP_VOLUME* *RESTORE_VOLUME* *CREATE_USER* *DELETE_USER* *CREATE_DATABASE* *DELETE_DATABASE*
"startTime": "A String", # The time this operation actually started in UTC timezone in RFC 3339 format, for example *2012-11-15T16:19:00.094Z*.
+ "selfLink": "A String", # The URI of this resource.
+ "kind": "A String", # This is always *sql#operation*.
+ "insertTime": "A String", # The time this operation was enqueued in UTC timezone in RFC 3339 format, for example *2012-11-15T16:19:00.094Z*.
+ "targetProject": "A String", # The project ID of the target instance related to this operation.
+ "importContext": { # Database instance import context. # The context for import operation, if applicable.
+ "uri": "A String", # Path to the import file in Cloud Storage, in the form *gs: //bucketName/fileName*. Compressed gzip files (.gz) are supported // when *fileType* is *SQL*. The instance must have // write permissions to the bucket and read access to the file.
+ "fileType": "A String", # The file type for the specified uri. *SQL*: The file contains SQL statements. *CSV*: The file contains CSV data.
+ "csvImportOptions": { # Options for importing data as CSV.
+ "columns": [ # The columns to which CSV data is imported. If not specified, all columns of the database table are loaded with CSV data.
+ "A String",
+ ],
+ "table": "A String", # The table to which CSV data is imported.
+ },
+ "database": "A String", # The target database for the import. If *fileType* is *SQL*, this field is required only if the import file does not specify a database, and is overridden by any database specification in the import file. If *fileType* is *CSV*, one database must be specified.
+ "importUser": "A String", # The PostgreSQL user for this import operation. PostgreSQL instances only.
+ "kind": "A String", # This is always *sql#importContext*.
+ "bakImportOptions": { # Import parameters specific to SQL Server .BAK files
+ "encryptionOptions": {
+ "certPath": "A String", # Path to the Certificate (.cer) in Cloud Storage, in the form *gs://bucketName/fileName*. The instance must have write permissions to the bucket and read access to the file.
+ "pvkPassword": "A String", # Password that encrypts the private key
+ "pvkPath": "A String", # Path to the Certificate Private Key (.pvk) in Cloud Storage, in the form *gs://bucketName/fileName*. The instance must have write permissions to the bucket and read access to the file.
+ },
+ },
+ },
+ "user": "A String", # The email address of the user who initiated this operation.
+ "targetId": "A String", # Name of the database instance related to this operation.
}</pre>
</div>