docs: update generated docs (#1053)

Updates for both discovery docs and epydoc API Documentation

Fixes: #1049
diff --git a/docs/dyn/sqladmin_v1beta4.operations.html b/docs/dyn/sqladmin_v1beta4.operations.html
index 2b17a17..763a354 100644
--- a/docs/dyn/sqladmin_v1beta4.operations.html
+++ b/docs/dyn/sqladmin_v1beta4.operations.html
@@ -75,16 +75,24 @@
 <h1><a href="sqladmin_v1beta4.html">Cloud SQL Admin API</a> . <a href="sqladmin_v1beta4.operations.html">operations</a></h1>
 <h2>Instance Methods</h2>
 <p class="toc_element">
+  <code><a href="#close">close()</a></code></p>
+<p class="firstline">Close httplib2 connections.</p>
+<p class="toc_element">
   <code><a href="#get">get(project, operation, x__xgafv=None)</a></code></p>
 <p class="firstline">Retrieves an instance operation that has been performed on an instance.</p>
 <p class="toc_element">
-  <code><a href="#list">list(project, maxResults=None, instance=None, pageToken=None, x__xgafv=None)</a></code></p>
-<p class="firstline">Lists all instance operations that have been performed on the given Cloud</p>
+  <code><a href="#list">list(project, maxResults=None, pageToken=None, instance=None, x__xgafv=None)</a></code></p>
+<p class="firstline">Lists all instance operations that have been performed on the given Cloud SQL instance in the reverse chronological order of the start time.</p>
 <p class="toc_element">
   <code><a href="#list_next">list_next(previous_request, previous_response)</a></code></p>
 <p class="firstline">Retrieves the next page of results.</p>
 <h3>Method Details</h3>
 <div class="method">
+    <code class="details" id="close">close()</code>
+  <pre>Close httplib2 connections.</pre>
+</div>
+
+<div class="method">
     <code class="details" id="get">get(project, operation, x__xgafv=None)</code>
   <pre>Retrieves an instance operation that has been performed on an instance.
 
@@ -99,137 +107,82 @@
 Returns:
   An object of the form:
 
-    { # An Operation resource.&amp;nbsp;For successful operations that return an
-      # Operation resource, only the fields relevant to the operation are populated
-      # in the resource.
-    &quot;targetLink&quot;: &quot;A String&quot;,
-    &quot;operationType&quot;: &quot;A String&quot;, # The type of the operation. Valid values are &lt;code&gt;CREATE&lt;/code&gt;,
-        # &lt;code&gt;DELETE&lt;/code&gt;, &lt;code&gt;UPDATE&lt;/code&gt;, &lt;code&gt;RESTART&lt;/code&gt;,
-        # &lt;code&gt;IMPORT&lt;/code&gt;, &lt;code&gt;EXPORT&lt;/code&gt;, &lt;code&gt;BACKUP_VOLUME&lt;/code&gt;,
-        # &lt;code&gt;RESTORE_VOLUME&lt;/code&gt;, &lt;code&gt;CREATE_USER&lt;/code&gt;,
-        # &lt;code&gt;DELETE_USER&lt;/code&gt;, &lt;code&gt;CREATE_DATABASE&lt;/code&gt;,
-        # &lt;code&gt;DELETE_DATABASE&lt;/code&gt; .
-    &quot;error&quot;: { # Database instance operation errors list wrapper. # If errors occurred during processing of this operation, this field will be
-        # populated.
+    { # An Operation resource. For successful operations that return an Operation resource, only the fields relevant to the operation are populated in the resource.
+    &quot;kind&quot;: &quot;A String&quot;, # This is always *sql#operation*.
+    &quot;name&quot;: &quot;A String&quot;, # An identifier that uniquely identifies the operation. You can use this identifier to retrieve the Operations resource that has information about the operation.
+    &quot;endTime&quot;: &quot;A String&quot;, # The time this operation finished in UTC timezone in RFC 3339 format, for example *2012-11-15T16:19:00.094Z*.
+    &quot;error&quot;: { # Database instance operation errors list wrapper. # If errors occurred during processing of this operation, this field will be populated.
       &quot;errors&quot;: [ # The list of errors encountered while processing this operation.
         { # Database instance operation error.
-          &quot;kind&quot;: &quot;A String&quot;, # This is always &lt;code&gt;sql#operationError&lt;/code&gt;.
           &quot;code&quot;: &quot;A String&quot;, # Identifies the specific error that occurred.
+          &quot;kind&quot;: &quot;A String&quot;, # This is always *sql#operationError*.
           &quot;message&quot;: &quot;A String&quot;, # Additional information about the error encountered.
         },
       ],
-      &quot;kind&quot;: &quot;A String&quot;, # This is always &lt;code&gt;sql#operationErrors&lt;/code&gt;.
+      &quot;kind&quot;: &quot;A String&quot;, # This is always *sql#operationErrors*.
     },
-    &quot;kind&quot;: &quot;A String&quot;, # This is always &lt;code&gt;sql#operation&lt;/code&gt;.
-    &quot;importContext&quot;: { # Database instance import context. # The context for import operation, if applicable.
-      &quot;database&quot;: &quot;A String&quot;, # The target database for the import. If &lt;code&gt;fileType&lt;/code&gt; is
-          # &lt;code&gt;SQL&lt;/code&gt;, this field is required only if the import file does not
-          # specify a database, and is overridden by any database specification in the
-          # import file. If &lt;code&gt;fileType&lt;/code&gt; is &lt;code&gt;CSV&lt;/code&gt;, one database
-          # must be specified.
-      &quot;importUser&quot;: &quot;A String&quot;, # The PostgreSQL user for this import operation. PostgreSQL instances only.
-      &quot;bakImportOptions&quot;: { # Import parameters specific to SQL Server .BAK files
-        &quot;encryptionOptions&quot;: {
-          &quot;pvkPassword&quot;: &quot;A String&quot;, # Password that encrypts the private key
-          &quot;certPath&quot;: &quot;A String&quot;, # Path to the Certificate (.cer) in Cloud Storage, in the form
-              # &lt;code&gt;gs://bucketName/fileName&lt;/code&gt;. The instance must have
-              # write permissions to the bucket and read access to the file.
-          &quot;pvkPath&quot;: &quot;A String&quot;, # Path to the Certificate Private Key (.pvk)  in Cloud Storage, in the
-              # form &lt;code&gt;gs://bucketName/fileName&lt;/code&gt;. The instance must have
-              # write permissions to the bucket and read access to the file.
-        },
-      },
-      &quot;uri&quot;: &quot;A String&quot;, # Path to the import file in Cloud Storage, in the form
-          # &lt;code&gt;gs:
-          # //bucketName/fileName&lt;/code&gt;. Compressed gzip files (.gz) are supported
-          # // when &lt;code&gt;fileType&lt;/code&gt; is &lt;code&gt;SQL&lt;/code&gt;. The instance must have
-          # // write permissions to the bucket and read access to the file.
-      &quot;fileType&quot;: &quot;A String&quot;, # The file type for the specified uri. &lt;br&gt;&lt;code&gt;SQL&lt;/code&gt;: The file
-          # contains SQL statements. &lt;br&gt;&lt;code&gt;CSV&lt;/code&gt;: The file contains CSV data.
-      &quot;kind&quot;: &quot;A String&quot;, # This is always &lt;code&gt;sql#importContext&lt;/code&gt;.
-      &quot;csvImportOptions&quot;: { # Options for importing data as CSV.
-        &quot;table&quot;: &quot;A String&quot;, # The table to which CSV data is imported.
-        &quot;columns&quot;: [ # The columns to which CSV data is imported. If not specified, all columns
-            # of the database table are loaded with CSV data.
-          &quot;A String&quot;,
-        ],
-      },
-    },
-    &quot;status&quot;: &quot;A String&quot;, # The status of an operation. Valid values are &lt;code&gt;PENDING&lt;/code&gt;,
-        # &lt;code&gt;RUNNING&lt;/code&gt;, &lt;code&gt;DONE&lt;/code&gt;,
-        # &lt;code&gt;SQL_OPERATION_STATUS_UNSPECIFIED&lt;/code&gt;.
-    &quot;name&quot;: &quot;A String&quot;, # An identifier that uniquely identifies the operation. You can use this
-        # identifier to retrieve the Operations resource that has information about
-        # the operation.
+    &quot;status&quot;: &quot;A String&quot;, # The status of an operation. Valid values are: *PENDING* *RUNNING* *DONE* *SQL_OPERATION_STATUS_UNSPECIFIED*
+    &quot;selfLink&quot;: &quot;A String&quot;, # The URI of this resource.
+    &quot;startTime&quot;: &quot;A String&quot;, # The time this operation actually started in UTC timezone in RFC 3339 format, for example *2012-11-15T16:19:00.094Z*.
     &quot;exportContext&quot;: { # Database instance export context. # The context for export operation, if applicable.
-      &quot;fileType&quot;: &quot;A String&quot;, # The file type for the specified uri. &lt;br&gt;&lt;code&gt;SQL&lt;/code&gt;: The file
-          # contains SQL statements. &lt;br&gt;&lt;code&gt;CSV&lt;/code&gt;: The file contains CSV data.
-      &quot;uri&quot;: &quot;A String&quot;, # The path to the file in Google Cloud Storage where the export will be
-          # stored. The URI is in the form &lt;code&gt;gs:
-          # //bucketName/fileName&lt;/code&gt;. If the file already exists, the requests
-          # // succeeds, but the operation fails. If &lt;code&gt;fileType&lt;/code&gt; is
-          # // &lt;code&gt;SQL&lt;/code&gt; and the filename ends with .gz, the contents are
-          # // compressed.
       &quot;csvExportOptions&quot;: { # Options for exporting data as CSV.
         &quot;selectQuery&quot;: &quot;A String&quot;, # The select query used to extract the data.
       },
-      &quot;kind&quot;: &quot;A String&quot;, # This is always &lt;code&gt;sql#exportContext&lt;/code&gt;.
-      &quot;databases&quot;: [ # Databases to be exported. &lt;br /&gt; &lt;b&gt;MySQL instances:&lt;/b&gt; If
-          # &lt;code&gt;fileType&lt;/code&gt; is &lt;code&gt;SQL&lt;/code&gt; and no database is specified, all
-          # databases are exported, except for the &lt;code&gt;mysql&lt;/code&gt; system database.
-          # If &lt;code&gt;fileType&lt;/code&gt; is &lt;code&gt;CSV&lt;/code&gt;, you can specify one database,
-          # either by using this property or by using the
-          # &lt;code&gt;csvExportOptions.selectQuery&lt;/code&gt; property, which takes precedence
-          # over this property. &lt;br /&gt; &lt;b&gt;PostgreSQL instances:&lt;/b&gt; You must specify
-          # one database to be exported. If &lt;code&gt;fileType&lt;/code&gt; is &lt;code&gt;CSV&lt;/code&gt;,
-          # this database must match the one specified in the
-          # &lt;code&gt;csvExportOptions.selectQuery&lt;/code&gt; property.
+      &quot;fileType&quot;: &quot;A String&quot;, # The file type for the specified uri. *SQL*: The file contains SQL statements. *CSV*: The file contains CSV data.
+      &quot;kind&quot;: &quot;A String&quot;, # This is always *sql#exportContext*.
+      &quot;uri&quot;: &quot;A String&quot;, # The path to the file in Google Cloud Storage where the export will be stored. The URI is in the form *gs: //bucketName/fileName*. If the file already exists, the requests // succeeds, but the operation fails. If *fileType* is // *SQL* and the filename ends with .gz, the contents are // compressed.
+      &quot;databases&quot;: [ # Databases to be exported. *MySQL instances:* If *fileType* is *SQL* and no database is specified, all databases are exported, except for the *mysql* system database. If *fileType* is *CSV*, you can specify one database, either by using this property or by using the *csvExportOptions.selectQuery* property, which takes precedence over this property. *PostgreSQL instances:* You must specify one database to be exported. If *fileType* is *CSV*, this database must match the one specified in the *csvExportOptions.selectQuery* property.
         &quot;A String&quot;,
       ],
+      &quot;offload&quot;: True or False, # Option for export offload.
       &quot;sqlExportOptions&quot;: { # Options for exporting data as SQL statements.
-        &quot;tables&quot;: [ # Tables to export, or that were exported, from the specified database. If
-            # you specify tables, specify one and only one database. For PostgreSQL
-            # instances, you can specify only one table.
+        &quot;schemaOnly&quot;: True or False, # Export only schemas.
+        &quot;tables&quot;: [ # Tables to export, or that were exported, from the specified database. If you specify tables, specify one and only one database. For PostgreSQL instances, you can specify only one table.
           &quot;A String&quot;,
         ],
-        &quot;schemaOnly&quot;: True or False, # Export only schemas.
         &quot;mysqlExportOptions&quot;: { # Options for exporting from MySQL.
-          &quot;masterData&quot;: 42, # Option to include SQL statement required to set up replication.
-              # If set to &lt;code&gt;1&lt;/code&gt;, the dump file includes
-              #  a CHANGE MASTER TO statement with the binary log coordinates.
-              # If set to &lt;code&gt;2&lt;/code&gt;, the CHANGE MASTER TO statement is written as
-              #  a SQL comment, and has no effect.
-              # All other values are ignored.
+          &quot;masterData&quot;: 42, # Option to include SQL statement required to set up replication. If set to *1*, the dump file includes a CHANGE MASTER TO statement with the binary log coordinates. If set to *2*, the CHANGE MASTER TO statement is written as a SQL comment, and has no effect. All other values are ignored.
         },
       },
     },
-    &quot;targetId&quot;: &quot;A String&quot;, # Name of the database instance related to this operation.
-    &quot;endTime&quot;: &quot;A String&quot;, # The time this operation finished in UTC timezone in &lt;a
-        # href=&quot;https://tools.ietf.org/html/rfc3339&quot;&gt;RFC 3339&lt;/a&gt; format, for example
-        # &lt;code&gt;2012-11-15T16:19:00.094Z&lt;/code&gt;.
-    &quot;startTime&quot;: &quot;A String&quot;, # The time this operation actually started in UTC timezone in &lt;a
-        # href=&quot;https://tools.ietf.org/html/rfc3339&quot;&gt;RFC 3339&lt;/a&gt; format, for example
-        # &lt;code&gt;2012-11-15T16:19:00.094Z&lt;/code&gt;.
-    &quot;insertTime&quot;: &quot;A String&quot;, # The time this operation was enqueued in UTC timezone in &lt;a
-        # href=&quot;https://tools.ietf.org/html/rfc3339&quot;&gt;RFC 3339&lt;/a&gt; format, for example
-        # &lt;code&gt;2012-11-15T16:19:00.094Z&lt;/code&gt;.
     &quot;user&quot;: &quot;A String&quot;, # The email address of the user who initiated this operation.
+    &quot;targetId&quot;: &quot;A String&quot;, # Name of the database instance related to this operation.
+    &quot;insertTime&quot;: &quot;A String&quot;, # The time this operation was enqueued in UTC timezone in RFC 3339 format, for example *2012-11-15T16:19:00.094Z*.
+    &quot;operationType&quot;: &quot;A String&quot;, # The type of the operation. Valid values are: *CREATE* *DELETE* *UPDATE* *RESTART* *IMPORT* *EXPORT* *BACKUP_VOLUME* *RESTORE_VOLUME* *CREATE_USER* *DELETE_USER* *CREATE_DATABASE* *DELETE_DATABASE*
     &quot;targetProject&quot;: &quot;A String&quot;, # The project ID of the target instance related to this operation.
-    &quot;selfLink&quot;: &quot;A String&quot;, # The URI of this resource.
+    &quot;targetLink&quot;: &quot;A String&quot;,
+    &quot;importContext&quot;: { # Database instance import context. # The context for import operation, if applicable.
+      &quot;fileType&quot;: &quot;A String&quot;, # The file type for the specified uri. *SQL*: The file contains SQL statements. *CSV*: The file contains CSV data.
+      &quot;kind&quot;: &quot;A String&quot;, # This is always *sql#importContext*.
+      &quot;csvImportOptions&quot;: { # Options for importing data as CSV.
+        &quot;columns&quot;: [ # The columns to which CSV data is imported. If not specified, all columns of the database table are loaded with CSV data.
+          &quot;A String&quot;,
+        ],
+        &quot;table&quot;: &quot;A String&quot;, # The table to which CSV data is imported.
+      },
+      &quot;bakImportOptions&quot;: { # Import parameters specific to SQL Server .BAK files
+        &quot;encryptionOptions&quot;: {
+          &quot;pvkPath&quot;: &quot;A String&quot;, # Path to the Certificate Private Key (.pvk) in Cloud Storage, in the form *gs://bucketName/fileName*. The instance must have write permissions to the bucket and read access to the file.
+          &quot;certPath&quot;: &quot;A String&quot;, # Path to the Certificate (.cer) in Cloud Storage, in the form *gs://bucketName/fileName*. The instance must have write permissions to the bucket and read access to the file.
+          &quot;pvkPassword&quot;: &quot;A String&quot;, # Password that encrypts the private key
+        },
+      },
+      &quot;database&quot;: &quot;A String&quot;, # The target database for the import. If *fileType* is *SQL*, this field is required only if the import file does not specify a database, and is overridden by any database specification in the import file. If *fileType* is *CSV*, one database must be specified.
+      &quot;uri&quot;: &quot;A String&quot;, # Path to the import file in Cloud Storage, in the form *gs: //bucketName/fileName*. Compressed gzip files (.gz) are supported // when *fileType* is *SQL*. The instance must have // write permissions to the bucket and read access to the file.
+      &quot;importUser&quot;: &quot;A String&quot;, # The PostgreSQL user for this import operation. PostgreSQL instances only.
+    },
   }</pre>
 </div>
 
 <div class="method">
-    <code class="details" id="list">list(project, maxResults=None, instance=None, pageToken=None, x__xgafv=None)</code>
-  <pre>Lists all instance operations that have been performed on the given Cloud
-SQL instance in the reverse chronological order of the start time.
+    <code class="details" id="list">list(project, maxResults=None, pageToken=None, instance=None, x__xgafv=None)</code>
+  <pre>Lists all instance operations that have been performed on the given Cloud SQL instance in the reverse chronological order of the start time.
 
 Args:
   project: string, Project ID of the project that contains the instance. (required)
   maxResults: integer, Maximum number of operations per response.
+  pageToken: string, A previously-returned page token representing part of the larger set of results to view.
   instance: string, Cloud SQL instance ID. This does not include the project ID.
-  pageToken: string, A previously-returned page token representing part of the larger set of
-results to view.
   x__xgafv: string, V1 error format.
     Allowed values
       1 - v1 error format
@@ -239,127 +192,73 @@
   An object of the form:
 
     { # Database instance list operations response.
-    &quot;nextPageToken&quot;: &quot;A String&quot;, # The continuation token, used to page through large result sets. Provide
-        # this value in a subsequent request to return the next page of results.
-    &quot;kind&quot;: &quot;A String&quot;, # This is always &lt;code&gt;sql#operationsList&lt;/code&gt;.
+    &quot;kind&quot;: &quot;A String&quot;, # This is always *sql#operationsList*.
+    &quot;nextPageToken&quot;: &quot;A String&quot;, # The continuation token, used to page through large result sets. Provide this value in a subsequent request to return the next page of results.
     &quot;items&quot;: [ # List of operation resources.
-      { # An Operation resource.&amp;nbsp;For successful operations that return an
-          # Operation resource, only the fields relevant to the operation are populated
-          # in the resource.
-        &quot;targetLink&quot;: &quot;A String&quot;,
-        &quot;operationType&quot;: &quot;A String&quot;, # The type of the operation. Valid values are &lt;code&gt;CREATE&lt;/code&gt;,
-            # &lt;code&gt;DELETE&lt;/code&gt;, &lt;code&gt;UPDATE&lt;/code&gt;, &lt;code&gt;RESTART&lt;/code&gt;,
-            # &lt;code&gt;IMPORT&lt;/code&gt;, &lt;code&gt;EXPORT&lt;/code&gt;, &lt;code&gt;BACKUP_VOLUME&lt;/code&gt;,
-            # &lt;code&gt;RESTORE_VOLUME&lt;/code&gt;, &lt;code&gt;CREATE_USER&lt;/code&gt;,
-            # &lt;code&gt;DELETE_USER&lt;/code&gt;, &lt;code&gt;CREATE_DATABASE&lt;/code&gt;,
-            # &lt;code&gt;DELETE_DATABASE&lt;/code&gt; .
-        &quot;error&quot;: { # Database instance operation errors list wrapper. # If errors occurred during processing of this operation, this field will be
-            # populated.
+      { # An Operation resource. For successful operations that return an Operation resource, only the fields relevant to the operation are populated in the resource.
+        &quot;kind&quot;: &quot;A String&quot;, # This is always *sql#operation*.
+        &quot;name&quot;: &quot;A String&quot;, # An identifier that uniquely identifies the operation. You can use this identifier to retrieve the Operations resource that has information about the operation.
+        &quot;endTime&quot;: &quot;A String&quot;, # The time this operation finished in UTC timezone in RFC 3339 format, for example *2012-11-15T16:19:00.094Z*.
+        &quot;error&quot;: { # Database instance operation errors list wrapper. # If errors occurred during processing of this operation, this field will be populated.
           &quot;errors&quot;: [ # The list of errors encountered while processing this operation.
             { # Database instance operation error.
-              &quot;kind&quot;: &quot;A String&quot;, # This is always &lt;code&gt;sql#operationError&lt;/code&gt;.
               &quot;code&quot;: &quot;A String&quot;, # Identifies the specific error that occurred.
+              &quot;kind&quot;: &quot;A String&quot;, # This is always *sql#operationError*.
               &quot;message&quot;: &quot;A String&quot;, # Additional information about the error encountered.
             },
           ],
-          &quot;kind&quot;: &quot;A String&quot;, # This is always &lt;code&gt;sql#operationErrors&lt;/code&gt;.
+          &quot;kind&quot;: &quot;A String&quot;, # This is always *sql#operationErrors*.
         },
-        &quot;kind&quot;: &quot;A String&quot;, # This is always &lt;code&gt;sql#operation&lt;/code&gt;.
-        &quot;importContext&quot;: { # Database instance import context. # The context for import operation, if applicable.
-          &quot;database&quot;: &quot;A String&quot;, # The target database for the import. If &lt;code&gt;fileType&lt;/code&gt; is
-              # &lt;code&gt;SQL&lt;/code&gt;, this field is required only if the import file does not
-              # specify a database, and is overridden by any database specification in the
-              # import file. If &lt;code&gt;fileType&lt;/code&gt; is &lt;code&gt;CSV&lt;/code&gt;, one database
-              # must be specified.
-          &quot;importUser&quot;: &quot;A String&quot;, # The PostgreSQL user for this import operation. PostgreSQL instances only.
-          &quot;bakImportOptions&quot;: { # Import parameters specific to SQL Server .BAK files
-            &quot;encryptionOptions&quot;: {
-              &quot;pvkPassword&quot;: &quot;A String&quot;, # Password that encrypts the private key
-              &quot;certPath&quot;: &quot;A String&quot;, # Path to the Certificate (.cer) in Cloud Storage, in the form
-                  # &lt;code&gt;gs://bucketName/fileName&lt;/code&gt;. The instance must have
-                  # write permissions to the bucket and read access to the file.
-              &quot;pvkPath&quot;: &quot;A String&quot;, # Path to the Certificate Private Key (.pvk)  in Cloud Storage, in the
-                  # form &lt;code&gt;gs://bucketName/fileName&lt;/code&gt;. The instance must have
-                  # write permissions to the bucket and read access to the file.
-            },
-          },
-          &quot;uri&quot;: &quot;A String&quot;, # Path to the import file in Cloud Storage, in the form
-              # &lt;code&gt;gs:
-              # //bucketName/fileName&lt;/code&gt;. Compressed gzip files (.gz) are supported
-              # // when &lt;code&gt;fileType&lt;/code&gt; is &lt;code&gt;SQL&lt;/code&gt;. The instance must have
-              # // write permissions to the bucket and read access to the file.
-          &quot;fileType&quot;: &quot;A String&quot;, # The file type for the specified uri. &lt;br&gt;&lt;code&gt;SQL&lt;/code&gt;: The file
-              # contains SQL statements. &lt;br&gt;&lt;code&gt;CSV&lt;/code&gt;: The file contains CSV data.
-          &quot;kind&quot;: &quot;A String&quot;, # This is always &lt;code&gt;sql#importContext&lt;/code&gt;.
-          &quot;csvImportOptions&quot;: { # Options for importing data as CSV.
-            &quot;table&quot;: &quot;A String&quot;, # The table to which CSV data is imported.
-            &quot;columns&quot;: [ # The columns to which CSV data is imported. If not specified, all columns
-                # of the database table are loaded with CSV data.
-              &quot;A String&quot;,
-            ],
-          },
-        },
-        &quot;status&quot;: &quot;A String&quot;, # The status of an operation. Valid values are &lt;code&gt;PENDING&lt;/code&gt;,
-            # &lt;code&gt;RUNNING&lt;/code&gt;, &lt;code&gt;DONE&lt;/code&gt;,
-            # &lt;code&gt;SQL_OPERATION_STATUS_UNSPECIFIED&lt;/code&gt;.
-        &quot;name&quot;: &quot;A String&quot;, # An identifier that uniquely identifies the operation. You can use this
-            # identifier to retrieve the Operations resource that has information about
-            # the operation.
+        &quot;status&quot;: &quot;A String&quot;, # The status of an operation. Valid values are: *PENDING* *RUNNING* *DONE* *SQL_OPERATION_STATUS_UNSPECIFIED*
+        &quot;selfLink&quot;: &quot;A String&quot;, # The URI of this resource.
+        &quot;startTime&quot;: &quot;A String&quot;, # The time this operation actually started in UTC timezone in RFC 3339 format, for example *2012-11-15T16:19:00.094Z*.
         &quot;exportContext&quot;: { # Database instance export context. # The context for export operation, if applicable.
-          &quot;fileType&quot;: &quot;A String&quot;, # The file type for the specified uri. &lt;br&gt;&lt;code&gt;SQL&lt;/code&gt;: The file
-              # contains SQL statements. &lt;br&gt;&lt;code&gt;CSV&lt;/code&gt;: The file contains CSV data.
-          &quot;uri&quot;: &quot;A String&quot;, # The path to the file in Google Cloud Storage where the export will be
-              # stored. The URI is in the form &lt;code&gt;gs:
-              # //bucketName/fileName&lt;/code&gt;. If the file already exists, the requests
-              # // succeeds, but the operation fails. If &lt;code&gt;fileType&lt;/code&gt; is
-              # // &lt;code&gt;SQL&lt;/code&gt; and the filename ends with .gz, the contents are
-              # // compressed.
           &quot;csvExportOptions&quot;: { # Options for exporting data as CSV.
             &quot;selectQuery&quot;: &quot;A String&quot;, # The select query used to extract the data.
           },
-          &quot;kind&quot;: &quot;A String&quot;, # This is always &lt;code&gt;sql#exportContext&lt;/code&gt;.
-          &quot;databases&quot;: [ # Databases to be exported. &lt;br /&gt; &lt;b&gt;MySQL instances:&lt;/b&gt; If
-              # &lt;code&gt;fileType&lt;/code&gt; is &lt;code&gt;SQL&lt;/code&gt; and no database is specified, all
-              # databases are exported, except for the &lt;code&gt;mysql&lt;/code&gt; system database.
-              # If &lt;code&gt;fileType&lt;/code&gt; is &lt;code&gt;CSV&lt;/code&gt;, you can specify one database,
-              # either by using this property or by using the
-              # &lt;code&gt;csvExportOptions.selectQuery&lt;/code&gt; property, which takes precedence
-              # over this property. &lt;br /&gt; &lt;b&gt;PostgreSQL instances:&lt;/b&gt; You must specify
-              # one database to be exported. If &lt;code&gt;fileType&lt;/code&gt; is &lt;code&gt;CSV&lt;/code&gt;,
-              # this database must match the one specified in the
-              # &lt;code&gt;csvExportOptions.selectQuery&lt;/code&gt; property.
+          &quot;fileType&quot;: &quot;A String&quot;, # The file type for the specified uri. *SQL*: The file contains SQL statements. *CSV*: The file contains CSV data.
+          &quot;kind&quot;: &quot;A String&quot;, # This is always *sql#exportContext*.
+          &quot;uri&quot;: &quot;A String&quot;, # The path to the file in Google Cloud Storage where the export will be stored. The URI is in the form *gs: //bucketName/fileName*. If the file already exists, the requests // succeeds, but the operation fails. If *fileType* is // *SQL* and the filename ends with .gz, the contents are // compressed.
+          &quot;databases&quot;: [ # Databases to be exported. *MySQL instances:* If *fileType* is *SQL* and no database is specified, all databases are exported, except for the *mysql* system database. If *fileType* is *CSV*, you can specify one database, either by using this property or by using the *csvExportOptions.selectQuery* property, which takes precedence over this property. *PostgreSQL instances:* You must specify one database to be exported. If *fileType* is *CSV*, this database must match the one specified in the *csvExportOptions.selectQuery* property.
             &quot;A String&quot;,
           ],
+          &quot;offload&quot;: True or False, # Option for export offload.
           &quot;sqlExportOptions&quot;: { # Options for exporting data as SQL statements.
-            &quot;tables&quot;: [ # Tables to export, or that were exported, from the specified database. If
-                # you specify tables, specify one and only one database. For PostgreSQL
-                # instances, you can specify only one table.
+            &quot;schemaOnly&quot;: True or False, # Export only schemas.
+            &quot;tables&quot;: [ # Tables to export, or that were exported, from the specified database. If you specify tables, specify one and only one database. For PostgreSQL instances, you can specify only one table.
               &quot;A String&quot;,
             ],
-            &quot;schemaOnly&quot;: True or False, # Export only schemas.
             &quot;mysqlExportOptions&quot;: { # Options for exporting from MySQL.
-              &quot;masterData&quot;: 42, # Option to include SQL statement required to set up replication.
-                  # If set to &lt;code&gt;1&lt;/code&gt;, the dump file includes
-                  #  a CHANGE MASTER TO statement with the binary log coordinates.
-                  # If set to &lt;code&gt;2&lt;/code&gt;, the CHANGE MASTER TO statement is written as
-                  #  a SQL comment, and has no effect.
-                  # All other values are ignored.
+              &quot;masterData&quot;: 42, # Option to include SQL statement required to set up replication. If set to *1*, the dump file includes a CHANGE MASTER TO statement with the binary log coordinates. If set to *2*, the CHANGE MASTER TO statement is written as a SQL comment, and has no effect. All other values are ignored.
             },
           },
         },
-        &quot;targetId&quot;: &quot;A String&quot;, # Name of the database instance related to this operation.
-        &quot;endTime&quot;: &quot;A String&quot;, # The time this operation finished in UTC timezone in &lt;a
-            # href=&quot;https://tools.ietf.org/html/rfc3339&quot;&gt;RFC 3339&lt;/a&gt; format, for example
-            # &lt;code&gt;2012-11-15T16:19:00.094Z&lt;/code&gt;.
-        &quot;startTime&quot;: &quot;A String&quot;, # The time this operation actually started in UTC timezone in &lt;a
-            # href=&quot;https://tools.ietf.org/html/rfc3339&quot;&gt;RFC 3339&lt;/a&gt; format, for example
-            # &lt;code&gt;2012-11-15T16:19:00.094Z&lt;/code&gt;.
-        &quot;insertTime&quot;: &quot;A String&quot;, # The time this operation was enqueued in UTC timezone in &lt;a
-            # href=&quot;https://tools.ietf.org/html/rfc3339&quot;&gt;RFC 3339&lt;/a&gt; format, for example
-            # &lt;code&gt;2012-11-15T16:19:00.094Z&lt;/code&gt;.
         &quot;user&quot;: &quot;A String&quot;, # The email address of the user who initiated this operation.
+        &quot;targetId&quot;: &quot;A String&quot;, # Name of the database instance related to this operation.
+        &quot;insertTime&quot;: &quot;A String&quot;, # The time this operation was enqueued in UTC timezone in RFC 3339 format, for example *2012-11-15T16:19:00.094Z*.
+        &quot;operationType&quot;: &quot;A String&quot;, # The type of the operation. Valid values are: *CREATE* *DELETE* *UPDATE* *RESTART* *IMPORT* *EXPORT* *BACKUP_VOLUME* *RESTORE_VOLUME* *CREATE_USER* *DELETE_USER* *CREATE_DATABASE* *DELETE_DATABASE*
         &quot;targetProject&quot;: &quot;A String&quot;, # The project ID of the target instance related to this operation.
-        &quot;selfLink&quot;: &quot;A String&quot;, # The URI of this resource.
+        &quot;targetLink&quot;: &quot;A String&quot;,
+        &quot;importContext&quot;: { # Database instance import context. # The context for import operation, if applicable.
+          &quot;fileType&quot;: &quot;A String&quot;, # The file type for the specified uri. *SQL*: The file contains SQL statements. *CSV*: The file contains CSV data.
+          &quot;kind&quot;: &quot;A String&quot;, # This is always *sql#importContext*.
+          &quot;csvImportOptions&quot;: { # Options for importing data as CSV.
+            &quot;columns&quot;: [ # The columns to which CSV data is imported. If not specified, all columns of the database table are loaded with CSV data.
+              &quot;A String&quot;,
+            ],
+            &quot;table&quot;: &quot;A String&quot;, # The table to which CSV data is imported.
+          },
+          &quot;bakImportOptions&quot;: { # Import parameters specific to SQL Server .BAK files
+            &quot;encryptionOptions&quot;: {
+              &quot;pvkPath&quot;: &quot;A String&quot;, # Path to the Certificate Private Key (.pvk) in Cloud Storage, in the form *gs://bucketName/fileName*. The instance must have write permissions to the bucket and read access to the file.
+              &quot;certPath&quot;: &quot;A String&quot;, # Path to the Certificate (.cer) in Cloud Storage, in the form *gs://bucketName/fileName*. The instance must have write permissions to the bucket and read access to the file.
+              &quot;pvkPassword&quot;: &quot;A String&quot;, # Password that encrypts the private key
+            },
+          },
+          &quot;database&quot;: &quot;A String&quot;, # The target database for the import. If *fileType* is *SQL*, this field is required only if the import file does not specify a database, and is overridden by any database specification in the import file. If *fileType* is *CSV*, one database must be specified.
+          &quot;uri&quot;: &quot;A String&quot;, # Path to the import file in Cloud Storage, in the form *gs: //bucketName/fileName*. Compressed gzip files (.gz) are supported // when *fileType* is *SQL*. The instance must have // write permissions to the bucket and read access to the file.
+          &quot;importUser&quot;: &quot;A String&quot;, # The PostgreSQL user for this import operation. PostgreSQL instances only.
+        },
       },
     ],
   }</pre>