Update docs.
diff --git a/docs/dyn/bigquery_v2.jobs.html b/docs/dyn/bigquery_v2.jobs.html
index c6109f4..562a7e0 100644
--- a/docs/dyn/bigquery_v2.jobs.html
+++ b/docs/dyn/bigquery_v2.jobs.html
@@ -195,6 +195,7 @@
           ],
         },
         "query": { # [Pick one] Configures a query job.
+          "flattenResults": True or False, # [Experimental] Flattens all nested and repeated fields in the query results. The default value is true. allowLargeResults must be true if this is set to false.
           "useQueryCache": True or False, # [Optional] Whether to look for the result in the query cache. The query cache is a best-effort cache that will be flushed whenever tables in the query are modified. Moreover, the query cache is only available when a query does not have a destination table specified.
           "defaultDataset": { # [Optional] Specifies the default dataset to use for unqualified table names in the query.
             "projectId": "A String", # [Optional] The ID of the container project.
@@ -207,20 +208,27 @@
           },
           "priority": "A String", # [Optional] Specifies a priority for the query. Possible values include INTERACTIVE and BATCH. The default value is INTERACTIVE.
           "writeDisposition": "A String", # [Optional] Specifies the action that occurs if the destination table already exists. The following values are supported: WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data. WRITE_APPEND: If the table already exists, BigQuery appends the data to the table. WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result. The default value is WRITE_EMPTY. Each action is atomic and only occurs if BigQuery is able to complete the job successfully. Creation, truncation and append actions occur as one atomic update upon job completion.
-          "allowLargeResults": True or False, # If true, allows the query to produce arbitrarily large result tables at a slight cost in performance. Requires destination_table to be set.
+          "allowLargeResults": True or False, # If true, allows the query to produce arbitrarily large result tables at a slight cost in performance. Requires destinationTable to be set.
           "createDisposition": "A String", # [Optional] Specifies whether the job is allowed to create new tables. The following values are supported: CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table. CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result. The default value is CREATE_IF_NEEDED. Creation, truncation and append actions occur as one atomic update upon job completion.
           "query": "A String", # [Required] BigQuery SQL query to execute.
           "preserveNulls": True or False, # [Deprecated] This property is deprecated.
         },
         "copy": { # [Pick one] Copies a table.
           "createDisposition": "A String", # [Optional] Specifies whether the job is allowed to create new tables. The following values are supported: CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table. CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result. The default value is CREATE_IF_NEEDED. Creation, truncation and append actions occur as one atomic update upon job completion.
+          "sourceTables": [ # [Pick one] Source tables to copy.
+            {
+              "projectId": "A String", # [Required] ID of the project billed for storage of the table.
+              "tableId": "A String", # [Required] ID of the table.
+              "datasetId": "A String", # [Required] ID of the dataset containing the table.
+            },
+          ],
           "writeDisposition": "A String", # [Optional] Specifies the action that occurs if the destination table already exists. The following values are supported: WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data. WRITE_APPEND: If the table already exists, BigQuery appends the data to the table. WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result. The default value is WRITE_EMPTY. Each action is atomic and only occurs if BigQuery is able to complete the job successfully. Creation, truncation and append actions occur as one atomic update upon job completion.
           "destinationTable": { # [Required] The destination table
             "projectId": "A String", # [Required] ID of the project billed for storage of the table.
             "tableId": "A String", # [Required] ID of the table.
             "datasetId": "A String", # [Required] ID of the dataset containing the table.
           },
-          "sourceTable": { # [Required] Source table to copy.
+          "sourceTable": { # [Pick one] Source table to copy.
             "projectId": "A String", # [Required] ID of the project billed for storage of the table.
             "tableId": "A String", # [Required] ID of the table.
             "datasetId": "A String", # [Required] ID of the dataset containing the table.
@@ -228,8 +236,9 @@
         },
         "extract": { # [Pick one] Configures an extract job.
           "destinationUri": "A String", # [Pick one] DEPRECATED: Use destinationUris instead, passing only one URI as necessary. The fully-qualified Google Cloud Storage URI where the extracted table should be written.
+          "compression": "A String", # [Optional] The compression type to use for exported files. Possible values include GZIP and NONE. The default value is NONE.
           "fieldDelimiter": "A String", # [Optional] Delimiter to use between fields in the exported data. Default is ','
-          "destinationFormat": "A String", # [Experimental] Optional and defaults to CSV. Format with which files should be exported. To export to CSV, specify "CSV". Tables with nested or repeated fields cannot be exported as CSV. To export to newline-delimited JSON, specify "NEWLINE_DELIMITED_JSON".
+          "destinationFormat": "A String", # [Optional] The exported file format. Possible values include CSV, NEWLINE_DELIMITED_JSON and AVRO. The default value is CSV. Tables with nested or repeated fields cannot be exported as CSV.
           "printHeader": True or False, # [Optional] Whether to print out a header row in the results. Default is true.
           "destinationUris": [ # [Pick one] A list of fully-qualified Google Cloud Storage URIs where the extracted table should be written.
             "A String",
@@ -397,6 +406,7 @@
         ],
       },
       "query": { # [Pick one] Configures a query job.
+        "flattenResults": True or False, # [Experimental] Flattens all nested and repeated fields in the query results. The default value is true. allowLargeResults must be true if this is set to false.
         "useQueryCache": True or False, # [Optional] Whether to look for the result in the query cache. The query cache is a best-effort cache that will be flushed whenever tables in the query are modified. Moreover, the query cache is only available when a query does not have a destination table specified.
         "defaultDataset": { # [Optional] Specifies the default dataset to use for unqualified table names in the query.
           "projectId": "A String", # [Optional] The ID of the container project.
@@ -409,20 +419,27 @@
         },
         "priority": "A String", # [Optional] Specifies a priority for the query. Possible values include INTERACTIVE and BATCH. The default value is INTERACTIVE.
         "writeDisposition": "A String", # [Optional] Specifies the action that occurs if the destination table already exists. The following values are supported: WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data. WRITE_APPEND: If the table already exists, BigQuery appends the data to the table. WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result. The default value is WRITE_EMPTY. Each action is atomic and only occurs if BigQuery is able to complete the job successfully. Creation, truncation and append actions occur as one atomic update upon job completion.
-        "allowLargeResults": True or False, # If true, allows the query to produce arbitrarily large result tables at a slight cost in performance. Requires destination_table to be set.
+        "allowLargeResults": True or False, # If true, allows the query to produce arbitrarily large result tables at a slight cost in performance. Requires destinationTable to be set.
         "createDisposition": "A String", # [Optional] Specifies whether the job is allowed to create new tables. The following values are supported: CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table. CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result. The default value is CREATE_IF_NEEDED. Creation, truncation and append actions occur as one atomic update upon job completion.
         "query": "A String", # [Required] BigQuery SQL query to execute.
         "preserveNulls": True or False, # [Deprecated] This property is deprecated.
       },
       "copy": { # [Pick one] Copies a table.
         "createDisposition": "A String", # [Optional] Specifies whether the job is allowed to create new tables. The following values are supported: CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table. CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result. The default value is CREATE_IF_NEEDED. Creation, truncation and append actions occur as one atomic update upon job completion.
+        "sourceTables": [ # [Pick one] Source tables to copy.
+          {
+            "projectId": "A String", # [Required] ID of the project billed for storage of the table.
+            "tableId": "A String", # [Required] ID of the table.
+            "datasetId": "A String", # [Required] ID of the dataset containing the table.
+          },
+        ],
         "writeDisposition": "A String", # [Optional] Specifies the action that occurs if the destination table already exists. The following values are supported: WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data. WRITE_APPEND: If the table already exists, BigQuery appends the data to the table. WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result. The default value is WRITE_EMPTY. Each action is atomic and only occurs if BigQuery is able to complete the job successfully. Creation, truncation and append actions occur as one atomic update upon job completion.
         "destinationTable": { # [Required] The destination table
           "projectId": "A String", # [Required] ID of the project billed for storage of the table.
           "tableId": "A String", # [Required] ID of the table.
           "datasetId": "A String", # [Required] ID of the dataset containing the table.
         },
-        "sourceTable": { # [Required] Source table to copy.
+        "sourceTable": { # [Pick one] Source table to copy.
           "projectId": "A String", # [Required] ID of the project billed for storage of the table.
           "tableId": "A String", # [Required] ID of the table.
           "datasetId": "A String", # [Required] ID of the dataset containing the table.
@@ -430,8 +447,9 @@
       },
       "extract": { # [Pick one] Configures an extract job.
         "destinationUri": "A String", # [Pick one] DEPRECATED: Use destinationUris instead, passing only one URI as necessary. The fully-qualified Google Cloud Storage URI where the extracted table should be written.
+        "compression": "A String", # [Optional] The compression type to use for exported files. Possible values include GZIP and NONE. The default value is NONE.
         "fieldDelimiter": "A String", # [Optional] Delimiter to use between fields in the exported data. Default is ','
-        "destinationFormat": "A String", # [Experimental] Optional and defaults to CSV. Format with which files should be exported. To export to CSV, specify "CSV". Tables with nested or repeated fields cannot be exported as CSV. To export to newline-delimited JSON, specify "NEWLINE_DELIMITED_JSON".
+        "destinationFormat": "A String", # [Optional] The exported file format. Possible values include CSV, NEWLINE_DELIMITED_JSON and AVRO. The default value is CSV. Tables with nested or repeated fields cannot be exported as CSV.
         "printHeader": True or False, # [Optional] Whether to print out a header row in the results. Default is true.
         "destinationUris": [ # [Pick one] A list of fully-qualified Google Cloud Storage URIs where the extracted table should be written.
           "A String",
@@ -543,6 +561,7 @@
           ],
         },
         "query": { # [Pick one] Configures a query job.
+          "flattenResults": True or False, # [Experimental] Flattens all nested and repeated fields in the query results. The default value is true. allowLargeResults must be true if this is set to false.
           "useQueryCache": True or False, # [Optional] Whether to look for the result in the query cache. The query cache is a best-effort cache that will be flushed whenever tables in the query are modified. Moreover, the query cache is only available when a query does not have a destination table specified.
           "defaultDataset": { # [Optional] Specifies the default dataset to use for unqualified table names in the query.
             "projectId": "A String", # [Optional] The ID of the container project.
@@ -555,20 +574,27 @@
           },
           "priority": "A String", # [Optional] Specifies a priority for the query. Possible values include INTERACTIVE and BATCH. The default value is INTERACTIVE.
           "writeDisposition": "A String", # [Optional] Specifies the action that occurs if the destination table already exists. The following values are supported: WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data. WRITE_APPEND: If the table already exists, BigQuery appends the data to the table. WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result. The default value is WRITE_EMPTY. Each action is atomic and only occurs if BigQuery is able to complete the job successfully. Creation, truncation and append actions occur as one atomic update upon job completion.
-          "allowLargeResults": True or False, # If true, allows the query to produce arbitrarily large result tables at a slight cost in performance. Requires destination_table to be set.
+          "allowLargeResults": True or False, # If true, allows the query to produce arbitrarily large result tables at a slight cost in performance. Requires destinationTable to be set.
           "createDisposition": "A String", # [Optional] Specifies whether the job is allowed to create new tables. The following values are supported: CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table. CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result. The default value is CREATE_IF_NEEDED. Creation, truncation and append actions occur as one atomic update upon job completion.
           "query": "A String", # [Required] BigQuery SQL query to execute.
           "preserveNulls": True or False, # [Deprecated] This property is deprecated.
         },
         "copy": { # [Pick one] Copies a table.
           "createDisposition": "A String", # [Optional] Specifies whether the job is allowed to create new tables. The following values are supported: CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table. CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result. The default value is CREATE_IF_NEEDED. Creation, truncation and append actions occur as one atomic update upon job completion.
+          "sourceTables": [ # [Pick one] Source tables to copy.
+            {
+              "projectId": "A String", # [Required] ID of the project billed for storage of the table.
+              "tableId": "A String", # [Required] ID of the table.
+              "datasetId": "A String", # [Required] ID of the dataset containing the table.
+            },
+          ],
           "writeDisposition": "A String", # [Optional] Specifies the action that occurs if the destination table already exists. The following values are supported: WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data. WRITE_APPEND: If the table already exists, BigQuery appends the data to the table. WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result. The default value is WRITE_EMPTY. Each action is atomic and only occurs if BigQuery is able to complete the job successfully. Creation, truncation and append actions occur as one atomic update upon job completion.
           "destinationTable": { # [Required] The destination table
             "projectId": "A String", # [Required] ID of the project billed for storage of the table.
             "tableId": "A String", # [Required] ID of the table.
             "datasetId": "A String", # [Required] ID of the dataset containing the table.
           },
-          "sourceTable": { # [Required] Source table to copy.
+          "sourceTable": { # [Pick one] Source table to copy.
             "projectId": "A String", # [Required] ID of the project billed for storage of the table.
             "tableId": "A String", # [Required] ID of the table.
             "datasetId": "A String", # [Required] ID of the dataset containing the table.
@@ -576,8 +602,9 @@
         },
         "extract": { # [Pick one] Configures an extract job.
           "destinationUri": "A String", # [Pick one] DEPRECATED: Use destinationUris instead, passing only one URI as necessary. The fully-qualified Google Cloud Storage URI where the extracted table should be written.
+          "compression": "A String", # [Optional] The compression type to use for exported files. Possible values include GZIP and NONE. The default value is NONE.
           "fieldDelimiter": "A String", # [Optional] Delimiter to use between fields in the exported data. Default is ','
-          "destinationFormat": "A String", # [Experimental] Optional and defaults to CSV. Format with which files should be exported. To export to CSV, specify "CSV". Tables with nested or repeated fields cannot be exported as CSV. To export to newline-delimited JSON, specify "NEWLINE_DELIMITED_JSON".
+          "destinationFormat": "A String", # [Optional] The exported file format. Possible values include CSV, NEWLINE_DELIMITED_JSON and AVRO. The default value is CSV. Tables with nested or repeated fields cannot be exported as CSV.
           "printHeader": True or False, # [Optional] Whether to print out a header row in the results. Default is true.
           "destinationUris": [ # [Pick one] A list of fully-qualified Google Cloud Storage URIs where the extracted table should be written.
             "A String",
@@ -714,6 +741,7 @@
             ],
           },
           "query": { # [Pick one] Configures a query job.
+            "flattenResults": True or False, # [Experimental] Flattens all nested and repeated fields in the query results. The default value is true. allowLargeResults must be true if this is set to false.
             "useQueryCache": True or False, # [Optional] Whether to look for the result in the query cache. The query cache is a best-effort cache that will be flushed whenever tables in the query are modified. Moreover, the query cache is only available when a query does not have a destination table specified.
             "defaultDataset": { # [Optional] Specifies the default dataset to use for unqualified table names in the query.
               "projectId": "A String", # [Optional] The ID of the container project.
@@ -726,20 +754,27 @@
             },
             "priority": "A String", # [Optional] Specifies a priority for the query. Possible values include INTERACTIVE and BATCH. The default value is INTERACTIVE.
             "writeDisposition": "A String", # [Optional] Specifies the action that occurs if the destination table already exists. The following values are supported: WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data. WRITE_APPEND: If the table already exists, BigQuery appends the data to the table. WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result. The default value is WRITE_EMPTY. Each action is atomic and only occurs if BigQuery is able to complete the job successfully. Creation, truncation and append actions occur as one atomic update upon job completion.
-            "allowLargeResults": True or False, # If true, allows the query to produce arbitrarily large result tables at a slight cost in performance. Requires destination_table to be set.
+            "allowLargeResults": True or False, # If true, allows the query to produce arbitrarily large result tables at a slight cost in performance. Requires destinationTable to be set.
             "createDisposition": "A String", # [Optional] Specifies whether the job is allowed to create new tables. The following values are supported: CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table. CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result. The default value is CREATE_IF_NEEDED. Creation, truncation and append actions occur as one atomic update upon job completion.
             "query": "A String", # [Required] BigQuery SQL query to execute.
             "preserveNulls": True or False, # [Deprecated] This property is deprecated.
           },
           "copy": { # [Pick one] Copies a table.
             "createDisposition": "A String", # [Optional] Specifies whether the job is allowed to create new tables. The following values are supported: CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table. CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result. The default value is CREATE_IF_NEEDED. Creation, truncation and append actions occur as one atomic update upon job completion.
+            "sourceTables": [ # [Pick one] Source tables to copy.
+              {
+                "projectId": "A String", # [Required] ID of the project billed for storage of the table.
+                "tableId": "A String", # [Required] ID of the table.
+                "datasetId": "A String", # [Required] ID of the dataset containing the table.
+              },
+            ],
             "writeDisposition": "A String", # [Optional] Specifies the action that occurs if the destination table already exists. The following values are supported: WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data. WRITE_APPEND: If the table already exists, BigQuery appends the data to the table. WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result. The default value is WRITE_EMPTY. Each action is atomic and only occurs if BigQuery is able to complete the job successfully. Creation, truncation and append actions occur as one atomic update upon job completion.
             "destinationTable": { # [Required] The destination table
               "projectId": "A String", # [Required] ID of the project billed for storage of the table.
               "tableId": "A String", # [Required] ID of the table.
               "datasetId": "A String", # [Required] ID of the dataset containing the table.
             },
-            "sourceTable": { # [Required] Source table to copy.
+            "sourceTable": { # [Pick one] Source table to copy.
               "projectId": "A String", # [Required] ID of the project billed for storage of the table.
               "tableId": "A String", # [Required] ID of the table.
               "datasetId": "A String", # [Required] ID of the dataset containing the table.
@@ -747,8 +782,9 @@
           },
           "extract": { # [Pick one] Configures an extract job.
             "destinationUri": "A String", # [Pick one] DEPRECATED: Use destinationUris instead, passing only one URI as necessary. The fully-qualified Google Cloud Storage URI where the extracted table should be written.
+            "compression": "A String", # [Optional] The compression type to use for exported files. Possible values include GZIP and NONE. The default value is NONE.
             "fieldDelimiter": "A String", # [Optional] Delimiter to use between fields in the exported data. Default is ','
-            "destinationFormat": "A String", # [Experimental] Optional and defaults to CSV. Format with which files should be exported. To export to CSV, specify "CSV". Tables with nested or repeated fields cannot be exported as CSV. To export to newline-delimited JSON, specify "NEWLINE_DELIMITED_JSON".
+            "destinationFormat": "A String", # [Optional] The exported file format. Possible values include CSV, NEWLINE_DELIMITED_JSON and AVRO. The default value is CSV. Tables with nested or repeated fields cannot be exported as CSV.
             "printHeader": True or False, # [Optional] Whether to print out a header row in the results. Default is true.
             "destinationUris": [ # [Pick one] A list of fully-qualified Google Cloud Storage URIs where the extracted table should be written.
               "A String",