docs: update docs (#916)
* fix: re-run script
* test: fix noxfile
diff --git a/docs/dyn/bigquerydatatransfer_v1.projects.locations.transferConfigs.html b/docs/dyn/bigquerydatatransfer_v1.projects.locations.transferConfigs.html
index f0a24fb..4acabdf 100644
--- a/docs/dyn/bigquerydatatransfer_v1.projects.locations.transferConfigs.html
+++ b/docs/dyn/bigquerydatatransfer_v1.projects.locations.transferConfigs.html
@@ -89,7 +89,7 @@
<code><a href="#get">get(name, x__xgafv=None)</a></code></p>
<p class="firstline">Returns information about a data transfer config.</p>
<p class="toc_element">
- <code><a href="#list">list(parent, pageToken=None, pageSize=None, dataSourceIds=None, x__xgafv=None)</a></code></p>
+ <code><a href="#list">list(parent, dataSourceIds=None, pageToken=None, pageSize=None, x__xgafv=None)</a></code></p>
<p class="firstline">Returns information about all data transfers in the project.</p>
<p class="toc_element">
<code><a href="#list_next">list_next(previous_request, previous_response)</a></code></p>
@@ -122,6 +122,38 @@
# When a new transfer configuration is created, the specified
# `destination_dataset_id` is created when needed and shared with the
# appropriate data source service account.
+ "dataRefreshWindowDays": 42, # The number of days to look back to automatically refresh the data.
+ # For example, if `data_refresh_window_days = 10`, then every day
+ # BigQuery reingests data for [today-10, today-1], rather than ingesting data
+ # for just [today-1].
+ # Only valid if the data source supports the feature. Set the value to 0
+ # to use the default value.
+ "dataSourceId": "A String", # Data source id. Cannot be changed once data transfer is created.
+ "scheduleOptions": { # Options customizing the data transfer schedule. # Options customizing the data transfer schedule.
+ "endTime": "A String", # Defines time to stop scheduling transfer runs. A transfer run cannot be
+ # scheduled at or after the end time. The end time can be changed at any
+ # moment. The time when a data transfer can be trigerred manually is not
+ # limited by this option.
+ "startTime": "A String", # Specifies time to start scheduling transfer runs. The first run will be
+ # scheduled at or after the start time according to a recurrence pattern
+ # defined in the schedule string. The start time can be changed at any
+ # moment. The time when a data transfer can be trigerred manually is not
+ # limited by this option.
+ "disableAutoScheduling": True or False, # If true, automatic scheduling of data transfer runs for this configuration
+ # will be disabled. The runs can be started on ad-hoc basis using
+ # StartManualTransferRuns API. When automatic scheduling is disabled, the
+ # TransferConfig.schedule field will be ignored.
+ },
+ "state": "A String", # Output only. State of the most recently updated transfer run.
+ "name": "A String", # The resource name of the transfer config.
+ # Transfer config names have the form of
+ # `projects/{project_id}/locations/{region}/transferConfigs/{config_id}`.
+ # The name is automatically generated based on the config_id specified in
+ # CreateTransferConfigRequest along with project_id and region. If config_id
+ # is not provided, usually a uuid, even though it is not guaranteed or
+ # required, will be generated for config_id.
+ "destinationDatasetId": "A String", # The BigQuery target dataset id.
+ "userId": "A String", # Deprecated. Unique ID of the user on whose behalf transfer is done.
"notificationPubsubTopic": "A String", # Pub/Sub topic where notifications will be sent after transfer runs
# associated with this transfer config finish.
"params": { # Data transfer specific parameters.
@@ -134,9 +166,9 @@
},
"datasetRegion": "A String", # Output only. Region in which BigQuery dataset is located.
"displayName": "A String", # User specified display name for the data transfer.
+ "nextRunTime": "A String", # Output only. Next time when data transfer will run.
"disabled": True or False, # Is this config disabled. When set to true, no runs are scheduled
# for a given transfer.
- "nextRunTime": "A String", # Output only. Next time when data transfer will run.
"schedule": "A String", # Data transfer schedule.
# If the data source does not support a custom schedule, this should be
# empty. If it is empty, the default value for the data source will be
@@ -150,38 +182,6 @@
# https://cloud.google.com/appengine/docs/flexible/python/scheduling-jobs-with-cron-yaml#the_schedule_format
# NOTE: the granularity should be at least 8 hours, or less frequent.
"updateTime": "A String", # Output only. Data transfer modification time. Ignored by server on input.
- "dataRefreshWindowDays": 42, # The number of days to look back to automatically refresh the data.
- # For example, if `data_refresh_window_days = 10`, then every day
- # BigQuery reingests data for [today-10, today-1], rather than ingesting data
- # for just [today-1].
- # Only valid if the data source supports the feature. Set the value to 0
- # to use the default value.
- "dataSourceId": "A String", # Data source id. Cannot be changed once data transfer is created.
- "scheduleOptions": { # Options customizing the data transfer schedule. # Options customizing the data transfer schedule.
- "disableAutoScheduling": True or False, # If true, automatic scheduling of data transfer runs for this configuration
- # will be disabled. The runs can be started on ad-hoc basis using
- # StartManualTransferRuns API. When automatic scheduling is disabled, the
- # TransferConfig.schedule field will be ignored.
- "endTime": "A String", # Defines time to stop scheduling transfer runs. A transfer run cannot be
- # scheduled at or after the end time. The end time can be changed at any
- # moment. The time when a data transfer can be trigerred manually is not
- # limited by this option.
- "startTime": "A String", # Specifies time to start scheduling transfer runs. The first run will be
- # scheduled at or after the start time according to a recurrence pattern
- # defined in the schedule string. The start time can be changed at any
- # moment. The time when a data transfer can be trigerred manually is not
- # limited by this option.
- },
- "state": "A String", # Output only. State of the most recently updated transfer run.
- "name": "A String", # The resource name of the transfer config.
- # Transfer config names have the form of
- # `projects/{project_id}/locations/{region}/transferConfigs/{config_id}`.
- # The name is automatically generated based on the config_id specified in
- # CreateTransferConfigRequest along with project_id and region. If config_id
- # is not provided, usually a uuid, even though it is not guaranteed or
- # required, will be generated for config_id.
- "destinationDatasetId": "A String", # The BigQuery target dataset id.
- "userId": "A String", # Deprecated. Unique ID of the user on whose behalf transfer is done.
}
serviceAccountName: string, Optional service account name. If this field is set, transfer config will
@@ -224,6 +224,38 @@
# When a new transfer configuration is created, the specified
# `destination_dataset_id` is created when needed and shared with the
# appropriate data source service account.
+ "dataRefreshWindowDays": 42, # The number of days to look back to automatically refresh the data.
+ # For example, if `data_refresh_window_days = 10`, then every day
+ # BigQuery reingests data for [today-10, today-1], rather than ingesting data
+ # for just [today-1].
+ # Only valid if the data source supports the feature. Set the value to 0
+ # to use the default value.
+ "dataSourceId": "A String", # Data source id. Cannot be changed once data transfer is created.
+ "scheduleOptions": { # Options customizing the data transfer schedule. # Options customizing the data transfer schedule.
+ "endTime": "A String", # Defines time to stop scheduling transfer runs. A transfer run cannot be
+ # scheduled at or after the end time. The end time can be changed at any
+ # moment. The time when a data transfer can be trigerred manually is not
+ # limited by this option.
+ "startTime": "A String", # Specifies time to start scheduling transfer runs. The first run will be
+ # scheduled at or after the start time according to a recurrence pattern
+ # defined in the schedule string. The start time can be changed at any
+ # moment. The time when a data transfer can be trigerred manually is not
+ # limited by this option.
+ "disableAutoScheduling": True or False, # If true, automatic scheduling of data transfer runs for this configuration
+ # will be disabled. The runs can be started on ad-hoc basis using
+ # StartManualTransferRuns API. When automatic scheduling is disabled, the
+ # TransferConfig.schedule field will be ignored.
+ },
+ "state": "A String", # Output only. State of the most recently updated transfer run.
+ "name": "A String", # The resource name of the transfer config.
+ # Transfer config names have the form of
+ # `projects/{project_id}/locations/{region}/transferConfigs/{config_id}`.
+ # The name is automatically generated based on the config_id specified in
+ # CreateTransferConfigRequest along with project_id and region. If config_id
+ # is not provided, usually a uuid, even though it is not guaranteed or
+ # required, will be generated for config_id.
+ "destinationDatasetId": "A String", # The BigQuery target dataset id.
+ "userId": "A String", # Deprecated. Unique ID of the user on whose behalf transfer is done.
"notificationPubsubTopic": "A String", # Pub/Sub topic where notifications will be sent after transfer runs
# associated with this transfer config finish.
"params": { # Data transfer specific parameters.
@@ -236,9 +268,9 @@
},
"datasetRegion": "A String", # Output only. Region in which BigQuery dataset is located.
"displayName": "A String", # User specified display name for the data transfer.
+ "nextRunTime": "A String", # Output only. Next time when data transfer will run.
"disabled": True or False, # Is this config disabled. When set to true, no runs are scheduled
# for a given transfer.
- "nextRunTime": "A String", # Output only. Next time when data transfer will run.
"schedule": "A String", # Data transfer schedule.
# If the data source does not support a custom schedule, this should be
# empty. If it is empty, the default value for the data source will be
@@ -252,38 +284,6 @@
# https://cloud.google.com/appengine/docs/flexible/python/scheduling-jobs-with-cron-yaml#the_schedule_format
# NOTE: the granularity should be at least 8 hours, or less frequent.
"updateTime": "A String", # Output only. Data transfer modification time. Ignored by server on input.
- "dataRefreshWindowDays": 42, # The number of days to look back to automatically refresh the data.
- # For example, if `data_refresh_window_days = 10`, then every day
- # BigQuery reingests data for [today-10, today-1], rather than ingesting data
- # for just [today-1].
- # Only valid if the data source supports the feature. Set the value to 0
- # to use the default value.
- "dataSourceId": "A String", # Data source id. Cannot be changed once data transfer is created.
- "scheduleOptions": { # Options customizing the data transfer schedule. # Options customizing the data transfer schedule.
- "disableAutoScheduling": True or False, # If true, automatic scheduling of data transfer runs for this configuration
- # will be disabled. The runs can be started on ad-hoc basis using
- # StartManualTransferRuns API. When automatic scheduling is disabled, the
- # TransferConfig.schedule field will be ignored.
- "endTime": "A String", # Defines time to stop scheduling transfer runs. A transfer run cannot be
- # scheduled at or after the end time. The end time can be changed at any
- # moment. The time when a data transfer can be trigerred manually is not
- # limited by this option.
- "startTime": "A String", # Specifies time to start scheduling transfer runs. The first run will be
- # scheduled at or after the start time according to a recurrence pattern
- # defined in the schedule string. The start time can be changed at any
- # moment. The time when a data transfer can be trigerred manually is not
- # limited by this option.
- },
- "state": "A String", # Output only. State of the most recently updated transfer run.
- "name": "A String", # The resource name of the transfer config.
- # Transfer config names have the form of
- # `projects/{project_id}/locations/{region}/transferConfigs/{config_id}`.
- # The name is automatically generated based on the config_id specified in
- # CreateTransferConfigRequest along with project_id and region. If config_id
- # is not provided, usually a uuid, even though it is not guaranteed or
- # required, will be generated for config_id.
- "destinationDatasetId": "A String", # The BigQuery target dataset id.
- "userId": "A String", # Deprecated. Unique ID of the user on whose behalf transfer is done.
}</pre>
</div>
@@ -338,6 +338,38 @@
# When a new transfer configuration is created, the specified
# `destination_dataset_id` is created when needed and shared with the
# appropriate data source service account.
+ "dataRefreshWindowDays": 42, # The number of days to look back to automatically refresh the data.
+ # For example, if `data_refresh_window_days = 10`, then every day
+ # BigQuery reingests data for [today-10, today-1], rather than ingesting data
+ # for just [today-1].
+ # Only valid if the data source supports the feature. Set the value to 0
+ # to use the default value.
+ "dataSourceId": "A String", # Data source id. Cannot be changed once data transfer is created.
+ "scheduleOptions": { # Options customizing the data transfer schedule. # Options customizing the data transfer schedule.
+ "endTime": "A String", # Defines time to stop scheduling transfer runs. A transfer run cannot be
+ # scheduled at or after the end time. The end time can be changed at any
+ # moment. The time when a data transfer can be trigerred manually is not
+ # limited by this option.
+ "startTime": "A String", # Specifies time to start scheduling transfer runs. The first run will be
+ # scheduled at or after the start time according to a recurrence pattern
+ # defined in the schedule string. The start time can be changed at any
+ # moment. The time when a data transfer can be trigerred manually is not
+ # limited by this option.
+ "disableAutoScheduling": True or False, # If true, automatic scheduling of data transfer runs for this configuration
+ # will be disabled. The runs can be started on ad-hoc basis using
+ # StartManualTransferRuns API. When automatic scheduling is disabled, the
+ # TransferConfig.schedule field will be ignored.
+ },
+ "state": "A String", # Output only. State of the most recently updated transfer run.
+ "name": "A String", # The resource name of the transfer config.
+ # Transfer config names have the form of
+ # `projects/{project_id}/locations/{region}/transferConfigs/{config_id}`.
+ # The name is automatically generated based on the config_id specified in
+ # CreateTransferConfigRequest along with project_id and region. If config_id
+ # is not provided, usually a uuid, even though it is not guaranteed or
+ # required, will be generated for config_id.
+ "destinationDatasetId": "A String", # The BigQuery target dataset id.
+ "userId": "A String", # Deprecated. Unique ID of the user on whose behalf transfer is done.
"notificationPubsubTopic": "A String", # Pub/Sub topic where notifications will be sent after transfer runs
# associated with this transfer config finish.
"params": { # Data transfer specific parameters.
@@ -350,9 +382,9 @@
},
"datasetRegion": "A String", # Output only. Region in which BigQuery dataset is located.
"displayName": "A String", # User specified display name for the data transfer.
+ "nextRunTime": "A String", # Output only. Next time when data transfer will run.
"disabled": True or False, # Is this config disabled. When set to true, no runs are scheduled
# for a given transfer.
- "nextRunTime": "A String", # Output only. Next time when data transfer will run.
"schedule": "A String", # Data transfer schedule.
# If the data source does not support a custom schedule, this should be
# empty. If it is empty, the default value for the data source will be
@@ -366,56 +398,24 @@
# https://cloud.google.com/appengine/docs/flexible/python/scheduling-jobs-with-cron-yaml#the_schedule_format
# NOTE: the granularity should be at least 8 hours, or less frequent.
"updateTime": "A String", # Output only. Data transfer modification time. Ignored by server on input.
- "dataRefreshWindowDays": 42, # The number of days to look back to automatically refresh the data.
- # For example, if `data_refresh_window_days = 10`, then every day
- # BigQuery reingests data for [today-10, today-1], rather than ingesting data
- # for just [today-1].
- # Only valid if the data source supports the feature. Set the value to 0
- # to use the default value.
- "dataSourceId": "A String", # Data source id. Cannot be changed once data transfer is created.
- "scheduleOptions": { # Options customizing the data transfer schedule. # Options customizing the data transfer schedule.
- "disableAutoScheduling": True or False, # If true, automatic scheduling of data transfer runs for this configuration
- # will be disabled. The runs can be started on ad-hoc basis using
- # StartManualTransferRuns API. When automatic scheduling is disabled, the
- # TransferConfig.schedule field will be ignored.
- "endTime": "A String", # Defines time to stop scheduling transfer runs. A transfer run cannot be
- # scheduled at or after the end time. The end time can be changed at any
- # moment. The time when a data transfer can be trigerred manually is not
- # limited by this option.
- "startTime": "A String", # Specifies time to start scheduling transfer runs. The first run will be
- # scheduled at or after the start time according to a recurrence pattern
- # defined in the schedule string. The start time can be changed at any
- # moment. The time when a data transfer can be trigerred manually is not
- # limited by this option.
- },
- "state": "A String", # Output only. State of the most recently updated transfer run.
- "name": "A String", # The resource name of the transfer config.
- # Transfer config names have the form of
- # `projects/{project_id}/locations/{region}/transferConfigs/{config_id}`.
- # The name is automatically generated based on the config_id specified in
- # CreateTransferConfigRequest along with project_id and region. If config_id
- # is not provided, usually a uuid, even though it is not guaranteed or
- # required, will be generated for config_id.
- "destinationDatasetId": "A String", # The BigQuery target dataset id.
- "userId": "A String", # Deprecated. Unique ID of the user on whose behalf transfer is done.
}</pre>
</div>
<div class="method">
- <code class="details" id="list">list(parent, pageToken=None, pageSize=None, dataSourceIds=None, x__xgafv=None)</code>
+ <code class="details" id="list">list(parent, dataSourceIds=None, pageToken=None, pageSize=None, x__xgafv=None)</code>
<pre>Returns information about all data transfers in the project.
Args:
parent: string, Required. The BigQuery project id for which data sources
should be returned: `projects/{project_id}` or
`projects/{project_id}/locations/{location_id}` (required)
+ dataSourceIds: string, When specified, only configurations of requested data sources are returned. (repeated)
pageToken: string, Pagination token, which can be used to request a specific page
of `ListTransfersRequest` list results. For multiple-page
results, `ListTransfersResponse` outputs
a `next_page` token, which can be used as the
`page_token` value to request the next page of list results.
pageSize: integer, Page size. The default page size is the maximum value of 1000 results.
- dataSourceIds: string, When specified, only configurations of requested data sources are returned. (repeated)
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
@@ -425,10 +425,6 @@
An object of the form:
{ # The returned list of pipelines in the project.
- "nextPageToken": "A String", # Output only. The next-pagination token. For multiple-page list results,
- # this token can be used as the
- # `ListTransferConfigsRequest.page_token`
- # to request the next page of list results.
"transferConfigs": [ # Output only. The stored pipeline transfer configurations.
{ # Represents a data transfer configuration. A transfer configuration
# contains all metadata needed to perform a data transfer. For example,
@@ -436,6 +432,38 @@
# When a new transfer configuration is created, the specified
# `destination_dataset_id` is created when needed and shared with the
# appropriate data source service account.
+ "dataRefreshWindowDays": 42, # The number of days to look back to automatically refresh the data.
+ # For example, if `data_refresh_window_days = 10`, then every day
+ # BigQuery reingests data for [today-10, today-1], rather than ingesting data
+ # for just [today-1].
+ # Only valid if the data source supports the feature. Set the value to 0
+ # to use the default value.
+ "dataSourceId": "A String", # Data source id. Cannot be changed once data transfer is created.
+ "scheduleOptions": { # Options customizing the data transfer schedule. # Options customizing the data transfer schedule.
+ "endTime": "A String", # Defines time to stop scheduling transfer runs. A transfer run cannot be
+ # scheduled at or after the end time. The end time can be changed at any
+ # moment. The time when a data transfer can be trigerred manually is not
+ # limited by this option.
+ "startTime": "A String", # Specifies time to start scheduling transfer runs. The first run will be
+ # scheduled at or after the start time according to a recurrence pattern
+ # defined in the schedule string. The start time can be changed at any
+ # moment. The time when a data transfer can be trigerred manually is not
+ # limited by this option.
+ "disableAutoScheduling": True or False, # If true, automatic scheduling of data transfer runs for this configuration
+ # will be disabled. The runs can be started on ad-hoc basis using
+ # StartManualTransferRuns API. When automatic scheduling is disabled, the
+ # TransferConfig.schedule field will be ignored.
+ },
+ "state": "A String", # Output only. State of the most recently updated transfer run.
+ "name": "A String", # The resource name of the transfer config.
+ # Transfer config names have the form of
+ # `projects/{project_id}/locations/{region}/transferConfigs/{config_id}`.
+ # The name is automatically generated based on the config_id specified in
+ # CreateTransferConfigRequest along with project_id and region. If config_id
+ # is not provided, usually a uuid, even though it is not guaranteed or
+ # required, will be generated for config_id.
+ "destinationDatasetId": "A String", # The BigQuery target dataset id.
+ "userId": "A String", # Deprecated. Unique ID of the user on whose behalf transfer is done.
"notificationPubsubTopic": "A String", # Pub/Sub topic where notifications will be sent after transfer runs
# associated with this transfer config finish.
"params": { # Data transfer specific parameters.
@@ -448,9 +476,9 @@
},
"datasetRegion": "A String", # Output only. Region in which BigQuery dataset is located.
"displayName": "A String", # User specified display name for the data transfer.
+ "nextRunTime": "A String", # Output only. Next time when data transfer will run.
"disabled": True or False, # Is this config disabled. When set to true, no runs are scheduled
# for a given transfer.
- "nextRunTime": "A String", # Output only. Next time when data transfer will run.
"schedule": "A String", # Data transfer schedule.
# If the data source does not support a custom schedule, this should be
# empty. If it is empty, the default value for the data source will be
@@ -464,40 +492,12 @@
# https://cloud.google.com/appengine/docs/flexible/python/scheduling-jobs-with-cron-yaml#the_schedule_format
# NOTE: the granularity should be at least 8 hours, or less frequent.
"updateTime": "A String", # Output only. Data transfer modification time. Ignored by server on input.
- "dataRefreshWindowDays": 42, # The number of days to look back to automatically refresh the data.
- # For example, if `data_refresh_window_days = 10`, then every day
- # BigQuery reingests data for [today-10, today-1], rather than ingesting data
- # for just [today-1].
- # Only valid if the data source supports the feature. Set the value to 0
- # to use the default value.
- "dataSourceId": "A String", # Data source id. Cannot be changed once data transfer is created.
- "scheduleOptions": { # Options customizing the data transfer schedule. # Options customizing the data transfer schedule.
- "disableAutoScheduling": True or False, # If true, automatic scheduling of data transfer runs for this configuration
- # will be disabled. The runs can be started on ad-hoc basis using
- # StartManualTransferRuns API. When automatic scheduling is disabled, the
- # TransferConfig.schedule field will be ignored.
- "endTime": "A String", # Defines time to stop scheduling transfer runs. A transfer run cannot be
- # scheduled at or after the end time. The end time can be changed at any
- # moment. The time when a data transfer can be trigerred manually is not
- # limited by this option.
- "startTime": "A String", # Specifies time to start scheduling transfer runs. The first run will be
- # scheduled at or after the start time according to a recurrence pattern
- # defined in the schedule string. The start time can be changed at any
- # moment. The time when a data transfer can be trigerred manually is not
- # limited by this option.
- },
- "state": "A String", # Output only. State of the most recently updated transfer run.
- "name": "A String", # The resource name of the transfer config.
- # Transfer config names have the form of
- # `projects/{project_id}/locations/{region}/transferConfigs/{config_id}`.
- # The name is automatically generated based on the config_id specified in
- # CreateTransferConfigRequest along with project_id and region. If config_id
- # is not provided, usually a uuid, even though it is not guaranteed or
- # required, will be generated for config_id.
- "destinationDatasetId": "A String", # The BigQuery target dataset id.
- "userId": "A String", # Deprecated. Unique ID of the user on whose behalf transfer is done.
},
],
+ "nextPageToken": "A String", # Output only. The next-pagination token. For multiple-page list results,
+ # this token can be used as the
+ # `ListTransferConfigsRequest.page_token`
+ # to request the next page of list results.
}</pre>
</div>
@@ -537,6 +537,38 @@
# When a new transfer configuration is created, the specified
# `destination_dataset_id` is created when needed and shared with the
# appropriate data source service account.
+ "dataRefreshWindowDays": 42, # The number of days to look back to automatically refresh the data.
+ # For example, if `data_refresh_window_days = 10`, then every day
+ # BigQuery reingests data for [today-10, today-1], rather than ingesting data
+ # for just [today-1].
+ # Only valid if the data source supports the feature. Set the value to 0
+ # to use the default value.
+ "dataSourceId": "A String", # Data source id. Cannot be changed once data transfer is created.
+ "scheduleOptions": { # Options customizing the data transfer schedule. # Options customizing the data transfer schedule.
+ "endTime": "A String", # Defines time to stop scheduling transfer runs. A transfer run cannot be
+ # scheduled at or after the end time. The end time can be changed at any
+ # moment. The time when a data transfer can be trigerred manually is not
+ # limited by this option.
+ "startTime": "A String", # Specifies time to start scheduling transfer runs. The first run will be
+ # scheduled at or after the start time according to a recurrence pattern
+ # defined in the schedule string. The start time can be changed at any
+ # moment. The time when a data transfer can be trigerred manually is not
+ # limited by this option.
+ "disableAutoScheduling": True or False, # If true, automatic scheduling of data transfer runs for this configuration
+ # will be disabled. The runs can be started on ad-hoc basis using
+ # StartManualTransferRuns API. When automatic scheduling is disabled, the
+ # TransferConfig.schedule field will be ignored.
+ },
+ "state": "A String", # Output only. State of the most recently updated transfer run.
+ "name": "A String", # The resource name of the transfer config.
+ # Transfer config names have the form of
+ # `projects/{project_id}/locations/{region}/transferConfigs/{config_id}`.
+ # The name is automatically generated based on the config_id specified in
+ # CreateTransferConfigRequest along with project_id and region. If config_id
+ # is not provided, usually a uuid, even though it is not guaranteed or
+ # required, will be generated for config_id.
+ "destinationDatasetId": "A String", # The BigQuery target dataset id.
+ "userId": "A String", # Deprecated. Unique ID of the user on whose behalf transfer is done.
"notificationPubsubTopic": "A String", # Pub/Sub topic where notifications will be sent after transfer runs
# associated with this transfer config finish.
"params": { # Data transfer specific parameters.
@@ -549,9 +581,9 @@
},
"datasetRegion": "A String", # Output only. Region in which BigQuery dataset is located.
"displayName": "A String", # User specified display name for the data transfer.
+ "nextRunTime": "A String", # Output only. Next time when data transfer will run.
"disabled": True or False, # Is this config disabled. When set to true, no runs are scheduled
# for a given transfer.
- "nextRunTime": "A String", # Output only. Next time when data transfer will run.
"schedule": "A String", # Data transfer schedule.
# If the data source does not support a custom schedule, this should be
# empty. If it is empty, the default value for the data source will be
@@ -565,38 +597,6 @@
# https://cloud.google.com/appengine/docs/flexible/python/scheduling-jobs-with-cron-yaml#the_schedule_format
# NOTE: the granularity should be at least 8 hours, or less frequent.
"updateTime": "A String", # Output only. Data transfer modification time. Ignored by server on input.
- "dataRefreshWindowDays": 42, # The number of days to look back to automatically refresh the data.
- # For example, if `data_refresh_window_days = 10`, then every day
- # BigQuery reingests data for [today-10, today-1], rather than ingesting data
- # for just [today-1].
- # Only valid if the data source supports the feature. Set the value to 0
- # to use the default value.
- "dataSourceId": "A String", # Data source id. Cannot be changed once data transfer is created.
- "scheduleOptions": { # Options customizing the data transfer schedule. # Options customizing the data transfer schedule.
- "disableAutoScheduling": True or False, # If true, automatic scheduling of data transfer runs for this configuration
- # will be disabled. The runs can be started on ad-hoc basis using
- # StartManualTransferRuns API. When automatic scheduling is disabled, the
- # TransferConfig.schedule field will be ignored.
- "endTime": "A String", # Defines time to stop scheduling transfer runs. A transfer run cannot be
- # scheduled at or after the end time. The end time can be changed at any
- # moment. The time when a data transfer can be trigerred manually is not
- # limited by this option.
- "startTime": "A String", # Specifies time to start scheduling transfer runs. The first run will be
- # scheduled at or after the start time according to a recurrence pattern
- # defined in the schedule string. The start time can be changed at any
- # moment. The time when a data transfer can be trigerred manually is not
- # limited by this option.
- },
- "state": "A String", # Output only. State of the most recently updated transfer run.
- "name": "A String", # The resource name of the transfer config.
- # Transfer config names have the form of
- # `projects/{project_id}/locations/{region}/transferConfigs/{config_id}`.
- # The name is automatically generated based on the config_id specified in
- # CreateTransferConfigRequest along with project_id and region. If config_id
- # is not provided, usually a uuid, even though it is not guaranteed or
- # required, will be generated for config_id.
- "destinationDatasetId": "A String", # The BigQuery target dataset id.
- "userId": "A String", # Deprecated. Unique ID of the user on whose behalf transfer is done.
}
versionInfo: string, Optional version info. If users want to find a very recent access token,
@@ -641,6 +641,38 @@
# When a new transfer configuration is created, the specified
# `destination_dataset_id` is created when needed and shared with the
# appropriate data source service account.
+ "dataRefreshWindowDays": 42, # The number of days to look back to automatically refresh the data.
+ # For example, if `data_refresh_window_days = 10`, then every day
+ # BigQuery reingests data for [today-10, today-1], rather than ingesting data
+ # for just [today-1].
+ # Only valid if the data source supports the feature. Set the value to 0
+ # to use the default value.
+ "dataSourceId": "A String", # Data source id. Cannot be changed once data transfer is created.
+ "scheduleOptions": { # Options customizing the data transfer schedule. # Options customizing the data transfer schedule.
+ "endTime": "A String", # Defines time to stop scheduling transfer runs. A transfer run cannot be
+ # scheduled at or after the end time. The end time can be changed at any
+ # moment. The time when a data transfer can be trigerred manually is not
+ # limited by this option.
+ "startTime": "A String", # Specifies time to start scheduling transfer runs. The first run will be
+ # scheduled at or after the start time according to a recurrence pattern
+ # defined in the schedule string. The start time can be changed at any
+ # moment. The time when a data transfer can be trigerred manually is not
+ # limited by this option.
+ "disableAutoScheduling": True or False, # If true, automatic scheduling of data transfer runs for this configuration
+ # will be disabled. The runs can be started on ad-hoc basis using
+ # StartManualTransferRuns API. When automatic scheduling is disabled, the
+ # TransferConfig.schedule field will be ignored.
+ },
+ "state": "A String", # Output only. State of the most recently updated transfer run.
+ "name": "A String", # The resource name of the transfer config.
+ # Transfer config names have the form of
+ # `projects/{project_id}/locations/{region}/transferConfigs/{config_id}`.
+ # The name is automatically generated based on the config_id specified in
+ # CreateTransferConfigRequest along with project_id and region. If config_id
+ # is not provided, usually a uuid, even though it is not guaranteed or
+ # required, will be generated for config_id.
+ "destinationDatasetId": "A String", # The BigQuery target dataset id.
+ "userId": "A String", # Deprecated. Unique ID of the user on whose behalf transfer is done.
"notificationPubsubTopic": "A String", # Pub/Sub topic where notifications will be sent after transfer runs
# associated with this transfer config finish.
"params": { # Data transfer specific parameters.
@@ -653,9 +685,9 @@
},
"datasetRegion": "A String", # Output only. Region in which BigQuery dataset is located.
"displayName": "A String", # User specified display name for the data transfer.
+ "nextRunTime": "A String", # Output only. Next time when data transfer will run.
"disabled": True or False, # Is this config disabled. When set to true, no runs are scheduled
# for a given transfer.
- "nextRunTime": "A String", # Output only. Next time when data transfer will run.
"schedule": "A String", # Data transfer schedule.
# If the data source does not support a custom schedule, this should be
# empty. If it is empty, the default value for the data source will be
@@ -669,38 +701,6 @@
# https://cloud.google.com/appengine/docs/flexible/python/scheduling-jobs-with-cron-yaml#the_schedule_format
# NOTE: the granularity should be at least 8 hours, or less frequent.
"updateTime": "A String", # Output only. Data transfer modification time. Ignored by server on input.
- "dataRefreshWindowDays": 42, # The number of days to look back to automatically refresh the data.
- # For example, if `data_refresh_window_days = 10`, then every day
- # BigQuery reingests data for [today-10, today-1], rather than ingesting data
- # for just [today-1].
- # Only valid if the data source supports the feature. Set the value to 0
- # to use the default value.
- "dataSourceId": "A String", # Data source id. Cannot be changed once data transfer is created.
- "scheduleOptions": { # Options customizing the data transfer schedule. # Options customizing the data transfer schedule.
- "disableAutoScheduling": True or False, # If true, automatic scheduling of data transfer runs for this configuration
- # will be disabled. The runs can be started on ad-hoc basis using
- # StartManualTransferRuns API. When automatic scheduling is disabled, the
- # TransferConfig.schedule field will be ignored.
- "endTime": "A String", # Defines time to stop scheduling transfer runs. A transfer run cannot be
- # scheduled at or after the end time. The end time can be changed at any
- # moment. The time when a data transfer can be trigerred manually is not
- # limited by this option.
- "startTime": "A String", # Specifies time to start scheduling transfer runs. The first run will be
- # scheduled at or after the start time according to a recurrence pattern
- # defined in the schedule string. The start time can be changed at any
- # moment. The time when a data transfer can be trigerred manually is not
- # limited by this option.
- },
- "state": "A String", # Output only. State of the most recently updated transfer run.
- "name": "A String", # The resource name of the transfer config.
- # Transfer config names have the form of
- # `projects/{project_id}/locations/{region}/transferConfigs/{config_id}`.
- # The name is automatically generated based on the config_id specified in
- # CreateTransferConfigRequest along with project_id and region. If config_id
- # is not provided, usually a uuid, even though it is not guaranteed or
- # required, will be generated for config_id.
- "destinationDatasetId": "A String", # The BigQuery target dataset id.
- "userId": "A String", # Deprecated. Unique ID of the user on whose behalf transfer is done.
}</pre>
</div>
@@ -768,6 +768,7 @@
#
# You can find out more about this error model and how to work with it in the
# [API Design Guide](https://cloud.google.com/apis/design/errors).
+ "code": 42, # The status code, which should be an enum value of google.rpc.Code.
"message": "A String", # A developer-facing error message, which should be in English. Any
# user-facing error message should be localized and sent in the
# google.rpc.Status.details field, or localized by the client.
@@ -777,7 +778,6 @@
"a_key": "", # Properties of the object. Contains field @type with type URL.
},
],
- "code": 42, # The status code, which should be an enum value of google.rpc.Code.
},
"schedule": "A String", # Output only. Describes the schedule of this transfer run if it was
# created as part of a regular schedule. For batch transfer runs that are
@@ -809,14 +809,14 @@
{ # A request to start manual transfer runs.
"requestedTimeRange": { # A specification for a time range, this will request transfer runs with # Time range for the transfer runs that should be started.
# run_time between start_time (inclusive) and end_time (exclusive).
- "startTime": "A String", # Start time of the range of transfer runs. For example,
- # `"2017-05-25T00:00:00+00:00"`. The start_time must be strictly less than
- # the end_time. Creates transfer runs where run_time is in the range betwen
- # start_time (inclusive) and end_time (exlusive).
"endTime": "A String", # End time of the range of transfer runs. For example,
# `"2017-05-30T00:00:00+00:00"`. The end_time must not be in the future.
# Creates transfer runs where run_time is in the range betwen start_time
# (inclusive) and end_time (exlusive).
+ "startTime": "A String", # Start time of the range of transfer runs. For example,
+ # `"2017-05-25T00:00:00+00:00"`. The start_time must be strictly less than
+ # the end_time. Creates transfer runs where run_time is in the range betwen
+ # start_time (inclusive) and end_time (exlusive).
},
"requestedRunTime": "A String", # Specific run_time for a transfer run to be started. The
# requested_run_time must not be in the future.
@@ -864,6 +864,7 @@
#
# You can find out more about this error model and how to work with it in the
# [API Design Guide](https://cloud.google.com/apis/design/errors).
+ "code": 42, # The status code, which should be an enum value of google.rpc.Code.
"message": "A String", # A developer-facing error message, which should be in English. Any
# user-facing error message should be localized and sent in the
# google.rpc.Status.details field, or localized by the client.
@@ -873,7 +874,6 @@
"a_key": "", # Properties of the object. Contains field @type with type URL.
},
],
- "code": 42, # The status code, which should be an enum value of google.rpc.Code.
},
"schedule": "A String", # Output only. Describes the schedule of this transfer run if it was
# created as part of a regular schedule. For batch transfer runs that are