blob: aaa07bea9a4f7caf06374d9e34537133c5ca3ff6 [file] [log] [blame]
Bu Sun Kim65020912020-05-20 12:08:20 -07001<html><body>
2<style>
3
4body, h1, h2, h3, div, span, p, pre, a {
5 margin: 0;
6 padding: 0;
7 border: 0;
8 font-weight: inherit;
9 font-style: inherit;
10 font-size: 100%;
11 font-family: inherit;
12 vertical-align: baseline;
13}
14
15body {
16 font-size: 13px;
17 padding: 1em;
18}
19
20h1 {
21 font-size: 26px;
22 margin-bottom: 1em;
23}
24
25h2 {
26 font-size: 24px;
27 margin-bottom: 1em;
28}
29
30h3 {
31 font-size: 20px;
32 margin-bottom: 1em;
33 margin-top: 1em;
34}
35
36pre, code {
37 line-height: 1.5;
38 font-family: Monaco, 'DejaVu Sans Mono', 'Bitstream Vera Sans Mono', 'Lucida Console', monospace;
39}
40
41pre {
42 margin-top: 0.5em;
43}
44
45h1, h2, h3, p {
46 font-family: Arial, sans serif;
47}
48
49h1, h2, h3 {
50 border-bottom: solid #CCC 1px;
51}
52
53.toc_element {
54 margin-top: 0.5em;
55}
56
57.firstline {
58 margin-left: 2 em;
59}
60
61.method {
62 margin-top: 1em;
63 border: solid 1px #CCC;
64 padding: 1em;
65 background: #EEE;
66}
67
68.details {
69 font-weight: bold;
70 font-size: 14px;
71}
72
73</style>
74
75<h1><a href="dataflow_v1b3.html">Dataflow API</a> . <a href="dataflow_v1b3.projects.html">projects</a> . <a href="dataflow_v1b3.projects.locations.html">locations</a> . <a href="dataflow_v1b3.projects.locations.flexTemplates.html">flexTemplates</a></h1>
76<h2>Instance Methods</h2>
77<p class="toc_element">
78 <code><a href="#launch">launch(projectId, location, body=None, x__xgafv=None)</a></code></p>
79<p class="firstline">Launch a job with a FlexTemplate.</p>
80<h3>Method Details</h3>
81<div class="method">
82 <code class="details" id="launch">launch(projectId, location, body=None, x__xgafv=None)</code>
83 <pre>Launch a job with a FlexTemplate.
84
85Args:
86 projectId: string, Required. The ID of the Cloud Platform project that the job belongs to. (required)
87 location: string, Required. The [regional endpoint]
88(https://cloud.google.com/dataflow/docs/concepts/regional-endpoints) to
89which to direct the request. E.g., us-central1, us-west1. (required)
90 body: object, The request body.
91 The object takes the form of:
92
93{ # A request to launch a Cloud Dataflow job from a FlexTemplate.
94 &quot;validateOnly&quot;: True or False, # If true, the request is validated but not actually executed.
95 # Defaults to false.
96 &quot;launchParameter&quot;: { # Launch FlexTemplate Parameter. # Required. Parameter to launch a job form Flex Template.
97 &quot;containerSpec&quot;: { # Container Spec. # Spec about the container image to launch.
98 &quot;metadata&quot;: { # Metadata describing a template. # Metadata describing a template including description and validation rules.
99 &quot;name&quot;: &quot;A String&quot;, # Required. The name of the template.
100 &quot;parameters&quot;: [ # The parameters for the template.
101 { # Metadata for a specific parameter.
102 &quot;label&quot;: &quot;A String&quot;, # Required. The label to display for the parameter.
103 &quot;paramType&quot;: &quot;A String&quot;, # Optional. The type of the parameter.
104 # Used for selecting input picker.
105 &quot;helpText&quot;: &quot;A String&quot;, # Required. The help text to display for the parameter.
106 &quot;name&quot;: &quot;A String&quot;, # Required. The name of the parameter.
107 &quot;regexes&quot;: [ # Optional. Regexes that the parameter must match.
108 &quot;A String&quot;,
109 ],
110 &quot;isOptional&quot;: True or False, # Optional. Whether the parameter is optional. Defaults to false.
111 },
112 ],
113 &quot;description&quot;: &quot;A String&quot;, # Optional. A description of the template.
114 },
115 &quot;sdkInfo&quot;: { # SDK Information. # Required. SDK info of the Flex Template.
116 &quot;language&quot;: &quot;A String&quot;, # Required. The SDK Language.
117 &quot;version&quot;: &quot;A String&quot;, # Optional. The SDK version.
118 },
119 &quot;image&quot;: &quot;A String&quot;, # Name of the docker container image. E.g., gcr.io/project/some-image
120 },
121 &quot;containerSpecGcsPath&quot;: &quot;A String&quot;, # Gcs path to a file with json serialized ContainerSpec as content.
122 &quot;parameters&quot;: { # The parameters for FlexTemplate.
123 # Ex. {&quot;num_workers&quot;:&quot;5&quot;}
124 &quot;a_key&quot;: &quot;A String&quot;,
125 },
126 &quot;jobName&quot;: &quot;A String&quot;, # Required. The job name to use for the created job.
127 },
128 }
129
130 x__xgafv: string, V1 error format.
131 Allowed values
132 1 - v1 error format
133 2 - v2 error format
134
135Returns:
136 An object of the form:
137
138 { # Response to the request to launch a job from Flex Template.
139 &quot;job&quot;: { # Defines a job to be run by the Cloud Dataflow service. # The job that was launched, if the request was not a dry run and
140 # the job was successfully launched.
141 &quot;clientRequestId&quot;: &quot;A String&quot;, # The client&#x27;s unique identifier of the job, re-used across retried attempts.
142 # If this field is set, the service will ensure its uniqueness.
143 # The request to create a job will fail if the service has knowledge of a
144 # previously submitted job with the same client&#x27;s ID and job name.
145 # The caller may use this field to ensure idempotence of job
146 # creation across retried attempts to create a job.
147 # By default, the field is empty and, in that case, the service ignores it.
148 &quot;id&quot;: &quot;A String&quot;, # The unique ID of this job.
149 #
150 # This field is set by the Cloud Dataflow service when the Job is
151 # created, and is immutable for the life of the job.
152 &quot;currentStateTime&quot;: &quot;A String&quot;, # The timestamp associated with the current state.
153 &quot;transformNameMapping&quot;: { # The map of transform name prefixes of the job to be replaced to the
154 # corresponding name prefixes of the new job.
155 &quot;a_key&quot;: &quot;A String&quot;,
156 },
157 &quot;environment&quot;: { # Describes the environment in which a Dataflow Job runs. # The environment for the job.
158 &quot;sdkPipelineOptions&quot;: { # The Cloud Dataflow SDK pipeline options specified by the user. These
159 # options are passed through the service and are used to recreate the
160 # SDK pipeline options on the worker in a language agnostic and platform
161 # independent way.
162 &quot;a_key&quot;: &quot;&quot;, # Properties of the object.
163 },
164 &quot;flexResourceSchedulingGoal&quot;: &quot;A String&quot;, # Which Flexible Resource Scheduling mode to run in.
165 &quot;workerPools&quot;: [ # The worker pools. At least one &quot;harness&quot; worker pool must be
166 # specified in order for the job to have workers.
167 { # Describes one particular pool of Cloud Dataflow workers to be
168 # instantiated by the Cloud Dataflow service in order to perform the
169 # computations required by a job. Note that a workflow job may use
170 # multiple pools, in order to match the various computational
171 # requirements of the various stages of the job.
172 &quot;defaultPackageSet&quot;: &quot;A String&quot;, # The default package set to install. This allows the service to
173 # select a default set of packages which are useful to worker
174 # harnesses written in a particular language.
175 &quot;network&quot;: &quot;A String&quot;, # Network to which VMs will be assigned. If empty or unspecified,
176 # the service will use the network &quot;default&quot;.
177 &quot;zone&quot;: &quot;A String&quot;, # Zone to run the worker pools in. If empty or unspecified, the service
178 # will attempt to choose a reasonable default.
179 &quot;numWorkers&quot;: 42, # Number of Google Compute Engine workers in this pool needed to
180 # execute the job. If zero or unspecified, the service will
181 # attempt to choose a reasonable default.
182 &quot;numThreadsPerWorker&quot;: 42, # The number of threads per worker harness. If empty or unspecified, the
183 # service will choose a number of threads (according to the number of cores
184 # on the selected machine type for batch, or 1 by convention for streaming).
185 &quot;diskSourceImage&quot;: &quot;A String&quot;, # Fully qualified source image for disks.
186 &quot;packages&quot;: [ # Packages to be installed on workers.
187 { # The packages that must be installed in order for a worker to run the
188 # steps of the Cloud Dataflow job that will be assigned to its worker
189 # pool.
190 #
191 # This is the mechanism by which the Cloud Dataflow SDK causes code to
192 # be loaded onto the workers. For example, the Cloud Dataflow Java SDK
193 # might use this to install jars containing the user&#x27;s code and all of the
194 # various dependencies (libraries, data files, etc.) required in order
195 # for that code to run.
196 &quot;location&quot;: &quot;A String&quot;, # The resource to read the package from. The supported resource type is:
197 #
198 # Google Cloud Storage:
199 #
200 # storage.googleapis.com/{bucket}
201 # bucket.storage.googleapis.com/
202 &quot;name&quot;: &quot;A String&quot;, # The name of the package.
203 },
204 ],
205 &quot;teardownPolicy&quot;: &quot;A String&quot;, # Sets the policy for determining when to turndown worker pool.
206 # Allowed values are: `TEARDOWN_ALWAYS`, `TEARDOWN_ON_SUCCESS`, and
207 # `TEARDOWN_NEVER`.
208 # `TEARDOWN_ALWAYS` means workers are always torn down regardless of whether
209 # the job succeeds. `TEARDOWN_ON_SUCCESS` means workers are torn down
210 # if the job succeeds. `TEARDOWN_NEVER` means the workers are never torn
211 # down.
212 #
213 # If the workers are not torn down by the service, they will
214 # continue to run and use Google Compute Engine VM resources in the
215 # user&#x27;s project until they are explicitly terminated by the user.
216 # Because of this, Google recommends using the `TEARDOWN_ALWAYS`
217 # policy except for small, manually supervised test jobs.
218 #
219 # If unknown or unspecified, the service will attempt to choose a reasonable
220 # default.
221 &quot;onHostMaintenance&quot;: &quot;A String&quot;, # The action to take on host maintenance, as defined by the Google
222 # Compute Engine API.
223 &quot;poolArgs&quot;: { # Extra arguments for this worker pool.
224 &quot;a_key&quot;: &quot;&quot;, # Properties of the object. Contains field @type with type URL.
225 },
226 &quot;diskSizeGb&quot;: 42, # Size of root disk for VMs, in GB. If zero or unspecified, the service will
227 # attempt to choose a reasonable default.
228 &quot;workerHarnessContainerImage&quot;: &quot;A String&quot;, # Required. Docker container image that executes the Cloud Dataflow worker
229 # harness, residing in Google Container Registry.
230 #
231 # Deprecated for the Fn API path. Use sdk_harness_container_images instead.
232 &quot;diskType&quot;: &quot;A String&quot;, # Type of root disk for VMs. If empty or unspecified, the service will
233 # attempt to choose a reasonable default.
234 &quot;machineType&quot;: &quot;A String&quot;, # Machine type (e.g. &quot;n1-standard-1&quot;). If empty or unspecified, the
235 # service will attempt to choose a reasonable default.
236 &quot;kind&quot;: &quot;A String&quot;, # The kind of the worker pool; currently only `harness` and `shuffle`
237 # are supported.
238 &quot;dataDisks&quot;: [ # Data disks that are used by a VM in this workflow.
239 { # Describes the data disk used by a workflow job.
240 &quot;sizeGb&quot;: 42, # Size of disk in GB. If zero or unspecified, the service will
241 # attempt to choose a reasonable default.
242 &quot;diskType&quot;: &quot;A String&quot;, # Disk storage type, as defined by Google Compute Engine. This
243 # must be a disk type appropriate to the project and zone in which
244 # the workers will run. If unknown or unspecified, the service
245 # will attempt to choose a reasonable default.
246 #
247 # For example, the standard persistent disk type is a resource name
248 # typically ending in &quot;pd-standard&quot;. If SSD persistent disks are
249 # available, the resource name typically ends with &quot;pd-ssd&quot;. The
250 # actual valid values are defined the Google Compute Engine API,
251 # not by the Cloud Dataflow API; consult the Google Compute Engine
252 # documentation for more information about determining the set of
253 # available disk types for a particular project and zone.
254 #
255 # Google Compute Engine Disk types are local to a particular
256 # project in a particular zone, and so the resource name will
257 # typically look something like this:
258 #
259 # compute.googleapis.com/projects/project-id/zones/zone/diskTypes/pd-standard
260 &quot;mountPoint&quot;: &quot;A String&quot;, # Directory in a VM where disk is mounted.
261 },
262 ],
263 &quot;sdkHarnessContainerImages&quot;: [ # Set of SDK harness containers needed to execute this pipeline. This will
264 # only be set in the Fn API path. For non-cross-language pipelines this
265 # should have only one entry. Cross-language pipelines will have two or more
266 # entries.
267 { # Defines a SDK harness container for executing Dataflow pipelines.
268 &quot;containerImage&quot;: &quot;A String&quot;, # A docker container image that resides in Google Container Registry.
269 &quot;useSingleCorePerContainer&quot;: True or False, # If true, recommends the Dataflow service to use only one core per SDK
270 # container instance with this image. If false (or unset) recommends using
271 # more than one core per SDK container instance with this image for
272 # efficiency. Note that Dataflow service may choose to override this property
273 # if needed.
274 },
275 ],
276 &quot;subnetwork&quot;: &quot;A String&quot;, # Subnetwork to which VMs will be assigned, if desired. Expected to be of
277 # the form &quot;regions/REGION/subnetworks/SUBNETWORK&quot;.
278 &quot;ipConfiguration&quot;: &quot;A String&quot;, # Configuration for VM IPs.
279 &quot;taskrunnerSettings&quot;: { # Taskrunner configuration settings. # Settings passed through to Google Compute Engine workers when
280 # using the standard Dataflow task runner. Users should ignore
281 # this field.
282 &quot;alsologtostderr&quot;: True or False, # Whether to also send taskrunner log info to stderr.
283 &quot;taskGroup&quot;: &quot;A String&quot;, # The UNIX group ID on the worker VM to use for tasks launched by
284 # taskrunner; e.g. &quot;wheel&quot;.
285 &quot;harnessCommand&quot;: &quot;A String&quot;, # The command to launch the worker harness.
286 &quot;logDir&quot;: &quot;A String&quot;, # The directory on the VM to store logs.
287 &quot;oauthScopes&quot;: [ # The OAuth2 scopes to be requested by the taskrunner in order to
288 # access the Cloud Dataflow API.
289 &quot;A String&quot;,
290 ],
291 &quot;dataflowApiVersion&quot;: &quot;A String&quot;, # The API version of endpoint, e.g. &quot;v1b3&quot;
292 &quot;logUploadLocation&quot;: &quot;A String&quot;, # Indicates where to put logs. If this is not specified, the logs
293 # will not be uploaded.
294 #
295 # The supported resource type is:
296 #
297 # Google Cloud Storage:
298 # storage.googleapis.com/{bucket}/{object}
299 # bucket.storage.googleapis.com/{object}
300 &quot;streamingWorkerMainClass&quot;: &quot;A String&quot;, # The streaming worker main class name.
301 &quot;workflowFileName&quot;: &quot;A String&quot;, # The file to store the workflow in.
302 &quot;baseTaskDir&quot;: &quot;A String&quot;, # The location on the worker for task-specific subdirectories.
303 &quot;tempStoragePrefix&quot;: &quot;A String&quot;, # The prefix of the resources the taskrunner should use for
304 # temporary storage.
305 #
306 # The supported resource type is:
307 #
308 # Google Cloud Storage:
309 # storage.googleapis.com/{bucket}/{object}
310 # bucket.storage.googleapis.com/{object}
311 &quot;commandlinesFileName&quot;: &quot;A String&quot;, # The file to store preprocessing commands in.
312 &quot;languageHint&quot;: &quot;A String&quot;, # The suggested backend language.
313 &quot;baseUrl&quot;: &quot;A String&quot;, # The base URL for the taskrunner to use when accessing Google Cloud APIs.
314 #
315 # When workers access Google Cloud APIs, they logically do so via
316 # relative URLs. If this field is specified, it supplies the base
317 # URL to use for resolving these relative URLs. The normative
318 # algorithm used is defined by RFC 1808, &quot;Relative Uniform Resource
319 # Locators&quot;.
320 #
321 # If not specified, the default value is &quot;http://www.googleapis.com/&quot;
322 &quot;logToSerialconsole&quot;: True or False, # Whether to send taskrunner log info to Google Compute Engine VM serial
323 # console.
324 &quot;continueOnException&quot;: True or False, # Whether to continue taskrunner if an exception is hit.
325 &quot;parallelWorkerSettings&quot;: { # Provides data to pass through to the worker harness. # The settings to pass to the parallel worker harness.
326 &quot;baseUrl&quot;: &quot;A String&quot;, # The base URL for accessing Google Cloud APIs.
327 #
328 # When workers access Google Cloud APIs, they logically do so via
329 # relative URLs. If this field is specified, it supplies the base
330 # URL to use for resolving these relative URLs. The normative
331 # algorithm used is defined by RFC 1808, &quot;Relative Uniform Resource
332 # Locators&quot;.
333 #
334 # If not specified, the default value is &quot;http://www.googleapis.com/&quot;
335 &quot;reportingEnabled&quot;: True or False, # Whether to send work progress updates to the service.
336 &quot;servicePath&quot;: &quot;A String&quot;, # The Cloud Dataflow service path relative to the root URL, for example,
337 # &quot;dataflow/v1b3/projects&quot;.
338 &quot;shuffleServicePath&quot;: &quot;A String&quot;, # The Shuffle service path relative to the root URL, for example,
339 # &quot;shuffle/v1beta1&quot;.
340 &quot;workerId&quot;: &quot;A String&quot;, # The ID of the worker running this pipeline.
341 &quot;tempStoragePrefix&quot;: &quot;A String&quot;, # The prefix of the resources the system should use for temporary
342 # storage.
343 #
344 # The supported resource type is:
345 #
346 # Google Cloud Storage:
347 #
348 # storage.googleapis.com/{bucket}/{object}
349 # bucket.storage.googleapis.com/{object}
350 },
351 &quot;vmId&quot;: &quot;A String&quot;, # The ID string of the VM.
352 &quot;taskUser&quot;: &quot;A String&quot;, # The UNIX user ID on the worker VM to use for tasks launched by
353 # taskrunner; e.g. &quot;root&quot;.
354 },
355 &quot;autoscalingSettings&quot;: { # Settings for WorkerPool autoscaling. # Settings for autoscaling of this WorkerPool.
356 &quot;maxNumWorkers&quot;: 42, # The maximum number of workers to cap scaling at.
357 &quot;algorithm&quot;: &quot;A String&quot;, # The algorithm to use for autoscaling.
358 },
359 &quot;metadata&quot;: { # Metadata to set on the Google Compute Engine VMs.
360 &quot;a_key&quot;: &quot;A String&quot;,
361 },
362 },
363 ],
364 &quot;dataset&quot;: &quot;A String&quot;, # The dataset for the current project where various workflow
365 # related tables are stored.
366 #
367 # The supported resource type is:
368 #
369 # Google BigQuery:
370 # bigquery.googleapis.com/{dataset}
371 &quot;internalExperiments&quot;: { # Experimental settings.
372 &quot;a_key&quot;: &quot;&quot;, # Properties of the object. Contains field @type with type URL.
373 },
374 &quot;workerRegion&quot;: &quot;A String&quot;, # The Compute Engine region
375 # (https://cloud.google.com/compute/docs/regions-zones/regions-zones) in
376 # which worker processing should occur, e.g. &quot;us-west1&quot;. Mutually exclusive
377 # with worker_zone. If neither worker_region nor worker_zone is specified,
378 # default to the control plane&#x27;s region.
379 &quot;serviceKmsKeyName&quot;: &quot;A String&quot;, # If set, contains the Cloud KMS key identifier used to encrypt data
380 # at rest, AKA a Customer Managed Encryption Key (CMEK).
381 #
382 # Format:
383 # projects/PROJECT_ID/locations/LOCATION/keyRings/KEY_RING/cryptoKeys/KEY
384 &quot;userAgent&quot;: { # A description of the process that generated the request.
385 &quot;a_key&quot;: &quot;&quot;, # Properties of the object.
386 },
387 &quot;workerZone&quot;: &quot;A String&quot;, # The Compute Engine zone
388 # (https://cloud.google.com/compute/docs/regions-zones/regions-zones) in
389 # which worker processing should occur, e.g. &quot;us-west1-a&quot;. Mutually exclusive
390 # with worker_region. If neither worker_region nor worker_zone is specified,
391 # a zone in the control plane&#x27;s region is chosen based on available capacity.
392 &quot;clusterManagerApiService&quot;: &quot;A String&quot;, # The type of cluster manager API to use. If unknown or
393 # unspecified, the service will attempt to choose a reasonable
394 # default. This should be in the form of the API service name,
395 # e.g. &quot;compute.googleapis.com&quot;.
396 &quot;tempStoragePrefix&quot;: &quot;A String&quot;, # The prefix of the resources the system should use for temporary
397 # storage. The system will append the suffix &quot;/temp-{JOBNAME} to
398 # this resource prefix, where {JOBNAME} is the value of the
399 # job_name field. The resulting bucket and object prefix is used
400 # as the prefix of the resources used to store temporary data
401 # needed during the job execution. NOTE: This will override the
402 # value in taskrunner_settings.
403 # The supported resource type is:
404 #
405 # Google Cloud Storage:
406 #
407 # storage.googleapis.com/{bucket}/{object}
408 # bucket.storage.googleapis.com/{object}
409 &quot;experiments&quot;: [ # The list of experiments to enable.
410 &quot;A String&quot;,
411 ],
412 &quot;version&quot;: { # A structure describing which components and their versions of the service
413 # are required in order to run the job.
414 &quot;a_key&quot;: &quot;&quot;, # Properties of the object.
415 },
416 &quot;serviceAccountEmail&quot;: &quot;A String&quot;, # Identity to run virtual machines as. Defaults to the default account.
417 },
418 &quot;stageStates&quot;: [ # This field may be mutated by the Cloud Dataflow service;
419 # callers cannot mutate it.
420 { # A message describing the state of a particular execution stage.
421 &quot;executionStageName&quot;: &quot;A String&quot;, # The name of the execution stage.
422 &quot;currentStateTime&quot;: &quot;A String&quot;, # The time at which the stage transitioned to this state.
423 &quot;executionStageState&quot;: &quot;A String&quot;, # Executions stage states allow the same set of values as JobState.
424 },
425 ],
426 &quot;jobMetadata&quot;: { # Metadata available primarily for filtering jobs. Will be included in the # This field is populated by the Dataflow service to support filtering jobs
427 # by the metadata values provided here. Populated for ListJobs and all GetJob
428 # views SUMMARY and higher.
429 # ListJob response and Job SUMMARY view.
430 &quot;bigTableDetails&quot;: [ # Identification of a BigTable source used in the Dataflow job.
431 { # Metadata for a BigTable connector used by the job.
432 &quot;tableId&quot;: &quot;A String&quot;, # TableId accessed in the connection.
433 &quot;projectId&quot;: &quot;A String&quot;, # ProjectId accessed in the connection.
434 &quot;instanceId&quot;: &quot;A String&quot;, # InstanceId accessed in the connection.
435 },
436 ],
437 &quot;spannerDetails&quot;: [ # Identification of a Spanner source used in the Dataflow job.
438 { # Metadata for a Spanner connector used by the job.
439 &quot;databaseId&quot;: &quot;A String&quot;, # DatabaseId accessed in the connection.
440 &quot;instanceId&quot;: &quot;A String&quot;, # InstanceId accessed in the connection.
441 &quot;projectId&quot;: &quot;A String&quot;, # ProjectId accessed in the connection.
442 },
443 ],
444 &quot;datastoreDetails&quot;: [ # Identification of a Datastore source used in the Dataflow job.
445 { # Metadata for a Datastore connector used by the job.
446 &quot;projectId&quot;: &quot;A String&quot;, # ProjectId accessed in the connection.
447 &quot;namespace&quot;: &quot;A String&quot;, # Namespace used in the connection.
448 },
449 ],
450 &quot;sdkVersion&quot;: { # The version of the SDK used to run the job. # The SDK version used to run the job.
451 &quot;versionDisplayName&quot;: &quot;A String&quot;, # A readable string describing the version of the SDK.
452 &quot;sdkSupportStatus&quot;: &quot;A String&quot;, # The support status for this SDK version.
453 &quot;version&quot;: &quot;A String&quot;, # The version of the SDK used to run the job.
454 },
455 &quot;bigqueryDetails&quot;: [ # Identification of a BigQuery source used in the Dataflow job.
456 { # Metadata for a BigQuery connector used by the job.
457 &quot;table&quot;: &quot;A String&quot;, # Table accessed in the connection.
458 &quot;dataset&quot;: &quot;A String&quot;, # Dataset accessed in the connection.
459 &quot;projectId&quot;: &quot;A String&quot;, # Project accessed in the connection.
460 &quot;query&quot;: &quot;A String&quot;, # Query used to access data in the connection.
461 },
462 ],
463 &quot;fileDetails&quot;: [ # Identification of a File source used in the Dataflow job.
464 { # Metadata for a File connector used by the job.
465 &quot;filePattern&quot;: &quot;A String&quot;, # File Pattern used to access files by the connector.
466 },
467 ],
468 &quot;pubsubDetails&quot;: [ # Identification of a PubSub source used in the Dataflow job.
469 { # Metadata for a PubSub connector used by the job.
470 &quot;subscription&quot;: &quot;A String&quot;, # Subscription used in the connection.
471 &quot;topic&quot;: &quot;A String&quot;, # Topic accessed in the connection.
472 },
473 ],
474 },
475 &quot;createdFromSnapshotId&quot;: &quot;A String&quot;, # If this is specified, the job&#x27;s initial state is populated from the given
476 # snapshot.
477 &quot;projectId&quot;: &quot;A String&quot;, # The ID of the Cloud Platform project that the job belongs to.
478 &quot;type&quot;: &quot;A String&quot;, # The type of Cloud Dataflow job.
479 &quot;pipelineDescription&quot;: { # A descriptive representation of submitted pipeline as well as the executed # Preliminary field: The format of this data may change at any time.
480 # A description of the user pipeline and stages through which it is executed.
481 # Created by Cloud Dataflow service. Only retrieved with
482 # JOB_VIEW_DESCRIPTION or JOB_VIEW_ALL.
483 # form. This data is provided by the Dataflow service for ease of visualizing
484 # the pipeline and interpreting Dataflow provided metrics.
485 &quot;executionPipelineStage&quot;: [ # Description of each stage of execution of the pipeline.
486 { # Description of the composing transforms, names/ids, and input/outputs of a
487 # stage of execution. Some composing transforms and sources may have been
488 # generated by the Dataflow service during execution planning.
489 &quot;id&quot;: &quot;A String&quot;, # Dataflow service generated id for this stage.
490 &quot;componentTransform&quot;: [ # Transforms that comprise this execution stage.
491 { # Description of a transform executed as part of an execution stage.
492 &quot;originalTransform&quot;: &quot;A String&quot;, # User name for the original user transform with which this transform is
493 # most closely associated.
494 &quot;name&quot;: &quot;A String&quot;, # Dataflow service generated name for this source.
495 &quot;userName&quot;: &quot;A String&quot;, # Human-readable name for this transform; may be user or system generated.
496 },
497 ],
498 &quot;componentSource&quot;: [ # Collections produced and consumed by component transforms of this stage.
499 { # Description of an interstitial value between transforms in an execution
500 # stage.
501 &quot;name&quot;: &quot;A String&quot;, # Dataflow service generated name for this source.
502 &quot;userName&quot;: &quot;A String&quot;, # Human-readable name for this transform; may be user or system generated.
503 &quot;originalTransformOrCollection&quot;: &quot;A String&quot;, # User name for the original user transform or collection with which this
504 # source is most closely associated.
505 },
506 ],
507 &quot;kind&quot;: &quot;A String&quot;, # Type of tranform this stage is executing.
508 &quot;outputSource&quot;: [ # Output sources for this stage.
509 { # Description of an input or output of an execution stage.
510 &quot;originalTransformOrCollection&quot;: &quot;A String&quot;, # User name for the original user transform or collection with which this
511 # source is most closely associated.
512 &quot;name&quot;: &quot;A String&quot;, # Dataflow service generated name for this source.
513 &quot;sizeBytes&quot;: &quot;A String&quot;, # Size of the source, if measurable.
514 &quot;userName&quot;: &quot;A String&quot;, # Human-readable name for this source; may be user or system generated.
515 },
516 ],
517 &quot;name&quot;: &quot;A String&quot;, # Dataflow service generated name for this stage.
518 &quot;inputSource&quot;: [ # Input sources for this stage.
519 { # Description of an input or output of an execution stage.
520 &quot;originalTransformOrCollection&quot;: &quot;A String&quot;, # User name for the original user transform or collection with which this
521 # source is most closely associated.
522 &quot;name&quot;: &quot;A String&quot;, # Dataflow service generated name for this source.
523 &quot;sizeBytes&quot;: &quot;A String&quot;, # Size of the source, if measurable.
524 &quot;userName&quot;: &quot;A String&quot;, # Human-readable name for this source; may be user or system generated.
525 },
526 ],
527 },
528 ],
529 &quot;originalPipelineTransform&quot;: [ # Description of each transform in the pipeline and collections between them.
530 { # Description of the type, names/ids, and input/outputs for a transform.
531 &quot;kind&quot;: &quot;A String&quot;, # Type of transform.
532 &quot;inputCollectionName&quot;: [ # User names for all collection inputs to this transform.
533 &quot;A String&quot;,
534 ],
535 &quot;name&quot;: &quot;A String&quot;, # User provided name for this transform instance.
536 &quot;id&quot;: &quot;A String&quot;, # SDK generated id of this transform instance.
537 &quot;displayData&quot;: [ # Transform-specific display data.
538 { # Data provided with a pipeline or transform to provide descriptive info.
539 &quot;timestampValue&quot;: &quot;A String&quot;, # Contains value if the data is of timestamp type.
540 &quot;boolValue&quot;: True or False, # Contains value if the data is of a boolean type.
541 &quot;javaClassValue&quot;: &quot;A String&quot;, # Contains value if the data is of java class type.
542 &quot;strValue&quot;: &quot;A String&quot;, # Contains value if the data is of string type.
543 &quot;int64Value&quot;: &quot;A String&quot;, # Contains value if the data is of int64 type.
544 &quot;durationValue&quot;: &quot;A String&quot;, # Contains value if the data is of duration type.
545 &quot;namespace&quot;: &quot;A String&quot;, # The namespace for the key. This is usually a class name or programming
546 # language namespace (i.e. python module) which defines the display data.
547 # This allows a dax monitoring system to specially handle the data
548 # and perform custom rendering.
549 &quot;floatValue&quot;: 3.14, # Contains value if the data is of float type.
550 &quot;key&quot;: &quot;A String&quot;, # The key identifying the display data.
551 # This is intended to be used as a label for the display data
552 # when viewed in a dax monitoring system.
553 &quot;shortStrValue&quot;: &quot;A String&quot;, # A possible additional shorter value to display.
554 # For example a java_class_name_value of com.mypackage.MyDoFn
555 # will be stored with MyDoFn as the short_str_value and
556 # com.mypackage.MyDoFn as the java_class_name value.
557 # short_str_value can be displayed and java_class_name_value
558 # will be displayed as a tooltip.
559 &quot;url&quot;: &quot;A String&quot;, # An optional full URL.
560 &quot;label&quot;: &quot;A String&quot;, # An optional label to display in a dax UI for the element.
561 },
562 ],
563 &quot;outputCollectionName&quot;: [ # User names for all collection outputs to this transform.
564 &quot;A String&quot;,
565 ],
566 },
567 ],
568 &quot;displayData&quot;: [ # Pipeline level display data.
569 { # Data provided with a pipeline or transform to provide descriptive info.
570 &quot;timestampValue&quot;: &quot;A String&quot;, # Contains value if the data is of timestamp type.
571 &quot;boolValue&quot;: True or False, # Contains value if the data is of a boolean type.
572 &quot;javaClassValue&quot;: &quot;A String&quot;, # Contains value if the data is of java class type.
573 &quot;strValue&quot;: &quot;A String&quot;, # Contains value if the data is of string type.
574 &quot;int64Value&quot;: &quot;A String&quot;, # Contains value if the data is of int64 type.
575 &quot;durationValue&quot;: &quot;A String&quot;, # Contains value if the data is of duration type.
576 &quot;namespace&quot;: &quot;A String&quot;, # The namespace for the key. This is usually a class name or programming
577 # language namespace (i.e. python module) which defines the display data.
578 # This allows a dax monitoring system to specially handle the data
579 # and perform custom rendering.
580 &quot;floatValue&quot;: 3.14, # Contains value if the data is of float type.
581 &quot;key&quot;: &quot;A String&quot;, # The key identifying the display data.
582 # This is intended to be used as a label for the display data
583 # when viewed in a dax monitoring system.
584 &quot;shortStrValue&quot;: &quot;A String&quot;, # A possible additional shorter value to display.
585 # For example a java_class_name_value of com.mypackage.MyDoFn
586 # will be stored with MyDoFn as the short_str_value and
587 # com.mypackage.MyDoFn as the java_class_name value.
588 # short_str_value can be displayed and java_class_name_value
589 # will be displayed as a tooltip.
590 &quot;url&quot;: &quot;A String&quot;, # An optional full URL.
591 &quot;label&quot;: &quot;A String&quot;, # An optional label to display in a dax UI for the element.
592 },
593 ],
594 },
595 &quot;replaceJobId&quot;: &quot;A String&quot;, # If this job is an update of an existing job, this field is the job ID
596 # of the job it replaced.
597 #
598 # When sending a `CreateJobRequest`, you can update a job by specifying it
599 # here. The job named here is stopped, and its intermediate state is
600 # transferred to this job.
601 &quot;tempFiles&quot;: [ # A set of files the system should be aware of that are used
602 # for temporary storage. These temporary files will be
603 # removed on job completion.
604 # No duplicates are allowed.
605 # No file patterns are supported.
606 #
607 # The supported files are:
608 #
609 # Google Cloud Storage:
610 #
611 # storage.googleapis.com/{bucket}/{object}
612 # bucket.storage.googleapis.com/{object}
613 &quot;A String&quot;,
614 ],
615 &quot;name&quot;: &quot;A String&quot;, # The user-specified Cloud Dataflow job name.
616 #
617 # Only one Job with a given name may exist in a project at any
618 # given time. If a caller attempts to create a Job with the same
619 # name as an already-existing Job, the attempt returns the
620 # existing Job.
621 #
622 # The name must match the regular expression
623 # `[a-z]([-a-z0-9]{0,38}[a-z0-9])?`
624 &quot;steps&quot;: [ # Exactly one of step or steps_location should be specified.
625 #
626 # The top-level steps that constitute the entire job.
627 { # Defines a particular step within a Cloud Dataflow job.
628 #
629 # A job consists of multiple steps, each of which performs some
630 # specific operation as part of the overall job. Data is typically
631 # passed from one step to another as part of the job.
632 #
633 # Here&#x27;s an example of a sequence of steps which together implement a
634 # Map-Reduce job:
635 #
636 # * Read a collection of data from some source, parsing the
637 # collection&#x27;s elements.
638 #
639 # * Validate the elements.
640 #
641 # * Apply a user-defined function to map each element to some value
642 # and extract an element-specific key value.
643 #
644 # * Group elements with the same key into a single element with
645 # that key, transforming a multiply-keyed collection into a
646 # uniquely-keyed collection.
647 #
648 # * Write the elements out to some data sink.
649 #
650 # Note that the Cloud Dataflow service may be used to run many different
651 # types of jobs, not just Map-Reduce.
652 &quot;name&quot;: &quot;A String&quot;, # The name that identifies the step. This must be unique for each
653 # step with respect to all other steps in the Cloud Dataflow job.
654 &quot;kind&quot;: &quot;A String&quot;, # The kind of step in the Cloud Dataflow job.
655 &quot;properties&quot;: { # Named properties associated with the step. Each kind of
656 # predefined step has its own required set of properties.
657 # Must be provided on Create. Only retrieved with JOB_VIEW_ALL.
658 &quot;a_key&quot;: &quot;&quot;, # Properties of the object.
659 },
660 },
661 ],
662 &quot;replacedByJobId&quot;: &quot;A String&quot;, # If another job is an update of this job (and thus, this job is in
663 # `JOB_STATE_UPDATED`), this field contains the ID of that job.
664 &quot;executionInfo&quot;: { # Additional information about how a Cloud Dataflow job will be executed that # Deprecated.
665 # isn&#x27;t contained in the submitted job.
666 &quot;stages&quot;: { # A mapping from each stage to the information about that stage.
667 &quot;a_key&quot;: { # Contains information about how a particular
668 # google.dataflow.v1beta3.Step will be executed.
669 &quot;stepName&quot;: [ # The steps associated with the execution stage.
670 # Note that stages may have several steps, and that a given step
671 # might be run by more than one stage.
672 &quot;A String&quot;,
673 ],
674 },
675 },
676 },
677 &quot;currentState&quot;: &quot;A String&quot;, # The current state of the job.
678 #
679 # Jobs are created in the `JOB_STATE_STOPPED` state unless otherwise
680 # specified.
681 #
682 # A job in the `JOB_STATE_RUNNING` state may asynchronously enter a
683 # terminal state. After a job has reached a terminal state, no
684 # further state updates may be made.
685 #
686 # This field may be mutated by the Cloud Dataflow service;
687 # callers cannot mutate it.
688 &quot;location&quot;: &quot;A String&quot;, # The [regional endpoint]
689 # (https://cloud.google.com/dataflow/docs/concepts/regional-endpoints) that
690 # contains this job.
691 &quot;startTime&quot;: &quot;A String&quot;, # The timestamp when the job was started (transitioned to JOB_STATE_PENDING).
692 # Flexible resource scheduling jobs are started with some delay after job
693 # creation, so start_time is unset before start and is updated when the
694 # job is started by the Cloud Dataflow service. For other jobs, start_time
695 # always equals to create_time and is immutable and set by the Cloud Dataflow
696 # service.
697 &quot;stepsLocation&quot;: &quot;A String&quot;, # The GCS location where the steps are stored.
698 &quot;labels&quot;: { # User-defined labels for this job.
699 #
700 # The labels map can contain no more than 64 entries. Entries of the labels
701 # map are UTF8 strings that comply with the following restrictions:
702 #
703 # * Keys must conform to regexp: \p{Ll}\p{Lo}{0,62}
704 # * Values must conform to regexp: [\p{Ll}\p{Lo}\p{N}_-]{0,63}
705 # * Both keys and values are additionally constrained to be &lt;= 128 bytes in
706 # size.
707 &quot;a_key&quot;: &quot;A String&quot;,
708 },
709 &quot;createTime&quot;: &quot;A String&quot;, # The timestamp when the job was initially created. Immutable and set by the
710 # Cloud Dataflow service.
711 &quot;requestedState&quot;: &quot;A String&quot;, # The job&#x27;s requested state.
712 #
713 # `UpdateJob` may be used to switch between the `JOB_STATE_STOPPED` and
714 # `JOB_STATE_RUNNING` states, by setting requested_state. `UpdateJob` may
715 # also be used to directly set a job&#x27;s requested state to
716 # `JOB_STATE_CANCELLED` or `JOB_STATE_DONE`, irrevocably terminating the
717 # job if it has not already reached a terminal state.
718 },
719 }</pre>
720</div>
721
722</body></html>