blob: ab6e84e13dd41abb576e54ef6308587132431e3e [file] [log] [blame]
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -08001<html><body>
2<style>
3
4body, h1, h2, h3, div, span, p, pre, a {
5 margin: 0;
6 padding: 0;
7 border: 0;
8 font-weight: inherit;
9 font-style: inherit;
10 font-size: 100%;
11 font-family: inherit;
12 vertical-align: baseline;
13}
14
15body {
16 font-size: 13px;
17 padding: 1em;
18}
19
20h1 {
21 font-size: 26px;
22 margin-bottom: 1em;
23}
24
25h2 {
26 font-size: 24px;
27 margin-bottom: 1em;
28}
29
30h3 {
31 font-size: 20px;
32 margin-bottom: 1em;
33 margin-top: 1em;
34}
35
36pre, code {
37 line-height: 1.5;
38 font-family: Monaco, 'DejaVu Sans Mono', 'Bitstream Vera Sans Mono', 'Lucida Console', monospace;
39}
40
41pre {
42 margin-top: 0.5em;
43}
44
45h1, h2, h3, p {
46 font-family: Arial, sans serif;
47}
48
49h1, h2, h3 {
50 border-bottom: solid #CCC 1px;
51}
52
53.toc_element {
54 margin-top: 0.5em;
55}
56
57.firstline {
58 margin-left: 2 em;
59}
60
61.method {
62 margin-top: 1em;
63 border: solid 1px #CCC;
64 padding: 1em;
65 background: #EEE;
66}
67
68.details {
69 font-weight: bold;
70 font-size: 14px;
71}
72
73</style>
74
75<h1><a href="dataproc_v1.html">Google Cloud Dataproc API</a> . <a href="dataproc_v1.projects.html">projects</a> . <a href="dataproc_v1.projects.regions.html">regions</a> . <a href="dataproc_v1.projects.regions.jobs.html">jobs</a></h1>
76<h2>Instance Methods</h2>
77<p class="toc_element">
78 <code><a href="#cancel">cancel(projectId, region, jobId, body, x__xgafv=None)</a></code></p>
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -070079<p class="firstline">Starts a job cancellation request. To access the job resource after cancellation, call [regions/{region}/jobs.list](/dataproc/reference/rest/v1/projects.regions.jobs/list) or [regions/{region}/jobs.get](/dataproc/reference/rest/v1/projects.regions.jobs/get).</p>
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -080080<p class="toc_element">
81 <code><a href="#delete">delete(projectId, region, jobId, x__xgafv=None)</a></code></p>
82<p class="firstline">Deletes the job from the project. If the job is active, the delete fails, and the response returns `FAILED_PRECONDITION`.</p>
83<p class="toc_element">
84 <code><a href="#get">get(projectId, region, jobId, x__xgafv=None)</a></code></p>
85<p class="firstline">Gets the resource representation for a job in a project.</p>
86<p class="toc_element">
Jon Wayne Parrott692617a2017-01-06 09:58:29 -080087 <code><a href="#list">list(projectId, region, pageSize=None, x__xgafv=None, pageToken=None, clusterName=None, jobStateMatcher=None, filter=None)</a></code></p>
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -080088<p class="firstline">Lists regions/{region}/jobs in a project.</p>
89<p class="toc_element">
90 <code><a href="#list_next">list_next(previous_request, previous_response)</a></code></p>
91<p class="firstline">Retrieves the next page of results.</p>
92<p class="toc_element">
93 <code><a href="#submit">submit(projectId, region, body, x__xgafv=None)</a></code></p>
94<p class="firstline">Submits a job to a cluster.</p>
95<h3>Method Details</h3>
96<div class="method">
97 <code class="details" id="cancel">cancel(projectId, region, jobId, body, x__xgafv=None)</code>
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -070098 <pre>Starts a job cancellation request. To access the job resource after cancellation, call [regions/{region}/jobs.list](/dataproc/reference/rest/v1/projects.regions.jobs/list) or [regions/{region}/jobs.get](/dataproc/reference/rest/v1/projects.regions.jobs/get).
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -080099
100Args:
101 projectId: string, [Required] The ID of the Google Cloud Platform project that the job belongs to. (required)
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -0700102 region: string, [Required] The Cloud Dataproc region in which to handle the request. (required)
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800103 jobId: string, [Required] The job ID. (required)
104 body: object, The request body. (required)
105 The object takes the form of:
106
107{ # A request to cancel a job.
108 }
109
110 x__xgafv: string, V1 error format.
111
112Returns:
113 An object of the form:
114
115 { # A Cloud Dataproc job resource.
116 "status": { # Cloud Dataproc job status. # [Output-only] The job status. Additional application-specific status information may be contained in the type_job and yarn_applications fields.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800117 "state": "A String", # [Output-only] A state message specifying the overall job state.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800118 "stateStartTime": "A String", # [Output-only] The time when this state was entered.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800119 "details": "A String", # [Output-only] Optional job state details, such as an error description if the state is ERROR.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800120 },
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800121 "hadoopJob": { # A Cloud Dataproc job for running [Apache Hadoop MapReduce](https://hadoop.apache.org/docs/current/hadoop-mapreduce-client/hadoop-mapreduce-client-core/MapReduceTutorial.html) jobs on [Apache Hadoop YARN](https://hadoop.apache.org/docs/r2.7.1/hadoop-yarn/hadoop-yarn-site/YARN.html). # Job is a Hadoop job.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800122 "jarFileUris": [ # [Optional] Jar file URIs to add to the CLASSPATHs of the Hadoop driver and tasks.
123 "A String",
124 ],
125 "loggingConfig": { # The runtime logging config of the job. # [Optional] The runtime log config for job execution.
126 "driverLogLevels": { # The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
127 "a_key": "A String",
128 },
129 },
130 "args": [ # [Optional] The arguments to pass to the driver. Do not include arguments, such as `-libjars` or `-Dfoo=bar`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
131 "A String",
132 ],
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -0700133 "fileUris": [ # [Optional] HCFS (Hadoop Compatible Filesystem) URIs of files to be copied to the working directory of Hadoop drivers and distributed tasks. Useful for naively parallel tasks.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800134 "A String",
135 ],
136 "mainClass": "A String", # The name of the driver's main class. The jar file containing the class must be in the default CLASSPATH or specified in `jar_file_uris`.
137 "archiveUris": [ # [Optional] HCFS URIs of archives to be extracted in the working directory of Hadoop drivers and tasks. Supported file types: .jar, .tar, .tar.gz, .tgz, or .zip.
138 "A String",
139 ],
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -0700140 "mainJarFileUri": "A String", # The HCFS URI of the jar file containing the main class. Examples: 'gs://foo-bucket/analytics-binaries/extract-useful-metrics-mr.jar' 'hdfs:/tmp/test-samples/custom-wordcount.jar' 'file:///home/usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples.jar'
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800141 "properties": { # [Optional] A mapping of property names to values, used to configure Hadoop. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site and classes in user code.
142 "a_key": "A String",
143 },
144 },
145 "statusHistory": [ # [Output-only] The previous job status.
146 { # Cloud Dataproc job status.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800147 "state": "A String", # [Output-only] A state message specifying the overall job state.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800148 "stateStartTime": "A String", # [Output-only] The time when this state was entered.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800149 "details": "A String", # [Output-only] Optional job state details, such as an error description if the state is ERROR.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800150 },
151 ],
152 "placement": { # Cloud Dataproc job config. # [Required] Job information, including how, when, and where to run the job.
153 "clusterName": "A String", # [Required] The name of the cluster where the job will be submitted.
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -0700154 "clusterUuid": "A String", # [Output-only] A cluster UUID generated by the Cloud Dataproc service when the job is submitted.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800155 },
156 "reference": { # Encapsulates the full scoping used to reference a job. # [Optional] The fully qualified reference to the job, which can be used to obtain the equivalent REST path of the job resource. If this property is not specified when a job is created, the server generates a job_id.
157 "projectId": "A String", # [Required] The ID of the Google Cloud Platform project that the job belongs to.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800158 "jobId": "A String", # [Optional] The job ID, which must be unique within the project. The job ID is generated by the server upon job submission or provided by the user as a means to perform retries without creating duplicate jobs. The ID must contain only letters (a-z, A-Z), numbers (0-9), underscores (_), or hyphens (-). The maximum length is 512 characters.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800159 },
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800160 "sparkSqlJob": { # A Cloud Dataproc job for running [Apache Spark SQL](http://spark.apache.org/sql/) queries. # Job is a SparkSql job.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800161 "queryFileUri": "A String", # The HCFS URI of the script that contains SQL queries.
162 "scriptVariables": { # [Optional] Mapping of query variable names to values (equivalent to the Spark SQL command: SET `name="value";`).
163 "a_key": "A String",
164 },
165 "loggingConfig": { # The runtime logging config of the job. # [Optional] The runtime log config for job execution.
166 "driverLogLevels": { # The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
167 "a_key": "A String",
168 },
169 },
170 "jarFileUris": [ # [Optional] HCFS URIs of jar files to be added to the Spark CLASSPATH.
171 "A String",
172 ],
173 "queryList": { # A list of queries to run on a cluster. # A list of queries.
174 "queries": [ # [Required] The queries to execute. You do not need to terminate a query with a semicolon. Multiple queries can be specified in one string by separating each with a semicolon. Here is an example of an Cloud Dataproc API snippet that uses a QueryList to specify a HiveJob: "hiveJob": { "queryList": { "queries": [ "query1", "query2", "query3;query4", ] } }
175 "A String",
176 ],
177 },
178 "properties": { # [Optional] A mapping of property names to values, used to configure Spark SQL's SparkConf. Properties that conflict with values set by the Cloud Dataproc API may be overwritten.
179 "a_key": "A String",
180 },
181 },
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800182 "pigJob": { # A Cloud Dataproc job for running [Apache Pig](https://pig.apache.org/) queries on YARN. # Job is a Pig job.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800183 "queryFileUri": "A String", # The HCFS URI of the script that contains the Pig queries.
184 "scriptVariables": { # [Optional] Mapping of query variable names to values (equivalent to the Pig command: `name=[value]`).
185 "a_key": "A String",
186 },
187 "loggingConfig": { # The runtime logging config of the job. # [Optional] The runtime log config for job execution.
188 "driverLogLevels": { # The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
189 "a_key": "A String",
190 },
191 },
192 "jarFileUris": [ # [Optional] HCFS URIs of jar files to add to the CLASSPATH of the Pig Client and Hadoop MapReduce (MR) tasks. Can contain Pig UDFs.
193 "A String",
194 ],
195 "queryList": { # A list of queries to run on a cluster. # A list of queries.
196 "queries": [ # [Required] The queries to execute. You do not need to terminate a query with a semicolon. Multiple queries can be specified in one string by separating each with a semicolon. Here is an example of an Cloud Dataproc API snippet that uses a QueryList to specify a HiveJob: "hiveJob": { "queryList": { "queries": [ "query1", "query2", "query3;query4", ] } }
197 "A String",
198 ],
199 },
200 "continueOnFailure": True or False, # [Optional] Whether to continue executing queries if a query fails. The default value is `false`. Setting to `true` can be useful when executing independent parallel queries.
201 "properties": { # [Optional] A mapping of property names to values, used to configure Pig. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, /etc/pig/conf/pig.properties, and classes in user code.
202 "a_key": "A String",
203 },
204 },
205 "driverOutputResourceUri": "A String", # [Output-only] A URI pointing to the location of the stdout of the job's driver program.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800206 "labels": { # [Optional] The labels to associate with this job. Label **keys** must contain 1 to 63 characters, and must conform to [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt). Label **values** may be empty, but, if present, must contain 1 to 63 characters, and must conform to [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt). No more than 32 labels can be associated with a job.
207 "a_key": "A String",
208 },
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800209 "driverControlFilesUri": "A String", # [Output-only] If present, the location of miscellaneous control files which may be used as part of job setup and handling. If not present, control files may be placed in the same location as `driver_output_uri`.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800210 "sparkJob": { # A Cloud Dataproc job for running [Apache Spark](http://spark.apache.org/) applications on YARN. # Job is a Spark job.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800211 "jarFileUris": [ # [Optional] HCFS URIs of jar files to add to the CLASSPATHs of the Spark driver and tasks.
212 "A String",
213 ],
214 "loggingConfig": { # The runtime logging config of the job. # [Optional] The runtime log config for job execution.
215 "driverLogLevels": { # The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
216 "a_key": "A String",
217 },
218 },
219 "args": [ # [Optional] The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
220 "A String",
221 ],
222 "fileUris": [ # [Optional] HCFS URIs of files to be copied to the working directory of Spark drivers and distributed tasks. Useful for naively parallel tasks.
223 "A String",
224 ],
225 "mainClass": "A String", # The name of the driver's main class. The jar file that contains the class must be in the default CLASSPATH or specified in `jar_file_uris`.
226 "archiveUris": [ # [Optional] HCFS URIs of archives to be extracted in the working directory of Spark drivers and tasks. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
227 "A String",
228 ],
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -0700229 "mainJarFileUri": "A String", # The HCFS URI of the jar file that contains the main class.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800230 "properties": { # [Optional] A mapping of property names to values, used to configure Spark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
231 "a_key": "A String",
232 },
233 },
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800234 "yarnApplications": [ # [Output-only] The collection of YARN applications spun up by this job. **Beta** Feature: This report is available for testing purposes only. It may be changed before final release.
235 { # A YARN application created by a job. Application information is a subset of org.apache.hadoop.yarn.proto.YarnProtos.ApplicationReportProto. **Beta Feature**: This report is available for testing purposes only. It may be changed before final release.
236 "progress": 3.14, # [Required] The numerical progress of the application, from 1 to 100.
237 "state": "A String", # [Required] The application state.
238 "name": "A String", # [Required] The application name.
239 "trackingUrl": "A String", # [Optional] The HTTP URL of the ApplicationMaster, HistoryServer, or TimelineServer that provides application-specific information. The URL uses the internal hostname, and requires a proxy server for resolution and, possibly, access.
240 },
241 ],
242 "pysparkJob": { # A Cloud Dataproc job for running [Apache PySpark](https://spark.apache.org/docs/0.9.0/python-programming-guide.html) applications on YARN. # Job is a Pyspark job.
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -0700243 "mainPythonFileUri": "A String", # [Required] The HCFS URI of the main Python file to use as the driver. Must be a .py file.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800244 "args": [ # [Optional] The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
245 "A String",
246 ],
247 "loggingConfig": { # The runtime logging config of the job. # [Optional] The runtime log config for job execution.
248 "driverLogLevels": { # The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
249 "a_key": "A String",
250 },
251 },
252 "jarFileUris": [ # [Optional] HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
253 "A String",
254 ],
255 "fileUris": [ # [Optional] HCFS URIs of files to be copied to the working directory of Python drivers and distributed tasks. Useful for naively parallel tasks.
256 "A String",
257 ],
258 "archiveUris": [ # [Optional] HCFS URIs of archives to be extracted in the working directory of .jar, .tar, .tar.gz, .tgz, and .zip.
259 "A String",
260 ],
261 "pythonFileUris": [ # [Optional] HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
262 "A String",
263 ],
264 "properties": { # [Optional] A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
265 "a_key": "A String",
266 },
267 },
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800268 "hiveJob": { # A Cloud Dataproc job for running [Apache Hive](https://hive.apache.org/) queries on YARN. # Job is a Hive job.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800269 "queryFileUri": "A String", # The HCFS URI of the script that contains Hive queries.
270 "scriptVariables": { # [Optional] Mapping of query variable names to values (equivalent to the Hive command: `SET name="value";`).
271 "a_key": "A String",
272 },
273 "jarFileUris": [ # [Optional] HCFS URIs of jar files to add to the CLASSPATH of the Hive server and Hadoop MapReduce (MR) tasks. Can contain Hive SerDes and UDFs.
274 "A String",
275 ],
276 "queryList": { # A list of queries to run on a cluster. # A list of queries.
277 "queries": [ # [Required] The queries to execute. You do not need to terminate a query with a semicolon. Multiple queries can be specified in one string by separating each with a semicolon. Here is an example of an Cloud Dataproc API snippet that uses a QueryList to specify a HiveJob: "hiveJob": { "queryList": { "queries": [ "query1", "query2", "query3;query4", ] } }
278 "A String",
279 ],
280 },
281 "continueOnFailure": True or False, # [Optional] Whether to continue executing queries if a query fails. The default value is `false`. Setting to `true` can be useful when executing independent parallel queries.
282 "properties": { # [Optional] A mapping of property names and values, used to configure Hive. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, /etc/hive/conf/hive-site.xml, and classes in user code.
283 "a_key": "A String",
284 },
285 },
286 }</pre>
287</div>
288
289<div class="method">
290 <code class="details" id="delete">delete(projectId, region, jobId, x__xgafv=None)</code>
291 <pre>Deletes the job from the project. If the job is active, the delete fails, and the response returns `FAILED_PRECONDITION`.
292
293Args:
294 projectId: string, [Required] The ID of the Google Cloud Platform project that the job belongs to. (required)
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -0700295 region: string, [Required] The Cloud Dataproc region in which to handle the request. (required)
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800296 jobId: string, [Required] The job ID. (required)
297 x__xgafv: string, V1 error format.
298
299Returns:
300 An object of the form:
301
302 { # A generic empty message that you can re-use to avoid defining duplicated empty messages in your APIs. A typical example is to use it as the request or the response type of an API method. For instance: service Foo { rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty); } The JSON representation for `Empty` is empty JSON object `{}`.
303 }</pre>
304</div>
305
306<div class="method">
307 <code class="details" id="get">get(projectId, region, jobId, x__xgafv=None)</code>
308 <pre>Gets the resource representation for a job in a project.
309
310Args:
311 projectId: string, [Required] The ID of the Google Cloud Platform project that the job belongs to. (required)
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -0700312 region: string, [Required] The Cloud Dataproc region in which to handle the request. (required)
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800313 jobId: string, [Required] The job ID. (required)
314 x__xgafv: string, V1 error format.
315
316Returns:
317 An object of the form:
318
319 { # A Cloud Dataproc job resource.
320 "status": { # Cloud Dataproc job status. # [Output-only] The job status. Additional application-specific status information may be contained in the type_job and yarn_applications fields.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800321 "state": "A String", # [Output-only] A state message specifying the overall job state.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800322 "stateStartTime": "A String", # [Output-only] The time when this state was entered.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800323 "details": "A String", # [Output-only] Optional job state details, such as an error description if the state is ERROR.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800324 },
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800325 "hadoopJob": { # A Cloud Dataproc job for running [Apache Hadoop MapReduce](https://hadoop.apache.org/docs/current/hadoop-mapreduce-client/hadoop-mapreduce-client-core/MapReduceTutorial.html) jobs on [Apache Hadoop YARN](https://hadoop.apache.org/docs/r2.7.1/hadoop-yarn/hadoop-yarn-site/YARN.html). # Job is a Hadoop job.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800326 "jarFileUris": [ # [Optional] Jar file URIs to add to the CLASSPATHs of the Hadoop driver and tasks.
327 "A String",
328 ],
329 "loggingConfig": { # The runtime logging config of the job. # [Optional] The runtime log config for job execution.
330 "driverLogLevels": { # The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
331 "a_key": "A String",
332 },
333 },
334 "args": [ # [Optional] The arguments to pass to the driver. Do not include arguments, such as `-libjars` or `-Dfoo=bar`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
335 "A String",
336 ],
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -0700337 "fileUris": [ # [Optional] HCFS (Hadoop Compatible Filesystem) URIs of files to be copied to the working directory of Hadoop drivers and distributed tasks. Useful for naively parallel tasks.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800338 "A String",
339 ],
340 "mainClass": "A String", # The name of the driver's main class. The jar file containing the class must be in the default CLASSPATH or specified in `jar_file_uris`.
341 "archiveUris": [ # [Optional] HCFS URIs of archives to be extracted in the working directory of Hadoop drivers and tasks. Supported file types: .jar, .tar, .tar.gz, .tgz, or .zip.
342 "A String",
343 ],
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -0700344 "mainJarFileUri": "A String", # The HCFS URI of the jar file containing the main class. Examples: 'gs://foo-bucket/analytics-binaries/extract-useful-metrics-mr.jar' 'hdfs:/tmp/test-samples/custom-wordcount.jar' 'file:///home/usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples.jar'
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800345 "properties": { # [Optional] A mapping of property names to values, used to configure Hadoop. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site and classes in user code.
346 "a_key": "A String",
347 },
348 },
349 "statusHistory": [ # [Output-only] The previous job status.
350 { # Cloud Dataproc job status.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800351 "state": "A String", # [Output-only] A state message specifying the overall job state.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800352 "stateStartTime": "A String", # [Output-only] The time when this state was entered.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800353 "details": "A String", # [Output-only] Optional job state details, such as an error description if the state is ERROR.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800354 },
355 ],
356 "placement": { # Cloud Dataproc job config. # [Required] Job information, including how, when, and where to run the job.
357 "clusterName": "A String", # [Required] The name of the cluster where the job will be submitted.
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -0700358 "clusterUuid": "A String", # [Output-only] A cluster UUID generated by the Cloud Dataproc service when the job is submitted.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800359 },
360 "reference": { # Encapsulates the full scoping used to reference a job. # [Optional] The fully qualified reference to the job, which can be used to obtain the equivalent REST path of the job resource. If this property is not specified when a job is created, the server generates a job_id.
361 "projectId": "A String", # [Required] The ID of the Google Cloud Platform project that the job belongs to.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800362 "jobId": "A String", # [Optional] The job ID, which must be unique within the project. The job ID is generated by the server upon job submission or provided by the user as a means to perform retries without creating duplicate jobs. The ID must contain only letters (a-z, A-Z), numbers (0-9), underscores (_), or hyphens (-). The maximum length is 512 characters.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800363 },
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800364 "sparkSqlJob": { # A Cloud Dataproc job for running [Apache Spark SQL](http://spark.apache.org/sql/) queries. # Job is a SparkSql job.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800365 "queryFileUri": "A String", # The HCFS URI of the script that contains SQL queries.
366 "scriptVariables": { # [Optional] Mapping of query variable names to values (equivalent to the Spark SQL command: SET `name="value";`).
367 "a_key": "A String",
368 },
369 "loggingConfig": { # The runtime logging config of the job. # [Optional] The runtime log config for job execution.
370 "driverLogLevels": { # The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
371 "a_key": "A String",
372 },
373 },
374 "jarFileUris": [ # [Optional] HCFS URIs of jar files to be added to the Spark CLASSPATH.
375 "A String",
376 ],
377 "queryList": { # A list of queries to run on a cluster. # A list of queries.
378 "queries": [ # [Required] The queries to execute. You do not need to terminate a query with a semicolon. Multiple queries can be specified in one string by separating each with a semicolon. Here is an example of an Cloud Dataproc API snippet that uses a QueryList to specify a HiveJob: "hiveJob": { "queryList": { "queries": [ "query1", "query2", "query3;query4", ] } }
379 "A String",
380 ],
381 },
382 "properties": { # [Optional] A mapping of property names to values, used to configure Spark SQL's SparkConf. Properties that conflict with values set by the Cloud Dataproc API may be overwritten.
383 "a_key": "A String",
384 },
385 },
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800386 "pigJob": { # A Cloud Dataproc job for running [Apache Pig](https://pig.apache.org/) queries on YARN. # Job is a Pig job.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800387 "queryFileUri": "A String", # The HCFS URI of the script that contains the Pig queries.
388 "scriptVariables": { # [Optional] Mapping of query variable names to values (equivalent to the Pig command: `name=[value]`).
389 "a_key": "A String",
390 },
391 "loggingConfig": { # The runtime logging config of the job. # [Optional] The runtime log config for job execution.
392 "driverLogLevels": { # The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
393 "a_key": "A String",
394 },
395 },
396 "jarFileUris": [ # [Optional] HCFS URIs of jar files to add to the CLASSPATH of the Pig Client and Hadoop MapReduce (MR) tasks. Can contain Pig UDFs.
397 "A String",
398 ],
399 "queryList": { # A list of queries to run on a cluster. # A list of queries.
400 "queries": [ # [Required] The queries to execute. You do not need to terminate a query with a semicolon. Multiple queries can be specified in one string by separating each with a semicolon. Here is an example of an Cloud Dataproc API snippet that uses a QueryList to specify a HiveJob: "hiveJob": { "queryList": { "queries": [ "query1", "query2", "query3;query4", ] } }
401 "A String",
402 ],
403 },
404 "continueOnFailure": True or False, # [Optional] Whether to continue executing queries if a query fails. The default value is `false`. Setting to `true` can be useful when executing independent parallel queries.
405 "properties": { # [Optional] A mapping of property names to values, used to configure Pig. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, /etc/pig/conf/pig.properties, and classes in user code.
406 "a_key": "A String",
407 },
408 },
409 "driverOutputResourceUri": "A String", # [Output-only] A URI pointing to the location of the stdout of the job's driver program.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800410 "labels": { # [Optional] The labels to associate with this job. Label **keys** must contain 1 to 63 characters, and must conform to [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt). Label **values** may be empty, but, if present, must contain 1 to 63 characters, and must conform to [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt). No more than 32 labels can be associated with a job.
411 "a_key": "A String",
412 },
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800413 "driverControlFilesUri": "A String", # [Output-only] If present, the location of miscellaneous control files which may be used as part of job setup and handling. If not present, control files may be placed in the same location as `driver_output_uri`.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800414 "sparkJob": { # A Cloud Dataproc job for running [Apache Spark](http://spark.apache.org/) applications on YARN. # Job is a Spark job.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800415 "jarFileUris": [ # [Optional] HCFS URIs of jar files to add to the CLASSPATHs of the Spark driver and tasks.
416 "A String",
417 ],
418 "loggingConfig": { # The runtime logging config of the job. # [Optional] The runtime log config for job execution.
419 "driverLogLevels": { # The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
420 "a_key": "A String",
421 },
422 },
423 "args": [ # [Optional] The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
424 "A String",
425 ],
426 "fileUris": [ # [Optional] HCFS URIs of files to be copied to the working directory of Spark drivers and distributed tasks. Useful for naively parallel tasks.
427 "A String",
428 ],
429 "mainClass": "A String", # The name of the driver's main class. The jar file that contains the class must be in the default CLASSPATH or specified in `jar_file_uris`.
430 "archiveUris": [ # [Optional] HCFS URIs of archives to be extracted in the working directory of Spark drivers and tasks. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
431 "A String",
432 ],
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -0700433 "mainJarFileUri": "A String", # The HCFS URI of the jar file that contains the main class.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800434 "properties": { # [Optional] A mapping of property names to values, used to configure Spark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
435 "a_key": "A String",
436 },
437 },
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800438 "yarnApplications": [ # [Output-only] The collection of YARN applications spun up by this job. **Beta** Feature: This report is available for testing purposes only. It may be changed before final release.
439 { # A YARN application created by a job. Application information is a subset of org.apache.hadoop.yarn.proto.YarnProtos.ApplicationReportProto. **Beta Feature**: This report is available for testing purposes only. It may be changed before final release.
440 "progress": 3.14, # [Required] The numerical progress of the application, from 1 to 100.
441 "state": "A String", # [Required] The application state.
442 "name": "A String", # [Required] The application name.
443 "trackingUrl": "A String", # [Optional] The HTTP URL of the ApplicationMaster, HistoryServer, or TimelineServer that provides application-specific information. The URL uses the internal hostname, and requires a proxy server for resolution and, possibly, access.
444 },
445 ],
446 "pysparkJob": { # A Cloud Dataproc job for running [Apache PySpark](https://spark.apache.org/docs/0.9.0/python-programming-guide.html) applications on YARN. # Job is a Pyspark job.
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -0700447 "mainPythonFileUri": "A String", # [Required] The HCFS URI of the main Python file to use as the driver. Must be a .py file.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800448 "args": [ # [Optional] The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
449 "A String",
450 ],
451 "loggingConfig": { # The runtime logging config of the job. # [Optional] The runtime log config for job execution.
452 "driverLogLevels": { # The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
453 "a_key": "A String",
454 },
455 },
456 "jarFileUris": [ # [Optional] HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
457 "A String",
458 ],
459 "fileUris": [ # [Optional] HCFS URIs of files to be copied to the working directory of Python drivers and distributed tasks. Useful for naively parallel tasks.
460 "A String",
461 ],
462 "archiveUris": [ # [Optional] HCFS URIs of archives to be extracted in the working directory of .jar, .tar, .tar.gz, .tgz, and .zip.
463 "A String",
464 ],
465 "pythonFileUris": [ # [Optional] HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
466 "A String",
467 ],
468 "properties": { # [Optional] A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
469 "a_key": "A String",
470 },
471 },
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800472 "hiveJob": { # A Cloud Dataproc job for running [Apache Hive](https://hive.apache.org/) queries on YARN. # Job is a Hive job.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800473 "queryFileUri": "A String", # The HCFS URI of the script that contains Hive queries.
474 "scriptVariables": { # [Optional] Mapping of query variable names to values (equivalent to the Hive command: `SET name="value";`).
475 "a_key": "A String",
476 },
477 "jarFileUris": [ # [Optional] HCFS URIs of jar files to add to the CLASSPATH of the Hive server and Hadoop MapReduce (MR) tasks. Can contain Hive SerDes and UDFs.
478 "A String",
479 ],
480 "queryList": { # A list of queries to run on a cluster. # A list of queries.
481 "queries": [ # [Required] The queries to execute. You do not need to terminate a query with a semicolon. Multiple queries can be specified in one string by separating each with a semicolon. Here is an example of an Cloud Dataproc API snippet that uses a QueryList to specify a HiveJob: "hiveJob": { "queryList": { "queries": [ "query1", "query2", "query3;query4", ] } }
482 "A String",
483 ],
484 },
485 "continueOnFailure": True or False, # [Optional] Whether to continue executing queries if a query fails. The default value is `false`. Setting to `true` can be useful when executing independent parallel queries.
486 "properties": { # [Optional] A mapping of property names and values, used to configure Hive. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, /etc/hive/conf/hive-site.xml, and classes in user code.
487 "a_key": "A String",
488 },
489 },
490 }</pre>
491</div>
492
493<div class="method">
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800494 <code class="details" id="list">list(projectId, region, pageSize=None, x__xgafv=None, pageToken=None, clusterName=None, jobStateMatcher=None, filter=None)</code>
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800495 <pre>Lists regions/{region}/jobs in a project.
496
497Args:
498 projectId: string, [Required] The ID of the Google Cloud Platform project that the job belongs to. (required)
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -0700499 region: string, [Required] The Cloud Dataproc region in which to handle the request. (required)
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800500 pageSize: integer, [Optional] The number of results to return in each response.
501 x__xgafv: string, V1 error format.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800502 pageToken: string, [Optional] The page token, returned by a previous call, to request the next page of results.
503 clusterName: string, [Optional] If set, the returned jobs list includes only jobs that were submitted to the named cluster.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800504 jobStateMatcher: string, [Optional] Specifies enumerated categories of jobs to list (default = match ALL jobs).
505 filter: string, [Optional] A filter constraining the jobs to list. Filters are case-sensitive and have the following syntax: field:value] ... or [field = value] AND [field [= value]] ... where **field** is `status.state` or `labels.[KEY]`, and `[KEY]` is a label key. **value** can be `*` to match all values. `status.state` can be either `ACTIVE` or `INACTIVE`. Only the logical `AND` operator is supported; space-separated items are treated as having an implicit `AND` operator. Example valid filters are: status.state:ACTIVE labels.env:staging labels.starred:* and status.state = ACTIVE AND labels.env = staging AND labels.starred = *
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800506
507Returns:
508 An object of the form:
509
510 { # A list of jobs in a project.
511 "nextPageToken": "A String", # [Optional] This token is included in the response if there are more results to fetch. To fetch additional results, provide this value as the `page_token` in a subsequent ListJobsRequest.
512 "jobs": [ # [Output-only] Jobs list.
513 { # A Cloud Dataproc job resource.
514 "status": { # Cloud Dataproc job status. # [Output-only] The job status. Additional application-specific status information may be contained in the type_job and yarn_applications fields.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800515 "state": "A String", # [Output-only] A state message specifying the overall job state.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800516 "stateStartTime": "A String", # [Output-only] The time when this state was entered.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800517 "details": "A String", # [Output-only] Optional job state details, such as an error description if the state is ERROR.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800518 },
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800519 "hadoopJob": { # A Cloud Dataproc job for running [Apache Hadoop MapReduce](https://hadoop.apache.org/docs/current/hadoop-mapreduce-client/hadoop-mapreduce-client-core/MapReduceTutorial.html) jobs on [Apache Hadoop YARN](https://hadoop.apache.org/docs/r2.7.1/hadoop-yarn/hadoop-yarn-site/YARN.html). # Job is a Hadoop job.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800520 "jarFileUris": [ # [Optional] Jar file URIs to add to the CLASSPATHs of the Hadoop driver and tasks.
521 "A String",
522 ],
523 "loggingConfig": { # The runtime logging config of the job. # [Optional] The runtime log config for job execution.
524 "driverLogLevels": { # The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
525 "a_key": "A String",
526 },
527 },
528 "args": [ # [Optional] The arguments to pass to the driver. Do not include arguments, such as `-libjars` or `-Dfoo=bar`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
529 "A String",
530 ],
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -0700531 "fileUris": [ # [Optional] HCFS (Hadoop Compatible Filesystem) URIs of files to be copied to the working directory of Hadoop drivers and distributed tasks. Useful for naively parallel tasks.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800532 "A String",
533 ],
534 "mainClass": "A String", # The name of the driver's main class. The jar file containing the class must be in the default CLASSPATH or specified in `jar_file_uris`.
535 "archiveUris": [ # [Optional] HCFS URIs of archives to be extracted in the working directory of Hadoop drivers and tasks. Supported file types: .jar, .tar, .tar.gz, .tgz, or .zip.
536 "A String",
537 ],
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -0700538 "mainJarFileUri": "A String", # The HCFS URI of the jar file containing the main class. Examples: 'gs://foo-bucket/analytics-binaries/extract-useful-metrics-mr.jar' 'hdfs:/tmp/test-samples/custom-wordcount.jar' 'file:///home/usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples.jar'
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800539 "properties": { # [Optional] A mapping of property names to values, used to configure Hadoop. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site and classes in user code.
540 "a_key": "A String",
541 },
542 },
543 "statusHistory": [ # [Output-only] The previous job status.
544 { # Cloud Dataproc job status.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800545 "state": "A String", # [Output-only] A state message specifying the overall job state.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800546 "stateStartTime": "A String", # [Output-only] The time when this state was entered.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800547 "details": "A String", # [Output-only] Optional job state details, such as an error description if the state is ERROR.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800548 },
549 ],
550 "placement": { # Cloud Dataproc job config. # [Required] Job information, including how, when, and where to run the job.
551 "clusterName": "A String", # [Required] The name of the cluster where the job will be submitted.
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -0700552 "clusterUuid": "A String", # [Output-only] A cluster UUID generated by the Cloud Dataproc service when the job is submitted.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800553 },
554 "reference": { # Encapsulates the full scoping used to reference a job. # [Optional] The fully qualified reference to the job, which can be used to obtain the equivalent REST path of the job resource. If this property is not specified when a job is created, the server generates a job_id.
555 "projectId": "A String", # [Required] The ID of the Google Cloud Platform project that the job belongs to.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800556 "jobId": "A String", # [Optional] The job ID, which must be unique within the project. The job ID is generated by the server upon job submission or provided by the user as a means to perform retries without creating duplicate jobs. The ID must contain only letters (a-z, A-Z), numbers (0-9), underscores (_), or hyphens (-). The maximum length is 512 characters.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800557 },
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800558 "sparkSqlJob": { # A Cloud Dataproc job for running [Apache Spark SQL](http://spark.apache.org/sql/) queries. # Job is a SparkSql job.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800559 "queryFileUri": "A String", # The HCFS URI of the script that contains SQL queries.
560 "scriptVariables": { # [Optional] Mapping of query variable names to values (equivalent to the Spark SQL command: SET `name="value";`).
561 "a_key": "A String",
562 },
563 "loggingConfig": { # The runtime logging config of the job. # [Optional] The runtime log config for job execution.
564 "driverLogLevels": { # The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
565 "a_key": "A String",
566 },
567 },
568 "jarFileUris": [ # [Optional] HCFS URIs of jar files to be added to the Spark CLASSPATH.
569 "A String",
570 ],
571 "queryList": { # A list of queries to run on a cluster. # A list of queries.
572 "queries": [ # [Required] The queries to execute. You do not need to terminate a query with a semicolon. Multiple queries can be specified in one string by separating each with a semicolon. Here is an example of an Cloud Dataproc API snippet that uses a QueryList to specify a HiveJob: "hiveJob": { "queryList": { "queries": [ "query1", "query2", "query3;query4", ] } }
573 "A String",
574 ],
575 },
576 "properties": { # [Optional] A mapping of property names to values, used to configure Spark SQL's SparkConf. Properties that conflict with values set by the Cloud Dataproc API may be overwritten.
577 "a_key": "A String",
578 },
579 },
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800580 "pigJob": { # A Cloud Dataproc job for running [Apache Pig](https://pig.apache.org/) queries on YARN. # Job is a Pig job.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800581 "queryFileUri": "A String", # The HCFS URI of the script that contains the Pig queries.
582 "scriptVariables": { # [Optional] Mapping of query variable names to values (equivalent to the Pig command: `name=[value]`).
583 "a_key": "A String",
584 },
585 "loggingConfig": { # The runtime logging config of the job. # [Optional] The runtime log config for job execution.
586 "driverLogLevels": { # The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
587 "a_key": "A String",
588 },
589 },
590 "jarFileUris": [ # [Optional] HCFS URIs of jar files to add to the CLASSPATH of the Pig Client and Hadoop MapReduce (MR) tasks. Can contain Pig UDFs.
591 "A String",
592 ],
593 "queryList": { # A list of queries to run on a cluster. # A list of queries.
594 "queries": [ # [Required] The queries to execute. You do not need to terminate a query with a semicolon. Multiple queries can be specified in one string by separating each with a semicolon. Here is an example of an Cloud Dataproc API snippet that uses a QueryList to specify a HiveJob: "hiveJob": { "queryList": { "queries": [ "query1", "query2", "query3;query4", ] } }
595 "A String",
596 ],
597 },
598 "continueOnFailure": True or False, # [Optional] Whether to continue executing queries if a query fails. The default value is `false`. Setting to `true` can be useful when executing independent parallel queries.
599 "properties": { # [Optional] A mapping of property names to values, used to configure Pig. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, /etc/pig/conf/pig.properties, and classes in user code.
600 "a_key": "A String",
601 },
602 },
603 "driverOutputResourceUri": "A String", # [Output-only] A URI pointing to the location of the stdout of the job's driver program.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800604 "labels": { # [Optional] The labels to associate with this job. Label **keys** must contain 1 to 63 characters, and must conform to [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt). Label **values** may be empty, but, if present, must contain 1 to 63 characters, and must conform to [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt). No more than 32 labels can be associated with a job.
605 "a_key": "A String",
606 },
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800607 "driverControlFilesUri": "A String", # [Output-only] If present, the location of miscellaneous control files which may be used as part of job setup and handling. If not present, control files may be placed in the same location as `driver_output_uri`.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800608 "sparkJob": { # A Cloud Dataproc job for running [Apache Spark](http://spark.apache.org/) applications on YARN. # Job is a Spark job.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800609 "jarFileUris": [ # [Optional] HCFS URIs of jar files to add to the CLASSPATHs of the Spark driver and tasks.
610 "A String",
611 ],
612 "loggingConfig": { # The runtime logging config of the job. # [Optional] The runtime log config for job execution.
613 "driverLogLevels": { # The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
614 "a_key": "A String",
615 },
616 },
617 "args": [ # [Optional] The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
618 "A String",
619 ],
620 "fileUris": [ # [Optional] HCFS URIs of files to be copied to the working directory of Spark drivers and distributed tasks. Useful for naively parallel tasks.
621 "A String",
622 ],
623 "mainClass": "A String", # The name of the driver's main class. The jar file that contains the class must be in the default CLASSPATH or specified in `jar_file_uris`.
624 "archiveUris": [ # [Optional] HCFS URIs of archives to be extracted in the working directory of Spark drivers and tasks. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
625 "A String",
626 ],
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -0700627 "mainJarFileUri": "A String", # The HCFS URI of the jar file that contains the main class.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800628 "properties": { # [Optional] A mapping of property names to values, used to configure Spark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
629 "a_key": "A String",
630 },
631 },
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800632 "yarnApplications": [ # [Output-only] The collection of YARN applications spun up by this job. **Beta** Feature: This report is available for testing purposes only. It may be changed before final release.
633 { # A YARN application created by a job. Application information is a subset of org.apache.hadoop.yarn.proto.YarnProtos.ApplicationReportProto. **Beta Feature**: This report is available for testing purposes only. It may be changed before final release.
634 "progress": 3.14, # [Required] The numerical progress of the application, from 1 to 100.
635 "state": "A String", # [Required] The application state.
636 "name": "A String", # [Required] The application name.
637 "trackingUrl": "A String", # [Optional] The HTTP URL of the ApplicationMaster, HistoryServer, or TimelineServer that provides application-specific information. The URL uses the internal hostname, and requires a proxy server for resolution and, possibly, access.
638 },
639 ],
640 "pysparkJob": { # A Cloud Dataproc job for running [Apache PySpark](https://spark.apache.org/docs/0.9.0/python-programming-guide.html) applications on YARN. # Job is a Pyspark job.
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -0700641 "mainPythonFileUri": "A String", # [Required] The HCFS URI of the main Python file to use as the driver. Must be a .py file.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800642 "args": [ # [Optional] The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
643 "A String",
644 ],
645 "loggingConfig": { # The runtime logging config of the job. # [Optional] The runtime log config for job execution.
646 "driverLogLevels": { # The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
647 "a_key": "A String",
648 },
649 },
650 "jarFileUris": [ # [Optional] HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
651 "A String",
652 ],
653 "fileUris": [ # [Optional] HCFS URIs of files to be copied to the working directory of Python drivers and distributed tasks. Useful for naively parallel tasks.
654 "A String",
655 ],
656 "archiveUris": [ # [Optional] HCFS URIs of archives to be extracted in the working directory of .jar, .tar, .tar.gz, .tgz, and .zip.
657 "A String",
658 ],
659 "pythonFileUris": [ # [Optional] HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
660 "A String",
661 ],
662 "properties": { # [Optional] A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
663 "a_key": "A String",
664 },
665 },
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800666 "hiveJob": { # A Cloud Dataproc job for running [Apache Hive](https://hive.apache.org/) queries on YARN. # Job is a Hive job.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800667 "queryFileUri": "A String", # The HCFS URI of the script that contains Hive queries.
668 "scriptVariables": { # [Optional] Mapping of query variable names to values (equivalent to the Hive command: `SET name="value";`).
669 "a_key": "A String",
670 },
671 "jarFileUris": [ # [Optional] HCFS URIs of jar files to add to the CLASSPATH of the Hive server and Hadoop MapReduce (MR) tasks. Can contain Hive SerDes and UDFs.
672 "A String",
673 ],
674 "queryList": { # A list of queries to run on a cluster. # A list of queries.
675 "queries": [ # [Required] The queries to execute. You do not need to terminate a query with a semicolon. Multiple queries can be specified in one string by separating each with a semicolon. Here is an example of an Cloud Dataproc API snippet that uses a QueryList to specify a HiveJob: "hiveJob": { "queryList": { "queries": [ "query1", "query2", "query3;query4", ] } }
676 "A String",
677 ],
678 },
679 "continueOnFailure": True or False, # [Optional] Whether to continue executing queries if a query fails. The default value is `false`. Setting to `true` can be useful when executing independent parallel queries.
680 "properties": { # [Optional] A mapping of property names and values, used to configure Hive. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, /etc/hive/conf/hive-site.xml, and classes in user code.
681 "a_key": "A String",
682 },
683 },
684 },
685 ],
686 }</pre>
687</div>
688
689<div class="method">
690 <code class="details" id="list_next">list_next(previous_request, previous_response)</code>
691 <pre>Retrieves the next page of results.
692
693Args:
694 previous_request: The request for the previous page. (required)
695 previous_response: The response from the request for the previous page. (required)
696
697Returns:
698 A request object that you can call 'execute()' on to request the next
699 page. Returns None if there are no more items in the collection.
700 </pre>
701</div>
702
703<div class="method">
704 <code class="details" id="submit">submit(projectId, region, body, x__xgafv=None)</code>
705 <pre>Submits a job to a cluster.
706
707Args:
708 projectId: string, [Required] The ID of the Google Cloud Platform project that the job belongs to. (required)
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -0700709 region: string, [Required] The Cloud Dataproc region in which to handle the request. (required)
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800710 body: object, The request body. (required)
711 The object takes the form of:
712
713{ # A request to submit a job.
714 "job": { # A Cloud Dataproc job resource. # [Required] The job resource.
715 "status": { # Cloud Dataproc job status. # [Output-only] The job status. Additional application-specific status information may be contained in the type_job and yarn_applications fields.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800716 "state": "A String", # [Output-only] A state message specifying the overall job state.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800717 "stateStartTime": "A String", # [Output-only] The time when this state was entered.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800718 "details": "A String", # [Output-only] Optional job state details, such as an error description if the state is ERROR.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800719 },
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800720 "hadoopJob": { # A Cloud Dataproc job for running [Apache Hadoop MapReduce](https://hadoop.apache.org/docs/current/hadoop-mapreduce-client/hadoop-mapreduce-client-core/MapReduceTutorial.html) jobs on [Apache Hadoop YARN](https://hadoop.apache.org/docs/r2.7.1/hadoop-yarn/hadoop-yarn-site/YARN.html). # Job is a Hadoop job.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800721 "jarFileUris": [ # [Optional] Jar file URIs to add to the CLASSPATHs of the Hadoop driver and tasks.
722 "A String",
723 ],
724 "loggingConfig": { # The runtime logging config of the job. # [Optional] The runtime log config for job execution.
725 "driverLogLevels": { # The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
726 "a_key": "A String",
727 },
728 },
729 "args": [ # [Optional] The arguments to pass to the driver. Do not include arguments, such as `-libjars` or `-Dfoo=bar`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
730 "A String",
731 ],
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -0700732 "fileUris": [ # [Optional] HCFS (Hadoop Compatible Filesystem) URIs of files to be copied to the working directory of Hadoop drivers and distributed tasks. Useful for naively parallel tasks.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800733 "A String",
734 ],
735 "mainClass": "A String", # The name of the driver's main class. The jar file containing the class must be in the default CLASSPATH or specified in `jar_file_uris`.
736 "archiveUris": [ # [Optional] HCFS URIs of archives to be extracted in the working directory of Hadoop drivers and tasks. Supported file types: .jar, .tar, .tar.gz, .tgz, or .zip.
737 "A String",
738 ],
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -0700739 "mainJarFileUri": "A String", # The HCFS URI of the jar file containing the main class. Examples: 'gs://foo-bucket/analytics-binaries/extract-useful-metrics-mr.jar' 'hdfs:/tmp/test-samples/custom-wordcount.jar' 'file:///home/usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples.jar'
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800740 "properties": { # [Optional] A mapping of property names to values, used to configure Hadoop. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site and classes in user code.
741 "a_key": "A String",
742 },
743 },
744 "statusHistory": [ # [Output-only] The previous job status.
745 { # Cloud Dataproc job status.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800746 "state": "A String", # [Output-only] A state message specifying the overall job state.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800747 "stateStartTime": "A String", # [Output-only] The time when this state was entered.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800748 "details": "A String", # [Output-only] Optional job state details, such as an error description if the state is ERROR.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800749 },
750 ],
751 "placement": { # Cloud Dataproc job config. # [Required] Job information, including how, when, and where to run the job.
752 "clusterName": "A String", # [Required] The name of the cluster where the job will be submitted.
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -0700753 "clusterUuid": "A String", # [Output-only] A cluster UUID generated by the Cloud Dataproc service when the job is submitted.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800754 },
755 "reference": { # Encapsulates the full scoping used to reference a job. # [Optional] The fully qualified reference to the job, which can be used to obtain the equivalent REST path of the job resource. If this property is not specified when a job is created, the server generates a job_id.
756 "projectId": "A String", # [Required] The ID of the Google Cloud Platform project that the job belongs to.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800757 "jobId": "A String", # [Optional] The job ID, which must be unique within the project. The job ID is generated by the server upon job submission or provided by the user as a means to perform retries without creating duplicate jobs. The ID must contain only letters (a-z, A-Z), numbers (0-9), underscores (_), or hyphens (-). The maximum length is 512 characters.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800758 },
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800759 "sparkSqlJob": { # A Cloud Dataproc job for running [Apache Spark SQL](http://spark.apache.org/sql/) queries. # Job is a SparkSql job.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800760 "queryFileUri": "A String", # The HCFS URI of the script that contains SQL queries.
761 "scriptVariables": { # [Optional] Mapping of query variable names to values (equivalent to the Spark SQL command: SET `name="value";`).
762 "a_key": "A String",
763 },
764 "loggingConfig": { # The runtime logging config of the job. # [Optional] The runtime log config for job execution.
765 "driverLogLevels": { # The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
766 "a_key": "A String",
767 },
768 },
769 "jarFileUris": [ # [Optional] HCFS URIs of jar files to be added to the Spark CLASSPATH.
770 "A String",
771 ],
772 "queryList": { # A list of queries to run on a cluster. # A list of queries.
773 "queries": [ # [Required] The queries to execute. You do not need to terminate a query with a semicolon. Multiple queries can be specified in one string by separating each with a semicolon. Here is an example of an Cloud Dataproc API snippet that uses a QueryList to specify a HiveJob: "hiveJob": { "queryList": { "queries": [ "query1", "query2", "query3;query4", ] } }
774 "A String",
775 ],
776 },
777 "properties": { # [Optional] A mapping of property names to values, used to configure Spark SQL's SparkConf. Properties that conflict with values set by the Cloud Dataproc API may be overwritten.
778 "a_key": "A String",
779 },
780 },
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800781 "pigJob": { # A Cloud Dataproc job for running [Apache Pig](https://pig.apache.org/) queries on YARN. # Job is a Pig job.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800782 "queryFileUri": "A String", # The HCFS URI of the script that contains the Pig queries.
783 "scriptVariables": { # [Optional] Mapping of query variable names to values (equivalent to the Pig command: `name=[value]`).
784 "a_key": "A String",
785 },
786 "loggingConfig": { # The runtime logging config of the job. # [Optional] The runtime log config for job execution.
787 "driverLogLevels": { # The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
788 "a_key": "A String",
789 },
790 },
791 "jarFileUris": [ # [Optional] HCFS URIs of jar files to add to the CLASSPATH of the Pig Client and Hadoop MapReduce (MR) tasks. Can contain Pig UDFs.
792 "A String",
793 ],
794 "queryList": { # A list of queries to run on a cluster. # A list of queries.
795 "queries": [ # [Required] The queries to execute. You do not need to terminate a query with a semicolon. Multiple queries can be specified in one string by separating each with a semicolon. Here is an example of an Cloud Dataproc API snippet that uses a QueryList to specify a HiveJob: "hiveJob": { "queryList": { "queries": [ "query1", "query2", "query3;query4", ] } }
796 "A String",
797 ],
798 },
799 "continueOnFailure": True or False, # [Optional] Whether to continue executing queries if a query fails. The default value is `false`. Setting to `true` can be useful when executing independent parallel queries.
800 "properties": { # [Optional] A mapping of property names to values, used to configure Pig. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, /etc/pig/conf/pig.properties, and classes in user code.
801 "a_key": "A String",
802 },
803 },
804 "driverOutputResourceUri": "A String", # [Output-only] A URI pointing to the location of the stdout of the job's driver program.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800805 "labels": { # [Optional] The labels to associate with this job. Label **keys** must contain 1 to 63 characters, and must conform to [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt). Label **values** may be empty, but, if present, must contain 1 to 63 characters, and must conform to [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt). No more than 32 labels can be associated with a job.
806 "a_key": "A String",
807 },
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800808 "driverControlFilesUri": "A String", # [Output-only] If present, the location of miscellaneous control files which may be used as part of job setup and handling. If not present, control files may be placed in the same location as `driver_output_uri`.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800809 "sparkJob": { # A Cloud Dataproc job for running [Apache Spark](http://spark.apache.org/) applications on YARN. # Job is a Spark job.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800810 "jarFileUris": [ # [Optional] HCFS URIs of jar files to add to the CLASSPATHs of the Spark driver and tasks.
811 "A String",
812 ],
813 "loggingConfig": { # The runtime logging config of the job. # [Optional] The runtime log config for job execution.
814 "driverLogLevels": { # The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
815 "a_key": "A String",
816 },
817 },
818 "args": [ # [Optional] The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
819 "A String",
820 ],
821 "fileUris": [ # [Optional] HCFS URIs of files to be copied to the working directory of Spark drivers and distributed tasks. Useful for naively parallel tasks.
822 "A String",
823 ],
824 "mainClass": "A String", # The name of the driver's main class. The jar file that contains the class must be in the default CLASSPATH or specified in `jar_file_uris`.
825 "archiveUris": [ # [Optional] HCFS URIs of archives to be extracted in the working directory of Spark drivers and tasks. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
826 "A String",
827 ],
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -0700828 "mainJarFileUri": "A String", # The HCFS URI of the jar file that contains the main class.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800829 "properties": { # [Optional] A mapping of property names to values, used to configure Spark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
830 "a_key": "A String",
831 },
832 },
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800833 "yarnApplications": [ # [Output-only] The collection of YARN applications spun up by this job. **Beta** Feature: This report is available for testing purposes only. It may be changed before final release.
834 { # A YARN application created by a job. Application information is a subset of org.apache.hadoop.yarn.proto.YarnProtos.ApplicationReportProto. **Beta Feature**: This report is available for testing purposes only. It may be changed before final release.
835 "progress": 3.14, # [Required] The numerical progress of the application, from 1 to 100.
836 "state": "A String", # [Required] The application state.
837 "name": "A String", # [Required] The application name.
838 "trackingUrl": "A String", # [Optional] The HTTP URL of the ApplicationMaster, HistoryServer, or TimelineServer that provides application-specific information. The URL uses the internal hostname, and requires a proxy server for resolution and, possibly, access.
839 },
840 ],
841 "pysparkJob": { # A Cloud Dataproc job for running [Apache PySpark](https://spark.apache.org/docs/0.9.0/python-programming-guide.html) applications on YARN. # Job is a Pyspark job.
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -0700842 "mainPythonFileUri": "A String", # [Required] The HCFS URI of the main Python file to use as the driver. Must be a .py file.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800843 "args": [ # [Optional] The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
844 "A String",
845 ],
846 "loggingConfig": { # The runtime logging config of the job. # [Optional] The runtime log config for job execution.
847 "driverLogLevels": { # The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
848 "a_key": "A String",
849 },
850 },
851 "jarFileUris": [ # [Optional] HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
852 "A String",
853 ],
854 "fileUris": [ # [Optional] HCFS URIs of files to be copied to the working directory of Python drivers and distributed tasks. Useful for naively parallel tasks.
855 "A String",
856 ],
857 "archiveUris": [ # [Optional] HCFS URIs of archives to be extracted in the working directory of .jar, .tar, .tar.gz, .tgz, and .zip.
858 "A String",
859 ],
860 "pythonFileUris": [ # [Optional] HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
861 "A String",
862 ],
863 "properties": { # [Optional] A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
864 "a_key": "A String",
865 },
866 },
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800867 "hiveJob": { # A Cloud Dataproc job for running [Apache Hive](https://hive.apache.org/) queries on YARN. # Job is a Hive job.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800868 "queryFileUri": "A String", # The HCFS URI of the script that contains Hive queries.
869 "scriptVariables": { # [Optional] Mapping of query variable names to values (equivalent to the Hive command: `SET name="value";`).
870 "a_key": "A String",
871 },
872 "jarFileUris": [ # [Optional] HCFS URIs of jar files to add to the CLASSPATH of the Hive server and Hadoop MapReduce (MR) tasks. Can contain Hive SerDes and UDFs.
873 "A String",
874 ],
875 "queryList": { # A list of queries to run on a cluster. # A list of queries.
876 "queries": [ # [Required] The queries to execute. You do not need to terminate a query with a semicolon. Multiple queries can be specified in one string by separating each with a semicolon. Here is an example of an Cloud Dataproc API snippet that uses a QueryList to specify a HiveJob: "hiveJob": { "queryList": { "queries": [ "query1", "query2", "query3;query4", ] } }
877 "A String",
878 ],
879 },
880 "continueOnFailure": True or False, # [Optional] Whether to continue executing queries if a query fails. The default value is `false`. Setting to `true` can be useful when executing independent parallel queries.
881 "properties": { # [Optional] A mapping of property names and values, used to configure Hive. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, /etc/hive/conf/hive-site.xml, and classes in user code.
882 "a_key": "A String",
883 },
884 },
885 },
886 }
887
888 x__xgafv: string, V1 error format.
889
890Returns:
891 An object of the form:
892
893 { # A Cloud Dataproc job resource.
894 "status": { # Cloud Dataproc job status. # [Output-only] The job status. Additional application-specific status information may be contained in the type_job and yarn_applications fields.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800895 "state": "A String", # [Output-only] A state message specifying the overall job state.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800896 "stateStartTime": "A String", # [Output-only] The time when this state was entered.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800897 "details": "A String", # [Output-only] Optional job state details, such as an error description if the state is ERROR.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800898 },
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800899 "hadoopJob": { # A Cloud Dataproc job for running [Apache Hadoop MapReduce](https://hadoop.apache.org/docs/current/hadoop-mapreduce-client/hadoop-mapreduce-client-core/MapReduceTutorial.html) jobs on [Apache Hadoop YARN](https://hadoop.apache.org/docs/r2.7.1/hadoop-yarn/hadoop-yarn-site/YARN.html). # Job is a Hadoop job.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800900 "jarFileUris": [ # [Optional] Jar file URIs to add to the CLASSPATHs of the Hadoop driver and tasks.
901 "A String",
902 ],
903 "loggingConfig": { # The runtime logging config of the job. # [Optional] The runtime log config for job execution.
904 "driverLogLevels": { # The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
905 "a_key": "A String",
906 },
907 },
908 "args": [ # [Optional] The arguments to pass to the driver. Do not include arguments, such as `-libjars` or `-Dfoo=bar`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
909 "A String",
910 ],
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -0700911 "fileUris": [ # [Optional] HCFS (Hadoop Compatible Filesystem) URIs of files to be copied to the working directory of Hadoop drivers and distributed tasks. Useful for naively parallel tasks.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800912 "A String",
913 ],
914 "mainClass": "A String", # The name of the driver's main class. The jar file containing the class must be in the default CLASSPATH or specified in `jar_file_uris`.
915 "archiveUris": [ # [Optional] HCFS URIs of archives to be extracted in the working directory of Hadoop drivers and tasks. Supported file types: .jar, .tar, .tar.gz, .tgz, or .zip.
916 "A String",
917 ],
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -0700918 "mainJarFileUri": "A String", # The HCFS URI of the jar file containing the main class. Examples: 'gs://foo-bucket/analytics-binaries/extract-useful-metrics-mr.jar' 'hdfs:/tmp/test-samples/custom-wordcount.jar' 'file:///home/usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples.jar'
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800919 "properties": { # [Optional] A mapping of property names to values, used to configure Hadoop. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site and classes in user code.
920 "a_key": "A String",
921 },
922 },
923 "statusHistory": [ # [Output-only] The previous job status.
924 { # Cloud Dataproc job status.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800925 "state": "A String", # [Output-only] A state message specifying the overall job state.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800926 "stateStartTime": "A String", # [Output-only] The time when this state was entered.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800927 "details": "A String", # [Output-only] Optional job state details, such as an error description if the state is ERROR.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800928 },
929 ],
930 "placement": { # Cloud Dataproc job config. # [Required] Job information, including how, when, and where to run the job.
931 "clusterName": "A String", # [Required] The name of the cluster where the job will be submitted.
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -0700932 "clusterUuid": "A String", # [Output-only] A cluster UUID generated by the Cloud Dataproc service when the job is submitted.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800933 },
934 "reference": { # Encapsulates the full scoping used to reference a job. # [Optional] The fully qualified reference to the job, which can be used to obtain the equivalent REST path of the job resource. If this property is not specified when a job is created, the server generates a job_id.
935 "projectId": "A String", # [Required] The ID of the Google Cloud Platform project that the job belongs to.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800936 "jobId": "A String", # [Optional] The job ID, which must be unique within the project. The job ID is generated by the server upon job submission or provided by the user as a means to perform retries without creating duplicate jobs. The ID must contain only letters (a-z, A-Z), numbers (0-9), underscores (_), or hyphens (-). The maximum length is 512 characters.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800937 },
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800938 "sparkSqlJob": { # A Cloud Dataproc job for running [Apache Spark SQL](http://spark.apache.org/sql/) queries. # Job is a SparkSql job.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800939 "queryFileUri": "A String", # The HCFS URI of the script that contains SQL queries.
940 "scriptVariables": { # [Optional] Mapping of query variable names to values (equivalent to the Spark SQL command: SET `name="value";`).
941 "a_key": "A String",
942 },
943 "loggingConfig": { # The runtime logging config of the job. # [Optional] The runtime log config for job execution.
944 "driverLogLevels": { # The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
945 "a_key": "A String",
946 },
947 },
948 "jarFileUris": [ # [Optional] HCFS URIs of jar files to be added to the Spark CLASSPATH.
949 "A String",
950 ],
951 "queryList": { # A list of queries to run on a cluster. # A list of queries.
952 "queries": [ # [Required] The queries to execute. You do not need to terminate a query with a semicolon. Multiple queries can be specified in one string by separating each with a semicolon. Here is an example of an Cloud Dataproc API snippet that uses a QueryList to specify a HiveJob: "hiveJob": { "queryList": { "queries": [ "query1", "query2", "query3;query4", ] } }
953 "A String",
954 ],
955 },
956 "properties": { # [Optional] A mapping of property names to values, used to configure Spark SQL's SparkConf. Properties that conflict with values set by the Cloud Dataproc API may be overwritten.
957 "a_key": "A String",
958 },
959 },
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800960 "pigJob": { # A Cloud Dataproc job for running [Apache Pig](https://pig.apache.org/) queries on YARN. # Job is a Pig job.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800961 "queryFileUri": "A String", # The HCFS URI of the script that contains the Pig queries.
962 "scriptVariables": { # [Optional] Mapping of query variable names to values (equivalent to the Pig command: `name=[value]`).
963 "a_key": "A String",
964 },
965 "loggingConfig": { # The runtime logging config of the job. # [Optional] The runtime log config for job execution.
966 "driverLogLevels": { # The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
967 "a_key": "A String",
968 },
969 },
970 "jarFileUris": [ # [Optional] HCFS URIs of jar files to add to the CLASSPATH of the Pig Client and Hadoop MapReduce (MR) tasks. Can contain Pig UDFs.
971 "A String",
972 ],
973 "queryList": { # A list of queries to run on a cluster. # A list of queries.
974 "queries": [ # [Required] The queries to execute. You do not need to terminate a query with a semicolon. Multiple queries can be specified in one string by separating each with a semicolon. Here is an example of an Cloud Dataproc API snippet that uses a QueryList to specify a HiveJob: "hiveJob": { "queryList": { "queries": [ "query1", "query2", "query3;query4", ] } }
975 "A String",
976 ],
977 },
978 "continueOnFailure": True or False, # [Optional] Whether to continue executing queries if a query fails. The default value is `false`. Setting to `true` can be useful when executing independent parallel queries.
979 "properties": { # [Optional] A mapping of property names to values, used to configure Pig. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, /etc/pig/conf/pig.properties, and classes in user code.
980 "a_key": "A String",
981 },
982 },
983 "driverOutputResourceUri": "A String", # [Output-only] A URI pointing to the location of the stdout of the job's driver program.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800984 "labels": { # [Optional] The labels to associate with this job. Label **keys** must contain 1 to 63 characters, and must conform to [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt). Label **values** may be empty, but, if present, must contain 1 to 63 characters, and must conform to [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt). No more than 32 labels can be associated with a job.
985 "a_key": "A String",
986 },
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800987 "driverControlFilesUri": "A String", # [Output-only] If present, the location of miscellaneous control files which may be used as part of job setup and handling. If not present, control files may be placed in the same location as `driver_output_uri`.
Jon Wayne Parrott692617a2017-01-06 09:58:29 -0800988 "sparkJob": { # A Cloud Dataproc job for running [Apache Spark](http://spark.apache.org/) applications on YARN. # Job is a Spark job.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -0800989 "jarFileUris": [ # [Optional] HCFS URIs of jar files to add to the CLASSPATHs of the Spark driver and tasks.
990 "A String",
991 ],
992 "loggingConfig": { # The runtime logging config of the job. # [Optional] The runtime log config for job execution.
993 "driverLogLevels": { # The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
994 "a_key": "A String",
995 },
996 },
997 "args": [ # [Optional] The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
998 "A String",
999 ],
1000 "fileUris": [ # [Optional] HCFS URIs of files to be copied to the working directory of Spark drivers and distributed tasks. Useful for naively parallel tasks.
1001 "A String",
1002 ],
1003 "mainClass": "A String", # The name of the driver's main class. The jar file that contains the class must be in the default CLASSPATH or specified in `jar_file_uris`.
1004 "archiveUris": [ # [Optional] HCFS URIs of archives to be extracted in the working directory of Spark drivers and tasks. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
1005 "A String",
1006 ],
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -07001007 "mainJarFileUri": "A String", # The HCFS URI of the jar file that contains the main class.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -08001008 "properties": { # [Optional] A mapping of property names to values, used to configure Spark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
1009 "a_key": "A String",
1010 },
1011 },
Jon Wayne Parrott692617a2017-01-06 09:58:29 -08001012 "yarnApplications": [ # [Output-only] The collection of YARN applications spun up by this job. **Beta** Feature: This report is available for testing purposes only. It may be changed before final release.
1013 { # A YARN application created by a job. Application information is a subset of org.apache.hadoop.yarn.proto.YarnProtos.ApplicationReportProto. **Beta Feature**: This report is available for testing purposes only. It may be changed before final release.
1014 "progress": 3.14, # [Required] The numerical progress of the application, from 1 to 100.
1015 "state": "A String", # [Required] The application state.
1016 "name": "A String", # [Required] The application name.
1017 "trackingUrl": "A String", # [Optional] The HTTP URL of the ApplicationMaster, HistoryServer, or TimelineServer that provides application-specific information. The URL uses the internal hostname, and requires a proxy server for resolution and, possibly, access.
1018 },
1019 ],
1020 "pysparkJob": { # A Cloud Dataproc job for running [Apache PySpark](https://spark.apache.org/docs/0.9.0/python-programming-guide.html) applications on YARN. # Job is a Pyspark job.
Jon Wayne Parrott0a471d32016-05-19 10:54:38 -07001021 "mainPythonFileUri": "A String", # [Required] The HCFS URI of the main Python file to use as the driver. Must be a .py file.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -08001022 "args": [ # [Optional] The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
1023 "A String",
1024 ],
1025 "loggingConfig": { # The runtime logging config of the job. # [Optional] The runtime log config for job execution.
1026 "driverLogLevels": { # The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
1027 "a_key": "A String",
1028 },
1029 },
1030 "jarFileUris": [ # [Optional] HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
1031 "A String",
1032 ],
1033 "fileUris": [ # [Optional] HCFS URIs of files to be copied to the working directory of Python drivers and distributed tasks. Useful for naively parallel tasks.
1034 "A String",
1035 ],
1036 "archiveUris": [ # [Optional] HCFS URIs of archives to be extracted in the working directory of .jar, .tar, .tar.gz, .tgz, and .zip.
1037 "A String",
1038 ],
1039 "pythonFileUris": [ # [Optional] HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
1040 "A String",
1041 ],
1042 "properties": { # [Optional] A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
1043 "a_key": "A String",
1044 },
1045 },
Jon Wayne Parrott692617a2017-01-06 09:58:29 -08001046 "hiveJob": { # A Cloud Dataproc job for running [Apache Hive](https://hive.apache.org/) queries on YARN. # Job is a Hive job.
Jon Wayne Parrott36e41bc2016-02-19 16:02:29 -08001047 "queryFileUri": "A String", # The HCFS URI of the script that contains Hive queries.
1048 "scriptVariables": { # [Optional] Mapping of query variable names to values (equivalent to the Hive command: `SET name="value";`).
1049 "a_key": "A String",
1050 },
1051 "jarFileUris": [ # [Optional] HCFS URIs of jar files to add to the CLASSPATH of the Hive server and Hadoop MapReduce (MR) tasks. Can contain Hive SerDes and UDFs.
1052 "A String",
1053 ],
1054 "queryList": { # A list of queries to run on a cluster. # A list of queries.
1055 "queries": [ # [Required] The queries to execute. You do not need to terminate a query with a semicolon. Multiple queries can be specified in one string by separating each with a semicolon. Here is an example of an Cloud Dataproc API snippet that uses a QueryList to specify a HiveJob: "hiveJob": { "queryList": { "queries": [ "query1", "query2", "query3;query4", ] } }
1056 "A String",
1057 ],
1058 },
1059 "continueOnFailure": True or False, # [Optional] Whether to continue executing queries if a query fails. The default value is `false`. Setting to `true` can be useful when executing independent parallel queries.
1060 "properties": { # [Optional] A mapping of property names and values, used to configure Hive. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, /etc/hive/conf/hive-site.xml, and classes in user code.
1061 "a_key": "A String",
1062 },
1063 },
1064 }</pre>
1065</div>
1066
1067</body></html>