docs: docs update (#911)

Thank you for opening a Pull Request! Before submitting your PR, there are a few things you can do to make sure it goes smoothly:
- [ ] Make sure to open an issue as a [bug/issue](https://github.com/googleapis/google-api-python-client/issues/new/choose) before writing your code!  That way we can discuss the change, evaluate designs, and agree on the general idea
- [ ] Ensure the tests and linter pass
- [ ] Code coverage does not decrease (if any source code was changed)
- [ ] Appropriate docs were updated (if necessary)

Fixes #<issue_number_goes_here> 🦕
diff --git a/docs/dyn/ml_v1.projects.jobs.html b/docs/dyn/ml_v1.projects.jobs.html
index a45bbe5..979bc29 100644
--- a/docs/dyn/ml_v1.projects.jobs.html
+++ b/docs/dyn/ml_v1.projects.jobs.html
@@ -87,7 +87,7 @@
   <code><a href="#getIamPolicy">getIamPolicy(resource, options_requestedPolicyVersion=None, x__xgafv=None)</a></code></p>
 <p class="firstline">Gets the access control policy for a resource.</p>
 <p class="toc_element">
-  <code><a href="#list">list(parent, pageSize=None, pageToken=None, x__xgafv=None, filter=None)</a></code></p>
+  <code><a href="#list">list(parent, pageToken=None, pageSize=None, filter=None, x__xgafv=None)</a></code></p>
 <p class="firstline">Lists the jobs in the project.</p>
 <p class="toc_element">
   <code><a href="#list_next">list_next(previous_request, previous_response)</a></code></p>
@@ -144,699 +144,256 @@
     The object takes the form of:
 
 { # Represents a training or prediction job.
-  "errorMessage": "A String", # Output only. The details of a failure or a cancellation.
-  "trainingOutput": { # Represents results of a training job. Output only. # The current training job result.
-    "completedTrialCount": "A String", # The number of hyperparameter tuning trials that completed successfully.
-        # Only set for hyperparameter tuning jobs.
-    "trials": [ # Results for individual Hyperparameter trials.
-        # Only set for hyperparameter tuning jobs.
-      { # Represents the result of a single hyperparameter tuning trial from a
-          # training job. The TrainingOutput object that is returned on successful
-          # completion of a training job with hyperparameter tuning includes a list
-          # of HyperparameterOutput objects, one for each successful trial.
-        "startTime": "A String", # Output only. Start time for the trial.
-        "hyperparameters": { # The hyperparameters given to this trial.
-          "a_key": "A String",
-        },
-        "finalMetric": { # An observed value of a metric. # The final objective metric seen for this trial.
-          "trainingStep": "A String", # The global training step for this metric.
-          "objectiveValue": 3.14, # The objective value at this training step.
-        },
-        "state": "A String", # Output only. The detailed state of the trial.
-        "allMetrics": [ # All recorded object metrics for this trial. This field is not currently
-            # populated.
-          { # An observed value of a metric.
-            "trainingStep": "A String", # The global training step for this metric.
-            "objectiveValue": 3.14, # The objective value at this training step.
-          },
-        ],
-        "isTrialStoppedEarly": True or False, # True if the trial is stopped early.
-        "endTime": "A String", # Output only. End time for the trial.
-        "trialId": "A String", # The trial id for these results.
-        "builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
-            # Only set for trials of built-in algorithms jobs that have succeeded.
-          "framework": "A String", # Framework on which the built-in algorithm was trained.
-          "modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job
-              # saves the trained model. Only set for successful jobs that don't use
-              # hyperparameter tuning.
-          "runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was
-              # trained.
-          "pythonVersion": "A String", # Python version on which the built-in algorithm was trained.
-        },
-      },
-    ],
-    "isHyperparameterTuningJob": True or False, # Whether this job is a hyperparameter tuning job.
-    "isBuiltInAlgorithmJob": True or False, # Whether this job is a built-in Algorithm job.
-    "consumedMLUnits": 3.14, # The amount of ML units consumed by the job.
-    "hyperparameterMetricTag": "A String", # The TensorFlow summary tag name used for optimizing hyperparameter tuning
-        # trials. See
-        # [`HyperparameterSpec.hyperparameterMetricTag`](#HyperparameterSpec.FIELDS.hyperparameter_metric_tag)
-        # for more information. Only set for hyperparameter tuning jobs.
-    "builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
-        # Only set for built-in algorithms jobs.
-      "framework": "A String", # Framework on which the built-in algorithm was trained.
-      "modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job
-          # saves the trained model. Only set for successful jobs that don't use
-          # hyperparameter tuning.
-      "runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was
-          # trained.
-      "pythonVersion": "A String", # Python version on which the built-in algorithm was trained.
-    },
-  },
-  "predictionInput": { # Represents input parameters for a prediction job. # Input parameters to create a prediction job.
-    "modelName": "A String", # Use this field if you want to use the default version for the specified
-        # model. The string must use the following format:
-        #
-        # `"projects/YOUR_PROJECT/models/YOUR_MODEL"`
-    "signatureName": "A String", # Optional. The name of the signature defined in the SavedModel to use for
-        # this job. Please refer to
-        # [SavedModel](https://tensorflow.github.io/serving/serving_basic.html)
-        # for information about how to use signatures.
-        #
-        # Defaults to
-        # [DEFAULT_SERVING_SIGNATURE_DEF_KEY](https://www.tensorflow.org/api_docs/python/tf/saved_model/signature_constants)
-        # , which is "serving_default".
-    "runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for this batch
-        # prediction. If not set, AI Platform will pick the runtime version used
-        # during the CreateVersion request for this model version, or choose the
-        # latest stable version when model version information is not available
-        # such as when the model is specified by uri.
-    "batchSize": "A String", # Optional. Number of records per batch, defaults to 64.
-        # The service will buffer batch_size number of records in memory before
-        # invoking one Tensorflow prediction call internally. So take the record
-        # size and memory available into consideration when setting this parameter.
-    "maxWorkerCount": "A String", # Optional. The maximum number of workers to be used for parallel processing.
-        # Defaults to 10 if not specified.
-    "uri": "A String", # Use this field if you want to specify a Google Cloud Storage path for
-        # the model to use.
-    "outputPath": "A String", # Required. The output Google Cloud Storage location.
-    "dataFormat": "A String", # Required. The format of the input data files.
-    "versionName": "A String", # Use this field if you want to specify a version of the model to use. The
-        # string is formatted the same way as `model_version`, with the addition
-        # of the version information:
-        #
-        # `"projects/YOUR_PROJECT/models/YOUR_MODEL/versions/YOUR_VERSION"`
-    "region": "A String", # Required. The Google Compute Engine region to run the prediction job in.
-        # See the &lt;a href="/ml-engine/docs/tensorflow/regions"&gt;available regions&lt;/a&gt;
-        # for AI Platform services.
-    "inputPaths": [ # Required. The Cloud Storage location of the input data files. May contain
-        # &lt;a href="/storage/docs/gsutil/addlhelp/WildcardNames"&gt;wildcards&lt;/a&gt;.
-      "A String",
-    ],
-    "outputDataFormat": "A String", # Optional. Format of the output data files, defaults to JSON.
-  },
-  "labels": { # Optional. One or more labels that you can add, to organize your jobs.
-      # Each label is a key-value pair, where both the key and the value are
-      # arbitrary strings that you supply.
-      # For more information, see the documentation on
-      # &lt;a href="/ml-engine/docs/tensorflow/resource-labels"&gt;using labels&lt;/a&gt;.
-    "a_key": "A String",
-  },
-  "jobId": "A String", # Required. The user-specified id of the job.
-  "state": "A String", # Output only. The detailed state of a job.
-  "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help
-      # prevent simultaneous updates of a job from overwriting each other.
-      # It is strongly suggested that systems make use of the `etag` in the
-      # read-modify-write cycle to perform job updates in order to avoid race
-      # conditions: An `etag` is returned in the response to `GetJob`, and
-      # systems are expected to put that etag in the request to `UpdateJob` to
-      # ensure that their change will be applied to the same version of the job.
-  "startTime": "A String", # Output only. When the job processing was started.
-  "trainingInput": { # Represents input parameters for a training job. When using the gcloud command # Input parameters to create a training job.
-      # to submit your training job, you can specify the input parameters as
-      # command-line arguments and/or in a YAML configuration file referenced from
-      # the --config command-line argument. For details, see the guide to [submitting
-      # a training job](/ai-platform/training/docs/training-jobs).
-    "masterType": "A String", # Optional. Specifies the type of virtual machine to use for your training
-        # job's master worker. You must specify this field when `scaleTier` is set to
-        # `CUSTOM`.
-        #
-        # You can use certain Compute Engine machine types directly in this field.
-        # The following types are supported:
-        #
-        # - `n1-standard-4`
-        # - `n1-standard-8`
-        # - `n1-standard-16`
-        # - `n1-standard-32`
-        # - `n1-standard-64`
-        # - `n1-standard-96`
-        # - `n1-highmem-2`
-        # - `n1-highmem-4`
-        # - `n1-highmem-8`
-        # - `n1-highmem-16`
-        # - `n1-highmem-32`
-        # - `n1-highmem-64`
-        # - `n1-highmem-96`
-        # - `n1-highcpu-16`
-        # - `n1-highcpu-32`
-        # - `n1-highcpu-64`
-        # - `n1-highcpu-96`
-        #
-        # Learn more about [using Compute Engine machine
-        # types](/ml-engine/docs/machine-types#compute-engine-machine-types).
-        #
-        # Alternatively, you can use the following legacy machine types:
-        #
-        # - `standard`
-        # - `large_model`
-        # - `complex_model_s`
-        # - `complex_model_m`
-        # - `complex_model_l`
-        # - `standard_gpu`
-        # - `complex_model_m_gpu`
-        # - `complex_model_l_gpu`
-        # - `standard_p100`
-        # - `complex_model_m_p100`
-        # - `standard_v100`
-        # - `large_model_v100`
-        # - `complex_model_m_v100`
-        # - `complex_model_l_v100`
-        #
-        # Learn more about [using legacy machine
-        # types](/ml-engine/docs/machine-types#legacy-machine-types).
-        #
-        # Finally, if you want to use a TPU for training, specify `cloud_tpu` in this
-        # field. Learn more about the [special configuration options for training
-        # with
-        # TPUs](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine).
-    "jobDir": "A String", # Optional. A Google Cloud Storage path in which to store training outputs
-        # and other data needed for training. This path is passed to your TensorFlow
-        # program as the '--job-dir' command-line argument. The benefit of specifying
-        # this field is that Cloud ML validates the path for use in training.
-    "scheduling": { # All parameters related to scheduling of training jobs. # Optional. Scheduling options for a training job.
-      "maxRunningTime": "A String", # Optional. The maximum job running time, expressed in seconds. The field can
-          # contain up to nine fractional digits, terminated by `s`. By default there
-          # is no limit to the running time.
-          #
-          # If the training job is still running after this duration, AI Platform
-          # Training cancels it.
-          #
-          # For example, if you want to ensure your job runs for no more than 2 hours,
-          # set this field to `7200s` (2 hours * 60 minutes / hour * 60 seconds /
-          # minute).
-          #
-          # If you submit your training job using the `gcloud` tool, you can [provide
-          # this field in a `config.yaml`
-          # file](/ai-platform/training/docs/training-jobs#formatting_your_configuration_parameters).
-          # For example:
-          #
-          # ```yaml
-          # trainingInput:
-          #   ...
-          #   scheduling:
-          #     maxRunningTime: 7200s
-          #   ...
-          # ```
-    },
-    "parameterServerCount": "A String", # Optional. The number of parameter server replicas to use for the training
-        # job. Each replica in the cluster will be of the type specified in
-        # `parameter_server_type`.
-        #
-        # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
-        # set this value, you must also set `parameter_server_type`.
-        #
-        # The default value is zero.
-    "evaluatorCount": "A String", # Optional. The number of evaluator replicas to use for the training job.
-        # Each replica in the cluster will be of the type specified in
-        # `evaluator_type`.
-        #
-        # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
-        # set this value, you must also set `evaluator_type`.
-        #
-        # The default value is zero.
-    "workerType": "A String", # Optional. Specifies the type of virtual machine to use for your training
-        # job's worker nodes.
-        #
-        # The supported values are the same as those described in the entry for
-        # `masterType`.
-        #
-        # This value must be consistent with the category of machine type that
-        # `masterType` uses. In other words, both must be Compute Engine machine
-        # types or both must be legacy machine types.
-        #
-        # If you use `cloud_tpu` for this value, see special instructions for
-        # [configuring a custom TPU
-        # machine](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine).
-        #
-        # This value must be present when `scaleTier` is set to `CUSTOM` and
-        # `workerCount` is greater than zero.
-    "scaleTier": "A String", # Required. Specifies the machine types, the number of replicas for workers
-        # and parameter servers.
-    "packageUris": [ # Required. The Google Cloud Storage location of the packages with
-        # the training program and any additional dependencies.
-        # The maximum number of package URIs is 100.
-      "A String",
-    ],
-    "workerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for workers.
-        #
-        # You should only set `workerConfig.acceleratorConfig` if `workerType` is set
-        # to a Compute Engine machine type. [Learn about restrictions on accelerator
-        # configurations for
-        # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-        #
-        # Set `workerConfig.imageUri` only if you build a custom image for your
-        # worker. If `workerConfig.imageUri` has not been set, AI Platform uses
-        # the value of `masterConfig.imageUri`. Learn more about [configuring custom
-        # containers](/ai-platform/training/docs/distributed-training-containers).
-      "tpuTfVersion": "A String", # The AI Platform runtime version that includes a TensorFlow version matching
-          # the one used in the custom container. This field is required if the replica
-          # is a TPU worker that uses a custom container. Otherwise, do not specify
-          # this field. This must be a [runtime version that currently supports
-          # training with
-          # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
-          #
-          # Note that the version of TensorFlow included in a runtime version may
-          # differ from the numbering of the runtime version itself, because it may
-          # have a different [patch
-          # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
-          # In this field, you must specify the runtime version (TensorFlow minor
-          # version). For example, if your custom container runs TensorFlow `1.x.y`,
-          # specify `1.x`.
-      "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
-          # [Learn about restrictions on accelerator configurations for
-          # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-          # Note that the AcceleratorConfig can be used in both Jobs and Versions.
-          # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
-          # [accelerators for online
-          # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
-        "count": "A String", # The number of accelerators to attach to each machine running the job.
-        "type": "A String", # The type of accelerator to use.
-      },
-      "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
-          # Registry. Learn more about [configuring custom
-          # containers](/ai-platform/training/docs/distributed-training-containers).
-    },
-    "evaluatorConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for evaluators.
-        #
-        # You should only set `evaluatorConfig.acceleratorConfig` if
-        # `evaluatorType` is set to a Compute Engine machine type. [Learn
-        # about restrictions on accelerator configurations for
-        # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-        #
-        # Set `evaluatorConfig.imageUri` only if you build a custom image for
-        # your evaluator. If `evaluatorConfig.imageUri` has not been
-        # set, AI Platform uses the value of `masterConfig.imageUri`. Learn more about [configuring custom
-        # containers](/ai-platform/training/docs/distributed-training-containers).
-      "tpuTfVersion": "A String", # The AI Platform runtime version that includes a TensorFlow version matching
-          # the one used in the custom container. This field is required if the replica
-          # is a TPU worker that uses a custom container. Otherwise, do not specify
-          # this field. This must be a [runtime version that currently supports
-          # training with
-          # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
-          #
-          # Note that the version of TensorFlow included in a runtime version may
-          # differ from the numbering of the runtime version itself, because it may
-          # have a different [patch
-          # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
-          # In this field, you must specify the runtime version (TensorFlow minor
-          # version). For example, if your custom container runs TensorFlow `1.x.y`,
-          # specify `1.x`.
-      "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
-          # [Learn about restrictions on accelerator configurations for
-          # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-          # Note that the AcceleratorConfig can be used in both Jobs and Versions.
-          # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
-          # [accelerators for online
-          # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
-        "count": "A String", # The number of accelerators to attach to each machine running the job.
-        "type": "A String", # The type of accelerator to use.
-      },
-      "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
-          # Registry. Learn more about [configuring custom
-          # containers](/ai-platform/training/docs/distributed-training-containers).
-    },
-    "useChiefInTfConfig": True or False, # Optional. Use `chief` instead of `master` in the `TF_CONFIG` environment
-        # variable when training with a custom container. Defaults to `false`. [Learn
-        # more about this
-        # field.](/ai-platform/training/docs/distributed-training-details#chief-versus-master)
-        #
-        # This field has no effect for training jobs that don't use a custom
-        # container.
-    "masterConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for your master worker.
-        #
-        # You should only set `masterConfig.acceleratorConfig` if `masterType` is set
-        # to a Compute Engine machine type. Learn about [restrictions on accelerator
-        # configurations for
-        # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-        #
-        # Set `masterConfig.imageUri` only if you build a custom image. Only one of
-        # `masterConfig.imageUri` and `runtimeVersion` should be set. Learn more
-        # about [configuring custom
-        # containers](/ai-platform/training/docs/distributed-training-containers).
-      "tpuTfVersion": "A String", # The AI Platform runtime version that includes a TensorFlow version matching
-          # the one used in the custom container. This field is required if the replica
-          # is a TPU worker that uses a custom container. Otherwise, do not specify
-          # this field. This must be a [runtime version that currently supports
-          # training with
-          # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
-          #
-          # Note that the version of TensorFlow included in a runtime version may
-          # differ from the numbering of the runtime version itself, because it may
-          # have a different [patch
-          # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
-          # In this field, you must specify the runtime version (TensorFlow minor
-          # version). For example, if your custom container runs TensorFlow `1.x.y`,
-          # specify `1.x`.
-      "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
-          # [Learn about restrictions on accelerator configurations for
-          # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-          # Note that the AcceleratorConfig can be used in both Jobs and Versions.
-          # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
-          # [accelerators for online
-          # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
-        "count": "A String", # The number of accelerators to attach to each machine running the job.
-        "type": "A String", # The type of accelerator to use.
-      },
-      "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
-          # Registry. Learn more about [configuring custom
-          # containers](/ai-platform/training/docs/distributed-training-containers).
-    },
-    "runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for training. You must
-        # either specify this field or specify `masterConfig.imageUri`.
-        #
-        # For more information, see the [runtime version
-        # list](/ai-platform/training/docs/runtime-version-list) and learn [how to
-        # manage runtime versions](/ai-platform/training/docs/versioning).
-    "hyperparameters": { # Represents a set of hyperparameters to optimize. # Optional. The set of Hyperparameters to tune.
-      "maxTrials": 42, # Optional. How many training trials should be attempted to optimize
-          # the specified hyperparameters.
-          #
-          # Defaults to one.
-      "goal": "A String", # Required. The type of goal to use for tuning. Available types are
-          # `MAXIMIZE` and `MINIMIZE`.
-          #
-          # Defaults to `MAXIMIZE`.
-      "algorithm": "A String", # Optional. The search algorithm specified for the hyperparameter
-          # tuning job.
-          # Uses the default AI Platform hyperparameter tuning
-          # algorithm if unspecified.
-      "maxFailedTrials": 42, # Optional. The number of failed trials that need to be seen before failing
-          # the hyperparameter tuning job. You can specify this field to override the
-          # default failing criteria for AI Platform hyperparameter tuning jobs.
-          #
-          # Defaults to zero, which means the service decides when a hyperparameter
-          # job should fail.
-      "enableTrialEarlyStopping": True or False, # Optional. Indicates if the hyperparameter tuning job enables auto trial
-          # early stopping.
-      "resumePreviousJobId": "A String", # Optional. The prior hyperparameter tuning job id that users hope to
-          # continue with. The job id will be used to find the corresponding vizier
-          # study guid and resume the study.
-      "params": [ # Required. The set of parameters to tune.
-        { # Represents a single hyperparameter to optimize.
-          "maxValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
-              # should be unset if type is `CATEGORICAL`. This value should be integers if
-              # type is `INTEGER`.
-          "minValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
-              # should be unset if type is `CATEGORICAL`. This value should be integers if
-              # type is INTEGER.
-          "discreteValues": [ # Required if type is `DISCRETE`.
-              # A list of feasible points.
-              # The list should be in strictly increasing order. For instance, this
-              # parameter might have possible settings of 1.5, 2.5, and 4.0. This list
-              # should not contain more than 1,000 values.
-            3.14,
-          ],
-          "parameterName": "A String", # Required. The parameter name must be unique amongst all ParameterConfigs in
-              # a HyperparameterSpec message. E.g., "learning_rate".
-          "categoricalValues": [ # Required if type is `CATEGORICAL`. The list of possible categories.
-            "A String",
-          ],
-          "type": "A String", # Required. The type of the parameter.
-          "scaleType": "A String", # Optional. How the parameter should be scaled to the hypercube.
-              # Leave unset for categorical parameters.
-              # Some kind of scaling is strongly recommended for real or integral
-              # parameters (e.g., `UNIT_LINEAR_SCALE`).
-        },
-      ],
-      "hyperparameterMetricTag": "A String", # Optional. The TensorFlow summary tag name to use for optimizing trials. For
-          # current versions of TensorFlow, this tag name should exactly match what is
-          # shown in TensorBoard, including all scopes.  For versions of TensorFlow
-          # prior to 0.12, this should be only the tag passed to tf.Summary.
-          # By default, "training/hptuning/metric" will be used.
-      "maxParallelTrials": 42, # Optional. The number of training trials to run concurrently.
-          # You can reduce the time it takes to perform hyperparameter tuning by adding
-          # trials in parallel. However, each trail only benefits from the information
-          # gained in completed trials. That means that a trial does not get access to
-          # the results of trials running at the same time, which could reduce the
-          # quality of the overall optimization.
-          #
-          # Each trial will use the same scale tier and machine types.
-          #
-          # Defaults to one.
-    },
-    "args": [ # Optional. Command-line arguments passed to the training application when it
-        # starts. If your job uses a custom container, then the arguments are passed
-        # to the container's &lt;a class="external" target="_blank"
-        # href="https://docs.docker.com/engine/reference/builder/#entrypoint"&gt;
-        # `ENTRYPOINT`&lt;/a&gt; command.
-      "A String",
-    ],
-    "pythonModule": "A String", # Required. The Python module name to run after installing the packages.
-    "workerCount": "A String", # Optional. The number of worker replicas to use for the training job. Each
-        # replica in the cluster will be of the type specified in `worker_type`.
-        #
-        # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
-        # set this value, you must also set `worker_type`.
-        #
-        # The default value is zero.
-    "encryptionConfig": { # Represents a custom encryption key configuration that can be applied to # Optional. Options for using customer-managed encryption keys (CMEK) to
-        # protect resources created by a training job, instead of using Google's
-        # default encryption. If this is set, then all resources created by the
-        # training job will be encrypted with the customer-managed encryption key
-        # that you specify.
-        #
-        # [Learn how and when to use CMEK with AI Platform
-        # Training](/ai-platform/training/docs/cmek).
-        # a resource.
-      "kmsKeyName": "A String", # The Cloud KMS resource identifier of the customer-managed encryption key
-          # used to protect a resource, such as a training job. It has the following
-          # format:
-          # `projects/{PROJECT_ID}/locations/{REGION}/keyRings/{KEY_RING_NAME}/cryptoKeys/{KEY_NAME}`
-    },
-    "parameterServerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for parameter servers.
-        #
-        # You should only set `parameterServerConfig.acceleratorConfig` if
-        # `parameterServerType` is set to a Compute Engine machine type. [Learn
-        # about restrictions on accelerator configurations for
-        # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-        #
-        # Set `parameterServerConfig.imageUri` only if you build a custom image for
-        # your parameter server. If `parameterServerConfig.imageUri` has not been
-        # set, AI Platform uses the value of `masterConfig.imageUri`. Learn more about [configuring custom
-        # containers](/ai-platform/training/docs/distributed-training-containers).
-      "tpuTfVersion": "A String", # The AI Platform runtime version that includes a TensorFlow version matching
-          # the one used in the custom container. This field is required if the replica
-          # is a TPU worker that uses a custom container. Otherwise, do not specify
-          # this field. This must be a [runtime version that currently supports
-          # training with
-          # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
-          #
-          # Note that the version of TensorFlow included in a runtime version may
-          # differ from the numbering of the runtime version itself, because it may
-          # have a different [patch
-          # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
-          # In this field, you must specify the runtime version (TensorFlow minor
-          # version). For example, if your custom container runs TensorFlow `1.x.y`,
-          # specify `1.x`.
-      "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
-          # [Learn about restrictions on accelerator configurations for
-          # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-          # Note that the AcceleratorConfig can be used in both Jobs and Versions.
-          # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
-          # [accelerators for online
-          # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
-        "count": "A String", # The number of accelerators to attach to each machine running the job.
-        "type": "A String", # The type of accelerator to use.
-      },
-      "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
-          # Registry. Learn more about [configuring custom
-          # containers](/ai-platform/training/docs/distributed-training-containers).
-    },
-    "region": "A String", # Required. The region to run the training job in. See the [available
-        # regions](/ai-platform/training/docs/regions) for AI Platform Training.
-    "pythonVersion": "A String", # Optional. The version of Python used in training. You must either specify
-        # this field or specify `masterConfig.imageUri`.
-        #
-        # The following Python versions are available:
-        #
-        # * Python '3.7' is available when `runtime_version` is set to '1.15' or
-        #   later.
-        # * Python '3.5' is available when `runtime_version` is set to a version
-        #   from '1.4' to '1.14'.
-        # * Python '2.7' is available when `runtime_version` is set to '1.15' or
-        #   earlier.
-        #
-        # Read more about the Python versions available for [each runtime
-        # version](/ml-engine/docs/runtime-version-list).
-    "evaluatorType": "A String", # Optional. Specifies the type of virtual machine to use for your training
-        # job's evaluator nodes.
-        #
-        # The supported values are the same as those described in the entry for
-        # `masterType`.
-        #
-        # This value must be consistent with the category of machine type that
-        # `masterType` uses. In other words, both must be Compute Engine machine
-        # types or both must be legacy machine types.
-        #
-        # This value must be present when `scaleTier` is set to `CUSTOM` and
-        # `evaluatorCount` is greater than zero.
-    "parameterServerType": "A String", # Optional. Specifies the type of virtual machine to use for your training
-        # job's parameter server.
-        #
-        # The supported values are the same as those described in the entry for
-        # `master_type`.
-        #
-        # This value must be consistent with the category of machine type that
-        # `masterType` uses. In other words, both must be Compute Engine machine
-        # types or both must be legacy machine types.
-        #
-        # This value must be present when `scaleTier` is set to `CUSTOM` and
-        # `parameter_server_count` is greater than zero.
-  },
-  "endTime": "A String", # Output only. When the job processing was completed.
-  "predictionOutput": { # Represents results of a prediction job. # The current prediction job result.
-    "outputPath": "A String", # The output Google Cloud Storage location provided at the job creation time.
-    "nodeHours": 3.14, # Node hours used by the batch prediction job.
-    "predictionCount": "A String", # The number of generated predictions.
-    "errorCount": "A String", # The number of data instances which resulted in errors.
-  },
-  "createTime": "A String", # Output only. When the job was created.
-}
-
-  x__xgafv: string, V1 error format.
-    Allowed values
-      1 - v1 error format
-      2 - v2 error format
-
-Returns:
-  An object of the form:
-
-    { # Represents a training or prediction job.
-    "errorMessage": "A String", # Output only. The details of a failure or a cancellation.
-    "trainingOutput": { # Represents results of a training job. Output only. # The current training job result.
-      "completedTrialCount": "A String", # The number of hyperparameter tuning trials that completed successfully.
-          # Only set for hyperparameter tuning jobs.
-      "trials": [ # Results for individual Hyperparameter trials.
-          # Only set for hyperparameter tuning jobs.
-        { # Represents the result of a single hyperparameter tuning trial from a
-            # training job. The TrainingOutput object that is returned on successful
-            # completion of a training job with hyperparameter tuning includes a list
-            # of HyperparameterOutput objects, one for each successful trial.
-          "startTime": "A String", # Output only. Start time for the trial.
-          "hyperparameters": { # The hyperparameters given to this trial.
-            "a_key": "A String",
-          },
-          "finalMetric": { # An observed value of a metric. # The final objective metric seen for this trial.
-            "trainingStep": "A String", # The global training step for this metric.
-            "objectiveValue": 3.14, # The objective value at this training step.
-          },
-          "state": "A String", # Output only. The detailed state of the trial.
-          "allMetrics": [ # All recorded object metrics for this trial. This field is not currently
-              # populated.
-            { # An observed value of a metric.
-              "trainingStep": "A String", # The global training step for this metric.
-              "objectiveValue": 3.14, # The objective value at this training step.
-            },
-          ],
-          "isTrialStoppedEarly": True or False, # True if the trial is stopped early.
-          "endTime": "A String", # Output only. End time for the trial.
-          "trialId": "A String", # The trial id for these results.
-          "builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
-              # Only set for trials of built-in algorithms jobs that have succeeded.
-            "framework": "A String", # Framework on which the built-in algorithm was trained.
-            "modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job
-                # saves the trained model. Only set for successful jobs that don't use
-                # hyperparameter tuning.
-            "runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was
-                # trained.
-            "pythonVersion": "A String", # Python version on which the built-in algorithm was trained.
-          },
-        },
-      ],
-      "isHyperparameterTuningJob": True or False, # Whether this job is a hyperparameter tuning job.
-      "isBuiltInAlgorithmJob": True or False, # Whether this job is a built-in Algorithm job.
-      "consumedMLUnits": 3.14, # The amount of ML units consumed by the job.
-      "hyperparameterMetricTag": "A String", # The TensorFlow summary tag name used for optimizing hyperparameter tuning
-          # trials. See
-          # [`HyperparameterSpec.hyperparameterMetricTag`](#HyperparameterSpec.FIELDS.hyperparameter_metric_tag)
-          # for more information. Only set for hyperparameter tuning jobs.
-      "builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
-          # Only set for built-in algorithms jobs.
-        "framework": "A String", # Framework on which the built-in algorithm was trained.
-        "modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job
-            # saves the trained model. Only set for successful jobs that don't use
-            # hyperparameter tuning.
-        "runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was
-            # trained.
-        "pythonVersion": "A String", # Python version on which the built-in algorithm was trained.
-      },
-    },
-    "predictionInput": { # Represents input parameters for a prediction job. # Input parameters to create a prediction job.
-      "modelName": "A String", # Use this field if you want to use the default version for the specified
-          # model. The string must use the following format:
-          #
-          # `"projects/YOUR_PROJECT/models/YOUR_MODEL"`
-      "signatureName": "A String", # Optional. The name of the signature defined in the SavedModel to use for
-          # this job. Please refer to
-          # [SavedModel](https://tensorflow.github.io/serving/serving_basic.html)
-          # for information about how to use signatures.
-          #
-          # Defaults to
-          # [DEFAULT_SERVING_SIGNATURE_DEF_KEY](https://www.tensorflow.org/api_docs/python/tf/saved_model/signature_constants)
-          # , which is "serving_default".
-      "runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for this batch
-          # prediction. If not set, AI Platform will pick the runtime version used
-          # during the CreateVersion request for this model version, or choose the
-          # latest stable version when model version information is not available
-          # such as when the model is specified by uri.
-      "batchSize": "A String", # Optional. Number of records per batch, defaults to 64.
-          # The service will buffer batch_size number of records in memory before
-          # invoking one Tensorflow prediction call internally. So take the record
-          # size and memory available into consideration when setting this parameter.
-      "maxWorkerCount": "A String", # Optional. The maximum number of workers to be used for parallel processing.
-          # Defaults to 10 if not specified.
-      "uri": "A String", # Use this field if you want to specify a Google Cloud Storage path for
-          # the model to use.
-      "outputPath": "A String", # Required. The output Google Cloud Storage location.
-      "dataFormat": "A String", # Required. The format of the input data files.
-      "versionName": "A String", # Use this field if you want to specify a version of the model to use. The
-          # string is formatted the same way as `model_version`, with the addition
-          # of the version information:
-          #
-          # `"projects/YOUR_PROJECT/models/YOUR_MODEL/versions/YOUR_VERSION"`
-      "region": "A String", # Required. The Google Compute Engine region to run the prediction job in.
-          # See the &lt;a href="/ml-engine/docs/tensorflow/regions"&gt;available regions&lt;/a&gt;
-          # for AI Platform services.
-      "inputPaths": [ # Required. The Cloud Storage location of the input data files. May contain
-          # &lt;a href="/storage/docs/gsutil/addlhelp/WildcardNames"&gt;wildcards&lt;/a&gt;.
-        "A String",
-      ],
-      "outputDataFormat": "A String", # Optional. Format of the output data files, defaults to JSON.
-    },
-    "labels": { # Optional. One or more labels that you can add, to organize your jobs.
-        # Each label is a key-value pair, where both the key and the value are
-        # arbitrary strings that you supply.
-        # For more information, see the documentation on
-        # &lt;a href="/ml-engine/docs/tensorflow/resource-labels"&gt;using labels&lt;/a&gt;.
-      "a_key": "A String",
-    },
-    "jobId": "A String", # Required. The user-specified id of the job.
-    "state": "A String", # Output only. The detailed state of a job.
-    "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help
+    &quot;etag&quot;: &quot;A String&quot;, # `etag` is used for optimistic concurrency control as a way to help
         # prevent simultaneous updates of a job from overwriting each other.
         # It is strongly suggested that systems make use of the `etag` in the
         # read-modify-write cycle to perform job updates in order to avoid race
         # conditions: An `etag` is returned in the response to `GetJob`, and
         # systems are expected to put that etag in the request to `UpdateJob` to
         # ensure that their change will be applied to the same version of the job.
-    "startTime": "A String", # Output only. When the job processing was started.
-    "trainingInput": { # Represents input parameters for a training job. When using the gcloud command # Input parameters to create a training job.
+    &quot;trainingInput&quot;: { # Represents input parameters for a training job. When using the gcloud command # Input parameters to create a training job.
         # to submit your training job, you can specify the input parameters as
         # command-line arguments and/or in a YAML configuration file referenced from
         # the --config command-line argument. For details, see the guide to [submitting
         # a training job](/ai-platform/training/docs/training-jobs).
-      "masterType": "A String", # Optional. Specifies the type of virtual machine to use for your training
-          # job's master worker. You must specify this field when `scaleTier` is set to
+      &quot;parameterServerConfig&quot;: { # Represents the configuration for a replica in a cluster. # Optional. The configuration for parameter servers.
+          #
+          # You should only set `parameterServerConfig.acceleratorConfig` if
+          # `parameterServerType` is set to a Compute Engine machine type. [Learn
+          # about restrictions on accelerator configurations for
+          # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+          #
+          # Set `parameterServerConfig.imageUri` only if you build a custom image for
+          # your parameter server. If `parameterServerConfig.imageUri` has not been
+          # set, AI Platform uses the value of `masterConfig.imageUri`. Learn more about [configuring custom
+          # containers](/ai-platform/training/docs/distributed-training-containers).
+        &quot;tpuTfVersion&quot;: &quot;A String&quot;, # The AI Platform runtime version that includes a TensorFlow version matching
+            # the one used in the custom container. This field is required if the replica
+            # is a TPU worker that uses a custom container. Otherwise, do not specify
+            # this field. This must be a [runtime version that currently supports
+            # training with
+            # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
+            #
+            # Note that the version of TensorFlow included in a runtime version may
+            # differ from the numbering of the runtime version itself, because it may
+            # have a different [patch
+            # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
+            # In this field, you must specify the runtime version (TensorFlow minor
+            # version). For example, if your custom container runs TensorFlow `1.x.y`,
+            # specify `1.x`.
+        &quot;containerCommand&quot;: [ # The command with which the replica&#x27;s custom container is run.
+            # If provided, it will override default ENTRYPOINT of the docker image.
+            # If not provided, the docker image&#x27;s ENTRYPOINT is used.
+            # It cannot be set if custom container image is
+            # not provided.
+            # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+            # both cannot be set at the same time.
+          &quot;A String&quot;,
+        ],
+        &quot;imageUri&quot;: &quot;A String&quot;, # The Docker image to run on the replica. This image must be in Container
+            # Registry. Learn more about [configuring custom
+            # containers](/ai-platform/training/docs/distributed-training-containers).
+        &quot;containerArgs&quot;: [ # Arguments to the entrypoint command.
+            # The following rules apply for container_command and container_args:
+            # - If you do not supply command or args:
+            #   The defaults defined in the Docker image are used.
+            # - If you supply a command but no args:
+            #   The default EntryPoint and the default Cmd defined in the Docker image
+            #   are ignored. Your command is run without any arguments.
+            # - If you supply only args:
+            #   The default Entrypoint defined in the Docker image is run with the args
+            #   that you supplied.
+            # - If you supply a command and args:
+            #   The default Entrypoint and the default Cmd defined in the Docker image
+            #   are ignored. Your command is run with your args.
+            # It cannot be set if custom container image is
+            # not provided.
+            # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+            # both cannot be set at the same time.
+          &quot;A String&quot;,
+        ],
+        &quot;acceleratorConfig&quot;: { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
+            # [Learn about restrictions on accelerator configurations for
+            # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+            # Note that the AcceleratorConfig can be used in both Jobs and Versions.
+            # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
+            # [accelerators for online
+            # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
+          &quot;count&quot;: &quot;A String&quot;, # The number of accelerators to attach to each machine running the job.
+          &quot;type&quot;: &quot;A String&quot;, # The type of accelerator to use.
+        },
+      },
+      &quot;encryptionConfig&quot;: { # Represents a custom encryption key configuration that can be applied to # Optional. Options for using customer-managed encryption keys (CMEK) to
+          # protect resources created by a training job, instead of using Google&#x27;s
+          # default encryption. If this is set, then all resources created by the
+          # training job will be encrypted with the customer-managed encryption key
+          # that you specify.
+          #
+          # [Learn how and when to use CMEK with AI Platform
+          # Training](/ai-platform/training/docs/cmek).
+          # a resource.
+        &quot;kmsKeyName&quot;: &quot;A String&quot;, # The Cloud KMS resource identifier of the customer-managed encryption key
+            # used to protect a resource, such as a training job. It has the following
+            # format:
+            # `projects/{PROJECT_ID}/locations/{REGION}/keyRings/{KEY_RING_NAME}/cryptoKeys/{KEY_NAME}`
+      },
+      &quot;hyperparameters&quot;: { # Represents a set of hyperparameters to optimize. # Optional. The set of Hyperparameters to tune.
+        &quot;hyperparameterMetricTag&quot;: &quot;A String&quot;, # Optional. The TensorFlow summary tag name to use for optimizing trials. For
+            # current versions of TensorFlow, this tag name should exactly match what is
+            # shown in TensorBoard, including all scopes.  For versions of TensorFlow
+            # prior to 0.12, this should be only the tag passed to tf.Summary.
+            # By default, &quot;training/hptuning/metric&quot; will be used.
+        &quot;params&quot;: [ # Required. The set of parameters to tune.
+          { # Represents a single hyperparameter to optimize.
+            &quot;type&quot;: &quot;A String&quot;, # Required. The type of the parameter.
+            &quot;categoricalValues&quot;: [ # Required if type is `CATEGORICAL`. The list of possible categories.
+              &quot;A String&quot;,
+            ],
+            &quot;parameterName&quot;: &quot;A String&quot;, # Required. The parameter name must be unique amongst all ParameterConfigs in
+                # a HyperparameterSpec message. E.g., &quot;learning_rate&quot;.
+            &quot;minValue&quot;: 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
+                # should be unset if type is `CATEGORICAL`. This value should be integers if
+                # type is INTEGER.
+            &quot;discreteValues&quot;: [ # Required if type is `DISCRETE`.
+                # A list of feasible points.
+                # The list should be in strictly increasing order. For instance, this
+                # parameter might have possible settings of 1.5, 2.5, and 4.0. This list
+                # should not contain more than 1,000 values.
+              3.14,
+            ],
+            &quot;scaleType&quot;: &quot;A String&quot;, # Optional. How the parameter should be scaled to the hypercube.
+                # Leave unset for categorical parameters.
+                # Some kind of scaling is strongly recommended for real or integral
+                # parameters (e.g., `UNIT_LINEAR_SCALE`).
+            &quot;maxValue&quot;: 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
+                # should be unset if type is `CATEGORICAL`. This value should be integers if
+                # type is `INTEGER`.
+          },
+        ],
+        &quot;enableTrialEarlyStopping&quot;: True or False, # Optional. Indicates if the hyperparameter tuning job enables auto trial
+            # early stopping.
+        &quot;resumePreviousJobId&quot;: &quot;A String&quot;, # Optional. The prior hyperparameter tuning job id that users hope to
+            # continue with. The job id will be used to find the corresponding vizier
+            # study guid and resume the study.
+        &quot;maxParallelTrials&quot;: 42, # Optional. The number of training trials to run concurrently.
+            # You can reduce the time it takes to perform hyperparameter tuning by adding
+            # trials in parallel. However, each trail only benefits from the information
+            # gained in completed trials. That means that a trial does not get access to
+            # the results of trials running at the same time, which could reduce the
+            # quality of the overall optimization.
+            #
+            # Each trial will use the same scale tier and machine types.
+            #
+            # Defaults to one.
+        &quot;maxFailedTrials&quot;: 42, # Optional. The number of failed trials that need to be seen before failing
+            # the hyperparameter tuning job. You can specify this field to override the
+            # default failing criteria for AI Platform hyperparameter tuning jobs.
+            #
+            # Defaults to zero, which means the service decides when a hyperparameter
+            # job should fail.
+        &quot;goal&quot;: &quot;A String&quot;, # Required. The type of goal to use for tuning. Available types are
+            # `MAXIMIZE` and `MINIMIZE`.
+            #
+            # Defaults to `MAXIMIZE`.
+        &quot;maxTrials&quot;: 42, # Optional. How many training trials should be attempted to optimize
+            # the specified hyperparameters.
+            #
+            # Defaults to one.
+        &quot;algorithm&quot;: &quot;A String&quot;, # Optional. The search algorithm specified for the hyperparameter
+            # tuning job.
+            # Uses the default AI Platform hyperparameter tuning
+            # algorithm if unspecified.
+      },
+      &quot;workerConfig&quot;: { # Represents the configuration for a replica in a cluster. # Optional. The configuration for workers.
+          #
+          # You should only set `workerConfig.acceleratorConfig` if `workerType` is set
+          # to a Compute Engine machine type. [Learn about restrictions on accelerator
+          # configurations for
+          # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+          #
+          # Set `workerConfig.imageUri` only if you build a custom image for your
+          # worker. If `workerConfig.imageUri` has not been set, AI Platform uses
+          # the value of `masterConfig.imageUri`. Learn more about [configuring custom
+          # containers](/ai-platform/training/docs/distributed-training-containers).
+        &quot;tpuTfVersion&quot;: &quot;A String&quot;, # The AI Platform runtime version that includes a TensorFlow version matching
+            # the one used in the custom container. This field is required if the replica
+            # is a TPU worker that uses a custom container. Otherwise, do not specify
+            # this field. This must be a [runtime version that currently supports
+            # training with
+            # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
+            #
+            # Note that the version of TensorFlow included in a runtime version may
+            # differ from the numbering of the runtime version itself, because it may
+            # have a different [patch
+            # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
+            # In this field, you must specify the runtime version (TensorFlow minor
+            # version). For example, if your custom container runs TensorFlow `1.x.y`,
+            # specify `1.x`.
+        &quot;containerCommand&quot;: [ # The command with which the replica&#x27;s custom container is run.
+            # If provided, it will override default ENTRYPOINT of the docker image.
+            # If not provided, the docker image&#x27;s ENTRYPOINT is used.
+            # It cannot be set if custom container image is
+            # not provided.
+            # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+            # both cannot be set at the same time.
+          &quot;A String&quot;,
+        ],
+        &quot;imageUri&quot;: &quot;A String&quot;, # The Docker image to run on the replica. This image must be in Container
+            # Registry. Learn more about [configuring custom
+            # containers](/ai-platform/training/docs/distributed-training-containers).
+        &quot;containerArgs&quot;: [ # Arguments to the entrypoint command.
+            # The following rules apply for container_command and container_args:
+            # - If you do not supply command or args:
+            #   The defaults defined in the Docker image are used.
+            # - If you supply a command but no args:
+            #   The default EntryPoint and the default Cmd defined in the Docker image
+            #   are ignored. Your command is run without any arguments.
+            # - If you supply only args:
+            #   The default Entrypoint defined in the Docker image is run with the args
+            #   that you supplied.
+            # - If you supply a command and args:
+            #   The default Entrypoint and the default Cmd defined in the Docker image
+            #   are ignored. Your command is run with your args.
+            # It cannot be set if custom container image is
+            # not provided.
+            # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+            # both cannot be set at the same time.
+          &quot;A String&quot;,
+        ],
+        &quot;acceleratorConfig&quot;: { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
+            # [Learn about restrictions on accelerator configurations for
+            # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+            # Note that the AcceleratorConfig can be used in both Jobs and Versions.
+            # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
+            # [accelerators for online
+            # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
+          &quot;count&quot;: &quot;A String&quot;, # The number of accelerators to attach to each machine running the job.
+          &quot;type&quot;: &quot;A String&quot;, # The type of accelerator to use.
+        },
+      },
+      &quot;parameterServerCount&quot;: &quot;A String&quot;, # Optional. The number of parameter server replicas to use for the training
+          # job. Each replica in the cluster will be of the type specified in
+          # `parameter_server_type`.
+          #
+          # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
+          # set this value, you must also set `parameter_server_type`.
+          #
+          # The default value is zero.
+      &quot;packageUris&quot;: [ # Required. The Google Cloud Storage location of the packages with
+          # the training program and any additional dependencies.
+          # The maximum number of package URIs is 100.
+        &quot;A String&quot;,
+      ],
+      &quot;evaluatorCount&quot;: &quot;A String&quot;, # Optional. The number of evaluator replicas to use for the training job.
+          # Each replica in the cluster will be of the type specified in
+          # `evaluator_type`.
+          #
+          # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
+          # set this value, you must also set `evaluator_type`.
+          #
+          # The default value is zero.
+      &quot;masterType&quot;: &quot;A String&quot;, # Optional. Specifies the type of virtual machine to use for your training
+          # job&#x27;s master worker. You must specify this field when `scaleTier` is set to
           # `CUSTOM`.
           #
           # You can use certain Compute Engine machine types directly in this field.
@@ -887,14 +444,156 @@
           # field. Learn more about the [special configuration options for training
           # with
           # TPUs](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine).
-      "jobDir": "A String", # Optional. A Google Cloud Storage path in which to store training outputs
+      &quot;runtimeVersion&quot;: &quot;A String&quot;, # Optional. The AI Platform runtime version to use for training. You must
+          # either specify this field or specify `masterConfig.imageUri`.
+          #
+          # For more information, see the [runtime version
+          # list](/ai-platform/training/docs/runtime-version-list) and learn [how to
+          # manage runtime versions](/ai-platform/training/docs/versioning).
+      &quot;evaluatorType&quot;: &quot;A String&quot;, # Optional. Specifies the type of virtual machine to use for your training
+          # job&#x27;s evaluator nodes.
+          #
+          # The supported values are the same as those described in the entry for
+          # `masterType`.
+          #
+          # This value must be consistent with the category of machine type that
+          # `masterType` uses. In other words, both must be Compute Engine machine
+          # types or both must be legacy machine types.
+          #
+          # This value must be present when `scaleTier` is set to `CUSTOM` and
+          # `evaluatorCount` is greater than zero.
+      &quot;region&quot;: &quot;A String&quot;, # Required. The region to run the training job in. See the [available
+          # regions](/ai-platform/training/docs/regions) for AI Platform Training.
+      &quot;workerType&quot;: &quot;A String&quot;, # Optional. Specifies the type of virtual machine to use for your training
+          # job&#x27;s worker nodes.
+          #
+          # The supported values are the same as those described in the entry for
+          # `masterType`.
+          #
+          # This value must be consistent with the category of machine type that
+          # `masterType` uses. In other words, both must be Compute Engine machine
+          # types or both must be legacy machine types.
+          #
+          # If you use `cloud_tpu` for this value, see special instructions for
+          # [configuring a custom TPU
+          # machine](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine).
+          #
+          # This value must be present when `scaleTier` is set to `CUSTOM` and
+          # `workerCount` is greater than zero.
+      &quot;parameterServerType&quot;: &quot;A String&quot;, # Optional. Specifies the type of virtual machine to use for your training
+          # job&#x27;s parameter server.
+          #
+          # The supported values are the same as those described in the entry for
+          # `master_type`.
+          #
+          # This value must be consistent with the category of machine type that
+          # `masterType` uses. In other words, both must be Compute Engine machine
+          # types or both must be legacy machine types.
+          #
+          # This value must be present when `scaleTier` is set to `CUSTOM` and
+          # `parameter_server_count` is greater than zero.
+      &quot;masterConfig&quot;: { # Represents the configuration for a replica in a cluster. # Optional. The configuration for your master worker.
+          #
+          # You should only set `masterConfig.acceleratorConfig` if `masterType` is set
+          # to a Compute Engine machine type. Learn about [restrictions on accelerator
+          # configurations for
+          # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+          #
+          # Set `masterConfig.imageUri` only if you build a custom image. Only one of
+          # `masterConfig.imageUri` and `runtimeVersion` should be set. Learn more
+          # about [configuring custom
+          # containers](/ai-platform/training/docs/distributed-training-containers).
+        &quot;tpuTfVersion&quot;: &quot;A String&quot;, # The AI Platform runtime version that includes a TensorFlow version matching
+            # the one used in the custom container. This field is required if the replica
+            # is a TPU worker that uses a custom container. Otherwise, do not specify
+            # this field. This must be a [runtime version that currently supports
+            # training with
+            # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
+            #
+            # Note that the version of TensorFlow included in a runtime version may
+            # differ from the numbering of the runtime version itself, because it may
+            # have a different [patch
+            # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
+            # In this field, you must specify the runtime version (TensorFlow minor
+            # version). For example, if your custom container runs TensorFlow `1.x.y`,
+            # specify `1.x`.
+        &quot;containerCommand&quot;: [ # The command with which the replica&#x27;s custom container is run.
+            # If provided, it will override default ENTRYPOINT of the docker image.
+            # If not provided, the docker image&#x27;s ENTRYPOINT is used.
+            # It cannot be set if custom container image is
+            # not provided.
+            # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+            # both cannot be set at the same time.
+          &quot;A String&quot;,
+        ],
+        &quot;imageUri&quot;: &quot;A String&quot;, # The Docker image to run on the replica. This image must be in Container
+            # Registry. Learn more about [configuring custom
+            # containers](/ai-platform/training/docs/distributed-training-containers).
+        &quot;containerArgs&quot;: [ # Arguments to the entrypoint command.
+            # The following rules apply for container_command and container_args:
+            # - If you do not supply command or args:
+            #   The defaults defined in the Docker image are used.
+            # - If you supply a command but no args:
+            #   The default EntryPoint and the default Cmd defined in the Docker image
+            #   are ignored. Your command is run without any arguments.
+            # - If you supply only args:
+            #   The default Entrypoint defined in the Docker image is run with the args
+            #   that you supplied.
+            # - If you supply a command and args:
+            #   The default Entrypoint and the default Cmd defined in the Docker image
+            #   are ignored. Your command is run with your args.
+            # It cannot be set if custom container image is
+            # not provided.
+            # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+            # both cannot be set at the same time.
+          &quot;A String&quot;,
+        ],
+        &quot;acceleratorConfig&quot;: { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
+            # [Learn about restrictions on accelerator configurations for
+            # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+            # Note that the AcceleratorConfig can be used in both Jobs and Versions.
+            # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
+            # [accelerators for online
+            # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
+          &quot;count&quot;: &quot;A String&quot;, # The number of accelerators to attach to each machine running the job.
+          &quot;type&quot;: &quot;A String&quot;, # The type of accelerator to use.
+        },
+      },
+      &quot;scaleTier&quot;: &quot;A String&quot;, # Required. Specifies the machine types, the number of replicas for workers
+          # and parameter servers.
+      &quot;jobDir&quot;: &quot;A String&quot;, # Optional. A Google Cloud Storage path in which to store training outputs
           # and other data needed for training. This path is passed to your TensorFlow
-          # program as the '--job-dir' command-line argument. The benefit of specifying
+          # program as the &#x27;--job-dir&#x27; command-line argument. The benefit of specifying
           # this field is that Cloud ML validates the path for use in training.
-      "scheduling": { # All parameters related to scheduling of training jobs. # Optional. Scheduling options for a training job.
-        "maxRunningTime": "A String", # Optional. The maximum job running time, expressed in seconds. The field can
-            # contain up to nine fractional digits, terminated by `s`. By default there
-            # is no limit to the running time.
+      &quot;pythonVersion&quot;: &quot;A String&quot;, # Optional. The version of Python used in training. You must either specify
+          # this field or specify `masterConfig.imageUri`.
+          #
+          # The following Python versions are available:
+          #
+          # * Python &#x27;3.7&#x27; is available when `runtime_version` is set to &#x27;1.15&#x27; or
+          #   later.
+          # * Python &#x27;3.5&#x27; is available when `runtime_version` is set to a version
+          #   from &#x27;1.4&#x27; to &#x27;1.14&#x27;.
+          # * Python &#x27;2.7&#x27; is available when `runtime_version` is set to &#x27;1.15&#x27; or
+          #   earlier.
+          #
+          # Read more about the Python versions available for [each runtime
+          # version](/ml-engine/docs/runtime-version-list).
+      &quot;network&quot;: &quot;A String&quot;, # Optional. The full name of the Google Compute Engine
+          # [network](/compute/docs/networks-and-firewalls#networks) to which the Job
+          # is peered. For example, projects/12345/global/networks/myVPC. Format is of
+          # the form projects/{project}/global/networks/{network}. Where {project} is a
+          # project number, as in &#x27;12345&#x27;, and {network} is network name.&quot;.
+          #
+          # Private services access must already be configured for the network. If left
+          # unspecified, the Job is not peered with any network. Learn more -
+          # Connecting Job to user network over private
+          # IP.
+      &quot;scheduling&quot;: { # All parameters related to scheduling of training jobs. # Optional. Scheduling options for a training job.
+        &quot;maxWaitTime&quot;: &quot;A String&quot;,
+        &quot;maxRunningTime&quot;: &quot;A String&quot;, # Optional. The maximum job running time, expressed in seconds. The field can
+            # contain up to nine fractional digits, terminated by `s`. If not specified,
+            # this field defaults to `604800s` (seven days).
             #
             # If the training job is still running after this duration, AI Platform
             # Training cancels it.
@@ -916,85 +615,7 @@
             #   ...
             # ```
       },
-      "parameterServerCount": "A String", # Optional. The number of parameter server replicas to use for the training
-          # job. Each replica in the cluster will be of the type specified in
-          # `parameter_server_type`.
-          #
-          # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
-          # set this value, you must also set `parameter_server_type`.
-          #
-          # The default value is zero.
-      "evaluatorCount": "A String", # Optional. The number of evaluator replicas to use for the training job.
-          # Each replica in the cluster will be of the type specified in
-          # `evaluator_type`.
-          #
-          # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
-          # set this value, you must also set `evaluator_type`.
-          #
-          # The default value is zero.
-      "workerType": "A String", # Optional. Specifies the type of virtual machine to use for your training
-          # job's worker nodes.
-          #
-          # The supported values are the same as those described in the entry for
-          # `masterType`.
-          #
-          # This value must be consistent with the category of machine type that
-          # `masterType` uses. In other words, both must be Compute Engine machine
-          # types or both must be legacy machine types.
-          #
-          # If you use `cloud_tpu` for this value, see special instructions for
-          # [configuring a custom TPU
-          # machine](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine).
-          #
-          # This value must be present when `scaleTier` is set to `CUSTOM` and
-          # `workerCount` is greater than zero.
-      "scaleTier": "A String", # Required. Specifies the machine types, the number of replicas for workers
-          # and parameter servers.
-      "packageUris": [ # Required. The Google Cloud Storage location of the packages with
-          # the training program and any additional dependencies.
-          # The maximum number of package URIs is 100.
-        "A String",
-      ],
-      "workerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for workers.
-          #
-          # You should only set `workerConfig.acceleratorConfig` if `workerType` is set
-          # to a Compute Engine machine type. [Learn about restrictions on accelerator
-          # configurations for
-          # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-          #
-          # Set `workerConfig.imageUri` only if you build a custom image for your
-          # worker. If `workerConfig.imageUri` has not been set, AI Platform uses
-          # the value of `masterConfig.imageUri`. Learn more about [configuring custom
-          # containers](/ai-platform/training/docs/distributed-training-containers).
-        "tpuTfVersion": "A String", # The AI Platform runtime version that includes a TensorFlow version matching
-            # the one used in the custom container. This field is required if the replica
-            # is a TPU worker that uses a custom container. Otherwise, do not specify
-            # this field. This must be a [runtime version that currently supports
-            # training with
-            # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
-            #
-            # Note that the version of TensorFlow included in a runtime version may
-            # differ from the numbering of the runtime version itself, because it may
-            # have a different [patch
-            # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
-            # In this field, you must specify the runtime version (TensorFlow minor
-            # version). For example, if your custom container runs TensorFlow `1.x.y`,
-            # specify `1.x`.
-        "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
-            # [Learn about restrictions on accelerator configurations for
-            # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-            # Note that the AcceleratorConfig can be used in both Jobs and Versions.
-            # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
-            # [accelerators for online
-            # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
-          "count": "A String", # The number of accelerators to attach to each machine running the job.
-          "type": "A String", # The type of accelerator to use.
-        },
-        "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
-            # Registry. Learn more about [configuring custom
-            # containers](/ai-platform/training/docs/distributed-training-containers).
-      },
-      "evaluatorConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for evaluators.
+      &quot;evaluatorConfig&quot;: { # Represents the configuration for a replica in a cluster. # Optional. The configuration for evaluators.
           #
           # You should only set `evaluatorConfig.acceleratorConfig` if
           # `evaluatorType` is set to a Compute Engine machine type. [Learn
@@ -1005,7 +626,7 @@
           # your evaluator. If `evaluatorConfig.imageUri` has not been
           # set, AI Platform uses the value of `masterConfig.imageUri`. Learn more about [configuring custom
           # containers](/ai-platform/training/docs/distributed-training-containers).
-        "tpuTfVersion": "A String", # The AI Platform runtime version that includes a TensorFlow version matching
+        &quot;tpuTfVersion&quot;: &quot;A String&quot;, # The AI Platform runtime version that includes a TensorFlow version matching
             # the one used in the custom container. This field is required if the replica
             # is a TPU worker that uses a custom container. Otherwise, do not specify
             # this field. This must be a [runtime version that currently supports
@@ -1019,257 +640,882 @@
             # In this field, you must specify the runtime version (TensorFlow minor
             # version). For example, if your custom container runs TensorFlow `1.x.y`,
             # specify `1.x`.
-        "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
+        &quot;containerCommand&quot;: [ # The command with which the replica&#x27;s custom container is run.
+            # If provided, it will override default ENTRYPOINT of the docker image.
+            # If not provided, the docker image&#x27;s ENTRYPOINT is used.
+            # It cannot be set if custom container image is
+            # not provided.
+            # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+            # both cannot be set at the same time.
+          &quot;A String&quot;,
+        ],
+        &quot;imageUri&quot;: &quot;A String&quot;, # The Docker image to run on the replica. This image must be in Container
+            # Registry. Learn more about [configuring custom
+            # containers](/ai-platform/training/docs/distributed-training-containers).
+        &quot;containerArgs&quot;: [ # Arguments to the entrypoint command.
+            # The following rules apply for container_command and container_args:
+            # - If you do not supply command or args:
+            #   The defaults defined in the Docker image are used.
+            # - If you supply a command but no args:
+            #   The default EntryPoint and the default Cmd defined in the Docker image
+            #   are ignored. Your command is run without any arguments.
+            # - If you supply only args:
+            #   The default Entrypoint defined in the Docker image is run with the args
+            #   that you supplied.
+            # - If you supply a command and args:
+            #   The default Entrypoint and the default Cmd defined in the Docker image
+            #   are ignored. Your command is run with your args.
+            # It cannot be set if custom container image is
+            # not provided.
+            # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+            # both cannot be set at the same time.
+          &quot;A String&quot;,
+        ],
+        &quot;acceleratorConfig&quot;: { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
             # [Learn about restrictions on accelerator configurations for
             # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
             # Note that the AcceleratorConfig can be used in both Jobs and Versions.
             # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
             # [accelerators for online
             # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
-          "count": "A String", # The number of accelerators to attach to each machine running the job.
-          "type": "A String", # The type of accelerator to use.
+          &quot;count&quot;: &quot;A String&quot;, # The number of accelerators to attach to each machine running the job.
+          &quot;type&quot;: &quot;A String&quot;, # The type of accelerator to use.
         },
-        "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
-            # Registry. Learn more about [configuring custom
-            # containers](/ai-platform/training/docs/distributed-training-containers).
       },
-      "useChiefInTfConfig": True or False, # Optional. Use `chief` instead of `master` in the `TF_CONFIG` environment
+      &quot;useChiefInTfConfig&quot;: True or False, # Optional. Use `chief` instead of `master` in the `TF_CONFIG` environment
           # variable when training with a custom container. Defaults to `false`. [Learn
           # more about this
           # field.](/ai-platform/training/docs/distributed-training-details#chief-versus-master)
           #
-          # This field has no effect for training jobs that don't use a custom
+          # This field has no effect for training jobs that don&#x27;t use a custom
           # container.
-      "masterConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for your master worker.
-          #
-          # You should only set `masterConfig.acceleratorConfig` if `masterType` is set
-          # to a Compute Engine machine type. Learn about [restrictions on accelerator
-          # configurations for
-          # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-          #
-          # Set `masterConfig.imageUri` only if you build a custom image. Only one of
-          # `masterConfig.imageUri` and `runtimeVersion` should be set. Learn more
-          # about [configuring custom
-          # containers](/ai-platform/training/docs/distributed-training-containers).
-        "tpuTfVersion": "A String", # The AI Platform runtime version that includes a TensorFlow version matching
-            # the one used in the custom container. This field is required if the replica
-            # is a TPU worker that uses a custom container. Otherwise, do not specify
-            # this field. This must be a [runtime version that currently supports
-            # training with
-            # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
-            #
-            # Note that the version of TensorFlow included in a runtime version may
-            # differ from the numbering of the runtime version itself, because it may
-            # have a different [patch
-            # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
-            # In this field, you must specify the runtime version (TensorFlow minor
-            # version). For example, if your custom container runs TensorFlow `1.x.y`,
-            # specify `1.x`.
-        "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
-            # [Learn about restrictions on accelerator configurations for
-            # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-            # Note that the AcceleratorConfig can be used in both Jobs and Versions.
-            # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
-            # [accelerators for online
-            # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
-          "count": "A String", # The number of accelerators to attach to each machine running the job.
-          "type": "A String", # The type of accelerator to use.
-        },
-        "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
-            # Registry. Learn more about [configuring custom
-            # containers](/ai-platform/training/docs/distributed-training-containers).
-      },
-      "runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for training. You must
-          # either specify this field or specify `masterConfig.imageUri`.
-          #
-          # For more information, see the [runtime version
-          # list](/ai-platform/training/docs/runtime-version-list) and learn [how to
-          # manage runtime versions](/ai-platform/training/docs/versioning).
-      "hyperparameters": { # Represents a set of hyperparameters to optimize. # Optional. The set of Hyperparameters to tune.
-        "maxTrials": 42, # Optional. How many training trials should be attempted to optimize
-            # the specified hyperparameters.
-            #
-            # Defaults to one.
-        "goal": "A String", # Required. The type of goal to use for tuning. Available types are
-            # `MAXIMIZE` and `MINIMIZE`.
-            #
-            # Defaults to `MAXIMIZE`.
-        "algorithm": "A String", # Optional. The search algorithm specified for the hyperparameter
-            # tuning job.
-            # Uses the default AI Platform hyperparameter tuning
-            # algorithm if unspecified.
-        "maxFailedTrials": 42, # Optional. The number of failed trials that need to be seen before failing
-            # the hyperparameter tuning job. You can specify this field to override the
-            # default failing criteria for AI Platform hyperparameter tuning jobs.
-            #
-            # Defaults to zero, which means the service decides when a hyperparameter
-            # job should fail.
-        "enableTrialEarlyStopping": True or False, # Optional. Indicates if the hyperparameter tuning job enables auto trial
-            # early stopping.
-        "resumePreviousJobId": "A String", # Optional. The prior hyperparameter tuning job id that users hope to
-            # continue with. The job id will be used to find the corresponding vizier
-            # study guid and resume the study.
-        "params": [ # Required. The set of parameters to tune.
-          { # Represents a single hyperparameter to optimize.
-            "maxValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
-                # should be unset if type is `CATEGORICAL`. This value should be integers if
-                # type is `INTEGER`.
-            "minValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
-                # should be unset if type is `CATEGORICAL`. This value should be integers if
-                # type is INTEGER.
-            "discreteValues": [ # Required if type is `DISCRETE`.
-                # A list of feasible points.
-                # The list should be in strictly increasing order. For instance, this
-                # parameter might have possible settings of 1.5, 2.5, and 4.0. This list
-                # should not contain more than 1,000 values.
-              3.14,
-            ],
-            "parameterName": "A String", # Required. The parameter name must be unique amongst all ParameterConfigs in
-                # a HyperparameterSpec message. E.g., "learning_rate".
-            "categoricalValues": [ # Required if type is `CATEGORICAL`. The list of possible categories.
-              "A String",
-            ],
-            "type": "A String", # Required. The type of the parameter.
-            "scaleType": "A String", # Optional. How the parameter should be scaled to the hypercube.
-                # Leave unset for categorical parameters.
-                # Some kind of scaling is strongly recommended for real or integral
-                # parameters (e.g., `UNIT_LINEAR_SCALE`).
-          },
-        ],
-        "hyperparameterMetricTag": "A String", # Optional. The TensorFlow summary tag name to use for optimizing trials. For
-            # current versions of TensorFlow, this tag name should exactly match what is
-            # shown in TensorBoard, including all scopes.  For versions of TensorFlow
-            # prior to 0.12, this should be only the tag passed to tf.Summary.
-            # By default, "training/hptuning/metric" will be used.
-        "maxParallelTrials": 42, # Optional. The number of training trials to run concurrently.
-            # You can reduce the time it takes to perform hyperparameter tuning by adding
-            # trials in parallel. However, each trail only benefits from the information
-            # gained in completed trials. That means that a trial does not get access to
-            # the results of trials running at the same time, which could reduce the
-            # quality of the overall optimization.
-            #
-            # Each trial will use the same scale tier and machine types.
-            #
-            # Defaults to one.
-      },
-      "args": [ # Optional. Command-line arguments passed to the training application when it
-          # starts. If your job uses a custom container, then the arguments are passed
-          # to the container's &lt;a class="external" target="_blank"
-          # href="https://docs.docker.com/engine/reference/builder/#entrypoint"&gt;
-          # `ENTRYPOINT`&lt;/a&gt; command.
-        "A String",
-      ],
-      "pythonModule": "A String", # Required. The Python module name to run after installing the packages.
-      "workerCount": "A String", # Optional. The number of worker replicas to use for the training job. Each
+      &quot;workerCount&quot;: &quot;A String&quot;, # Optional. The number of worker replicas to use for the training job. Each
           # replica in the cluster will be of the type specified in `worker_type`.
           #
           # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
           # set this value, you must also set `worker_type`.
           #
           # The default value is zero.
-      "encryptionConfig": { # Represents a custom encryption key configuration that can be applied to # Optional. Options for using customer-managed encryption keys (CMEK) to
-          # protect resources created by a training job, instead of using Google's
-          # default encryption. If this is set, then all resources created by the
-          # training job will be encrypted with the customer-managed encryption key
-          # that you specify.
-          #
-          # [Learn how and when to use CMEK with AI Platform
-          # Training](/ai-platform/training/docs/cmek).
-          # a resource.
-        "kmsKeyName": "A String", # The Cloud KMS resource identifier of the customer-managed encryption key
-            # used to protect a resource, such as a training job. It has the following
-            # format:
-            # `projects/{PROJECT_ID}/locations/{REGION}/keyRings/{KEY_RING_NAME}/cryptoKeys/{KEY_NAME}`
-      },
-      "parameterServerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for parameter servers.
-          #
-          # You should only set `parameterServerConfig.acceleratorConfig` if
-          # `parameterServerType` is set to a Compute Engine machine type. [Learn
-          # about restrictions on accelerator configurations for
-          # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-          #
-          # Set `parameterServerConfig.imageUri` only if you build a custom image for
-          # your parameter server. If `parameterServerConfig.imageUri` has not been
-          # set, AI Platform uses the value of `masterConfig.imageUri`. Learn more about [configuring custom
-          # containers](/ai-platform/training/docs/distributed-training-containers).
-        "tpuTfVersion": "A String", # The AI Platform runtime version that includes a TensorFlow version matching
-            # the one used in the custom container. This field is required if the replica
-            # is a TPU worker that uses a custom container. Otherwise, do not specify
-            # this field. This must be a [runtime version that currently supports
-            # training with
-            # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
-            #
-            # Note that the version of TensorFlow included in a runtime version may
-            # differ from the numbering of the runtime version itself, because it may
-            # have a different [patch
-            # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
-            # In this field, you must specify the runtime version (TensorFlow minor
-            # version). For example, if your custom container runs TensorFlow `1.x.y`,
-            # specify `1.x`.
-        "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
-            # [Learn about restrictions on accelerator configurations for
-            # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-            # Note that the AcceleratorConfig can be used in both Jobs and Versions.
-            # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
-            # [accelerators for online
-            # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
-          "count": "A String", # The number of accelerators to attach to each machine running the job.
-          "type": "A String", # The type of accelerator to use.
+      &quot;pythonModule&quot;: &quot;A String&quot;, # Required. The Python module name to run after installing the packages.
+      &quot;args&quot;: [ # Optional. Command-line arguments passed to the training application when it
+          # starts. If your job uses a custom container, then the arguments are passed
+          # to the container&#x27;s &lt;a class=&quot;external&quot; target=&quot;_blank&quot;
+          # href=&quot;https://docs.docker.com/engine/reference/builder/#entrypoint&quot;&gt;
+          # `ENTRYPOINT`&lt;/a&gt; command.
+        &quot;A String&quot;,
+      ],
+    },
+    &quot;state&quot;: &quot;A String&quot;, # Output only. The detailed state of a job.
+    &quot;jobId&quot;: &quot;A String&quot;, # Required. The user-specified id of the job.
+    &quot;endTime&quot;: &quot;A String&quot;, # Output only. When the job processing was completed.
+    &quot;startTime&quot;: &quot;A String&quot;, # Output only. When the job processing was started.
+    &quot;predictionOutput&quot;: { # Represents results of a prediction job. # The current prediction job result.
+      &quot;errorCount&quot;: &quot;A String&quot;, # The number of data instances which resulted in errors.
+      &quot;outputPath&quot;: &quot;A String&quot;, # The output Google Cloud Storage location provided at the job creation time.
+      &quot;nodeHours&quot;: 3.14, # Node hours used by the batch prediction job.
+      &quot;predictionCount&quot;: &quot;A String&quot;, # The number of generated predictions.
+    },
+    &quot;trainingOutput&quot;: { # Represents results of a training job. Output only. # The current training job result.
+      &quot;trials&quot;: [ # Results for individual Hyperparameter trials.
+          # Only set for hyperparameter tuning jobs.
+        { # Represents the result of a single hyperparameter tuning trial from a
+            # training job. The TrainingOutput object that is returned on successful
+            # completion of a training job with hyperparameter tuning includes a list
+            # of HyperparameterOutput objects, one for each successful trial.
+          &quot;allMetrics&quot;: [ # All recorded object metrics for this trial. This field is not currently
+              # populated.
+            { # An observed value of a metric.
+              &quot;objectiveValue&quot;: 3.14, # The objective value at this training step.
+              &quot;trainingStep&quot;: &quot;A String&quot;, # The global training step for this metric.
+            },
+          ],
+          &quot;hyperparameters&quot;: { # The hyperparameters given to this trial.
+            &quot;a_key&quot;: &quot;A String&quot;,
+          },
+          &quot;trialId&quot;: &quot;A String&quot;, # The trial id for these results.
+          &quot;endTime&quot;: &quot;A String&quot;, # Output only. End time for the trial.
+          &quot;isTrialStoppedEarly&quot;: True or False, # True if the trial is stopped early.
+          &quot;startTime&quot;: &quot;A String&quot;, # Output only. Start time for the trial.
+          &quot;finalMetric&quot;: { # An observed value of a metric. # The final objective metric seen for this trial.
+            &quot;objectiveValue&quot;: 3.14, # The objective value at this training step.
+            &quot;trainingStep&quot;: &quot;A String&quot;, # The global training step for this metric.
+          },
+          &quot;builtInAlgorithmOutput&quot;: { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
+              # Only set for trials of built-in algorithms jobs that have succeeded.
+            &quot;framework&quot;: &quot;A String&quot;, # Framework on which the built-in algorithm was trained.
+            &quot;modelPath&quot;: &quot;A String&quot;, # The Cloud Storage path to the `model/` directory where the training job
+                # saves the trained model. Only set for successful jobs that don&#x27;t use
+                # hyperparameter tuning.
+            &quot;pythonVersion&quot;: &quot;A String&quot;, # Python version on which the built-in algorithm was trained.
+            &quot;runtimeVersion&quot;: &quot;A String&quot;, # AI Platform runtime version on which the built-in algorithm was
+                # trained.
+          },
+          &quot;state&quot;: &quot;A String&quot;, # Output only. The detailed state of the trial.
         },
-        "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
-            # Registry. Learn more about [configuring custom
-            # containers](/ai-platform/training/docs/distributed-training-containers).
+      ],
+      &quot;hyperparameterMetricTag&quot;: &quot;A String&quot;, # The TensorFlow summary tag name used for optimizing hyperparameter tuning
+          # trials. See
+          # [`HyperparameterSpec.hyperparameterMetricTag`](#HyperparameterSpec.FIELDS.hyperparameter_metric_tag)
+          # for more information. Only set for hyperparameter tuning jobs.
+      &quot;completedTrialCount&quot;: &quot;A String&quot;, # The number of hyperparameter tuning trials that completed successfully.
+          # Only set for hyperparameter tuning jobs.
+      &quot;isHyperparameterTuningJob&quot;: True or False, # Whether this job is a hyperparameter tuning job.
+      &quot;consumedMLUnits&quot;: 3.14, # The amount of ML units consumed by the job.
+      &quot;isBuiltInAlgorithmJob&quot;: True or False, # Whether this job is a built-in Algorithm job.
+      &quot;builtInAlgorithmOutput&quot;: { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
+          # Only set for built-in algorithms jobs.
+        &quot;framework&quot;: &quot;A String&quot;, # Framework on which the built-in algorithm was trained.
+        &quot;modelPath&quot;: &quot;A String&quot;, # The Cloud Storage path to the `model/` directory where the training job
+            # saves the trained model. Only set for successful jobs that don&#x27;t use
+            # hyperparameter tuning.
+        &quot;pythonVersion&quot;: &quot;A String&quot;, # Python version on which the built-in algorithm was trained.
+        &quot;runtimeVersion&quot;: &quot;A String&quot;, # AI Platform runtime version on which the built-in algorithm was
+            # trained.
       },
-      "region": "A String", # Required. The region to run the training job in. See the [available
-          # regions](/ai-platform/training/docs/regions) for AI Platform Training.
-      "pythonVersion": "A String", # Optional. The version of Python used in training. You must either specify
-          # this field or specify `masterConfig.imageUri`.
-          #
-          # The following Python versions are available:
-          #
-          # * Python '3.7' is available when `runtime_version` is set to '1.15' or
-          #   later.
-          # * Python '3.5' is available when `runtime_version` is set to a version
-          #   from '1.4' to '1.14'.
-          # * Python '2.7' is available when `runtime_version` is set to '1.15' or
-          #   earlier.
-          #
-          # Read more about the Python versions available for [each runtime
-          # version](/ml-engine/docs/runtime-version-list).
-      "evaluatorType": "A String", # Optional. Specifies the type of virtual machine to use for your training
-          # job's evaluator nodes.
-          #
-          # The supported values are the same as those described in the entry for
-          # `masterType`.
-          #
-          # This value must be consistent with the category of machine type that
-          # `masterType` uses. In other words, both must be Compute Engine machine
-          # types or both must be legacy machine types.
-          #
-          # This value must be present when `scaleTier` is set to `CUSTOM` and
-          # `evaluatorCount` is greater than zero.
-      "parameterServerType": "A String", # Optional. Specifies the type of virtual machine to use for your training
-          # job's parameter server.
-          #
-          # The supported values are the same as those described in the entry for
-          # `master_type`.
-          #
-          # This value must be consistent with the category of machine type that
-          # `masterType` uses. In other words, both must be Compute Engine machine
-          # types or both must be legacy machine types.
-          #
-          # This value must be present when `scaleTier` is set to `CUSTOM` and
-          # `parameter_server_count` is greater than zero.
     },
-    "endTime": "A String", # Output only. When the job processing was completed.
-    "predictionOutput": { # Represents results of a prediction job. # The current prediction job result.
-      "outputPath": "A String", # The output Google Cloud Storage location provided at the job creation time.
-      "nodeHours": 3.14, # Node hours used by the batch prediction job.
-      "predictionCount": "A String", # The number of generated predictions.
-      "errorCount": "A String", # The number of data instances which resulted in errors.
+    &quot;createTime&quot;: &quot;A String&quot;, # Output only. When the job was created.
+    &quot;labels&quot;: { # Optional. One or more labels that you can add, to organize your jobs.
+        # Each label is a key-value pair, where both the key and the value are
+        # arbitrary strings that you supply.
+        # For more information, see the documentation on
+        # &lt;a href=&quot;/ml-engine/docs/tensorflow/resource-labels&quot;&gt;using labels&lt;/a&gt;.
+      &quot;a_key&quot;: &quot;A String&quot;,
     },
-    "createTime": "A String", # Output only. When the job was created.
-  }</pre>
+    &quot;predictionInput&quot;: { # Represents input parameters for a prediction job. # Input parameters to create a prediction job.
+      &quot;outputPath&quot;: &quot;A String&quot;, # Required. The output Google Cloud Storage location.
+      &quot;outputDataFormat&quot;: &quot;A String&quot;, # Optional. Format of the output data files, defaults to JSON.
+      &quot;dataFormat&quot;: &quot;A String&quot;, # Required. The format of the input data files.
+      &quot;batchSize&quot;: &quot;A String&quot;, # Optional. Number of records per batch, defaults to 64.
+          # The service will buffer batch_size number of records in memory before
+          # invoking one Tensorflow prediction call internally. So take the record
+          # size and memory available into consideration when setting this parameter.
+      &quot;runtimeVersion&quot;: &quot;A String&quot;, # Optional. The AI Platform runtime version to use for this batch
+          # prediction. If not set, AI Platform will pick the runtime version used
+          # during the CreateVersion request for this model version, or choose the
+          # latest stable version when model version information is not available
+          # such as when the model is specified by uri.
+      &quot;inputPaths&quot;: [ # Required. The Cloud Storage location of the input data files. May contain
+          # &lt;a href=&quot;/storage/docs/gsutil/addlhelp/WildcardNames&quot;&gt;wildcards&lt;/a&gt;.
+        &quot;A String&quot;,
+      ],
+      &quot;region&quot;: &quot;A String&quot;, # Required. The Google Compute Engine region to run the prediction job in.
+          # See the &lt;a href=&quot;/ml-engine/docs/tensorflow/regions&quot;&gt;available regions&lt;/a&gt;
+          # for AI Platform services.
+      &quot;versionName&quot;: &quot;A String&quot;, # Use this field if you want to specify a version of the model to use. The
+          # string is formatted the same way as `model_version`, with the addition
+          # of the version information:
+          #
+          # `&quot;projects/YOUR_PROJECT/models/YOUR_MODEL/versions/YOUR_VERSION&quot;`
+      &quot;modelName&quot;: &quot;A String&quot;, # Use this field if you want to use the default version for the specified
+          # model. The string must use the following format:
+          #
+          # `&quot;projects/YOUR_PROJECT/models/YOUR_MODEL&quot;`
+      &quot;uri&quot;: &quot;A String&quot;, # Use this field if you want to specify a Google Cloud Storage path for
+          # the model to use.
+      &quot;maxWorkerCount&quot;: &quot;A String&quot;, # Optional. The maximum number of workers to be used for parallel processing.
+          # Defaults to 10 if not specified.
+      &quot;signatureName&quot;: &quot;A String&quot;, # Optional. The name of the signature defined in the SavedModel to use for
+          # this job. Please refer to
+          # [SavedModel](https://tensorflow.github.io/serving/serving_basic.html)
+          # for information about how to use signatures.
+          #
+          # Defaults to
+          # [DEFAULT_SERVING_SIGNATURE_DEF_KEY](https://www.tensorflow.org/api_docs/python/tf/saved_model/signature_constants)
+          # , which is &quot;serving_default&quot;.
+    },
+    &quot;errorMessage&quot;: &quot;A String&quot;, # Output only. The details of a failure or a cancellation.
+  }
+
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # Represents a training or prediction job.
+      &quot;etag&quot;: &quot;A String&quot;, # `etag` is used for optimistic concurrency control as a way to help
+          # prevent simultaneous updates of a job from overwriting each other.
+          # It is strongly suggested that systems make use of the `etag` in the
+          # read-modify-write cycle to perform job updates in order to avoid race
+          # conditions: An `etag` is returned in the response to `GetJob`, and
+          # systems are expected to put that etag in the request to `UpdateJob` to
+          # ensure that their change will be applied to the same version of the job.
+      &quot;trainingInput&quot;: { # Represents input parameters for a training job. When using the gcloud command # Input parameters to create a training job.
+          # to submit your training job, you can specify the input parameters as
+          # command-line arguments and/or in a YAML configuration file referenced from
+          # the --config command-line argument. For details, see the guide to [submitting
+          # a training job](/ai-platform/training/docs/training-jobs).
+        &quot;parameterServerConfig&quot;: { # Represents the configuration for a replica in a cluster. # Optional. The configuration for parameter servers.
+            #
+            # You should only set `parameterServerConfig.acceleratorConfig` if
+            # `parameterServerType` is set to a Compute Engine machine type. [Learn
+            # about restrictions on accelerator configurations for
+            # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+            #
+            # Set `parameterServerConfig.imageUri` only if you build a custom image for
+            # your parameter server. If `parameterServerConfig.imageUri` has not been
+            # set, AI Platform uses the value of `masterConfig.imageUri`. Learn more about [configuring custom
+            # containers](/ai-platform/training/docs/distributed-training-containers).
+          &quot;tpuTfVersion&quot;: &quot;A String&quot;, # The AI Platform runtime version that includes a TensorFlow version matching
+              # the one used in the custom container. This field is required if the replica
+              # is a TPU worker that uses a custom container. Otherwise, do not specify
+              # this field. This must be a [runtime version that currently supports
+              # training with
+              # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
+              #
+              # Note that the version of TensorFlow included in a runtime version may
+              # differ from the numbering of the runtime version itself, because it may
+              # have a different [patch
+              # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
+              # In this field, you must specify the runtime version (TensorFlow minor
+              # version). For example, if your custom container runs TensorFlow `1.x.y`,
+              # specify `1.x`.
+          &quot;containerCommand&quot;: [ # The command with which the replica&#x27;s custom container is run.
+              # If provided, it will override default ENTRYPOINT of the docker image.
+              # If not provided, the docker image&#x27;s ENTRYPOINT is used.
+              # It cannot be set if custom container image is
+              # not provided.
+              # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+              # both cannot be set at the same time.
+            &quot;A String&quot;,
+          ],
+          &quot;imageUri&quot;: &quot;A String&quot;, # The Docker image to run on the replica. This image must be in Container
+              # Registry. Learn more about [configuring custom
+              # containers](/ai-platform/training/docs/distributed-training-containers).
+          &quot;containerArgs&quot;: [ # Arguments to the entrypoint command.
+              # The following rules apply for container_command and container_args:
+              # - If you do not supply command or args:
+              #   The defaults defined in the Docker image are used.
+              # - If you supply a command but no args:
+              #   The default EntryPoint and the default Cmd defined in the Docker image
+              #   are ignored. Your command is run without any arguments.
+              # - If you supply only args:
+              #   The default Entrypoint defined in the Docker image is run with the args
+              #   that you supplied.
+              # - If you supply a command and args:
+              #   The default Entrypoint and the default Cmd defined in the Docker image
+              #   are ignored. Your command is run with your args.
+              # It cannot be set if custom container image is
+              # not provided.
+              # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+              # both cannot be set at the same time.
+            &quot;A String&quot;,
+          ],
+          &quot;acceleratorConfig&quot;: { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
+              # [Learn about restrictions on accelerator configurations for
+              # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+              # Note that the AcceleratorConfig can be used in both Jobs and Versions.
+              # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
+              # [accelerators for online
+              # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
+            &quot;count&quot;: &quot;A String&quot;, # The number of accelerators to attach to each machine running the job.
+            &quot;type&quot;: &quot;A String&quot;, # The type of accelerator to use.
+          },
+        },
+        &quot;encryptionConfig&quot;: { # Represents a custom encryption key configuration that can be applied to # Optional. Options for using customer-managed encryption keys (CMEK) to
+            # protect resources created by a training job, instead of using Google&#x27;s
+            # default encryption. If this is set, then all resources created by the
+            # training job will be encrypted with the customer-managed encryption key
+            # that you specify.
+            #
+            # [Learn how and when to use CMEK with AI Platform
+            # Training](/ai-platform/training/docs/cmek).
+            # a resource.
+          &quot;kmsKeyName&quot;: &quot;A String&quot;, # The Cloud KMS resource identifier of the customer-managed encryption key
+              # used to protect a resource, such as a training job. It has the following
+              # format:
+              # `projects/{PROJECT_ID}/locations/{REGION}/keyRings/{KEY_RING_NAME}/cryptoKeys/{KEY_NAME}`
+        },
+        &quot;hyperparameters&quot;: { # Represents a set of hyperparameters to optimize. # Optional. The set of Hyperparameters to tune.
+          &quot;hyperparameterMetricTag&quot;: &quot;A String&quot;, # Optional. The TensorFlow summary tag name to use for optimizing trials. For
+              # current versions of TensorFlow, this tag name should exactly match what is
+              # shown in TensorBoard, including all scopes.  For versions of TensorFlow
+              # prior to 0.12, this should be only the tag passed to tf.Summary.
+              # By default, &quot;training/hptuning/metric&quot; will be used.
+          &quot;params&quot;: [ # Required. The set of parameters to tune.
+            { # Represents a single hyperparameter to optimize.
+              &quot;type&quot;: &quot;A String&quot;, # Required. The type of the parameter.
+              &quot;categoricalValues&quot;: [ # Required if type is `CATEGORICAL`. The list of possible categories.
+                &quot;A String&quot;,
+              ],
+              &quot;parameterName&quot;: &quot;A String&quot;, # Required. The parameter name must be unique amongst all ParameterConfigs in
+                  # a HyperparameterSpec message. E.g., &quot;learning_rate&quot;.
+              &quot;minValue&quot;: 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
+                  # should be unset if type is `CATEGORICAL`. This value should be integers if
+                  # type is INTEGER.
+              &quot;discreteValues&quot;: [ # Required if type is `DISCRETE`.
+                  # A list of feasible points.
+                  # The list should be in strictly increasing order. For instance, this
+                  # parameter might have possible settings of 1.5, 2.5, and 4.0. This list
+                  # should not contain more than 1,000 values.
+                3.14,
+              ],
+              &quot;scaleType&quot;: &quot;A String&quot;, # Optional. How the parameter should be scaled to the hypercube.
+                  # Leave unset for categorical parameters.
+                  # Some kind of scaling is strongly recommended for real or integral
+                  # parameters (e.g., `UNIT_LINEAR_SCALE`).
+              &quot;maxValue&quot;: 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
+                  # should be unset if type is `CATEGORICAL`. This value should be integers if
+                  # type is `INTEGER`.
+            },
+          ],
+          &quot;enableTrialEarlyStopping&quot;: True or False, # Optional. Indicates if the hyperparameter tuning job enables auto trial
+              # early stopping.
+          &quot;resumePreviousJobId&quot;: &quot;A String&quot;, # Optional. The prior hyperparameter tuning job id that users hope to
+              # continue with. The job id will be used to find the corresponding vizier
+              # study guid and resume the study.
+          &quot;maxParallelTrials&quot;: 42, # Optional. The number of training trials to run concurrently.
+              # You can reduce the time it takes to perform hyperparameter tuning by adding
+              # trials in parallel. However, each trail only benefits from the information
+              # gained in completed trials. That means that a trial does not get access to
+              # the results of trials running at the same time, which could reduce the
+              # quality of the overall optimization.
+              #
+              # Each trial will use the same scale tier and machine types.
+              #
+              # Defaults to one.
+          &quot;maxFailedTrials&quot;: 42, # Optional. The number of failed trials that need to be seen before failing
+              # the hyperparameter tuning job. You can specify this field to override the
+              # default failing criteria for AI Platform hyperparameter tuning jobs.
+              #
+              # Defaults to zero, which means the service decides when a hyperparameter
+              # job should fail.
+          &quot;goal&quot;: &quot;A String&quot;, # Required. The type of goal to use for tuning. Available types are
+              # `MAXIMIZE` and `MINIMIZE`.
+              #
+              # Defaults to `MAXIMIZE`.
+          &quot;maxTrials&quot;: 42, # Optional. How many training trials should be attempted to optimize
+              # the specified hyperparameters.
+              #
+              # Defaults to one.
+          &quot;algorithm&quot;: &quot;A String&quot;, # Optional. The search algorithm specified for the hyperparameter
+              # tuning job.
+              # Uses the default AI Platform hyperparameter tuning
+              # algorithm if unspecified.
+        },
+        &quot;workerConfig&quot;: { # Represents the configuration for a replica in a cluster. # Optional. The configuration for workers.
+            #
+            # You should only set `workerConfig.acceleratorConfig` if `workerType` is set
+            # to a Compute Engine machine type. [Learn about restrictions on accelerator
+            # configurations for
+            # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+            #
+            # Set `workerConfig.imageUri` only if you build a custom image for your
+            # worker. If `workerConfig.imageUri` has not been set, AI Platform uses
+            # the value of `masterConfig.imageUri`. Learn more about [configuring custom
+            # containers](/ai-platform/training/docs/distributed-training-containers).
+          &quot;tpuTfVersion&quot;: &quot;A String&quot;, # The AI Platform runtime version that includes a TensorFlow version matching
+              # the one used in the custom container. This field is required if the replica
+              # is a TPU worker that uses a custom container. Otherwise, do not specify
+              # this field. This must be a [runtime version that currently supports
+              # training with
+              # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
+              #
+              # Note that the version of TensorFlow included in a runtime version may
+              # differ from the numbering of the runtime version itself, because it may
+              # have a different [patch
+              # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
+              # In this field, you must specify the runtime version (TensorFlow minor
+              # version). For example, if your custom container runs TensorFlow `1.x.y`,
+              # specify `1.x`.
+          &quot;containerCommand&quot;: [ # The command with which the replica&#x27;s custom container is run.
+              # If provided, it will override default ENTRYPOINT of the docker image.
+              # If not provided, the docker image&#x27;s ENTRYPOINT is used.
+              # It cannot be set if custom container image is
+              # not provided.
+              # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+              # both cannot be set at the same time.
+            &quot;A String&quot;,
+          ],
+          &quot;imageUri&quot;: &quot;A String&quot;, # The Docker image to run on the replica. This image must be in Container
+              # Registry. Learn more about [configuring custom
+              # containers](/ai-platform/training/docs/distributed-training-containers).
+          &quot;containerArgs&quot;: [ # Arguments to the entrypoint command.
+              # The following rules apply for container_command and container_args:
+              # - If you do not supply command or args:
+              #   The defaults defined in the Docker image are used.
+              # - If you supply a command but no args:
+              #   The default EntryPoint and the default Cmd defined in the Docker image
+              #   are ignored. Your command is run without any arguments.
+              # - If you supply only args:
+              #   The default Entrypoint defined in the Docker image is run with the args
+              #   that you supplied.
+              # - If you supply a command and args:
+              #   The default Entrypoint and the default Cmd defined in the Docker image
+              #   are ignored. Your command is run with your args.
+              # It cannot be set if custom container image is
+              # not provided.
+              # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+              # both cannot be set at the same time.
+            &quot;A String&quot;,
+          ],
+          &quot;acceleratorConfig&quot;: { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
+              # [Learn about restrictions on accelerator configurations for
+              # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+              # Note that the AcceleratorConfig can be used in both Jobs and Versions.
+              # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
+              # [accelerators for online
+              # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
+            &quot;count&quot;: &quot;A String&quot;, # The number of accelerators to attach to each machine running the job.
+            &quot;type&quot;: &quot;A String&quot;, # The type of accelerator to use.
+          },
+        },
+        &quot;parameterServerCount&quot;: &quot;A String&quot;, # Optional. The number of parameter server replicas to use for the training
+            # job. Each replica in the cluster will be of the type specified in
+            # `parameter_server_type`.
+            #
+            # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
+            # set this value, you must also set `parameter_server_type`.
+            #
+            # The default value is zero.
+        &quot;packageUris&quot;: [ # Required. The Google Cloud Storage location of the packages with
+            # the training program and any additional dependencies.
+            # The maximum number of package URIs is 100.
+          &quot;A String&quot;,
+        ],
+        &quot;evaluatorCount&quot;: &quot;A String&quot;, # Optional. The number of evaluator replicas to use for the training job.
+            # Each replica in the cluster will be of the type specified in
+            # `evaluator_type`.
+            #
+            # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
+            # set this value, you must also set `evaluator_type`.
+            #
+            # The default value is zero.
+        &quot;masterType&quot;: &quot;A String&quot;, # Optional. Specifies the type of virtual machine to use for your training
+            # job&#x27;s master worker. You must specify this field when `scaleTier` is set to
+            # `CUSTOM`.
+            #
+            # You can use certain Compute Engine machine types directly in this field.
+            # The following types are supported:
+            #
+            # - `n1-standard-4`
+            # - `n1-standard-8`
+            # - `n1-standard-16`
+            # - `n1-standard-32`
+            # - `n1-standard-64`
+            # - `n1-standard-96`
+            # - `n1-highmem-2`
+            # - `n1-highmem-4`
+            # - `n1-highmem-8`
+            # - `n1-highmem-16`
+            # - `n1-highmem-32`
+            # - `n1-highmem-64`
+            # - `n1-highmem-96`
+            # - `n1-highcpu-16`
+            # - `n1-highcpu-32`
+            # - `n1-highcpu-64`
+            # - `n1-highcpu-96`
+            #
+            # Learn more about [using Compute Engine machine
+            # types](/ml-engine/docs/machine-types#compute-engine-machine-types).
+            #
+            # Alternatively, you can use the following legacy machine types:
+            #
+            # - `standard`
+            # - `large_model`
+            # - `complex_model_s`
+            # - `complex_model_m`
+            # - `complex_model_l`
+            # - `standard_gpu`
+            # - `complex_model_m_gpu`
+            # - `complex_model_l_gpu`
+            # - `standard_p100`
+            # - `complex_model_m_p100`
+            # - `standard_v100`
+            # - `large_model_v100`
+            # - `complex_model_m_v100`
+            # - `complex_model_l_v100`
+            #
+            # Learn more about [using legacy machine
+            # types](/ml-engine/docs/machine-types#legacy-machine-types).
+            #
+            # Finally, if you want to use a TPU for training, specify `cloud_tpu` in this
+            # field. Learn more about the [special configuration options for training
+            # with
+            # TPUs](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine).
+        &quot;runtimeVersion&quot;: &quot;A String&quot;, # Optional. The AI Platform runtime version to use for training. You must
+            # either specify this field or specify `masterConfig.imageUri`.
+            #
+            # For more information, see the [runtime version
+            # list](/ai-platform/training/docs/runtime-version-list) and learn [how to
+            # manage runtime versions](/ai-platform/training/docs/versioning).
+        &quot;evaluatorType&quot;: &quot;A String&quot;, # Optional. Specifies the type of virtual machine to use for your training
+            # job&#x27;s evaluator nodes.
+            #
+            # The supported values are the same as those described in the entry for
+            # `masterType`.
+            #
+            # This value must be consistent with the category of machine type that
+            # `masterType` uses. In other words, both must be Compute Engine machine
+            # types or both must be legacy machine types.
+            #
+            # This value must be present when `scaleTier` is set to `CUSTOM` and
+            # `evaluatorCount` is greater than zero.
+        &quot;region&quot;: &quot;A String&quot;, # Required. The region to run the training job in. See the [available
+            # regions](/ai-platform/training/docs/regions) for AI Platform Training.
+        &quot;workerType&quot;: &quot;A String&quot;, # Optional. Specifies the type of virtual machine to use for your training
+            # job&#x27;s worker nodes.
+            #
+            # The supported values are the same as those described in the entry for
+            # `masterType`.
+            #
+            # This value must be consistent with the category of machine type that
+            # `masterType` uses. In other words, both must be Compute Engine machine
+            # types or both must be legacy machine types.
+            #
+            # If you use `cloud_tpu` for this value, see special instructions for
+            # [configuring a custom TPU
+            # machine](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine).
+            #
+            # This value must be present when `scaleTier` is set to `CUSTOM` and
+            # `workerCount` is greater than zero.
+        &quot;parameterServerType&quot;: &quot;A String&quot;, # Optional. Specifies the type of virtual machine to use for your training
+            # job&#x27;s parameter server.
+            #
+            # The supported values are the same as those described in the entry for
+            # `master_type`.
+            #
+            # This value must be consistent with the category of machine type that
+            # `masterType` uses. In other words, both must be Compute Engine machine
+            # types or both must be legacy machine types.
+            #
+            # This value must be present when `scaleTier` is set to `CUSTOM` and
+            # `parameter_server_count` is greater than zero.
+        &quot;masterConfig&quot;: { # Represents the configuration for a replica in a cluster. # Optional. The configuration for your master worker.
+            #
+            # You should only set `masterConfig.acceleratorConfig` if `masterType` is set
+            # to a Compute Engine machine type. Learn about [restrictions on accelerator
+            # configurations for
+            # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+            #
+            # Set `masterConfig.imageUri` only if you build a custom image. Only one of
+            # `masterConfig.imageUri` and `runtimeVersion` should be set. Learn more
+            # about [configuring custom
+            # containers](/ai-platform/training/docs/distributed-training-containers).
+          &quot;tpuTfVersion&quot;: &quot;A String&quot;, # The AI Platform runtime version that includes a TensorFlow version matching
+              # the one used in the custom container. This field is required if the replica
+              # is a TPU worker that uses a custom container. Otherwise, do not specify
+              # this field. This must be a [runtime version that currently supports
+              # training with
+              # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
+              #
+              # Note that the version of TensorFlow included in a runtime version may
+              # differ from the numbering of the runtime version itself, because it may
+              # have a different [patch
+              # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
+              # In this field, you must specify the runtime version (TensorFlow minor
+              # version). For example, if your custom container runs TensorFlow `1.x.y`,
+              # specify `1.x`.
+          &quot;containerCommand&quot;: [ # The command with which the replica&#x27;s custom container is run.
+              # If provided, it will override default ENTRYPOINT of the docker image.
+              # If not provided, the docker image&#x27;s ENTRYPOINT is used.
+              # It cannot be set if custom container image is
+              # not provided.
+              # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+              # both cannot be set at the same time.
+            &quot;A String&quot;,
+          ],
+          &quot;imageUri&quot;: &quot;A String&quot;, # The Docker image to run on the replica. This image must be in Container
+              # Registry. Learn more about [configuring custom
+              # containers](/ai-platform/training/docs/distributed-training-containers).
+          &quot;containerArgs&quot;: [ # Arguments to the entrypoint command.
+              # The following rules apply for container_command and container_args:
+              # - If you do not supply command or args:
+              #   The defaults defined in the Docker image are used.
+              # - If you supply a command but no args:
+              #   The default EntryPoint and the default Cmd defined in the Docker image
+              #   are ignored. Your command is run without any arguments.
+              # - If you supply only args:
+              #   The default Entrypoint defined in the Docker image is run with the args
+              #   that you supplied.
+              # - If you supply a command and args:
+              #   The default Entrypoint and the default Cmd defined in the Docker image
+              #   are ignored. Your command is run with your args.
+              # It cannot be set if custom container image is
+              # not provided.
+              # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+              # both cannot be set at the same time.
+            &quot;A String&quot;,
+          ],
+          &quot;acceleratorConfig&quot;: { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
+              # [Learn about restrictions on accelerator configurations for
+              # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+              # Note that the AcceleratorConfig can be used in both Jobs and Versions.
+              # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
+              # [accelerators for online
+              # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
+            &quot;count&quot;: &quot;A String&quot;, # The number of accelerators to attach to each machine running the job.
+            &quot;type&quot;: &quot;A String&quot;, # The type of accelerator to use.
+          },
+        },
+        &quot;scaleTier&quot;: &quot;A String&quot;, # Required. Specifies the machine types, the number of replicas for workers
+            # and parameter servers.
+        &quot;jobDir&quot;: &quot;A String&quot;, # Optional. A Google Cloud Storage path in which to store training outputs
+            # and other data needed for training. This path is passed to your TensorFlow
+            # program as the &#x27;--job-dir&#x27; command-line argument. The benefit of specifying
+            # this field is that Cloud ML validates the path for use in training.
+        &quot;pythonVersion&quot;: &quot;A String&quot;, # Optional. The version of Python used in training. You must either specify
+            # this field or specify `masterConfig.imageUri`.
+            #
+            # The following Python versions are available:
+            #
+            # * Python &#x27;3.7&#x27; is available when `runtime_version` is set to &#x27;1.15&#x27; or
+            #   later.
+            # * Python &#x27;3.5&#x27; is available when `runtime_version` is set to a version
+            #   from &#x27;1.4&#x27; to &#x27;1.14&#x27;.
+            # * Python &#x27;2.7&#x27; is available when `runtime_version` is set to &#x27;1.15&#x27; or
+            #   earlier.
+            #
+            # Read more about the Python versions available for [each runtime
+            # version](/ml-engine/docs/runtime-version-list).
+        &quot;network&quot;: &quot;A String&quot;, # Optional. The full name of the Google Compute Engine
+            # [network](/compute/docs/networks-and-firewalls#networks) to which the Job
+            # is peered. For example, projects/12345/global/networks/myVPC. Format is of
+            # the form projects/{project}/global/networks/{network}. Where {project} is a
+            # project number, as in &#x27;12345&#x27;, and {network} is network name.&quot;.
+            #
+            # Private services access must already be configured for the network. If left
+            # unspecified, the Job is not peered with any network. Learn more -
+            # Connecting Job to user network over private
+            # IP.
+        &quot;scheduling&quot;: { # All parameters related to scheduling of training jobs. # Optional. Scheduling options for a training job.
+          &quot;maxWaitTime&quot;: &quot;A String&quot;,
+          &quot;maxRunningTime&quot;: &quot;A String&quot;, # Optional. The maximum job running time, expressed in seconds. The field can
+              # contain up to nine fractional digits, terminated by `s`. If not specified,
+              # this field defaults to `604800s` (seven days).
+              #
+              # If the training job is still running after this duration, AI Platform
+              # Training cancels it.
+              #
+              # For example, if you want to ensure your job runs for no more than 2 hours,
+              # set this field to `7200s` (2 hours * 60 minutes / hour * 60 seconds /
+              # minute).
+              #
+              # If you submit your training job using the `gcloud` tool, you can [provide
+              # this field in a `config.yaml`
+              # file](/ai-platform/training/docs/training-jobs#formatting_your_configuration_parameters).
+              # For example:
+              #
+              # ```yaml
+              # trainingInput:
+              #   ...
+              #   scheduling:
+              #     maxRunningTime: 7200s
+              #   ...
+              # ```
+        },
+        &quot;evaluatorConfig&quot;: { # Represents the configuration for a replica in a cluster. # Optional. The configuration for evaluators.
+            #
+            # You should only set `evaluatorConfig.acceleratorConfig` if
+            # `evaluatorType` is set to a Compute Engine machine type. [Learn
+            # about restrictions on accelerator configurations for
+            # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+            #
+            # Set `evaluatorConfig.imageUri` only if you build a custom image for
+            # your evaluator. If `evaluatorConfig.imageUri` has not been
+            # set, AI Platform uses the value of `masterConfig.imageUri`. Learn more about [configuring custom
+            # containers](/ai-platform/training/docs/distributed-training-containers).
+          &quot;tpuTfVersion&quot;: &quot;A String&quot;, # The AI Platform runtime version that includes a TensorFlow version matching
+              # the one used in the custom container. This field is required if the replica
+              # is a TPU worker that uses a custom container. Otherwise, do not specify
+              # this field. This must be a [runtime version that currently supports
+              # training with
+              # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
+              #
+              # Note that the version of TensorFlow included in a runtime version may
+              # differ from the numbering of the runtime version itself, because it may
+              # have a different [patch
+              # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
+              # In this field, you must specify the runtime version (TensorFlow minor
+              # version). For example, if your custom container runs TensorFlow `1.x.y`,
+              # specify `1.x`.
+          &quot;containerCommand&quot;: [ # The command with which the replica&#x27;s custom container is run.
+              # If provided, it will override default ENTRYPOINT of the docker image.
+              # If not provided, the docker image&#x27;s ENTRYPOINT is used.
+              # It cannot be set if custom container image is
+              # not provided.
+              # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+              # both cannot be set at the same time.
+            &quot;A String&quot;,
+          ],
+          &quot;imageUri&quot;: &quot;A String&quot;, # The Docker image to run on the replica. This image must be in Container
+              # Registry. Learn more about [configuring custom
+              # containers](/ai-platform/training/docs/distributed-training-containers).
+          &quot;containerArgs&quot;: [ # Arguments to the entrypoint command.
+              # The following rules apply for container_command and container_args:
+              # - If you do not supply command or args:
+              #   The defaults defined in the Docker image are used.
+              # - If you supply a command but no args:
+              #   The default EntryPoint and the default Cmd defined in the Docker image
+              #   are ignored. Your command is run without any arguments.
+              # - If you supply only args:
+              #   The default Entrypoint defined in the Docker image is run with the args
+              #   that you supplied.
+              # - If you supply a command and args:
+              #   The default Entrypoint and the default Cmd defined in the Docker image
+              #   are ignored. Your command is run with your args.
+              # It cannot be set if custom container image is
+              # not provided.
+              # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+              # both cannot be set at the same time.
+            &quot;A String&quot;,
+          ],
+          &quot;acceleratorConfig&quot;: { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
+              # [Learn about restrictions on accelerator configurations for
+              # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+              # Note that the AcceleratorConfig can be used in both Jobs and Versions.
+              # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
+              # [accelerators for online
+              # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
+            &quot;count&quot;: &quot;A String&quot;, # The number of accelerators to attach to each machine running the job.
+            &quot;type&quot;: &quot;A String&quot;, # The type of accelerator to use.
+          },
+        },
+        &quot;useChiefInTfConfig&quot;: True or False, # Optional. Use `chief` instead of `master` in the `TF_CONFIG` environment
+            # variable when training with a custom container. Defaults to `false`. [Learn
+            # more about this
+            # field.](/ai-platform/training/docs/distributed-training-details#chief-versus-master)
+            #
+            # This field has no effect for training jobs that don&#x27;t use a custom
+            # container.
+        &quot;workerCount&quot;: &quot;A String&quot;, # Optional. The number of worker replicas to use for the training job. Each
+            # replica in the cluster will be of the type specified in `worker_type`.
+            #
+            # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
+            # set this value, you must also set `worker_type`.
+            #
+            # The default value is zero.
+        &quot;pythonModule&quot;: &quot;A String&quot;, # Required. The Python module name to run after installing the packages.
+        &quot;args&quot;: [ # Optional. Command-line arguments passed to the training application when it
+            # starts. If your job uses a custom container, then the arguments are passed
+            # to the container&#x27;s &lt;a class=&quot;external&quot; target=&quot;_blank&quot;
+            # href=&quot;https://docs.docker.com/engine/reference/builder/#entrypoint&quot;&gt;
+            # `ENTRYPOINT`&lt;/a&gt; command.
+          &quot;A String&quot;,
+        ],
+      },
+      &quot;state&quot;: &quot;A String&quot;, # Output only. The detailed state of a job.
+      &quot;jobId&quot;: &quot;A String&quot;, # Required. The user-specified id of the job.
+      &quot;endTime&quot;: &quot;A String&quot;, # Output only. When the job processing was completed.
+      &quot;startTime&quot;: &quot;A String&quot;, # Output only. When the job processing was started.
+      &quot;predictionOutput&quot;: { # Represents results of a prediction job. # The current prediction job result.
+        &quot;errorCount&quot;: &quot;A String&quot;, # The number of data instances which resulted in errors.
+        &quot;outputPath&quot;: &quot;A String&quot;, # The output Google Cloud Storage location provided at the job creation time.
+        &quot;nodeHours&quot;: 3.14, # Node hours used by the batch prediction job.
+        &quot;predictionCount&quot;: &quot;A String&quot;, # The number of generated predictions.
+      },
+      &quot;trainingOutput&quot;: { # Represents results of a training job. Output only. # The current training job result.
+        &quot;trials&quot;: [ # Results for individual Hyperparameter trials.
+            # Only set for hyperparameter tuning jobs.
+          { # Represents the result of a single hyperparameter tuning trial from a
+              # training job. The TrainingOutput object that is returned on successful
+              # completion of a training job with hyperparameter tuning includes a list
+              # of HyperparameterOutput objects, one for each successful trial.
+            &quot;allMetrics&quot;: [ # All recorded object metrics for this trial. This field is not currently
+                # populated.
+              { # An observed value of a metric.
+                &quot;objectiveValue&quot;: 3.14, # The objective value at this training step.
+                &quot;trainingStep&quot;: &quot;A String&quot;, # The global training step for this metric.
+              },
+            ],
+            &quot;hyperparameters&quot;: { # The hyperparameters given to this trial.
+              &quot;a_key&quot;: &quot;A String&quot;,
+            },
+            &quot;trialId&quot;: &quot;A String&quot;, # The trial id for these results.
+            &quot;endTime&quot;: &quot;A String&quot;, # Output only. End time for the trial.
+            &quot;isTrialStoppedEarly&quot;: True or False, # True if the trial is stopped early.
+            &quot;startTime&quot;: &quot;A String&quot;, # Output only. Start time for the trial.
+            &quot;finalMetric&quot;: { # An observed value of a metric. # The final objective metric seen for this trial.
+              &quot;objectiveValue&quot;: 3.14, # The objective value at this training step.
+              &quot;trainingStep&quot;: &quot;A String&quot;, # The global training step for this metric.
+            },
+            &quot;builtInAlgorithmOutput&quot;: { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
+                # Only set for trials of built-in algorithms jobs that have succeeded.
+              &quot;framework&quot;: &quot;A String&quot;, # Framework on which the built-in algorithm was trained.
+              &quot;modelPath&quot;: &quot;A String&quot;, # The Cloud Storage path to the `model/` directory where the training job
+                  # saves the trained model. Only set for successful jobs that don&#x27;t use
+                  # hyperparameter tuning.
+              &quot;pythonVersion&quot;: &quot;A String&quot;, # Python version on which the built-in algorithm was trained.
+              &quot;runtimeVersion&quot;: &quot;A String&quot;, # AI Platform runtime version on which the built-in algorithm was
+                  # trained.
+            },
+            &quot;state&quot;: &quot;A String&quot;, # Output only. The detailed state of the trial.
+          },
+        ],
+        &quot;hyperparameterMetricTag&quot;: &quot;A String&quot;, # The TensorFlow summary tag name used for optimizing hyperparameter tuning
+            # trials. See
+            # [`HyperparameterSpec.hyperparameterMetricTag`](#HyperparameterSpec.FIELDS.hyperparameter_metric_tag)
+            # for more information. Only set for hyperparameter tuning jobs.
+        &quot;completedTrialCount&quot;: &quot;A String&quot;, # The number of hyperparameter tuning trials that completed successfully.
+            # Only set for hyperparameter tuning jobs.
+        &quot;isHyperparameterTuningJob&quot;: True or False, # Whether this job is a hyperparameter tuning job.
+        &quot;consumedMLUnits&quot;: 3.14, # The amount of ML units consumed by the job.
+        &quot;isBuiltInAlgorithmJob&quot;: True or False, # Whether this job is a built-in Algorithm job.
+        &quot;builtInAlgorithmOutput&quot;: { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
+            # Only set for built-in algorithms jobs.
+          &quot;framework&quot;: &quot;A String&quot;, # Framework on which the built-in algorithm was trained.
+          &quot;modelPath&quot;: &quot;A String&quot;, # The Cloud Storage path to the `model/` directory where the training job
+              # saves the trained model. Only set for successful jobs that don&#x27;t use
+              # hyperparameter tuning.
+          &quot;pythonVersion&quot;: &quot;A String&quot;, # Python version on which the built-in algorithm was trained.
+          &quot;runtimeVersion&quot;: &quot;A String&quot;, # AI Platform runtime version on which the built-in algorithm was
+              # trained.
+        },
+      },
+      &quot;createTime&quot;: &quot;A String&quot;, # Output only. When the job was created.
+      &quot;labels&quot;: { # Optional. One or more labels that you can add, to organize your jobs.
+          # Each label is a key-value pair, where both the key and the value are
+          # arbitrary strings that you supply.
+          # For more information, see the documentation on
+          # &lt;a href=&quot;/ml-engine/docs/tensorflow/resource-labels&quot;&gt;using labels&lt;/a&gt;.
+        &quot;a_key&quot;: &quot;A String&quot;,
+      },
+      &quot;predictionInput&quot;: { # Represents input parameters for a prediction job. # Input parameters to create a prediction job.
+        &quot;outputPath&quot;: &quot;A String&quot;, # Required. The output Google Cloud Storage location.
+        &quot;outputDataFormat&quot;: &quot;A String&quot;, # Optional. Format of the output data files, defaults to JSON.
+        &quot;dataFormat&quot;: &quot;A String&quot;, # Required. The format of the input data files.
+        &quot;batchSize&quot;: &quot;A String&quot;, # Optional. Number of records per batch, defaults to 64.
+            # The service will buffer batch_size number of records in memory before
+            # invoking one Tensorflow prediction call internally. So take the record
+            # size and memory available into consideration when setting this parameter.
+        &quot;runtimeVersion&quot;: &quot;A String&quot;, # Optional. The AI Platform runtime version to use for this batch
+            # prediction. If not set, AI Platform will pick the runtime version used
+            # during the CreateVersion request for this model version, or choose the
+            # latest stable version when model version information is not available
+            # such as when the model is specified by uri.
+        &quot;inputPaths&quot;: [ # Required. The Cloud Storage location of the input data files. May contain
+            # &lt;a href=&quot;/storage/docs/gsutil/addlhelp/WildcardNames&quot;&gt;wildcards&lt;/a&gt;.
+          &quot;A String&quot;,
+        ],
+        &quot;region&quot;: &quot;A String&quot;, # Required. The Google Compute Engine region to run the prediction job in.
+            # See the &lt;a href=&quot;/ml-engine/docs/tensorflow/regions&quot;&gt;available regions&lt;/a&gt;
+            # for AI Platform services.
+        &quot;versionName&quot;: &quot;A String&quot;, # Use this field if you want to specify a version of the model to use. The
+            # string is formatted the same way as `model_version`, with the addition
+            # of the version information:
+            #
+            # `&quot;projects/YOUR_PROJECT/models/YOUR_MODEL/versions/YOUR_VERSION&quot;`
+        &quot;modelName&quot;: &quot;A String&quot;, # Use this field if you want to use the default version for the specified
+            # model. The string must use the following format:
+            #
+            # `&quot;projects/YOUR_PROJECT/models/YOUR_MODEL&quot;`
+        &quot;uri&quot;: &quot;A String&quot;, # Use this field if you want to specify a Google Cloud Storage path for
+            # the model to use.
+        &quot;maxWorkerCount&quot;: &quot;A String&quot;, # Optional. The maximum number of workers to be used for parallel processing.
+            # Defaults to 10 if not specified.
+        &quot;signatureName&quot;: &quot;A String&quot;, # Optional. The name of the signature defined in the SavedModel to use for
+            # this job. Please refer to
+            # [SavedModel](https://tensorflow.github.io/serving/serving_basic.html)
+            # for information about how to use signatures.
+            #
+            # Defaults to
+            # [DEFAULT_SERVING_SIGNATURE_DEF_KEY](https://www.tensorflow.org/api_docs/python/tf/saved_model/signature_constants)
+            # , which is &quot;serving_default&quot;.
+      },
+      &quot;errorMessage&quot;: &quot;A String&quot;, # Output only. The details of a failure or a cancellation.
+    }</pre>
 </div>
 
 <div class="method">
@@ -1287,564 +1533,687 @@
   An object of the form:
 
     { # Represents a training or prediction job.
-    "errorMessage": "A String", # Output only. The details of a failure or a cancellation.
-    "trainingOutput": { # Represents results of a training job. Output only. # The current training job result.
-      "completedTrialCount": "A String", # The number of hyperparameter tuning trials that completed successfully.
-          # Only set for hyperparameter tuning jobs.
-      "trials": [ # Results for individual Hyperparameter trials.
-          # Only set for hyperparameter tuning jobs.
-        { # Represents the result of a single hyperparameter tuning trial from a
-            # training job. The TrainingOutput object that is returned on successful
-            # completion of a training job with hyperparameter tuning includes a list
-            # of HyperparameterOutput objects, one for each successful trial.
-          "startTime": "A String", # Output only. Start time for the trial.
-          "hyperparameters": { # The hyperparameters given to this trial.
-            "a_key": "A String",
+      &quot;etag&quot;: &quot;A String&quot;, # `etag` is used for optimistic concurrency control as a way to help
+          # prevent simultaneous updates of a job from overwriting each other.
+          # It is strongly suggested that systems make use of the `etag` in the
+          # read-modify-write cycle to perform job updates in order to avoid race
+          # conditions: An `etag` is returned in the response to `GetJob`, and
+          # systems are expected to put that etag in the request to `UpdateJob` to
+          # ensure that their change will be applied to the same version of the job.
+      &quot;trainingInput&quot;: { # Represents input parameters for a training job. When using the gcloud command # Input parameters to create a training job.
+          # to submit your training job, you can specify the input parameters as
+          # command-line arguments and/or in a YAML configuration file referenced from
+          # the --config command-line argument. For details, see the guide to [submitting
+          # a training job](/ai-platform/training/docs/training-jobs).
+        &quot;parameterServerConfig&quot;: { # Represents the configuration for a replica in a cluster. # Optional. The configuration for parameter servers.
+            #
+            # You should only set `parameterServerConfig.acceleratorConfig` if
+            # `parameterServerType` is set to a Compute Engine machine type. [Learn
+            # about restrictions on accelerator configurations for
+            # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+            #
+            # Set `parameterServerConfig.imageUri` only if you build a custom image for
+            # your parameter server. If `parameterServerConfig.imageUri` has not been
+            # set, AI Platform uses the value of `masterConfig.imageUri`. Learn more about [configuring custom
+            # containers](/ai-platform/training/docs/distributed-training-containers).
+          &quot;tpuTfVersion&quot;: &quot;A String&quot;, # The AI Platform runtime version that includes a TensorFlow version matching
+              # the one used in the custom container. This field is required if the replica
+              # is a TPU worker that uses a custom container. Otherwise, do not specify
+              # this field. This must be a [runtime version that currently supports
+              # training with
+              # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
+              #
+              # Note that the version of TensorFlow included in a runtime version may
+              # differ from the numbering of the runtime version itself, because it may
+              # have a different [patch
+              # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
+              # In this field, you must specify the runtime version (TensorFlow minor
+              # version). For example, if your custom container runs TensorFlow `1.x.y`,
+              # specify `1.x`.
+          &quot;containerCommand&quot;: [ # The command with which the replica&#x27;s custom container is run.
+              # If provided, it will override default ENTRYPOINT of the docker image.
+              # If not provided, the docker image&#x27;s ENTRYPOINT is used.
+              # It cannot be set if custom container image is
+              # not provided.
+              # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+              # both cannot be set at the same time.
+            &quot;A String&quot;,
+          ],
+          &quot;imageUri&quot;: &quot;A String&quot;, # The Docker image to run on the replica. This image must be in Container
+              # Registry. Learn more about [configuring custom
+              # containers](/ai-platform/training/docs/distributed-training-containers).
+          &quot;containerArgs&quot;: [ # Arguments to the entrypoint command.
+              # The following rules apply for container_command and container_args:
+              # - If you do not supply command or args:
+              #   The defaults defined in the Docker image are used.
+              # - If you supply a command but no args:
+              #   The default EntryPoint and the default Cmd defined in the Docker image
+              #   are ignored. Your command is run without any arguments.
+              # - If you supply only args:
+              #   The default Entrypoint defined in the Docker image is run with the args
+              #   that you supplied.
+              # - If you supply a command and args:
+              #   The default Entrypoint and the default Cmd defined in the Docker image
+              #   are ignored. Your command is run with your args.
+              # It cannot be set if custom container image is
+              # not provided.
+              # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+              # both cannot be set at the same time.
+            &quot;A String&quot;,
+          ],
+          &quot;acceleratorConfig&quot;: { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
+              # [Learn about restrictions on accelerator configurations for
+              # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+              # Note that the AcceleratorConfig can be used in both Jobs and Versions.
+              # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
+              # [accelerators for online
+              # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
+            &quot;count&quot;: &quot;A String&quot;, # The number of accelerators to attach to each machine running the job.
+            &quot;type&quot;: &quot;A String&quot;, # The type of accelerator to use.
           },
-          "finalMetric": { # An observed value of a metric. # The final objective metric seen for this trial.
-            "trainingStep": "A String", # The global training step for this metric.
-            "objectiveValue": 3.14, # The objective value at this training step.
-          },
-          "state": "A String", # Output only. The detailed state of the trial.
-          "allMetrics": [ # All recorded object metrics for this trial. This field is not currently
-              # populated.
-            { # An observed value of a metric.
-              "trainingStep": "A String", # The global training step for this metric.
-              "objectiveValue": 3.14, # The objective value at this training step.
+        },
+        &quot;encryptionConfig&quot;: { # Represents a custom encryption key configuration that can be applied to # Optional. Options for using customer-managed encryption keys (CMEK) to
+            # protect resources created by a training job, instead of using Google&#x27;s
+            # default encryption. If this is set, then all resources created by the
+            # training job will be encrypted with the customer-managed encryption key
+            # that you specify.
+            #
+            # [Learn how and when to use CMEK with AI Platform
+            # Training](/ai-platform/training/docs/cmek).
+            # a resource.
+          &quot;kmsKeyName&quot;: &quot;A String&quot;, # The Cloud KMS resource identifier of the customer-managed encryption key
+              # used to protect a resource, such as a training job. It has the following
+              # format:
+              # `projects/{PROJECT_ID}/locations/{REGION}/keyRings/{KEY_RING_NAME}/cryptoKeys/{KEY_NAME}`
+        },
+        &quot;hyperparameters&quot;: { # Represents a set of hyperparameters to optimize. # Optional. The set of Hyperparameters to tune.
+          &quot;hyperparameterMetricTag&quot;: &quot;A String&quot;, # Optional. The TensorFlow summary tag name to use for optimizing trials. For
+              # current versions of TensorFlow, this tag name should exactly match what is
+              # shown in TensorBoard, including all scopes.  For versions of TensorFlow
+              # prior to 0.12, this should be only the tag passed to tf.Summary.
+              # By default, &quot;training/hptuning/metric&quot; will be used.
+          &quot;params&quot;: [ # Required. The set of parameters to tune.
+            { # Represents a single hyperparameter to optimize.
+              &quot;type&quot;: &quot;A String&quot;, # Required. The type of the parameter.
+              &quot;categoricalValues&quot;: [ # Required if type is `CATEGORICAL`. The list of possible categories.
+                &quot;A String&quot;,
+              ],
+              &quot;parameterName&quot;: &quot;A String&quot;, # Required. The parameter name must be unique amongst all ParameterConfigs in
+                  # a HyperparameterSpec message. E.g., &quot;learning_rate&quot;.
+              &quot;minValue&quot;: 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
+                  # should be unset if type is `CATEGORICAL`. This value should be integers if
+                  # type is INTEGER.
+              &quot;discreteValues&quot;: [ # Required if type is `DISCRETE`.
+                  # A list of feasible points.
+                  # The list should be in strictly increasing order. For instance, this
+                  # parameter might have possible settings of 1.5, 2.5, and 4.0. This list
+                  # should not contain more than 1,000 values.
+                3.14,
+              ],
+              &quot;scaleType&quot;: &quot;A String&quot;, # Optional. How the parameter should be scaled to the hypercube.
+                  # Leave unset for categorical parameters.
+                  # Some kind of scaling is strongly recommended for real or integral
+                  # parameters (e.g., `UNIT_LINEAR_SCALE`).
+              &quot;maxValue&quot;: 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
+                  # should be unset if type is `CATEGORICAL`. This value should be integers if
+                  # type is `INTEGER`.
             },
           ],
-          "isTrialStoppedEarly": True or False, # True if the trial is stopped early.
-          "endTime": "A String", # Output only. End time for the trial.
-          "trialId": "A String", # The trial id for these results.
-          "builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
-              # Only set for trials of built-in algorithms jobs that have succeeded.
-            "framework": "A String", # Framework on which the built-in algorithm was trained.
-            "modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job
-                # saves the trained model. Only set for successful jobs that don't use
-                # hyperparameter tuning.
-            "runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was
-                # trained.
-            "pythonVersion": "A String", # Python version on which the built-in algorithm was trained.
+          &quot;enableTrialEarlyStopping&quot;: True or False, # Optional. Indicates if the hyperparameter tuning job enables auto trial
+              # early stopping.
+          &quot;resumePreviousJobId&quot;: &quot;A String&quot;, # Optional. The prior hyperparameter tuning job id that users hope to
+              # continue with. The job id will be used to find the corresponding vizier
+              # study guid and resume the study.
+          &quot;maxParallelTrials&quot;: 42, # Optional. The number of training trials to run concurrently.
+              # You can reduce the time it takes to perform hyperparameter tuning by adding
+              # trials in parallel. However, each trail only benefits from the information
+              # gained in completed trials. That means that a trial does not get access to
+              # the results of trials running at the same time, which could reduce the
+              # quality of the overall optimization.
+              #
+              # Each trial will use the same scale tier and machine types.
+              #
+              # Defaults to one.
+          &quot;maxFailedTrials&quot;: 42, # Optional. The number of failed trials that need to be seen before failing
+              # the hyperparameter tuning job. You can specify this field to override the
+              # default failing criteria for AI Platform hyperparameter tuning jobs.
+              #
+              # Defaults to zero, which means the service decides when a hyperparameter
+              # job should fail.
+          &quot;goal&quot;: &quot;A String&quot;, # Required. The type of goal to use for tuning. Available types are
+              # `MAXIMIZE` and `MINIMIZE`.
+              #
+              # Defaults to `MAXIMIZE`.
+          &quot;maxTrials&quot;: 42, # Optional. How many training trials should be attempted to optimize
+              # the specified hyperparameters.
+              #
+              # Defaults to one.
+          &quot;algorithm&quot;: &quot;A String&quot;, # Optional. The search algorithm specified for the hyperparameter
+              # tuning job.
+              # Uses the default AI Platform hyperparameter tuning
+              # algorithm if unspecified.
+        },
+        &quot;workerConfig&quot;: { # Represents the configuration for a replica in a cluster. # Optional. The configuration for workers.
+            #
+            # You should only set `workerConfig.acceleratorConfig` if `workerType` is set
+            # to a Compute Engine machine type. [Learn about restrictions on accelerator
+            # configurations for
+            # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+            #
+            # Set `workerConfig.imageUri` only if you build a custom image for your
+            # worker. If `workerConfig.imageUri` has not been set, AI Platform uses
+            # the value of `masterConfig.imageUri`. Learn more about [configuring custom
+            # containers](/ai-platform/training/docs/distributed-training-containers).
+          &quot;tpuTfVersion&quot;: &quot;A String&quot;, # The AI Platform runtime version that includes a TensorFlow version matching
+              # the one used in the custom container. This field is required if the replica
+              # is a TPU worker that uses a custom container. Otherwise, do not specify
+              # this field. This must be a [runtime version that currently supports
+              # training with
+              # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
+              #
+              # Note that the version of TensorFlow included in a runtime version may
+              # differ from the numbering of the runtime version itself, because it may
+              # have a different [patch
+              # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
+              # In this field, you must specify the runtime version (TensorFlow minor
+              # version). For example, if your custom container runs TensorFlow `1.x.y`,
+              # specify `1.x`.
+          &quot;containerCommand&quot;: [ # The command with which the replica&#x27;s custom container is run.
+              # If provided, it will override default ENTRYPOINT of the docker image.
+              # If not provided, the docker image&#x27;s ENTRYPOINT is used.
+              # It cannot be set if custom container image is
+              # not provided.
+              # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+              # both cannot be set at the same time.
+            &quot;A String&quot;,
+          ],
+          &quot;imageUri&quot;: &quot;A String&quot;, # The Docker image to run on the replica. This image must be in Container
+              # Registry. Learn more about [configuring custom
+              # containers](/ai-platform/training/docs/distributed-training-containers).
+          &quot;containerArgs&quot;: [ # Arguments to the entrypoint command.
+              # The following rules apply for container_command and container_args:
+              # - If you do not supply command or args:
+              #   The defaults defined in the Docker image are used.
+              # - If you supply a command but no args:
+              #   The default EntryPoint and the default Cmd defined in the Docker image
+              #   are ignored. Your command is run without any arguments.
+              # - If you supply only args:
+              #   The default Entrypoint defined in the Docker image is run with the args
+              #   that you supplied.
+              # - If you supply a command and args:
+              #   The default Entrypoint and the default Cmd defined in the Docker image
+              #   are ignored. Your command is run with your args.
+              # It cannot be set if custom container image is
+              # not provided.
+              # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+              # both cannot be set at the same time.
+            &quot;A String&quot;,
+          ],
+          &quot;acceleratorConfig&quot;: { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
+              # [Learn about restrictions on accelerator configurations for
+              # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+              # Note that the AcceleratorConfig can be used in both Jobs and Versions.
+              # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
+              # [accelerators for online
+              # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
+            &quot;count&quot;: &quot;A String&quot;, # The number of accelerators to attach to each machine running the job.
+            &quot;type&quot;: &quot;A String&quot;, # The type of accelerator to use.
           },
         },
-      ],
-      "isHyperparameterTuningJob": True or False, # Whether this job is a hyperparameter tuning job.
-      "isBuiltInAlgorithmJob": True or False, # Whether this job is a built-in Algorithm job.
-      "consumedMLUnits": 3.14, # The amount of ML units consumed by the job.
-      "hyperparameterMetricTag": "A String", # The TensorFlow summary tag name used for optimizing hyperparameter tuning
-          # trials. See
-          # [`HyperparameterSpec.hyperparameterMetricTag`](#HyperparameterSpec.FIELDS.hyperparameter_metric_tag)
-          # for more information. Only set for hyperparameter tuning jobs.
-      "builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
-          # Only set for built-in algorithms jobs.
-        "framework": "A String", # Framework on which the built-in algorithm was trained.
-        "modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job
-            # saves the trained model. Only set for successful jobs that don't use
-            # hyperparameter tuning.
-        "runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was
-            # trained.
-        "pythonVersion": "A String", # Python version on which the built-in algorithm was trained.
-      },
-    },
-    "predictionInput": { # Represents input parameters for a prediction job. # Input parameters to create a prediction job.
-      "modelName": "A String", # Use this field if you want to use the default version for the specified
-          # model. The string must use the following format:
-          #
-          # `"projects/YOUR_PROJECT/models/YOUR_MODEL"`
-      "signatureName": "A String", # Optional. The name of the signature defined in the SavedModel to use for
-          # this job. Please refer to
-          # [SavedModel](https://tensorflow.github.io/serving/serving_basic.html)
-          # for information about how to use signatures.
-          #
-          # Defaults to
-          # [DEFAULT_SERVING_SIGNATURE_DEF_KEY](https://www.tensorflow.org/api_docs/python/tf/saved_model/signature_constants)
-          # , which is "serving_default".
-      "runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for this batch
-          # prediction. If not set, AI Platform will pick the runtime version used
-          # during the CreateVersion request for this model version, or choose the
-          # latest stable version when model version information is not available
-          # such as when the model is specified by uri.
-      "batchSize": "A String", # Optional. Number of records per batch, defaults to 64.
-          # The service will buffer batch_size number of records in memory before
-          # invoking one Tensorflow prediction call internally. So take the record
-          # size and memory available into consideration when setting this parameter.
-      "maxWorkerCount": "A String", # Optional. The maximum number of workers to be used for parallel processing.
-          # Defaults to 10 if not specified.
-      "uri": "A String", # Use this field if you want to specify a Google Cloud Storage path for
-          # the model to use.
-      "outputPath": "A String", # Required. The output Google Cloud Storage location.
-      "dataFormat": "A String", # Required. The format of the input data files.
-      "versionName": "A String", # Use this field if you want to specify a version of the model to use. The
-          # string is formatted the same way as `model_version`, with the addition
-          # of the version information:
-          #
-          # `"projects/YOUR_PROJECT/models/YOUR_MODEL/versions/YOUR_VERSION"`
-      "region": "A String", # Required. The Google Compute Engine region to run the prediction job in.
-          # See the &lt;a href="/ml-engine/docs/tensorflow/regions"&gt;available regions&lt;/a&gt;
-          # for AI Platform services.
-      "inputPaths": [ # Required. The Cloud Storage location of the input data files. May contain
-          # &lt;a href="/storage/docs/gsutil/addlhelp/WildcardNames"&gt;wildcards&lt;/a&gt;.
-        "A String",
-      ],
-      "outputDataFormat": "A String", # Optional. Format of the output data files, defaults to JSON.
-    },
-    "labels": { # Optional. One or more labels that you can add, to organize your jobs.
-        # Each label is a key-value pair, where both the key and the value are
-        # arbitrary strings that you supply.
-        # For more information, see the documentation on
-        # &lt;a href="/ml-engine/docs/tensorflow/resource-labels"&gt;using labels&lt;/a&gt;.
-      "a_key": "A String",
-    },
-    "jobId": "A String", # Required. The user-specified id of the job.
-    "state": "A String", # Output only. The detailed state of a job.
-    "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help
-        # prevent simultaneous updates of a job from overwriting each other.
-        # It is strongly suggested that systems make use of the `etag` in the
-        # read-modify-write cycle to perform job updates in order to avoid race
-        # conditions: An `etag` is returned in the response to `GetJob`, and
-        # systems are expected to put that etag in the request to `UpdateJob` to
-        # ensure that their change will be applied to the same version of the job.
-    "startTime": "A String", # Output only. When the job processing was started.
-    "trainingInput": { # Represents input parameters for a training job. When using the gcloud command # Input parameters to create a training job.
-        # to submit your training job, you can specify the input parameters as
-        # command-line arguments and/or in a YAML configuration file referenced from
-        # the --config command-line argument. For details, see the guide to [submitting
-        # a training job](/ai-platform/training/docs/training-jobs).
-      "masterType": "A String", # Optional. Specifies the type of virtual machine to use for your training
-          # job's master worker. You must specify this field when `scaleTier` is set to
-          # `CUSTOM`.
-          #
-          # You can use certain Compute Engine machine types directly in this field.
-          # The following types are supported:
-          #
-          # - `n1-standard-4`
-          # - `n1-standard-8`
-          # - `n1-standard-16`
-          # - `n1-standard-32`
-          # - `n1-standard-64`
-          # - `n1-standard-96`
-          # - `n1-highmem-2`
-          # - `n1-highmem-4`
-          # - `n1-highmem-8`
-          # - `n1-highmem-16`
-          # - `n1-highmem-32`
-          # - `n1-highmem-64`
-          # - `n1-highmem-96`
-          # - `n1-highcpu-16`
-          # - `n1-highcpu-32`
-          # - `n1-highcpu-64`
-          # - `n1-highcpu-96`
-          #
-          # Learn more about [using Compute Engine machine
-          # types](/ml-engine/docs/machine-types#compute-engine-machine-types).
-          #
-          # Alternatively, you can use the following legacy machine types:
-          #
-          # - `standard`
-          # - `large_model`
-          # - `complex_model_s`
-          # - `complex_model_m`
-          # - `complex_model_l`
-          # - `standard_gpu`
-          # - `complex_model_m_gpu`
-          # - `complex_model_l_gpu`
-          # - `standard_p100`
-          # - `complex_model_m_p100`
-          # - `standard_v100`
-          # - `large_model_v100`
-          # - `complex_model_m_v100`
-          # - `complex_model_l_v100`
-          #
-          # Learn more about [using legacy machine
-          # types](/ml-engine/docs/machine-types#legacy-machine-types).
-          #
-          # Finally, if you want to use a TPU for training, specify `cloud_tpu` in this
-          # field. Learn more about the [special configuration options for training
-          # with
-          # TPUs](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine).
-      "jobDir": "A String", # Optional. A Google Cloud Storage path in which to store training outputs
-          # and other data needed for training. This path is passed to your TensorFlow
-          # program as the '--job-dir' command-line argument. The benefit of specifying
-          # this field is that Cloud ML validates the path for use in training.
-      "scheduling": { # All parameters related to scheduling of training jobs. # Optional. Scheduling options for a training job.
-        "maxRunningTime": "A String", # Optional. The maximum job running time, expressed in seconds. The field can
-            # contain up to nine fractional digits, terminated by `s`. By default there
-            # is no limit to the running time.
+        &quot;parameterServerCount&quot;: &quot;A String&quot;, # Optional. The number of parameter server replicas to use for the training
+            # job. Each replica in the cluster will be of the type specified in
+            # `parameter_server_type`.
             #
-            # If the training job is still running after this duration, AI Platform
-            # Training cancels it.
+            # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
+            # set this value, you must also set `parameter_server_type`.
             #
-            # For example, if you want to ensure your job runs for no more than 2 hours,
-            # set this field to `7200s` (2 hours * 60 minutes / hour * 60 seconds /
-            # minute).
+            # The default value is zero.
+        &quot;packageUris&quot;: [ # Required. The Google Cloud Storage location of the packages with
+            # the training program and any additional dependencies.
+            # The maximum number of package URIs is 100.
+          &quot;A String&quot;,
+        ],
+        &quot;evaluatorCount&quot;: &quot;A String&quot;, # Optional. The number of evaluator replicas to use for the training job.
+            # Each replica in the cluster will be of the type specified in
+            # `evaluator_type`.
             #
-            # If you submit your training job using the `gcloud` tool, you can [provide
-            # this field in a `config.yaml`
-            # file](/ai-platform/training/docs/training-jobs#formatting_your_configuration_parameters).
-            # For example:
+            # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
+            # set this value, you must also set `evaluator_type`.
             #
-            # ```yaml
-            # trainingInput:
-            #   ...
-            #   scheduling:
-            #     maxRunningTime: 7200s
-            #   ...
-            # ```
-      },
-      "parameterServerCount": "A String", # Optional. The number of parameter server replicas to use for the training
-          # job. Each replica in the cluster will be of the type specified in
-          # `parameter_server_type`.
-          #
-          # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
-          # set this value, you must also set `parameter_server_type`.
-          #
-          # The default value is zero.
-      "evaluatorCount": "A String", # Optional. The number of evaluator replicas to use for the training job.
-          # Each replica in the cluster will be of the type specified in
-          # `evaluator_type`.
-          #
-          # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
-          # set this value, you must also set `evaluator_type`.
-          #
-          # The default value is zero.
-      "workerType": "A String", # Optional. Specifies the type of virtual machine to use for your training
-          # job's worker nodes.
-          #
-          # The supported values are the same as those described in the entry for
-          # `masterType`.
-          #
-          # This value must be consistent with the category of machine type that
-          # `masterType` uses. In other words, both must be Compute Engine machine
-          # types or both must be legacy machine types.
-          #
-          # If you use `cloud_tpu` for this value, see special instructions for
-          # [configuring a custom TPU
-          # machine](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine).
-          #
-          # This value must be present when `scaleTier` is set to `CUSTOM` and
-          # `workerCount` is greater than zero.
-      "scaleTier": "A String", # Required. Specifies the machine types, the number of replicas for workers
-          # and parameter servers.
-      "packageUris": [ # Required. The Google Cloud Storage location of the packages with
-          # the training program and any additional dependencies.
-          # The maximum number of package URIs is 100.
-        "A String",
-      ],
-      "workerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for workers.
-          #
-          # You should only set `workerConfig.acceleratorConfig` if `workerType` is set
-          # to a Compute Engine machine type. [Learn about restrictions on accelerator
-          # configurations for
-          # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-          #
-          # Set `workerConfig.imageUri` only if you build a custom image for your
-          # worker. If `workerConfig.imageUri` has not been set, AI Platform uses
-          # the value of `masterConfig.imageUri`. Learn more about [configuring custom
-          # containers](/ai-platform/training/docs/distributed-training-containers).
-        "tpuTfVersion": "A String", # The AI Platform runtime version that includes a TensorFlow version matching
-            # the one used in the custom container. This field is required if the replica
-            # is a TPU worker that uses a custom container. Otherwise, do not specify
-            # this field. This must be a [runtime version that currently supports
-            # training with
-            # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
+            # The default value is zero.
+        &quot;masterType&quot;: &quot;A String&quot;, # Optional. Specifies the type of virtual machine to use for your training
+            # job&#x27;s master worker. You must specify this field when `scaleTier` is set to
+            # `CUSTOM`.
             #
-            # Note that the version of TensorFlow included in a runtime version may
-            # differ from the numbering of the runtime version itself, because it may
-            # have a different [patch
-            # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
-            # In this field, you must specify the runtime version (TensorFlow minor
-            # version). For example, if your custom container runs TensorFlow `1.x.y`,
-            # specify `1.x`.
-        "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
-            # [Learn about restrictions on accelerator configurations for
+            # You can use certain Compute Engine machine types directly in this field.
+            # The following types are supported:
+            #
+            # - `n1-standard-4`
+            # - `n1-standard-8`
+            # - `n1-standard-16`
+            # - `n1-standard-32`
+            # - `n1-standard-64`
+            # - `n1-standard-96`
+            # - `n1-highmem-2`
+            # - `n1-highmem-4`
+            # - `n1-highmem-8`
+            # - `n1-highmem-16`
+            # - `n1-highmem-32`
+            # - `n1-highmem-64`
+            # - `n1-highmem-96`
+            # - `n1-highcpu-16`
+            # - `n1-highcpu-32`
+            # - `n1-highcpu-64`
+            # - `n1-highcpu-96`
+            #
+            # Learn more about [using Compute Engine machine
+            # types](/ml-engine/docs/machine-types#compute-engine-machine-types).
+            #
+            # Alternatively, you can use the following legacy machine types:
+            #
+            # - `standard`
+            # - `large_model`
+            # - `complex_model_s`
+            # - `complex_model_m`
+            # - `complex_model_l`
+            # - `standard_gpu`
+            # - `complex_model_m_gpu`
+            # - `complex_model_l_gpu`
+            # - `standard_p100`
+            # - `complex_model_m_p100`
+            # - `standard_v100`
+            # - `large_model_v100`
+            # - `complex_model_m_v100`
+            # - `complex_model_l_v100`
+            #
+            # Learn more about [using legacy machine
+            # types](/ml-engine/docs/machine-types#legacy-machine-types).
+            #
+            # Finally, if you want to use a TPU for training, specify `cloud_tpu` in this
+            # field. Learn more about the [special configuration options for training
+            # with
+            # TPUs](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine).
+        &quot;runtimeVersion&quot;: &quot;A String&quot;, # Optional. The AI Platform runtime version to use for training. You must
+            # either specify this field or specify `masterConfig.imageUri`.
+            #
+            # For more information, see the [runtime version
+            # list](/ai-platform/training/docs/runtime-version-list) and learn [how to
+            # manage runtime versions](/ai-platform/training/docs/versioning).
+        &quot;evaluatorType&quot;: &quot;A String&quot;, # Optional. Specifies the type of virtual machine to use for your training
+            # job&#x27;s evaluator nodes.
+            #
+            # The supported values are the same as those described in the entry for
+            # `masterType`.
+            #
+            # This value must be consistent with the category of machine type that
+            # `masterType` uses. In other words, both must be Compute Engine machine
+            # types or both must be legacy machine types.
+            #
+            # This value must be present when `scaleTier` is set to `CUSTOM` and
+            # `evaluatorCount` is greater than zero.
+        &quot;region&quot;: &quot;A String&quot;, # Required. The region to run the training job in. See the [available
+            # regions](/ai-platform/training/docs/regions) for AI Platform Training.
+        &quot;workerType&quot;: &quot;A String&quot;, # Optional. Specifies the type of virtual machine to use for your training
+            # job&#x27;s worker nodes.
+            #
+            # The supported values are the same as those described in the entry for
+            # `masterType`.
+            #
+            # This value must be consistent with the category of machine type that
+            # `masterType` uses. In other words, both must be Compute Engine machine
+            # types or both must be legacy machine types.
+            #
+            # If you use `cloud_tpu` for this value, see special instructions for
+            # [configuring a custom TPU
+            # machine](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine).
+            #
+            # This value must be present when `scaleTier` is set to `CUSTOM` and
+            # `workerCount` is greater than zero.
+        &quot;parameterServerType&quot;: &quot;A String&quot;, # Optional. Specifies the type of virtual machine to use for your training
+            # job&#x27;s parameter server.
+            #
+            # The supported values are the same as those described in the entry for
+            # `master_type`.
+            #
+            # This value must be consistent with the category of machine type that
+            # `masterType` uses. In other words, both must be Compute Engine machine
+            # types or both must be legacy machine types.
+            #
+            # This value must be present when `scaleTier` is set to `CUSTOM` and
+            # `parameter_server_count` is greater than zero.
+        &quot;masterConfig&quot;: { # Represents the configuration for a replica in a cluster. # Optional. The configuration for your master worker.
+            #
+            # You should only set `masterConfig.acceleratorConfig` if `masterType` is set
+            # to a Compute Engine machine type. Learn about [restrictions on accelerator
+            # configurations for
             # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-            # Note that the AcceleratorConfig can be used in both Jobs and Versions.
-            # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
-            # [accelerators for online
-            # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
-          "count": "A String", # The number of accelerators to attach to each machine running the job.
-          "type": "A String", # The type of accelerator to use.
-        },
-        "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
-            # Registry. Learn more about [configuring custom
-            # containers](/ai-platform/training/docs/distributed-training-containers).
-      },
-      "evaluatorConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for evaluators.
-          #
-          # You should only set `evaluatorConfig.acceleratorConfig` if
-          # `evaluatorType` is set to a Compute Engine machine type. [Learn
-          # about restrictions on accelerator configurations for
-          # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-          #
-          # Set `evaluatorConfig.imageUri` only if you build a custom image for
-          # your evaluator. If `evaluatorConfig.imageUri` has not been
-          # set, AI Platform uses the value of `masterConfig.imageUri`. Learn more about [configuring custom
-          # containers](/ai-platform/training/docs/distributed-training-containers).
-        "tpuTfVersion": "A String", # The AI Platform runtime version that includes a TensorFlow version matching
-            # the one used in the custom container. This field is required if the replica
-            # is a TPU worker that uses a custom container. Otherwise, do not specify
-            # this field. This must be a [runtime version that currently supports
-            # training with
-            # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
             #
-            # Note that the version of TensorFlow included in a runtime version may
-            # differ from the numbering of the runtime version itself, because it may
-            # have a different [patch
-            # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
-            # In this field, you must specify the runtime version (TensorFlow minor
-            # version). For example, if your custom container runs TensorFlow `1.x.y`,
-            # specify `1.x`.
-        "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
-            # [Learn about restrictions on accelerator configurations for
+            # Set `masterConfig.imageUri` only if you build a custom image. Only one of
+            # `masterConfig.imageUri` and `runtimeVersion` should be set. Learn more
+            # about [configuring custom
+            # containers](/ai-platform/training/docs/distributed-training-containers).
+          &quot;tpuTfVersion&quot;: &quot;A String&quot;, # The AI Platform runtime version that includes a TensorFlow version matching
+              # the one used in the custom container. This field is required if the replica
+              # is a TPU worker that uses a custom container. Otherwise, do not specify
+              # this field. This must be a [runtime version that currently supports
+              # training with
+              # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
+              #
+              # Note that the version of TensorFlow included in a runtime version may
+              # differ from the numbering of the runtime version itself, because it may
+              # have a different [patch
+              # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
+              # In this field, you must specify the runtime version (TensorFlow minor
+              # version). For example, if your custom container runs TensorFlow `1.x.y`,
+              # specify `1.x`.
+          &quot;containerCommand&quot;: [ # The command with which the replica&#x27;s custom container is run.
+              # If provided, it will override default ENTRYPOINT of the docker image.
+              # If not provided, the docker image&#x27;s ENTRYPOINT is used.
+              # It cannot be set if custom container image is
+              # not provided.
+              # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+              # both cannot be set at the same time.
+            &quot;A String&quot;,
+          ],
+          &quot;imageUri&quot;: &quot;A String&quot;, # The Docker image to run on the replica. This image must be in Container
+              # Registry. Learn more about [configuring custom
+              # containers](/ai-platform/training/docs/distributed-training-containers).
+          &quot;containerArgs&quot;: [ # Arguments to the entrypoint command.
+              # The following rules apply for container_command and container_args:
+              # - If you do not supply command or args:
+              #   The defaults defined in the Docker image are used.
+              # - If you supply a command but no args:
+              #   The default EntryPoint and the default Cmd defined in the Docker image
+              #   are ignored. Your command is run without any arguments.
+              # - If you supply only args:
+              #   The default Entrypoint defined in the Docker image is run with the args
+              #   that you supplied.
+              # - If you supply a command and args:
+              #   The default Entrypoint and the default Cmd defined in the Docker image
+              #   are ignored. Your command is run with your args.
+              # It cannot be set if custom container image is
+              # not provided.
+              # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+              # both cannot be set at the same time.
+            &quot;A String&quot;,
+          ],
+          &quot;acceleratorConfig&quot;: { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
+              # [Learn about restrictions on accelerator configurations for
+              # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+              # Note that the AcceleratorConfig can be used in both Jobs and Versions.
+              # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
+              # [accelerators for online
+              # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
+            &quot;count&quot;: &quot;A String&quot;, # The number of accelerators to attach to each machine running the job.
+            &quot;type&quot;: &quot;A String&quot;, # The type of accelerator to use.
+          },
+        },
+        &quot;scaleTier&quot;: &quot;A String&quot;, # Required. Specifies the machine types, the number of replicas for workers
+            # and parameter servers.
+        &quot;jobDir&quot;: &quot;A String&quot;, # Optional. A Google Cloud Storage path in which to store training outputs
+            # and other data needed for training. This path is passed to your TensorFlow
+            # program as the &#x27;--job-dir&#x27; command-line argument. The benefit of specifying
+            # this field is that Cloud ML validates the path for use in training.
+        &quot;pythonVersion&quot;: &quot;A String&quot;, # Optional. The version of Python used in training. You must either specify
+            # this field or specify `masterConfig.imageUri`.
+            #
+            # The following Python versions are available:
+            #
+            # * Python &#x27;3.7&#x27; is available when `runtime_version` is set to &#x27;1.15&#x27; or
+            #   later.
+            # * Python &#x27;3.5&#x27; is available when `runtime_version` is set to a version
+            #   from &#x27;1.4&#x27; to &#x27;1.14&#x27;.
+            # * Python &#x27;2.7&#x27; is available when `runtime_version` is set to &#x27;1.15&#x27; or
+            #   earlier.
+            #
+            # Read more about the Python versions available for [each runtime
+            # version](/ml-engine/docs/runtime-version-list).
+        &quot;network&quot;: &quot;A String&quot;, # Optional. The full name of the Google Compute Engine
+            # [network](/compute/docs/networks-and-firewalls#networks) to which the Job
+            # is peered. For example, projects/12345/global/networks/myVPC. Format is of
+            # the form projects/{project}/global/networks/{network}. Where {project} is a
+            # project number, as in &#x27;12345&#x27;, and {network} is network name.&quot;.
+            #
+            # Private services access must already be configured for the network. If left
+            # unspecified, the Job is not peered with any network. Learn more -
+            # Connecting Job to user network over private
+            # IP.
+        &quot;scheduling&quot;: { # All parameters related to scheduling of training jobs. # Optional. Scheduling options for a training job.
+          &quot;maxWaitTime&quot;: &quot;A String&quot;,
+          &quot;maxRunningTime&quot;: &quot;A String&quot;, # Optional. The maximum job running time, expressed in seconds. The field can
+              # contain up to nine fractional digits, terminated by `s`. If not specified,
+              # this field defaults to `604800s` (seven days).
+              #
+              # If the training job is still running after this duration, AI Platform
+              # Training cancels it.
+              #
+              # For example, if you want to ensure your job runs for no more than 2 hours,
+              # set this field to `7200s` (2 hours * 60 minutes / hour * 60 seconds /
+              # minute).
+              #
+              # If you submit your training job using the `gcloud` tool, you can [provide
+              # this field in a `config.yaml`
+              # file](/ai-platform/training/docs/training-jobs#formatting_your_configuration_parameters).
+              # For example:
+              #
+              # ```yaml
+              # trainingInput:
+              #   ...
+              #   scheduling:
+              #     maxRunningTime: 7200s
+              #   ...
+              # ```
+        },
+        &quot;evaluatorConfig&quot;: { # Represents the configuration for a replica in a cluster. # Optional. The configuration for evaluators.
+            #
+            # You should only set `evaluatorConfig.acceleratorConfig` if
+            # `evaluatorType` is set to a Compute Engine machine type. [Learn
+            # about restrictions on accelerator configurations for
             # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-            # Note that the AcceleratorConfig can be used in both Jobs and Versions.
-            # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
-            # [accelerators for online
-            # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
-          "count": "A String", # The number of accelerators to attach to each machine running the job.
-          "type": "A String", # The type of accelerator to use.
-        },
-        "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
-            # Registry. Learn more about [configuring custom
+            #
+            # Set `evaluatorConfig.imageUri` only if you build a custom image for
+            # your evaluator. If `evaluatorConfig.imageUri` has not been
+            # set, AI Platform uses the value of `masterConfig.imageUri`. Learn more about [configuring custom
             # containers](/ai-platform/training/docs/distributed-training-containers).
-      },
-      "useChiefInTfConfig": True or False, # Optional. Use `chief` instead of `master` in the `TF_CONFIG` environment
-          # variable when training with a custom container. Defaults to `false`. [Learn
-          # more about this
-          # field.](/ai-platform/training/docs/distributed-training-details#chief-versus-master)
-          #
-          # This field has no effect for training jobs that don't use a custom
-          # container.
-      "masterConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for your master worker.
-          #
-          # You should only set `masterConfig.acceleratorConfig` if `masterType` is set
-          # to a Compute Engine machine type. Learn about [restrictions on accelerator
-          # configurations for
-          # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-          #
-          # Set `masterConfig.imageUri` only if you build a custom image. Only one of
-          # `masterConfig.imageUri` and `runtimeVersion` should be set. Learn more
-          # about [configuring custom
-          # containers](/ai-platform/training/docs/distributed-training-containers).
-        "tpuTfVersion": "A String", # The AI Platform runtime version that includes a TensorFlow version matching
-            # the one used in the custom container. This field is required if the replica
-            # is a TPU worker that uses a custom container. Otherwise, do not specify
-            # this field. This must be a [runtime version that currently supports
-            # training with
-            # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
-            #
-            # Note that the version of TensorFlow included in a runtime version may
-            # differ from the numbering of the runtime version itself, because it may
-            # have a different [patch
-            # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
-            # In this field, you must specify the runtime version (TensorFlow minor
-            # version). For example, if your custom container runs TensorFlow `1.x.y`,
-            # specify `1.x`.
-        "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
-            # [Learn about restrictions on accelerator configurations for
-            # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-            # Note that the AcceleratorConfig can be used in both Jobs and Versions.
-            # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
-            # [accelerators for online
-            # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
-          "count": "A String", # The number of accelerators to attach to each machine running the job.
-          "type": "A String", # The type of accelerator to use.
+          &quot;tpuTfVersion&quot;: &quot;A String&quot;, # The AI Platform runtime version that includes a TensorFlow version matching
+              # the one used in the custom container. This field is required if the replica
+              # is a TPU worker that uses a custom container. Otherwise, do not specify
+              # this field. This must be a [runtime version that currently supports
+              # training with
+              # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
+              #
+              # Note that the version of TensorFlow included in a runtime version may
+              # differ from the numbering of the runtime version itself, because it may
+              # have a different [patch
+              # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
+              # In this field, you must specify the runtime version (TensorFlow minor
+              # version). For example, if your custom container runs TensorFlow `1.x.y`,
+              # specify `1.x`.
+          &quot;containerCommand&quot;: [ # The command with which the replica&#x27;s custom container is run.
+              # If provided, it will override default ENTRYPOINT of the docker image.
+              # If not provided, the docker image&#x27;s ENTRYPOINT is used.
+              # It cannot be set if custom container image is
+              # not provided.
+              # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+              # both cannot be set at the same time.
+            &quot;A String&quot;,
+          ],
+          &quot;imageUri&quot;: &quot;A String&quot;, # The Docker image to run on the replica. This image must be in Container
+              # Registry. Learn more about [configuring custom
+              # containers](/ai-platform/training/docs/distributed-training-containers).
+          &quot;containerArgs&quot;: [ # Arguments to the entrypoint command.
+              # The following rules apply for container_command and container_args:
+              # - If you do not supply command or args:
+              #   The defaults defined in the Docker image are used.
+              # - If you supply a command but no args:
+              #   The default EntryPoint and the default Cmd defined in the Docker image
+              #   are ignored. Your command is run without any arguments.
+              # - If you supply only args:
+              #   The default Entrypoint defined in the Docker image is run with the args
+              #   that you supplied.
+              # - If you supply a command and args:
+              #   The default Entrypoint and the default Cmd defined in the Docker image
+              #   are ignored. Your command is run with your args.
+              # It cannot be set if custom container image is
+              # not provided.
+              # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+              # both cannot be set at the same time.
+            &quot;A String&quot;,
+          ],
+          &quot;acceleratorConfig&quot;: { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
+              # [Learn about restrictions on accelerator configurations for
+              # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+              # Note that the AcceleratorConfig can be used in both Jobs and Versions.
+              # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
+              # [accelerators for online
+              # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
+            &quot;count&quot;: &quot;A String&quot;, # The number of accelerators to attach to each machine running the job.
+            &quot;type&quot;: &quot;A String&quot;, # The type of accelerator to use.
+          },
         },
-        "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
-            # Registry. Learn more about [configuring custom
-            # containers](/ai-platform/training/docs/distributed-training-containers).
+        &quot;useChiefInTfConfig&quot;: True or False, # Optional. Use `chief` instead of `master` in the `TF_CONFIG` environment
+            # variable when training with a custom container. Defaults to `false`. [Learn
+            # more about this
+            # field.](/ai-platform/training/docs/distributed-training-details#chief-versus-master)
+            #
+            # This field has no effect for training jobs that don&#x27;t use a custom
+            # container.
+        &quot;workerCount&quot;: &quot;A String&quot;, # Optional. The number of worker replicas to use for the training job. Each
+            # replica in the cluster will be of the type specified in `worker_type`.
+            #
+            # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
+            # set this value, you must also set `worker_type`.
+            #
+            # The default value is zero.
+        &quot;pythonModule&quot;: &quot;A String&quot;, # Required. The Python module name to run after installing the packages.
+        &quot;args&quot;: [ # Optional. Command-line arguments passed to the training application when it
+            # starts. If your job uses a custom container, then the arguments are passed
+            # to the container&#x27;s &lt;a class=&quot;external&quot; target=&quot;_blank&quot;
+            # href=&quot;https://docs.docker.com/engine/reference/builder/#entrypoint&quot;&gt;
+            # `ENTRYPOINT`&lt;/a&gt; command.
+          &quot;A String&quot;,
+        ],
       },
-      "runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for training. You must
-          # either specify this field or specify `masterConfig.imageUri`.
-          #
-          # For more information, see the [runtime version
-          # list](/ai-platform/training/docs/runtime-version-list) and learn [how to
-          # manage runtime versions](/ai-platform/training/docs/versioning).
-      "hyperparameters": { # Represents a set of hyperparameters to optimize. # Optional. The set of Hyperparameters to tune.
-        "maxTrials": 42, # Optional. How many training trials should be attempted to optimize
-            # the specified hyperparameters.
-            #
-            # Defaults to one.
-        "goal": "A String", # Required. The type of goal to use for tuning. Available types are
-            # `MAXIMIZE` and `MINIMIZE`.
-            #
-            # Defaults to `MAXIMIZE`.
-        "algorithm": "A String", # Optional. The search algorithm specified for the hyperparameter
-            # tuning job.
-            # Uses the default AI Platform hyperparameter tuning
-            # algorithm if unspecified.
-        "maxFailedTrials": 42, # Optional. The number of failed trials that need to be seen before failing
-            # the hyperparameter tuning job. You can specify this field to override the
-            # default failing criteria for AI Platform hyperparameter tuning jobs.
-            #
-            # Defaults to zero, which means the service decides when a hyperparameter
-            # job should fail.
-        "enableTrialEarlyStopping": True or False, # Optional. Indicates if the hyperparameter tuning job enables auto trial
-            # early stopping.
-        "resumePreviousJobId": "A String", # Optional. The prior hyperparameter tuning job id that users hope to
-            # continue with. The job id will be used to find the corresponding vizier
-            # study guid and resume the study.
-        "params": [ # Required. The set of parameters to tune.
-          { # Represents a single hyperparameter to optimize.
-            "maxValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
-                # should be unset if type is `CATEGORICAL`. This value should be integers if
-                # type is `INTEGER`.
-            "minValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
-                # should be unset if type is `CATEGORICAL`. This value should be integers if
-                # type is INTEGER.
-            "discreteValues": [ # Required if type is `DISCRETE`.
-                # A list of feasible points.
-                # The list should be in strictly increasing order. For instance, this
-                # parameter might have possible settings of 1.5, 2.5, and 4.0. This list
-                # should not contain more than 1,000 values.
-              3.14,
+      &quot;state&quot;: &quot;A String&quot;, # Output only. The detailed state of a job.
+      &quot;jobId&quot;: &quot;A String&quot;, # Required. The user-specified id of the job.
+      &quot;endTime&quot;: &quot;A String&quot;, # Output only. When the job processing was completed.
+      &quot;startTime&quot;: &quot;A String&quot;, # Output only. When the job processing was started.
+      &quot;predictionOutput&quot;: { # Represents results of a prediction job. # The current prediction job result.
+        &quot;errorCount&quot;: &quot;A String&quot;, # The number of data instances which resulted in errors.
+        &quot;outputPath&quot;: &quot;A String&quot;, # The output Google Cloud Storage location provided at the job creation time.
+        &quot;nodeHours&quot;: 3.14, # Node hours used by the batch prediction job.
+        &quot;predictionCount&quot;: &quot;A String&quot;, # The number of generated predictions.
+      },
+      &quot;trainingOutput&quot;: { # Represents results of a training job. Output only. # The current training job result.
+        &quot;trials&quot;: [ # Results for individual Hyperparameter trials.
+            # Only set for hyperparameter tuning jobs.
+          { # Represents the result of a single hyperparameter tuning trial from a
+              # training job. The TrainingOutput object that is returned on successful
+              # completion of a training job with hyperparameter tuning includes a list
+              # of HyperparameterOutput objects, one for each successful trial.
+            &quot;allMetrics&quot;: [ # All recorded object metrics for this trial. This field is not currently
+                # populated.
+              { # An observed value of a metric.
+                &quot;objectiveValue&quot;: 3.14, # The objective value at this training step.
+                &quot;trainingStep&quot;: &quot;A String&quot;, # The global training step for this metric.
+              },
             ],
-            "parameterName": "A String", # Required. The parameter name must be unique amongst all ParameterConfigs in
-                # a HyperparameterSpec message. E.g., "learning_rate".
-            "categoricalValues": [ # Required if type is `CATEGORICAL`. The list of possible categories.
-              "A String",
-            ],
-            "type": "A String", # Required. The type of the parameter.
-            "scaleType": "A String", # Optional. How the parameter should be scaled to the hypercube.
-                # Leave unset for categorical parameters.
-                # Some kind of scaling is strongly recommended for real or integral
-                # parameters (e.g., `UNIT_LINEAR_SCALE`).
+            &quot;hyperparameters&quot;: { # The hyperparameters given to this trial.
+              &quot;a_key&quot;: &quot;A String&quot;,
+            },
+            &quot;trialId&quot;: &quot;A String&quot;, # The trial id for these results.
+            &quot;endTime&quot;: &quot;A String&quot;, # Output only. End time for the trial.
+            &quot;isTrialStoppedEarly&quot;: True or False, # True if the trial is stopped early.
+            &quot;startTime&quot;: &quot;A String&quot;, # Output only. Start time for the trial.
+            &quot;finalMetric&quot;: { # An observed value of a metric. # The final objective metric seen for this trial.
+              &quot;objectiveValue&quot;: 3.14, # The objective value at this training step.
+              &quot;trainingStep&quot;: &quot;A String&quot;, # The global training step for this metric.
+            },
+            &quot;builtInAlgorithmOutput&quot;: { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
+                # Only set for trials of built-in algorithms jobs that have succeeded.
+              &quot;framework&quot;: &quot;A String&quot;, # Framework on which the built-in algorithm was trained.
+              &quot;modelPath&quot;: &quot;A String&quot;, # The Cloud Storage path to the `model/` directory where the training job
+                  # saves the trained model. Only set for successful jobs that don&#x27;t use
+                  # hyperparameter tuning.
+              &quot;pythonVersion&quot;: &quot;A String&quot;, # Python version on which the built-in algorithm was trained.
+              &quot;runtimeVersion&quot;: &quot;A String&quot;, # AI Platform runtime version on which the built-in algorithm was
+                  # trained.
+            },
+            &quot;state&quot;: &quot;A String&quot;, # Output only. The detailed state of the trial.
           },
         ],
-        "hyperparameterMetricTag": "A String", # Optional. The TensorFlow summary tag name to use for optimizing trials. For
-            # current versions of TensorFlow, this tag name should exactly match what is
-            # shown in TensorBoard, including all scopes.  For versions of TensorFlow
-            # prior to 0.12, this should be only the tag passed to tf.Summary.
-            # By default, "training/hptuning/metric" will be used.
-        "maxParallelTrials": 42, # Optional. The number of training trials to run concurrently.
-            # You can reduce the time it takes to perform hyperparameter tuning by adding
-            # trials in parallel. However, each trail only benefits from the information
-            # gained in completed trials. That means that a trial does not get access to
-            # the results of trials running at the same time, which could reduce the
-            # quality of the overall optimization.
-            #
-            # Each trial will use the same scale tier and machine types.
-            #
-            # Defaults to one.
-      },
-      "args": [ # Optional. Command-line arguments passed to the training application when it
-          # starts. If your job uses a custom container, then the arguments are passed
-          # to the container's &lt;a class="external" target="_blank"
-          # href="https://docs.docker.com/engine/reference/builder/#entrypoint"&gt;
-          # `ENTRYPOINT`&lt;/a&gt; command.
-        "A String",
-      ],
-      "pythonModule": "A String", # Required. The Python module name to run after installing the packages.
-      "workerCount": "A String", # Optional. The number of worker replicas to use for the training job. Each
-          # replica in the cluster will be of the type specified in `worker_type`.
-          #
-          # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
-          # set this value, you must also set `worker_type`.
-          #
-          # The default value is zero.
-      "encryptionConfig": { # Represents a custom encryption key configuration that can be applied to # Optional. Options for using customer-managed encryption keys (CMEK) to
-          # protect resources created by a training job, instead of using Google's
-          # default encryption. If this is set, then all resources created by the
-          # training job will be encrypted with the customer-managed encryption key
-          # that you specify.
-          #
-          # [Learn how and when to use CMEK with AI Platform
-          # Training](/ai-platform/training/docs/cmek).
-          # a resource.
-        "kmsKeyName": "A String", # The Cloud KMS resource identifier of the customer-managed encryption key
-            # used to protect a resource, such as a training job. It has the following
-            # format:
-            # `projects/{PROJECT_ID}/locations/{REGION}/keyRings/{KEY_RING_NAME}/cryptoKeys/{KEY_NAME}`
-      },
-      "parameterServerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for parameter servers.
-          #
-          # You should only set `parameterServerConfig.acceleratorConfig` if
-          # `parameterServerType` is set to a Compute Engine machine type. [Learn
-          # about restrictions on accelerator configurations for
-          # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-          #
-          # Set `parameterServerConfig.imageUri` only if you build a custom image for
-          # your parameter server. If `parameterServerConfig.imageUri` has not been
-          # set, AI Platform uses the value of `masterConfig.imageUri`. Learn more about [configuring custom
-          # containers](/ai-platform/training/docs/distributed-training-containers).
-        "tpuTfVersion": "A String", # The AI Platform runtime version that includes a TensorFlow version matching
-            # the one used in the custom container. This field is required if the replica
-            # is a TPU worker that uses a custom container. Otherwise, do not specify
-            # this field. This must be a [runtime version that currently supports
-            # training with
-            # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
-            #
-            # Note that the version of TensorFlow included in a runtime version may
-            # differ from the numbering of the runtime version itself, because it may
-            # have a different [patch
-            # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
-            # In this field, you must specify the runtime version (TensorFlow minor
-            # version). For example, if your custom container runs TensorFlow `1.x.y`,
-            # specify `1.x`.
-        "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
-            # [Learn about restrictions on accelerator configurations for
-            # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-            # Note that the AcceleratorConfig can be used in both Jobs and Versions.
-            # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
-            # [accelerators for online
-            # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
-          "count": "A String", # The number of accelerators to attach to each machine running the job.
-          "type": "A String", # The type of accelerator to use.
+        &quot;hyperparameterMetricTag&quot;: &quot;A String&quot;, # The TensorFlow summary tag name used for optimizing hyperparameter tuning
+            # trials. See
+            # [`HyperparameterSpec.hyperparameterMetricTag`](#HyperparameterSpec.FIELDS.hyperparameter_metric_tag)
+            # for more information. Only set for hyperparameter tuning jobs.
+        &quot;completedTrialCount&quot;: &quot;A String&quot;, # The number of hyperparameter tuning trials that completed successfully.
+            # Only set for hyperparameter tuning jobs.
+        &quot;isHyperparameterTuningJob&quot;: True or False, # Whether this job is a hyperparameter tuning job.
+        &quot;consumedMLUnits&quot;: 3.14, # The amount of ML units consumed by the job.
+        &quot;isBuiltInAlgorithmJob&quot;: True or False, # Whether this job is a built-in Algorithm job.
+        &quot;builtInAlgorithmOutput&quot;: { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
+            # Only set for built-in algorithms jobs.
+          &quot;framework&quot;: &quot;A String&quot;, # Framework on which the built-in algorithm was trained.
+          &quot;modelPath&quot;: &quot;A String&quot;, # The Cloud Storage path to the `model/` directory where the training job
+              # saves the trained model. Only set for successful jobs that don&#x27;t use
+              # hyperparameter tuning.
+          &quot;pythonVersion&quot;: &quot;A String&quot;, # Python version on which the built-in algorithm was trained.
+          &quot;runtimeVersion&quot;: &quot;A String&quot;, # AI Platform runtime version on which the built-in algorithm was
+              # trained.
         },
-        "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
-            # Registry. Learn more about [configuring custom
-            # containers](/ai-platform/training/docs/distributed-training-containers).
       },
-      "region": "A String", # Required. The region to run the training job in. See the [available
-          # regions](/ai-platform/training/docs/regions) for AI Platform Training.
-      "pythonVersion": "A String", # Optional. The version of Python used in training. You must either specify
-          # this field or specify `masterConfig.imageUri`.
-          #
-          # The following Python versions are available:
-          #
-          # * Python '3.7' is available when `runtime_version` is set to '1.15' or
-          #   later.
-          # * Python '3.5' is available when `runtime_version` is set to a version
-          #   from '1.4' to '1.14'.
-          # * Python '2.7' is available when `runtime_version` is set to '1.15' or
-          #   earlier.
-          #
-          # Read more about the Python versions available for [each runtime
-          # version](/ml-engine/docs/runtime-version-list).
-      "evaluatorType": "A String", # Optional. Specifies the type of virtual machine to use for your training
-          # job's evaluator nodes.
-          #
-          # The supported values are the same as those described in the entry for
-          # `masterType`.
-          #
-          # This value must be consistent with the category of machine type that
-          # `masterType` uses. In other words, both must be Compute Engine machine
-          # types or both must be legacy machine types.
-          #
-          # This value must be present when `scaleTier` is set to `CUSTOM` and
-          # `evaluatorCount` is greater than zero.
-      "parameterServerType": "A String", # Optional. Specifies the type of virtual machine to use for your training
-          # job's parameter server.
-          #
-          # The supported values are the same as those described in the entry for
-          # `master_type`.
-          #
-          # This value must be consistent with the category of machine type that
-          # `masterType` uses. In other words, both must be Compute Engine machine
-          # types or both must be legacy machine types.
-          #
-          # This value must be present when `scaleTier` is set to `CUSTOM` and
-          # `parameter_server_count` is greater than zero.
-    },
-    "endTime": "A String", # Output only. When the job processing was completed.
-    "predictionOutput": { # Represents results of a prediction job. # The current prediction job result.
-      "outputPath": "A String", # The output Google Cloud Storage location provided at the job creation time.
-      "nodeHours": 3.14, # Node hours used by the batch prediction job.
-      "predictionCount": "A String", # The number of generated predictions.
-      "errorCount": "A String", # The number of data instances which resulted in errors.
-    },
-    "createTime": "A String", # Output only. When the job was created.
-  }</pre>
+      &quot;createTime&quot;: &quot;A String&quot;, # Output only. When the job was created.
+      &quot;labels&quot;: { # Optional. One or more labels that you can add, to organize your jobs.
+          # Each label is a key-value pair, where both the key and the value are
+          # arbitrary strings that you supply.
+          # For more information, see the documentation on
+          # &lt;a href=&quot;/ml-engine/docs/tensorflow/resource-labels&quot;&gt;using labels&lt;/a&gt;.
+        &quot;a_key&quot;: &quot;A String&quot;,
+      },
+      &quot;predictionInput&quot;: { # Represents input parameters for a prediction job. # Input parameters to create a prediction job.
+        &quot;outputPath&quot;: &quot;A String&quot;, # Required. The output Google Cloud Storage location.
+        &quot;outputDataFormat&quot;: &quot;A String&quot;, # Optional. Format of the output data files, defaults to JSON.
+        &quot;dataFormat&quot;: &quot;A String&quot;, # Required. The format of the input data files.
+        &quot;batchSize&quot;: &quot;A String&quot;, # Optional. Number of records per batch, defaults to 64.
+            # The service will buffer batch_size number of records in memory before
+            # invoking one Tensorflow prediction call internally. So take the record
+            # size and memory available into consideration when setting this parameter.
+        &quot;runtimeVersion&quot;: &quot;A String&quot;, # Optional. The AI Platform runtime version to use for this batch
+            # prediction. If not set, AI Platform will pick the runtime version used
+            # during the CreateVersion request for this model version, or choose the
+            # latest stable version when model version information is not available
+            # such as when the model is specified by uri.
+        &quot;inputPaths&quot;: [ # Required. The Cloud Storage location of the input data files. May contain
+            # &lt;a href=&quot;/storage/docs/gsutil/addlhelp/WildcardNames&quot;&gt;wildcards&lt;/a&gt;.
+          &quot;A String&quot;,
+        ],
+        &quot;region&quot;: &quot;A String&quot;, # Required. The Google Compute Engine region to run the prediction job in.
+            # See the &lt;a href=&quot;/ml-engine/docs/tensorflow/regions&quot;&gt;available regions&lt;/a&gt;
+            # for AI Platform services.
+        &quot;versionName&quot;: &quot;A String&quot;, # Use this field if you want to specify a version of the model to use. The
+            # string is formatted the same way as `model_version`, with the addition
+            # of the version information:
+            #
+            # `&quot;projects/YOUR_PROJECT/models/YOUR_MODEL/versions/YOUR_VERSION&quot;`
+        &quot;modelName&quot;: &quot;A String&quot;, # Use this field if you want to use the default version for the specified
+            # model. The string must use the following format:
+            #
+            # `&quot;projects/YOUR_PROJECT/models/YOUR_MODEL&quot;`
+        &quot;uri&quot;: &quot;A String&quot;, # Use this field if you want to specify a Google Cloud Storage path for
+            # the model to use.
+        &quot;maxWorkerCount&quot;: &quot;A String&quot;, # Optional. The maximum number of workers to be used for parallel processing.
+            # Defaults to 10 if not specified.
+        &quot;signatureName&quot;: &quot;A String&quot;, # Optional. The name of the signature defined in the SavedModel to use for
+            # this job. Please refer to
+            # [SavedModel](https://tensorflow.github.io/serving/serving_basic.html)
+            # for information about how to use signatures.
+            #
+            # Defaults to
+            # [DEFAULT_SERVING_SIGNATURE_DEF_KEY](https://www.tensorflow.org/api_docs/python/tf/saved_model/signature_constants)
+            # , which is &quot;serving_default&quot;.
+      },
+      &quot;errorMessage&quot;: &quot;A String&quot;, # Output only. The details of a failure or a cancellation.
+    }</pre>
 </div>
 
 <div class="method">
@@ -1864,6 +2233,10 @@
 Requests for policies with any conditional bindings must specify version 3.
 Policies without any conditional bindings may specify any valid value or
 leave the field unset.
+
+To learn which resources support conditions in their IAM policies, see the
+[IAM
+documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
   x__xgafv: string, V1 error format.
     Allowed values
       1 - v1 error format
@@ -1882,36 +2255,40 @@
       # permissions; each `role` can be an IAM predefined role or a user-created
       # custom role.
       #
-      # Optionally, a `binding` can specify a `condition`, which is a logical
-      # expression that allows access to a resource only if the expression evaluates
-      # to `true`. A condition can add constraints based on attributes of the
-      # request, the resource, or both.
+      # For some types of Google Cloud resources, a `binding` can also specify a
+      # `condition`, which is a logical expression that allows access to a resource
+      # only if the expression evaluates to `true`. A condition can add constraints
+      # based on attributes of the request, the resource, or both. To learn which
+      # resources support conditions in their IAM policies, see the
+      # [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
       #
       # **JSON example:**
       #
       #     {
-      #       "bindings": [
+      #       &quot;bindings&quot;: [
       #         {
-      #           "role": "roles/resourcemanager.organizationAdmin",
-      #           "members": [
-      #             "user:mike@example.com",
-      #             "group:admins@example.com",
-      #             "domain:google.com",
-      #             "serviceAccount:my-project-id@appspot.gserviceaccount.com"
+      #           &quot;role&quot;: &quot;roles/resourcemanager.organizationAdmin&quot;,
+      #           &quot;members&quot;: [
+      #             &quot;user:mike@example.com&quot;,
+      #             &quot;group:admins@example.com&quot;,
+      #             &quot;domain:google.com&quot;,
+      #             &quot;serviceAccount:my-project-id@appspot.gserviceaccount.com&quot;
       #           ]
       #         },
       #         {
-      #           "role": "roles/resourcemanager.organizationViewer",
-      #           "members": ["user:eve@example.com"],
-      #           "condition": {
-      #             "title": "expirable access",
-      #             "description": "Does not grant access after Sep 2020",
-      #             "expression": "request.time &lt; timestamp('2020-10-01T00:00:00.000Z')",
+      #           &quot;role&quot;: &quot;roles/resourcemanager.organizationViewer&quot;,
+      #           &quot;members&quot;: [
+      #             &quot;user:eve@example.com&quot;
+      #           ],
+      #           &quot;condition&quot;: {
+      #             &quot;title&quot;: &quot;expirable access&quot;,
+      #             &quot;description&quot;: &quot;Does not grant access after Sep 2020&quot;,
+      #             &quot;expression&quot;: &quot;request.time &lt; timestamp(&#x27;2020-10-01T00:00:00.000Z&#x27;)&quot;,
       #           }
       #         }
       #       ],
-      #       "etag": "BwWWja0YfJA=",
-      #       "version": 3
+      #       &quot;etag&quot;: &quot;BwWWja0YfJA=&quot;,
+      #       &quot;version&quot;: 3
       #     }
       #
       # **YAML example:**
@@ -1929,19 +2306,190 @@
       #       condition:
       #         title: expirable access
       #         description: Does not grant access after Sep 2020
-      #         expression: request.time &lt; timestamp('2020-10-01T00:00:00.000Z')
+      #         expression: request.time &lt; timestamp(&#x27;2020-10-01T00:00:00.000Z&#x27;)
       #     - etag: BwWWja0YfJA=
       #     - version: 3
       #
       # For a description of IAM and its features, see the
       # [IAM documentation](https://cloud.google.com/iam/docs/).
-    "bindings": [ # Associates a list of `members` to a `role`. Optionally, may specify a
+    &quot;etag&quot;: &quot;A String&quot;, # `etag` is used for optimistic concurrency control as a way to help
+        # prevent simultaneous updates of a policy from overwriting each other.
+        # It is strongly suggested that systems make use of the `etag` in the
+        # read-modify-write cycle to perform policy updates in order to avoid race
+        # conditions: An `etag` is returned in the response to `getIamPolicy`, and
+        # systems are expected to put that etag in the request to `setIamPolicy` to
+        # ensure that their change will be applied to the same version of the policy.
+        #
+        # **Important:** If you use IAM Conditions, you must include the `etag` field
+        # whenever you call `setIamPolicy`. If you omit this field, then IAM allows
+        # you to overwrite a version `3` policy with a version `1` policy, and all of
+        # the conditions in the version `3` policy are lost.
+    &quot;version&quot;: 42, # Specifies the format of the policy.
+        #
+        # Valid values are `0`, `1`, and `3`. Requests that specify an invalid value
+        # are rejected.
+        #
+        # Any operation that affects conditional role bindings must specify version
+        # `3`. This requirement applies to the following operations:
+        #
+        # * Getting a policy that includes a conditional role binding
+        # * Adding a conditional role binding to a policy
+        # * Changing a conditional role binding in a policy
+        # * Removing any role binding, with or without a condition, from a policy
+        #   that includes conditions
+        #
+        # **Important:** If you use IAM Conditions, you must include the `etag` field
+        # whenever you call `setIamPolicy`. If you omit this field, then IAM allows
+        # you to overwrite a version `3` policy with a version `1` policy, and all of
+        # the conditions in the version `3` policy are lost.
+        #
+        # If a policy does not include any conditions, operations on that policy may
+        # specify any valid version or leave the field unset.
+        #
+        # To learn which resources support conditions in their IAM policies, see the
+        # [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+    &quot;auditConfigs&quot;: [ # Specifies cloud audit logging configuration for this policy.
+      { # Specifies the audit configuration for a service.
+          # The configuration determines which permission types are logged, and what
+          # identities, if any, are exempted from logging.
+          # An AuditConfig must have one or more AuditLogConfigs.
+          #
+          # If there are AuditConfigs for both `allServices` and a specific service,
+          # the union of the two AuditConfigs is used for that service: the log_types
+          # specified in each AuditConfig are enabled, and the exempted_members in each
+          # AuditLogConfig are exempted.
+          #
+          # Example Policy with multiple AuditConfigs:
+          #
+          #     {
+          #       &quot;audit_configs&quot;: [
+          #         {
+          #           &quot;service&quot;: &quot;allServices&quot;
+          #           &quot;audit_log_configs&quot;: [
+          #             {
+          #               &quot;log_type&quot;: &quot;DATA_READ&quot;,
+          #               &quot;exempted_members&quot;: [
+          #                 &quot;user:jose@example.com&quot;
+          #               ]
+          #             },
+          #             {
+          #               &quot;log_type&quot;: &quot;DATA_WRITE&quot;,
+          #             },
+          #             {
+          #               &quot;log_type&quot;: &quot;ADMIN_READ&quot;,
+          #             }
+          #           ]
+          #         },
+          #         {
+          #           &quot;service&quot;: &quot;sampleservice.googleapis.com&quot;
+          #           &quot;audit_log_configs&quot;: [
+          #             {
+          #               &quot;log_type&quot;: &quot;DATA_READ&quot;,
+          #             },
+          #             {
+          #               &quot;log_type&quot;: &quot;DATA_WRITE&quot;,
+          #               &quot;exempted_members&quot;: [
+          #                 &quot;user:aliya@example.com&quot;
+          #               ]
+          #             }
+          #           ]
+          #         }
+          #       ]
+          #     }
+          #
+          # For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ
+          # logging. It also exempts jose@example.com from DATA_READ logging, and
+          # aliya@example.com from DATA_WRITE logging.
+        &quot;service&quot;: &quot;A String&quot;, # Specifies a service that will be enabled for audit logging.
+            # For example, `storage.googleapis.com`, `cloudsql.googleapis.com`.
+            # `allServices` is a special value that covers all services.
+        &quot;auditLogConfigs&quot;: [ # The configuration for logging of each type of permission.
+          { # Provides the configuration for logging a type of permissions.
+              # Example:
+              #
+              #     {
+              #       &quot;audit_log_configs&quot;: [
+              #         {
+              #           &quot;log_type&quot;: &quot;DATA_READ&quot;,
+              #           &quot;exempted_members&quot;: [
+              #             &quot;user:jose@example.com&quot;
+              #           ]
+              #         },
+              #         {
+              #           &quot;log_type&quot;: &quot;DATA_WRITE&quot;,
+              #         }
+              #       ]
+              #     }
+              #
+              # This enables &#x27;DATA_READ&#x27; and &#x27;DATA_WRITE&#x27; logging, while exempting
+              # jose@example.com from DATA_READ logging.
+            &quot;exemptedMembers&quot;: [ # Specifies the identities that do not cause logging for this type of
+                # permission.
+                # Follows the same format of Binding.members.
+              &quot;A String&quot;,
+            ],
+            &quot;logType&quot;: &quot;A String&quot;, # The log type that this config enables.
+          },
+        ],
+      },
+    ],
+    &quot;bindings&quot;: [ # Associates a list of `members` to a `role`. Optionally, may specify a
         # `condition` that determines how and when the `bindings` are applied. Each
         # of the `bindings` must contain at least one member.
       { # Associates `members` with a `role`.
-        "role": "A String", # Role that is assigned to `members`.
-            # For example, `roles/viewer`, `roles/editor`, or `roles/owner`.
-        "members": [ # Specifies the identities requesting access for a Cloud Platform resource.
+        &quot;condition&quot;: { # Represents a textual expression in the Common Expression Language (CEL) # The condition that is associated with this binding.
+            #
+            # If the condition evaluates to `true`, then this binding applies to the
+            # current request.
+            #
+            # If the condition evaluates to `false`, then this binding does not apply to
+            # the current request. However, a different role binding might grant the same
+            # role to one or more of the members in this binding.
+            #
+            # To learn which resources support conditions in their IAM policies, see the
+            # [IAM
+            # documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+            # syntax. CEL is a C-like expression language. The syntax and semantics of CEL
+            # are documented at https://github.com/google/cel-spec.
+            #
+            # Example (Comparison):
+            #
+            #     title: &quot;Summary size limit&quot;
+            #     description: &quot;Determines if a summary is less than 100 chars&quot;
+            #     expression: &quot;document.summary.size() &lt; 100&quot;
+            #
+            # Example (Equality):
+            #
+            #     title: &quot;Requestor is owner&quot;
+            #     description: &quot;Determines if requestor is the document owner&quot;
+            #     expression: &quot;document.owner == request.auth.claims.email&quot;
+            #
+            # Example (Logic):
+            #
+            #     title: &quot;Public documents&quot;
+            #     description: &quot;Determine whether the document should be publicly visible&quot;
+            #     expression: &quot;document.type != &#x27;private&#x27; &amp;&amp; document.type != &#x27;internal&#x27;&quot;
+            #
+            # Example (Data Manipulation):
+            #
+            #     title: &quot;Notification string&quot;
+            #     description: &quot;Create a notification string with a timestamp.&quot;
+            #     expression: &quot;&#x27;New message received at &#x27; + string(document.create_time)&quot;
+            #
+            # The exact variables and functions that may be referenced within an expression
+            # are determined by the service that evaluates it. See the service
+            # documentation for additional information.
+          &quot;title&quot;: &quot;A String&quot;, # Optional. Title for the expression, i.e. a short string describing
+              # its purpose. This can be used e.g. in UIs which allow to enter the
+              # expression.
+          &quot;location&quot;: &quot;A String&quot;, # Optional. String indicating the location of the expression for error
+              # reporting, e.g. a file name and a position in the file.
+          &quot;description&quot;: &quot;A String&quot;, # Optional. Description of the expression. This is a longer text which
+              # describes the expression, e.g. when hovered over it in a UI.
+          &quot;expression&quot;: &quot;A String&quot;, # Textual representation of an expression in Common Expression Language
+              # syntax.
+        },
+        &quot;members&quot;: [ # Specifies the identities requesting access for a Cloud Platform resource.
             # `members` can have the following values:
             #
             # * `allUsers`: A special identifier that represents anyone who is
@@ -1984,177 +2532,17 @@
             # * `domain:{domain}`: The G Suite domain (primary) that represents all the
             #    users of that domain. For example, `google.com` or `example.com`.
             #
-          "A String",
+          &quot;A String&quot;,
         ],
-        "condition": { # Represents a textual expression in the Common Expression Language (CEL) # The condition that is associated with this binding.
-            # NOTE: An unsatisfied condition will not allow user access via current
-            # binding. Different bindings, including their conditions, are examined
-            # independently.
-            # syntax. CEL is a C-like expression language. The syntax and semantics of CEL
-            # are documented at https://github.com/google/cel-spec.
-            #
-            # Example (Comparison):
-            #
-            #     title: "Summary size limit"
-            #     description: "Determines if a summary is less than 100 chars"
-            #     expression: "document.summary.size() &lt; 100"
-            #
-            # Example (Equality):
-            #
-            #     title: "Requestor is owner"
-            #     description: "Determines if requestor is the document owner"
-            #     expression: "document.owner == request.auth.claims.email"
-            #
-            # Example (Logic):
-            #
-            #     title: "Public documents"
-            #     description: "Determine whether the document should be publicly visible"
-            #     expression: "document.type != 'private' &amp;&amp; document.type != 'internal'"
-            #
-            # Example (Data Manipulation):
-            #
-            #     title: "Notification string"
-            #     description: "Create a notification string with a timestamp."
-            #     expression: "'New message received at ' + string(document.create_time)"
-            #
-            # The exact variables and functions that may be referenced within an expression
-            # are determined by the service that evaluates it. See the service
-            # documentation for additional information.
-          "description": "A String", # Optional. Description of the expression. This is a longer text which
-              # describes the expression, e.g. when hovered over it in a UI.
-          "expression": "A String", # Textual representation of an expression in Common Expression Language
-              # syntax.
-          "location": "A String", # Optional. String indicating the location of the expression for error
-              # reporting, e.g. a file name and a position in the file.
-          "title": "A String", # Optional. Title for the expression, i.e. a short string describing
-              # its purpose. This can be used e.g. in UIs which allow to enter the
-              # expression.
-        },
+        &quot;role&quot;: &quot;A String&quot;, # Role that is assigned to `members`.
+            # For example, `roles/viewer`, `roles/editor`, or `roles/owner`.
       },
     ],
-    "auditConfigs": [ # Specifies cloud audit logging configuration for this policy.
-      { # Specifies the audit configuration for a service.
-          # The configuration determines which permission types are logged, and what
-          # identities, if any, are exempted from logging.
-          # An AuditConfig must have one or more AuditLogConfigs.
-          #
-          # If there are AuditConfigs for both `allServices` and a specific service,
-          # the union of the two AuditConfigs is used for that service: the log_types
-          # specified in each AuditConfig are enabled, and the exempted_members in each
-          # AuditLogConfig are exempted.
-          #
-          # Example Policy with multiple AuditConfigs:
-          #
-          #     {
-          #       "audit_configs": [
-          #         {
-          #           "service": "allServices"
-          #           "audit_log_configs": [
-          #             {
-          #               "log_type": "DATA_READ",
-          #               "exempted_members": [
-          #                 "user:jose@example.com"
-          #               ]
-          #             },
-          #             {
-          #               "log_type": "DATA_WRITE",
-          #             },
-          #             {
-          #               "log_type": "ADMIN_READ",
-          #             }
-          #           ]
-          #         },
-          #         {
-          #           "service": "sampleservice.googleapis.com"
-          #           "audit_log_configs": [
-          #             {
-          #               "log_type": "DATA_READ",
-          #             },
-          #             {
-          #               "log_type": "DATA_WRITE",
-          #               "exempted_members": [
-          #                 "user:aliya@example.com"
-          #               ]
-          #             }
-          #           ]
-          #         }
-          #       ]
-          #     }
-          #
-          # For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ
-          # logging. It also exempts jose@example.com from DATA_READ logging, and
-          # aliya@example.com from DATA_WRITE logging.
-        "auditLogConfigs": [ # The configuration for logging of each type of permission.
-          { # Provides the configuration for logging a type of permissions.
-              # Example:
-              #
-              #     {
-              #       "audit_log_configs": [
-              #         {
-              #           "log_type": "DATA_READ",
-              #           "exempted_members": [
-              #             "user:jose@example.com"
-              #           ]
-              #         },
-              #         {
-              #           "log_type": "DATA_WRITE",
-              #         }
-              #       ]
-              #     }
-              #
-              # This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting
-              # jose@example.com from DATA_READ logging.
-            "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of
-                # permission.
-                # Follows the same format of Binding.members.
-              "A String",
-            ],
-            "logType": "A String", # The log type that this config enables.
-          },
-        ],
-        "service": "A String", # Specifies a service that will be enabled for audit logging.
-            # For example, `storage.googleapis.com`, `cloudsql.googleapis.com`.
-            # `allServices` is a special value that covers all services.
-      },
-    ],
-    "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help
-        # prevent simultaneous updates of a policy from overwriting each other.
-        # It is strongly suggested that systems make use of the `etag` in the
-        # read-modify-write cycle to perform policy updates in order to avoid race
-        # conditions: An `etag` is returned in the response to `getIamPolicy`, and
-        # systems are expected to put that etag in the request to `setIamPolicy` to
-        # ensure that their change will be applied to the same version of the policy.
-        #
-        # **Important:** If you use IAM Conditions, you must include the `etag` field
-        # whenever you call `setIamPolicy`. If you omit this field, then IAM allows
-        # you to overwrite a version `3` policy with a version `1` policy, and all of
-        # the conditions in the version `3` policy are lost.
-    "version": 42, # Specifies the format of the policy.
-        #
-        # Valid values are `0`, `1`, and `3`. Requests that specify an invalid value
-        # are rejected.
-        #
-        # Any operation that affects conditional role bindings must specify version
-        # `3`. This requirement applies to the following operations:
-        #
-        # * Getting a policy that includes a conditional role binding
-        # * Adding a conditional role binding to a policy
-        # * Changing a conditional role binding in a policy
-        # * Removing any role binding, with or without a condition, from a policy
-        #   that includes conditions
-        #
-        # **Important:** If you use IAM Conditions, you must include the `etag` field
-        # whenever you call `setIamPolicy`. If you omit this field, then IAM allows
-        # you to overwrite a version `3` policy with a version `1` policy, and all of
-        # the conditions in the version `3` policy are lost.
-        #
-        # If a policy does not include any conditions, operations on that policy may
-        # specify any valid version or leave the field unset.
   }</pre>
 </div>
 
 <div class="method">
-    <code class="details" id="list">list(parent, pageSize=None, pageToken=None, x__xgafv=None, filter=None)</code>
+    <code class="details" id="list">list(parent, pageToken=None, pageSize=None, filter=None, x__xgafv=None)</code>
   <pre>Lists the jobs in the project.
 
 If there are no jobs that match the request parameters, the list
@@ -2162,596 +2550,719 @@
 
 Args:
   parent: string, Required. The name of the project for which to list jobs. (required)
-  pageSize: integer, Optional. The number of jobs to retrieve per "page" of results. If there
-are more remaining results than this number, the response message will
-contain a valid value in the `next_page_token` field.
-
-The default value is 20, and the maximum page size is 100.
   pageToken: string, Optional. A page token to request the next page of results.
 
 You get the token from the `next_page_token` field of the response from
 the previous call.
+  pageSize: integer, Optional. The number of jobs to retrieve per &quot;page&quot; of results. If there
+are more remaining results than this number, the response message will
+contain a valid value in the `next_page_token` field.
+
+The default value is 20, and the maximum page size is 100.
+  filter: string, Optional. Specifies the subset of jobs to retrieve.
+You can filter on the value of one or more attributes of the job object.
+For example, retrieve jobs with a job identifier that starts with &#x27;census&#x27;:
+&lt;p&gt;&lt;code&gt;gcloud ai-platform jobs list --filter=&#x27;jobId:census*&#x27;&lt;/code&gt;
+&lt;p&gt;List all failed jobs with names that start with &#x27;rnn&#x27;:
+&lt;p&gt;&lt;code&gt;gcloud ai-platform jobs list --filter=&#x27;jobId:rnn*
+AND state:FAILED&#x27;&lt;/code&gt;
+&lt;p&gt;For more examples, see the guide to
+&lt;a href=&quot;/ml-engine/docs/tensorflow/monitor-training&quot;&gt;monitoring jobs&lt;/a&gt;.
   x__xgafv: string, V1 error format.
     Allowed values
       1 - v1 error format
       2 - v2 error format
-  filter: string, Optional. Specifies the subset of jobs to retrieve.
-You can filter on the value of one or more attributes of the job object.
-For example, retrieve jobs with a job identifier that starts with 'census':
-&lt;p&gt;&lt;code&gt;gcloud ai-platform jobs list --filter='jobId:census*'&lt;/code&gt;
-&lt;p&gt;List all failed jobs with names that start with 'rnn':
-&lt;p&gt;&lt;code&gt;gcloud ai-platform jobs list --filter='jobId:rnn*
-AND state:FAILED'&lt;/code&gt;
-&lt;p&gt;For more examples, see the guide to
-&lt;a href="/ml-engine/docs/tensorflow/monitor-training"&gt;monitoring jobs&lt;/a&gt;.
 
 Returns:
   An object of the form:
 
     { # Response message for the ListJobs method.
-    "nextPageToken": "A String", # Optional. Pass this token as the `page_token` field of the request for a
-        # subsequent call.
-    "jobs": [ # The list of jobs.
+    &quot;jobs&quot;: [ # The list of jobs.
       { # Represents a training or prediction job.
-        "errorMessage": "A String", # Output only. The details of a failure or a cancellation.
-        "trainingOutput": { # Represents results of a training job. Output only. # The current training job result.
-          "completedTrialCount": "A String", # The number of hyperparameter tuning trials that completed successfully.
-              # Only set for hyperparameter tuning jobs.
-          "trials": [ # Results for individual Hyperparameter trials.
-              # Only set for hyperparameter tuning jobs.
-            { # Represents the result of a single hyperparameter tuning trial from a
-                # training job. The TrainingOutput object that is returned on successful
-                # completion of a training job with hyperparameter tuning includes a list
-                # of HyperparameterOutput objects, one for each successful trial.
-              "startTime": "A String", # Output only. Start time for the trial.
-              "hyperparameters": { # The hyperparameters given to this trial.
-                "a_key": "A String",
+          &quot;etag&quot;: &quot;A String&quot;, # `etag` is used for optimistic concurrency control as a way to help
+              # prevent simultaneous updates of a job from overwriting each other.
+              # It is strongly suggested that systems make use of the `etag` in the
+              # read-modify-write cycle to perform job updates in order to avoid race
+              # conditions: An `etag` is returned in the response to `GetJob`, and
+              # systems are expected to put that etag in the request to `UpdateJob` to
+              # ensure that their change will be applied to the same version of the job.
+          &quot;trainingInput&quot;: { # Represents input parameters for a training job. When using the gcloud command # Input parameters to create a training job.
+              # to submit your training job, you can specify the input parameters as
+              # command-line arguments and/or in a YAML configuration file referenced from
+              # the --config command-line argument. For details, see the guide to [submitting
+              # a training job](/ai-platform/training/docs/training-jobs).
+            &quot;parameterServerConfig&quot;: { # Represents the configuration for a replica in a cluster. # Optional. The configuration for parameter servers.
+                #
+                # You should only set `parameterServerConfig.acceleratorConfig` if
+                # `parameterServerType` is set to a Compute Engine machine type. [Learn
+                # about restrictions on accelerator configurations for
+                # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+                #
+                # Set `parameterServerConfig.imageUri` only if you build a custom image for
+                # your parameter server. If `parameterServerConfig.imageUri` has not been
+                # set, AI Platform uses the value of `masterConfig.imageUri`. Learn more about [configuring custom
+                # containers](/ai-platform/training/docs/distributed-training-containers).
+              &quot;tpuTfVersion&quot;: &quot;A String&quot;, # The AI Platform runtime version that includes a TensorFlow version matching
+                  # the one used in the custom container. This field is required if the replica
+                  # is a TPU worker that uses a custom container. Otherwise, do not specify
+                  # this field. This must be a [runtime version that currently supports
+                  # training with
+                  # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
+                  #
+                  # Note that the version of TensorFlow included in a runtime version may
+                  # differ from the numbering of the runtime version itself, because it may
+                  # have a different [patch
+                  # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
+                  # In this field, you must specify the runtime version (TensorFlow minor
+                  # version). For example, if your custom container runs TensorFlow `1.x.y`,
+                  # specify `1.x`.
+              &quot;containerCommand&quot;: [ # The command with which the replica&#x27;s custom container is run.
+                  # If provided, it will override default ENTRYPOINT of the docker image.
+                  # If not provided, the docker image&#x27;s ENTRYPOINT is used.
+                  # It cannot be set if custom container image is
+                  # not provided.
+                  # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+                  # both cannot be set at the same time.
+                &quot;A String&quot;,
+              ],
+              &quot;imageUri&quot;: &quot;A String&quot;, # The Docker image to run on the replica. This image must be in Container
+                  # Registry. Learn more about [configuring custom
+                  # containers](/ai-platform/training/docs/distributed-training-containers).
+              &quot;containerArgs&quot;: [ # Arguments to the entrypoint command.
+                  # The following rules apply for container_command and container_args:
+                  # - If you do not supply command or args:
+                  #   The defaults defined in the Docker image are used.
+                  # - If you supply a command but no args:
+                  #   The default EntryPoint and the default Cmd defined in the Docker image
+                  #   are ignored. Your command is run without any arguments.
+                  # - If you supply only args:
+                  #   The default Entrypoint defined in the Docker image is run with the args
+                  #   that you supplied.
+                  # - If you supply a command and args:
+                  #   The default Entrypoint and the default Cmd defined in the Docker image
+                  #   are ignored. Your command is run with your args.
+                  # It cannot be set if custom container image is
+                  # not provided.
+                  # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+                  # both cannot be set at the same time.
+                &quot;A String&quot;,
+              ],
+              &quot;acceleratorConfig&quot;: { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
+                  # [Learn about restrictions on accelerator configurations for
+                  # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+                  # Note that the AcceleratorConfig can be used in both Jobs and Versions.
+                  # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
+                  # [accelerators for online
+                  # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
+                &quot;count&quot;: &quot;A String&quot;, # The number of accelerators to attach to each machine running the job.
+                &quot;type&quot;: &quot;A String&quot;, # The type of accelerator to use.
               },
-              "finalMetric": { # An observed value of a metric. # The final objective metric seen for this trial.
-                "trainingStep": "A String", # The global training step for this metric.
-                "objectiveValue": 3.14, # The objective value at this training step.
-              },
-              "state": "A String", # Output only. The detailed state of the trial.
-              "allMetrics": [ # All recorded object metrics for this trial. This field is not currently
-                  # populated.
-                { # An observed value of a metric.
-                  "trainingStep": "A String", # The global training step for this metric.
-                  "objectiveValue": 3.14, # The objective value at this training step.
+            },
+            &quot;encryptionConfig&quot;: { # Represents a custom encryption key configuration that can be applied to # Optional. Options for using customer-managed encryption keys (CMEK) to
+                # protect resources created by a training job, instead of using Google&#x27;s
+                # default encryption. If this is set, then all resources created by the
+                # training job will be encrypted with the customer-managed encryption key
+                # that you specify.
+                #
+                # [Learn how and when to use CMEK with AI Platform
+                # Training](/ai-platform/training/docs/cmek).
+                # a resource.
+              &quot;kmsKeyName&quot;: &quot;A String&quot;, # The Cloud KMS resource identifier of the customer-managed encryption key
+                  # used to protect a resource, such as a training job. It has the following
+                  # format:
+                  # `projects/{PROJECT_ID}/locations/{REGION}/keyRings/{KEY_RING_NAME}/cryptoKeys/{KEY_NAME}`
+            },
+            &quot;hyperparameters&quot;: { # Represents a set of hyperparameters to optimize. # Optional. The set of Hyperparameters to tune.
+              &quot;hyperparameterMetricTag&quot;: &quot;A String&quot;, # Optional. The TensorFlow summary tag name to use for optimizing trials. For
+                  # current versions of TensorFlow, this tag name should exactly match what is
+                  # shown in TensorBoard, including all scopes.  For versions of TensorFlow
+                  # prior to 0.12, this should be only the tag passed to tf.Summary.
+                  # By default, &quot;training/hptuning/metric&quot; will be used.
+              &quot;params&quot;: [ # Required. The set of parameters to tune.
+                { # Represents a single hyperparameter to optimize.
+                  &quot;type&quot;: &quot;A String&quot;, # Required. The type of the parameter.
+                  &quot;categoricalValues&quot;: [ # Required if type is `CATEGORICAL`. The list of possible categories.
+                    &quot;A String&quot;,
+                  ],
+                  &quot;parameterName&quot;: &quot;A String&quot;, # Required. The parameter name must be unique amongst all ParameterConfigs in
+                      # a HyperparameterSpec message. E.g., &quot;learning_rate&quot;.
+                  &quot;minValue&quot;: 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
+                      # should be unset if type is `CATEGORICAL`. This value should be integers if
+                      # type is INTEGER.
+                  &quot;discreteValues&quot;: [ # Required if type is `DISCRETE`.
+                      # A list of feasible points.
+                      # The list should be in strictly increasing order. For instance, this
+                      # parameter might have possible settings of 1.5, 2.5, and 4.0. This list
+                      # should not contain more than 1,000 values.
+                    3.14,
+                  ],
+                  &quot;scaleType&quot;: &quot;A String&quot;, # Optional. How the parameter should be scaled to the hypercube.
+                      # Leave unset for categorical parameters.
+                      # Some kind of scaling is strongly recommended for real or integral
+                      # parameters (e.g., `UNIT_LINEAR_SCALE`).
+                  &quot;maxValue&quot;: 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
+                      # should be unset if type is `CATEGORICAL`. This value should be integers if
+                      # type is `INTEGER`.
                 },
               ],
-              "isTrialStoppedEarly": True or False, # True if the trial is stopped early.
-              "endTime": "A String", # Output only. End time for the trial.
-              "trialId": "A String", # The trial id for these results.
-              "builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
-                  # Only set for trials of built-in algorithms jobs that have succeeded.
-                "framework": "A String", # Framework on which the built-in algorithm was trained.
-                "modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job
-                    # saves the trained model. Only set for successful jobs that don't use
-                    # hyperparameter tuning.
-                "runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was
-                    # trained.
-                "pythonVersion": "A String", # Python version on which the built-in algorithm was trained.
+              &quot;enableTrialEarlyStopping&quot;: True or False, # Optional. Indicates if the hyperparameter tuning job enables auto trial
+                  # early stopping.
+              &quot;resumePreviousJobId&quot;: &quot;A String&quot;, # Optional. The prior hyperparameter tuning job id that users hope to
+                  # continue with. The job id will be used to find the corresponding vizier
+                  # study guid and resume the study.
+              &quot;maxParallelTrials&quot;: 42, # Optional. The number of training trials to run concurrently.
+                  # You can reduce the time it takes to perform hyperparameter tuning by adding
+                  # trials in parallel. However, each trail only benefits from the information
+                  # gained in completed trials. That means that a trial does not get access to
+                  # the results of trials running at the same time, which could reduce the
+                  # quality of the overall optimization.
+                  #
+                  # Each trial will use the same scale tier and machine types.
+                  #
+                  # Defaults to one.
+              &quot;maxFailedTrials&quot;: 42, # Optional. The number of failed trials that need to be seen before failing
+                  # the hyperparameter tuning job. You can specify this field to override the
+                  # default failing criteria for AI Platform hyperparameter tuning jobs.
+                  #
+                  # Defaults to zero, which means the service decides when a hyperparameter
+                  # job should fail.
+              &quot;goal&quot;: &quot;A String&quot;, # Required. The type of goal to use for tuning. Available types are
+                  # `MAXIMIZE` and `MINIMIZE`.
+                  #
+                  # Defaults to `MAXIMIZE`.
+              &quot;maxTrials&quot;: 42, # Optional. How many training trials should be attempted to optimize
+                  # the specified hyperparameters.
+                  #
+                  # Defaults to one.
+              &quot;algorithm&quot;: &quot;A String&quot;, # Optional. The search algorithm specified for the hyperparameter
+                  # tuning job.
+                  # Uses the default AI Platform hyperparameter tuning
+                  # algorithm if unspecified.
+            },
+            &quot;workerConfig&quot;: { # Represents the configuration for a replica in a cluster. # Optional. The configuration for workers.
+                #
+                # You should only set `workerConfig.acceleratorConfig` if `workerType` is set
+                # to a Compute Engine machine type. [Learn about restrictions on accelerator
+                # configurations for
+                # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+                #
+                # Set `workerConfig.imageUri` only if you build a custom image for your
+                # worker. If `workerConfig.imageUri` has not been set, AI Platform uses
+                # the value of `masterConfig.imageUri`. Learn more about [configuring custom
+                # containers](/ai-platform/training/docs/distributed-training-containers).
+              &quot;tpuTfVersion&quot;: &quot;A String&quot;, # The AI Platform runtime version that includes a TensorFlow version matching
+                  # the one used in the custom container. This field is required if the replica
+                  # is a TPU worker that uses a custom container. Otherwise, do not specify
+                  # this field. This must be a [runtime version that currently supports
+                  # training with
+                  # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
+                  #
+                  # Note that the version of TensorFlow included in a runtime version may
+                  # differ from the numbering of the runtime version itself, because it may
+                  # have a different [patch
+                  # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
+                  # In this field, you must specify the runtime version (TensorFlow minor
+                  # version). For example, if your custom container runs TensorFlow `1.x.y`,
+                  # specify `1.x`.
+              &quot;containerCommand&quot;: [ # The command with which the replica&#x27;s custom container is run.
+                  # If provided, it will override default ENTRYPOINT of the docker image.
+                  # If not provided, the docker image&#x27;s ENTRYPOINT is used.
+                  # It cannot be set if custom container image is
+                  # not provided.
+                  # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+                  # both cannot be set at the same time.
+                &quot;A String&quot;,
+              ],
+              &quot;imageUri&quot;: &quot;A String&quot;, # The Docker image to run on the replica. This image must be in Container
+                  # Registry. Learn more about [configuring custom
+                  # containers](/ai-platform/training/docs/distributed-training-containers).
+              &quot;containerArgs&quot;: [ # Arguments to the entrypoint command.
+                  # The following rules apply for container_command and container_args:
+                  # - If you do not supply command or args:
+                  #   The defaults defined in the Docker image are used.
+                  # - If you supply a command but no args:
+                  #   The default EntryPoint and the default Cmd defined in the Docker image
+                  #   are ignored. Your command is run without any arguments.
+                  # - If you supply only args:
+                  #   The default Entrypoint defined in the Docker image is run with the args
+                  #   that you supplied.
+                  # - If you supply a command and args:
+                  #   The default Entrypoint and the default Cmd defined in the Docker image
+                  #   are ignored. Your command is run with your args.
+                  # It cannot be set if custom container image is
+                  # not provided.
+                  # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+                  # both cannot be set at the same time.
+                &quot;A String&quot;,
+              ],
+              &quot;acceleratorConfig&quot;: { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
+                  # [Learn about restrictions on accelerator configurations for
+                  # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+                  # Note that the AcceleratorConfig can be used in both Jobs and Versions.
+                  # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
+                  # [accelerators for online
+                  # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
+                &quot;count&quot;: &quot;A String&quot;, # The number of accelerators to attach to each machine running the job.
+                &quot;type&quot;: &quot;A String&quot;, # The type of accelerator to use.
               },
             },
-          ],
-          "isHyperparameterTuningJob": True or False, # Whether this job is a hyperparameter tuning job.
-          "isBuiltInAlgorithmJob": True or False, # Whether this job is a built-in Algorithm job.
-          "consumedMLUnits": 3.14, # The amount of ML units consumed by the job.
-          "hyperparameterMetricTag": "A String", # The TensorFlow summary tag name used for optimizing hyperparameter tuning
-              # trials. See
-              # [`HyperparameterSpec.hyperparameterMetricTag`](#HyperparameterSpec.FIELDS.hyperparameter_metric_tag)
-              # for more information. Only set for hyperparameter tuning jobs.
-          "builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
-              # Only set for built-in algorithms jobs.
-            "framework": "A String", # Framework on which the built-in algorithm was trained.
-            "modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job
-                # saves the trained model. Only set for successful jobs that don't use
-                # hyperparameter tuning.
-            "runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was
-                # trained.
-            "pythonVersion": "A String", # Python version on which the built-in algorithm was trained.
-          },
-        },
-        "predictionInput": { # Represents input parameters for a prediction job. # Input parameters to create a prediction job.
-          "modelName": "A String", # Use this field if you want to use the default version for the specified
-              # model. The string must use the following format:
-              #
-              # `"projects/YOUR_PROJECT/models/YOUR_MODEL"`
-          "signatureName": "A String", # Optional. The name of the signature defined in the SavedModel to use for
-              # this job. Please refer to
-              # [SavedModel](https://tensorflow.github.io/serving/serving_basic.html)
-              # for information about how to use signatures.
-              #
-              # Defaults to
-              # [DEFAULT_SERVING_SIGNATURE_DEF_KEY](https://www.tensorflow.org/api_docs/python/tf/saved_model/signature_constants)
-              # , which is "serving_default".
-          "runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for this batch
-              # prediction. If not set, AI Platform will pick the runtime version used
-              # during the CreateVersion request for this model version, or choose the
-              # latest stable version when model version information is not available
-              # such as when the model is specified by uri.
-          "batchSize": "A String", # Optional. Number of records per batch, defaults to 64.
-              # The service will buffer batch_size number of records in memory before
-              # invoking one Tensorflow prediction call internally. So take the record
-              # size and memory available into consideration when setting this parameter.
-          "maxWorkerCount": "A String", # Optional. The maximum number of workers to be used for parallel processing.
-              # Defaults to 10 if not specified.
-          "uri": "A String", # Use this field if you want to specify a Google Cloud Storage path for
-              # the model to use.
-          "outputPath": "A String", # Required. The output Google Cloud Storage location.
-          "dataFormat": "A String", # Required. The format of the input data files.
-          "versionName": "A String", # Use this field if you want to specify a version of the model to use. The
-              # string is formatted the same way as `model_version`, with the addition
-              # of the version information:
-              #
-              # `"projects/YOUR_PROJECT/models/YOUR_MODEL/versions/YOUR_VERSION"`
-          "region": "A String", # Required. The Google Compute Engine region to run the prediction job in.
-              # See the &lt;a href="/ml-engine/docs/tensorflow/regions"&gt;available regions&lt;/a&gt;
-              # for AI Platform services.
-          "inputPaths": [ # Required. The Cloud Storage location of the input data files. May contain
-              # &lt;a href="/storage/docs/gsutil/addlhelp/WildcardNames"&gt;wildcards&lt;/a&gt;.
-            "A String",
-          ],
-          "outputDataFormat": "A String", # Optional. Format of the output data files, defaults to JSON.
-        },
-        "labels": { # Optional. One or more labels that you can add, to organize your jobs.
-            # Each label is a key-value pair, where both the key and the value are
-            # arbitrary strings that you supply.
-            # For more information, see the documentation on
-            # &lt;a href="/ml-engine/docs/tensorflow/resource-labels"&gt;using labels&lt;/a&gt;.
-          "a_key": "A String",
-        },
-        "jobId": "A String", # Required. The user-specified id of the job.
-        "state": "A String", # Output only. The detailed state of a job.
-        "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help
-            # prevent simultaneous updates of a job from overwriting each other.
-            # It is strongly suggested that systems make use of the `etag` in the
-            # read-modify-write cycle to perform job updates in order to avoid race
-            # conditions: An `etag` is returned in the response to `GetJob`, and
-            # systems are expected to put that etag in the request to `UpdateJob` to
-            # ensure that their change will be applied to the same version of the job.
-        "startTime": "A String", # Output only. When the job processing was started.
-        "trainingInput": { # Represents input parameters for a training job. When using the gcloud command # Input parameters to create a training job.
-            # to submit your training job, you can specify the input parameters as
-            # command-line arguments and/or in a YAML configuration file referenced from
-            # the --config command-line argument. For details, see the guide to [submitting
-            # a training job](/ai-platform/training/docs/training-jobs).
-          "masterType": "A String", # Optional. Specifies the type of virtual machine to use for your training
-              # job's master worker. You must specify this field when `scaleTier` is set to
-              # `CUSTOM`.
-              #
-              # You can use certain Compute Engine machine types directly in this field.
-              # The following types are supported:
-              #
-              # - `n1-standard-4`
-              # - `n1-standard-8`
-              # - `n1-standard-16`
-              # - `n1-standard-32`
-              # - `n1-standard-64`
-              # - `n1-standard-96`
-              # - `n1-highmem-2`
-              # - `n1-highmem-4`
-              # - `n1-highmem-8`
-              # - `n1-highmem-16`
-              # - `n1-highmem-32`
-              # - `n1-highmem-64`
-              # - `n1-highmem-96`
-              # - `n1-highcpu-16`
-              # - `n1-highcpu-32`
-              # - `n1-highcpu-64`
-              # - `n1-highcpu-96`
-              #
-              # Learn more about [using Compute Engine machine
-              # types](/ml-engine/docs/machine-types#compute-engine-machine-types).
-              #
-              # Alternatively, you can use the following legacy machine types:
-              #
-              # - `standard`
-              # - `large_model`
-              # - `complex_model_s`
-              # - `complex_model_m`
-              # - `complex_model_l`
-              # - `standard_gpu`
-              # - `complex_model_m_gpu`
-              # - `complex_model_l_gpu`
-              # - `standard_p100`
-              # - `complex_model_m_p100`
-              # - `standard_v100`
-              # - `large_model_v100`
-              # - `complex_model_m_v100`
-              # - `complex_model_l_v100`
-              #
-              # Learn more about [using legacy machine
-              # types](/ml-engine/docs/machine-types#legacy-machine-types).
-              #
-              # Finally, if you want to use a TPU for training, specify `cloud_tpu` in this
-              # field. Learn more about the [special configuration options for training
-              # with
-              # TPUs](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine).
-          "jobDir": "A String", # Optional. A Google Cloud Storage path in which to store training outputs
-              # and other data needed for training. This path is passed to your TensorFlow
-              # program as the '--job-dir' command-line argument. The benefit of specifying
-              # this field is that Cloud ML validates the path for use in training.
-          "scheduling": { # All parameters related to scheduling of training jobs. # Optional. Scheduling options for a training job.
-            "maxRunningTime": "A String", # Optional. The maximum job running time, expressed in seconds. The field can
-                # contain up to nine fractional digits, terminated by `s`. By default there
-                # is no limit to the running time.
+            &quot;parameterServerCount&quot;: &quot;A String&quot;, # Optional. The number of parameter server replicas to use for the training
+                # job. Each replica in the cluster will be of the type specified in
+                # `parameter_server_type`.
                 #
-                # If the training job is still running after this duration, AI Platform
-                # Training cancels it.
+                # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
+                # set this value, you must also set `parameter_server_type`.
                 #
-                # For example, if you want to ensure your job runs for no more than 2 hours,
-                # set this field to `7200s` (2 hours * 60 minutes / hour * 60 seconds /
-                # minute).
+                # The default value is zero.
+            &quot;packageUris&quot;: [ # Required. The Google Cloud Storage location of the packages with
+                # the training program and any additional dependencies.
+                # The maximum number of package URIs is 100.
+              &quot;A String&quot;,
+            ],
+            &quot;evaluatorCount&quot;: &quot;A String&quot;, # Optional. The number of evaluator replicas to use for the training job.
+                # Each replica in the cluster will be of the type specified in
+                # `evaluator_type`.
                 #
-                # If you submit your training job using the `gcloud` tool, you can [provide
-                # this field in a `config.yaml`
-                # file](/ai-platform/training/docs/training-jobs#formatting_your_configuration_parameters).
-                # For example:
+                # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
+                # set this value, you must also set `evaluator_type`.
                 #
-                # ```yaml
-                # trainingInput:
-                #   ...
-                #   scheduling:
-                #     maxRunningTime: 7200s
-                #   ...
-                # ```
-          },
-          "parameterServerCount": "A String", # Optional. The number of parameter server replicas to use for the training
-              # job. Each replica in the cluster will be of the type specified in
-              # `parameter_server_type`.
-              #
-              # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
-              # set this value, you must also set `parameter_server_type`.
-              #
-              # The default value is zero.
-          "evaluatorCount": "A String", # Optional. The number of evaluator replicas to use for the training job.
-              # Each replica in the cluster will be of the type specified in
-              # `evaluator_type`.
-              #
-              # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
-              # set this value, you must also set `evaluator_type`.
-              #
-              # The default value is zero.
-          "workerType": "A String", # Optional. Specifies the type of virtual machine to use for your training
-              # job's worker nodes.
-              #
-              # The supported values are the same as those described in the entry for
-              # `masterType`.
-              #
-              # This value must be consistent with the category of machine type that
-              # `masterType` uses. In other words, both must be Compute Engine machine
-              # types or both must be legacy machine types.
-              #
-              # If you use `cloud_tpu` for this value, see special instructions for
-              # [configuring a custom TPU
-              # machine](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine).
-              #
-              # This value must be present when `scaleTier` is set to `CUSTOM` and
-              # `workerCount` is greater than zero.
-          "scaleTier": "A String", # Required. Specifies the machine types, the number of replicas for workers
-              # and parameter servers.
-          "packageUris": [ # Required. The Google Cloud Storage location of the packages with
-              # the training program and any additional dependencies.
-              # The maximum number of package URIs is 100.
-            "A String",
-          ],
-          "workerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for workers.
-              #
-              # You should only set `workerConfig.acceleratorConfig` if `workerType` is set
-              # to a Compute Engine machine type. [Learn about restrictions on accelerator
-              # configurations for
-              # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-              #
-              # Set `workerConfig.imageUri` only if you build a custom image for your
-              # worker. If `workerConfig.imageUri` has not been set, AI Platform uses
-              # the value of `masterConfig.imageUri`. Learn more about [configuring custom
-              # containers](/ai-platform/training/docs/distributed-training-containers).
-            "tpuTfVersion": "A String", # The AI Platform runtime version that includes a TensorFlow version matching
-                # the one used in the custom container. This field is required if the replica
-                # is a TPU worker that uses a custom container. Otherwise, do not specify
-                # this field. This must be a [runtime version that currently supports
-                # training with
-                # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
+                # The default value is zero.
+            &quot;masterType&quot;: &quot;A String&quot;, # Optional. Specifies the type of virtual machine to use for your training
+                # job&#x27;s master worker. You must specify this field when `scaleTier` is set to
+                # `CUSTOM`.
                 #
-                # Note that the version of TensorFlow included in a runtime version may
-                # differ from the numbering of the runtime version itself, because it may
-                # have a different [patch
-                # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
-                # In this field, you must specify the runtime version (TensorFlow minor
-                # version). For example, if your custom container runs TensorFlow `1.x.y`,
-                # specify `1.x`.
-            "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
-                # [Learn about restrictions on accelerator configurations for
+                # You can use certain Compute Engine machine types directly in this field.
+                # The following types are supported:
+                #
+                # - `n1-standard-4`
+                # - `n1-standard-8`
+                # - `n1-standard-16`
+                # - `n1-standard-32`
+                # - `n1-standard-64`
+                # - `n1-standard-96`
+                # - `n1-highmem-2`
+                # - `n1-highmem-4`
+                # - `n1-highmem-8`
+                # - `n1-highmem-16`
+                # - `n1-highmem-32`
+                # - `n1-highmem-64`
+                # - `n1-highmem-96`
+                # - `n1-highcpu-16`
+                # - `n1-highcpu-32`
+                # - `n1-highcpu-64`
+                # - `n1-highcpu-96`
+                #
+                # Learn more about [using Compute Engine machine
+                # types](/ml-engine/docs/machine-types#compute-engine-machine-types).
+                #
+                # Alternatively, you can use the following legacy machine types:
+                #
+                # - `standard`
+                # - `large_model`
+                # - `complex_model_s`
+                # - `complex_model_m`
+                # - `complex_model_l`
+                # - `standard_gpu`
+                # - `complex_model_m_gpu`
+                # - `complex_model_l_gpu`
+                # - `standard_p100`
+                # - `complex_model_m_p100`
+                # - `standard_v100`
+                # - `large_model_v100`
+                # - `complex_model_m_v100`
+                # - `complex_model_l_v100`
+                #
+                # Learn more about [using legacy machine
+                # types](/ml-engine/docs/machine-types#legacy-machine-types).
+                #
+                # Finally, if you want to use a TPU for training, specify `cloud_tpu` in this
+                # field. Learn more about the [special configuration options for training
+                # with
+                # TPUs](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine).
+            &quot;runtimeVersion&quot;: &quot;A String&quot;, # Optional. The AI Platform runtime version to use for training. You must
+                # either specify this field or specify `masterConfig.imageUri`.
+                #
+                # For more information, see the [runtime version
+                # list](/ai-platform/training/docs/runtime-version-list) and learn [how to
+                # manage runtime versions](/ai-platform/training/docs/versioning).
+            &quot;evaluatorType&quot;: &quot;A String&quot;, # Optional. Specifies the type of virtual machine to use for your training
+                # job&#x27;s evaluator nodes.
+                #
+                # The supported values are the same as those described in the entry for
+                # `masterType`.
+                #
+                # This value must be consistent with the category of machine type that
+                # `masterType` uses. In other words, both must be Compute Engine machine
+                # types or both must be legacy machine types.
+                #
+                # This value must be present when `scaleTier` is set to `CUSTOM` and
+                # `evaluatorCount` is greater than zero.
+            &quot;region&quot;: &quot;A String&quot;, # Required. The region to run the training job in. See the [available
+                # regions](/ai-platform/training/docs/regions) for AI Platform Training.
+            &quot;workerType&quot;: &quot;A String&quot;, # Optional. Specifies the type of virtual machine to use for your training
+                # job&#x27;s worker nodes.
+                #
+                # The supported values are the same as those described in the entry for
+                # `masterType`.
+                #
+                # This value must be consistent with the category of machine type that
+                # `masterType` uses. In other words, both must be Compute Engine machine
+                # types or both must be legacy machine types.
+                #
+                # If you use `cloud_tpu` for this value, see special instructions for
+                # [configuring a custom TPU
+                # machine](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine).
+                #
+                # This value must be present when `scaleTier` is set to `CUSTOM` and
+                # `workerCount` is greater than zero.
+            &quot;parameterServerType&quot;: &quot;A String&quot;, # Optional. Specifies the type of virtual machine to use for your training
+                # job&#x27;s parameter server.
+                #
+                # The supported values are the same as those described in the entry for
+                # `master_type`.
+                #
+                # This value must be consistent with the category of machine type that
+                # `masterType` uses. In other words, both must be Compute Engine machine
+                # types or both must be legacy machine types.
+                #
+                # This value must be present when `scaleTier` is set to `CUSTOM` and
+                # `parameter_server_count` is greater than zero.
+            &quot;masterConfig&quot;: { # Represents the configuration for a replica in a cluster. # Optional. The configuration for your master worker.
+                #
+                # You should only set `masterConfig.acceleratorConfig` if `masterType` is set
+                # to a Compute Engine machine type. Learn about [restrictions on accelerator
+                # configurations for
                 # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-                # Note that the AcceleratorConfig can be used in both Jobs and Versions.
-                # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
-                # [accelerators for online
-                # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
-              "count": "A String", # The number of accelerators to attach to each machine running the job.
-              "type": "A String", # The type of accelerator to use.
-            },
-            "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
-                # Registry. Learn more about [configuring custom
-                # containers](/ai-platform/training/docs/distributed-training-containers).
-          },
-          "evaluatorConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for evaluators.
-              #
-              # You should only set `evaluatorConfig.acceleratorConfig` if
-              # `evaluatorType` is set to a Compute Engine machine type. [Learn
-              # about restrictions on accelerator configurations for
-              # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-              #
-              # Set `evaluatorConfig.imageUri` only if you build a custom image for
-              # your evaluator. If `evaluatorConfig.imageUri` has not been
-              # set, AI Platform uses the value of `masterConfig.imageUri`. Learn more about [configuring custom
-              # containers](/ai-platform/training/docs/distributed-training-containers).
-            "tpuTfVersion": "A String", # The AI Platform runtime version that includes a TensorFlow version matching
-                # the one used in the custom container. This field is required if the replica
-                # is a TPU worker that uses a custom container. Otherwise, do not specify
-                # this field. This must be a [runtime version that currently supports
-                # training with
-                # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
                 #
-                # Note that the version of TensorFlow included in a runtime version may
-                # differ from the numbering of the runtime version itself, because it may
-                # have a different [patch
-                # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
-                # In this field, you must specify the runtime version (TensorFlow minor
-                # version). For example, if your custom container runs TensorFlow `1.x.y`,
-                # specify `1.x`.
-            "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
-                # [Learn about restrictions on accelerator configurations for
+                # Set `masterConfig.imageUri` only if you build a custom image. Only one of
+                # `masterConfig.imageUri` and `runtimeVersion` should be set. Learn more
+                # about [configuring custom
+                # containers](/ai-platform/training/docs/distributed-training-containers).
+              &quot;tpuTfVersion&quot;: &quot;A String&quot;, # The AI Platform runtime version that includes a TensorFlow version matching
+                  # the one used in the custom container. This field is required if the replica
+                  # is a TPU worker that uses a custom container. Otherwise, do not specify
+                  # this field. This must be a [runtime version that currently supports
+                  # training with
+                  # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
+                  #
+                  # Note that the version of TensorFlow included in a runtime version may
+                  # differ from the numbering of the runtime version itself, because it may
+                  # have a different [patch
+                  # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
+                  # In this field, you must specify the runtime version (TensorFlow minor
+                  # version). For example, if your custom container runs TensorFlow `1.x.y`,
+                  # specify `1.x`.
+              &quot;containerCommand&quot;: [ # The command with which the replica&#x27;s custom container is run.
+                  # If provided, it will override default ENTRYPOINT of the docker image.
+                  # If not provided, the docker image&#x27;s ENTRYPOINT is used.
+                  # It cannot be set if custom container image is
+                  # not provided.
+                  # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+                  # both cannot be set at the same time.
+                &quot;A String&quot;,
+              ],
+              &quot;imageUri&quot;: &quot;A String&quot;, # The Docker image to run on the replica. This image must be in Container
+                  # Registry. Learn more about [configuring custom
+                  # containers](/ai-platform/training/docs/distributed-training-containers).
+              &quot;containerArgs&quot;: [ # Arguments to the entrypoint command.
+                  # The following rules apply for container_command and container_args:
+                  # - If you do not supply command or args:
+                  #   The defaults defined in the Docker image are used.
+                  # - If you supply a command but no args:
+                  #   The default EntryPoint and the default Cmd defined in the Docker image
+                  #   are ignored. Your command is run without any arguments.
+                  # - If you supply only args:
+                  #   The default Entrypoint defined in the Docker image is run with the args
+                  #   that you supplied.
+                  # - If you supply a command and args:
+                  #   The default Entrypoint and the default Cmd defined in the Docker image
+                  #   are ignored. Your command is run with your args.
+                  # It cannot be set if custom container image is
+                  # not provided.
+                  # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+                  # both cannot be set at the same time.
+                &quot;A String&quot;,
+              ],
+              &quot;acceleratorConfig&quot;: { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
+                  # [Learn about restrictions on accelerator configurations for
+                  # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+                  # Note that the AcceleratorConfig can be used in both Jobs and Versions.
+                  # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
+                  # [accelerators for online
+                  # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
+                &quot;count&quot;: &quot;A String&quot;, # The number of accelerators to attach to each machine running the job.
+                &quot;type&quot;: &quot;A String&quot;, # The type of accelerator to use.
+              },
+            },
+            &quot;scaleTier&quot;: &quot;A String&quot;, # Required. Specifies the machine types, the number of replicas for workers
+                # and parameter servers.
+            &quot;jobDir&quot;: &quot;A String&quot;, # Optional. A Google Cloud Storage path in which to store training outputs
+                # and other data needed for training. This path is passed to your TensorFlow
+                # program as the &#x27;--job-dir&#x27; command-line argument. The benefit of specifying
+                # this field is that Cloud ML validates the path for use in training.
+            &quot;pythonVersion&quot;: &quot;A String&quot;, # Optional. The version of Python used in training. You must either specify
+                # this field or specify `masterConfig.imageUri`.
+                #
+                # The following Python versions are available:
+                #
+                # * Python &#x27;3.7&#x27; is available when `runtime_version` is set to &#x27;1.15&#x27; or
+                #   later.
+                # * Python &#x27;3.5&#x27; is available when `runtime_version` is set to a version
+                #   from &#x27;1.4&#x27; to &#x27;1.14&#x27;.
+                # * Python &#x27;2.7&#x27; is available when `runtime_version` is set to &#x27;1.15&#x27; or
+                #   earlier.
+                #
+                # Read more about the Python versions available for [each runtime
+                # version](/ml-engine/docs/runtime-version-list).
+            &quot;network&quot;: &quot;A String&quot;, # Optional. The full name of the Google Compute Engine
+                # [network](/compute/docs/networks-and-firewalls#networks) to which the Job
+                # is peered. For example, projects/12345/global/networks/myVPC. Format is of
+                # the form projects/{project}/global/networks/{network}. Where {project} is a
+                # project number, as in &#x27;12345&#x27;, and {network} is network name.&quot;.
+                #
+                # Private services access must already be configured for the network. If left
+                # unspecified, the Job is not peered with any network. Learn more -
+                # Connecting Job to user network over private
+                # IP.
+            &quot;scheduling&quot;: { # All parameters related to scheduling of training jobs. # Optional. Scheduling options for a training job.
+              &quot;maxWaitTime&quot;: &quot;A String&quot;,
+              &quot;maxRunningTime&quot;: &quot;A String&quot;, # Optional. The maximum job running time, expressed in seconds. The field can
+                  # contain up to nine fractional digits, terminated by `s`. If not specified,
+                  # this field defaults to `604800s` (seven days).
+                  #
+                  # If the training job is still running after this duration, AI Platform
+                  # Training cancels it.
+                  #
+                  # For example, if you want to ensure your job runs for no more than 2 hours,
+                  # set this field to `7200s` (2 hours * 60 minutes / hour * 60 seconds /
+                  # minute).
+                  #
+                  # If you submit your training job using the `gcloud` tool, you can [provide
+                  # this field in a `config.yaml`
+                  # file](/ai-platform/training/docs/training-jobs#formatting_your_configuration_parameters).
+                  # For example:
+                  #
+                  # ```yaml
+                  # trainingInput:
+                  #   ...
+                  #   scheduling:
+                  #     maxRunningTime: 7200s
+                  #   ...
+                  # ```
+            },
+            &quot;evaluatorConfig&quot;: { # Represents the configuration for a replica in a cluster. # Optional. The configuration for evaluators.
+                #
+                # You should only set `evaluatorConfig.acceleratorConfig` if
+                # `evaluatorType` is set to a Compute Engine machine type. [Learn
+                # about restrictions on accelerator configurations for
                 # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-                # Note that the AcceleratorConfig can be used in both Jobs and Versions.
-                # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
-                # [accelerators for online
-                # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
-              "count": "A String", # The number of accelerators to attach to each machine running the job.
-              "type": "A String", # The type of accelerator to use.
-            },
-            "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
-                # Registry. Learn more about [configuring custom
+                #
+                # Set `evaluatorConfig.imageUri` only if you build a custom image for
+                # your evaluator. If `evaluatorConfig.imageUri` has not been
+                # set, AI Platform uses the value of `masterConfig.imageUri`. Learn more about [configuring custom
                 # containers](/ai-platform/training/docs/distributed-training-containers).
-          },
-          "useChiefInTfConfig": True or False, # Optional. Use `chief` instead of `master` in the `TF_CONFIG` environment
-              # variable when training with a custom container. Defaults to `false`. [Learn
-              # more about this
-              # field.](/ai-platform/training/docs/distributed-training-details#chief-versus-master)
-              #
-              # This field has no effect for training jobs that don't use a custom
-              # container.
-          "masterConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for your master worker.
-              #
-              # You should only set `masterConfig.acceleratorConfig` if `masterType` is set
-              # to a Compute Engine machine type. Learn about [restrictions on accelerator
-              # configurations for
-              # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-              #
-              # Set `masterConfig.imageUri` only if you build a custom image. Only one of
-              # `masterConfig.imageUri` and `runtimeVersion` should be set. Learn more
-              # about [configuring custom
-              # containers](/ai-platform/training/docs/distributed-training-containers).
-            "tpuTfVersion": "A String", # The AI Platform runtime version that includes a TensorFlow version matching
-                # the one used in the custom container. This field is required if the replica
-                # is a TPU worker that uses a custom container. Otherwise, do not specify
-                # this field. This must be a [runtime version that currently supports
-                # training with
-                # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
-                #
-                # Note that the version of TensorFlow included in a runtime version may
-                # differ from the numbering of the runtime version itself, because it may
-                # have a different [patch
-                # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
-                # In this field, you must specify the runtime version (TensorFlow minor
-                # version). For example, if your custom container runs TensorFlow `1.x.y`,
-                # specify `1.x`.
-            "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
-                # [Learn about restrictions on accelerator configurations for
-                # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-                # Note that the AcceleratorConfig can be used in both Jobs and Versions.
-                # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
-                # [accelerators for online
-                # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
-              "count": "A String", # The number of accelerators to attach to each machine running the job.
-              "type": "A String", # The type of accelerator to use.
+              &quot;tpuTfVersion&quot;: &quot;A String&quot;, # The AI Platform runtime version that includes a TensorFlow version matching
+                  # the one used in the custom container. This field is required if the replica
+                  # is a TPU worker that uses a custom container. Otherwise, do not specify
+                  # this field. This must be a [runtime version that currently supports
+                  # training with
+                  # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
+                  #
+                  # Note that the version of TensorFlow included in a runtime version may
+                  # differ from the numbering of the runtime version itself, because it may
+                  # have a different [patch
+                  # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
+                  # In this field, you must specify the runtime version (TensorFlow minor
+                  # version). For example, if your custom container runs TensorFlow `1.x.y`,
+                  # specify `1.x`.
+              &quot;containerCommand&quot;: [ # The command with which the replica&#x27;s custom container is run.
+                  # If provided, it will override default ENTRYPOINT of the docker image.
+                  # If not provided, the docker image&#x27;s ENTRYPOINT is used.
+                  # It cannot be set if custom container image is
+                  # not provided.
+                  # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+                  # both cannot be set at the same time.
+                &quot;A String&quot;,
+              ],
+              &quot;imageUri&quot;: &quot;A String&quot;, # The Docker image to run on the replica. This image must be in Container
+                  # Registry. Learn more about [configuring custom
+                  # containers](/ai-platform/training/docs/distributed-training-containers).
+              &quot;containerArgs&quot;: [ # Arguments to the entrypoint command.
+                  # The following rules apply for container_command and container_args:
+                  # - If you do not supply command or args:
+                  #   The defaults defined in the Docker image are used.
+                  # - If you supply a command but no args:
+                  #   The default EntryPoint and the default Cmd defined in the Docker image
+                  #   are ignored. Your command is run without any arguments.
+                  # - If you supply only args:
+                  #   The default Entrypoint defined in the Docker image is run with the args
+                  #   that you supplied.
+                  # - If you supply a command and args:
+                  #   The default Entrypoint and the default Cmd defined in the Docker image
+                  #   are ignored. Your command is run with your args.
+                  # It cannot be set if custom container image is
+                  # not provided.
+                  # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+                  # both cannot be set at the same time.
+                &quot;A String&quot;,
+              ],
+              &quot;acceleratorConfig&quot;: { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
+                  # [Learn about restrictions on accelerator configurations for
+                  # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+                  # Note that the AcceleratorConfig can be used in both Jobs and Versions.
+                  # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
+                  # [accelerators for online
+                  # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
+                &quot;count&quot;: &quot;A String&quot;, # The number of accelerators to attach to each machine running the job.
+                &quot;type&quot;: &quot;A String&quot;, # The type of accelerator to use.
+              },
             },
-            "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
-                # Registry. Learn more about [configuring custom
-                # containers](/ai-platform/training/docs/distributed-training-containers).
+            &quot;useChiefInTfConfig&quot;: True or False, # Optional. Use `chief` instead of `master` in the `TF_CONFIG` environment
+                # variable when training with a custom container. Defaults to `false`. [Learn
+                # more about this
+                # field.](/ai-platform/training/docs/distributed-training-details#chief-versus-master)
+                #
+                # This field has no effect for training jobs that don&#x27;t use a custom
+                # container.
+            &quot;workerCount&quot;: &quot;A String&quot;, # Optional. The number of worker replicas to use for the training job. Each
+                # replica in the cluster will be of the type specified in `worker_type`.
+                #
+                # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
+                # set this value, you must also set `worker_type`.
+                #
+                # The default value is zero.
+            &quot;pythonModule&quot;: &quot;A String&quot;, # Required. The Python module name to run after installing the packages.
+            &quot;args&quot;: [ # Optional. Command-line arguments passed to the training application when it
+                # starts. If your job uses a custom container, then the arguments are passed
+                # to the container&#x27;s &lt;a class=&quot;external&quot; target=&quot;_blank&quot;
+                # href=&quot;https://docs.docker.com/engine/reference/builder/#entrypoint&quot;&gt;
+                # `ENTRYPOINT`&lt;/a&gt; command.
+              &quot;A String&quot;,
+            ],
           },
-          "runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for training. You must
-              # either specify this field or specify `masterConfig.imageUri`.
-              #
-              # For more information, see the [runtime version
-              # list](/ai-platform/training/docs/runtime-version-list) and learn [how to
-              # manage runtime versions](/ai-platform/training/docs/versioning).
-          "hyperparameters": { # Represents a set of hyperparameters to optimize. # Optional. The set of Hyperparameters to tune.
-            "maxTrials": 42, # Optional. How many training trials should be attempted to optimize
-                # the specified hyperparameters.
-                #
-                # Defaults to one.
-            "goal": "A String", # Required. The type of goal to use for tuning. Available types are
-                # `MAXIMIZE` and `MINIMIZE`.
-                #
-                # Defaults to `MAXIMIZE`.
-            "algorithm": "A String", # Optional. The search algorithm specified for the hyperparameter
-                # tuning job.
-                # Uses the default AI Platform hyperparameter tuning
-                # algorithm if unspecified.
-            "maxFailedTrials": 42, # Optional. The number of failed trials that need to be seen before failing
-                # the hyperparameter tuning job. You can specify this field to override the
-                # default failing criteria for AI Platform hyperparameter tuning jobs.
-                #
-                # Defaults to zero, which means the service decides when a hyperparameter
-                # job should fail.
-            "enableTrialEarlyStopping": True or False, # Optional. Indicates if the hyperparameter tuning job enables auto trial
-                # early stopping.
-            "resumePreviousJobId": "A String", # Optional. The prior hyperparameter tuning job id that users hope to
-                # continue with. The job id will be used to find the corresponding vizier
-                # study guid and resume the study.
-            "params": [ # Required. The set of parameters to tune.
-              { # Represents a single hyperparameter to optimize.
-                "maxValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
-                    # should be unset if type is `CATEGORICAL`. This value should be integers if
-                    # type is `INTEGER`.
-                "minValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
-                    # should be unset if type is `CATEGORICAL`. This value should be integers if
-                    # type is INTEGER.
-                "discreteValues": [ # Required if type is `DISCRETE`.
-                    # A list of feasible points.
-                    # The list should be in strictly increasing order. For instance, this
-                    # parameter might have possible settings of 1.5, 2.5, and 4.0. This list
-                    # should not contain more than 1,000 values.
-                  3.14,
+          &quot;state&quot;: &quot;A String&quot;, # Output only. The detailed state of a job.
+          &quot;jobId&quot;: &quot;A String&quot;, # Required. The user-specified id of the job.
+          &quot;endTime&quot;: &quot;A String&quot;, # Output only. When the job processing was completed.
+          &quot;startTime&quot;: &quot;A String&quot;, # Output only. When the job processing was started.
+          &quot;predictionOutput&quot;: { # Represents results of a prediction job. # The current prediction job result.
+            &quot;errorCount&quot;: &quot;A String&quot;, # The number of data instances which resulted in errors.
+            &quot;outputPath&quot;: &quot;A String&quot;, # The output Google Cloud Storage location provided at the job creation time.
+            &quot;nodeHours&quot;: 3.14, # Node hours used by the batch prediction job.
+            &quot;predictionCount&quot;: &quot;A String&quot;, # The number of generated predictions.
+          },
+          &quot;trainingOutput&quot;: { # Represents results of a training job. Output only. # The current training job result.
+            &quot;trials&quot;: [ # Results for individual Hyperparameter trials.
+                # Only set for hyperparameter tuning jobs.
+              { # Represents the result of a single hyperparameter tuning trial from a
+                  # training job. The TrainingOutput object that is returned on successful
+                  # completion of a training job with hyperparameter tuning includes a list
+                  # of HyperparameterOutput objects, one for each successful trial.
+                &quot;allMetrics&quot;: [ # All recorded object metrics for this trial. This field is not currently
+                    # populated.
+                  { # An observed value of a metric.
+                    &quot;objectiveValue&quot;: 3.14, # The objective value at this training step.
+                    &quot;trainingStep&quot;: &quot;A String&quot;, # The global training step for this metric.
+                  },
                 ],
-                "parameterName": "A String", # Required. The parameter name must be unique amongst all ParameterConfigs in
-                    # a HyperparameterSpec message. E.g., "learning_rate".
-                "categoricalValues": [ # Required if type is `CATEGORICAL`. The list of possible categories.
-                  "A String",
-                ],
-                "type": "A String", # Required. The type of the parameter.
-                "scaleType": "A String", # Optional. How the parameter should be scaled to the hypercube.
-                    # Leave unset for categorical parameters.
-                    # Some kind of scaling is strongly recommended for real or integral
-                    # parameters (e.g., `UNIT_LINEAR_SCALE`).
+                &quot;hyperparameters&quot;: { # The hyperparameters given to this trial.
+                  &quot;a_key&quot;: &quot;A String&quot;,
+                },
+                &quot;trialId&quot;: &quot;A String&quot;, # The trial id for these results.
+                &quot;endTime&quot;: &quot;A String&quot;, # Output only. End time for the trial.
+                &quot;isTrialStoppedEarly&quot;: True or False, # True if the trial is stopped early.
+                &quot;startTime&quot;: &quot;A String&quot;, # Output only. Start time for the trial.
+                &quot;finalMetric&quot;: { # An observed value of a metric. # The final objective metric seen for this trial.
+                  &quot;objectiveValue&quot;: 3.14, # The objective value at this training step.
+                  &quot;trainingStep&quot;: &quot;A String&quot;, # The global training step for this metric.
+                },
+                &quot;builtInAlgorithmOutput&quot;: { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
+                    # Only set for trials of built-in algorithms jobs that have succeeded.
+                  &quot;framework&quot;: &quot;A String&quot;, # Framework on which the built-in algorithm was trained.
+                  &quot;modelPath&quot;: &quot;A String&quot;, # The Cloud Storage path to the `model/` directory where the training job
+                      # saves the trained model. Only set for successful jobs that don&#x27;t use
+                      # hyperparameter tuning.
+                  &quot;pythonVersion&quot;: &quot;A String&quot;, # Python version on which the built-in algorithm was trained.
+                  &quot;runtimeVersion&quot;: &quot;A String&quot;, # AI Platform runtime version on which the built-in algorithm was
+                      # trained.
+                },
+                &quot;state&quot;: &quot;A String&quot;, # Output only. The detailed state of the trial.
               },
             ],
-            "hyperparameterMetricTag": "A String", # Optional. The TensorFlow summary tag name to use for optimizing trials. For
-                # current versions of TensorFlow, this tag name should exactly match what is
-                # shown in TensorBoard, including all scopes.  For versions of TensorFlow
-                # prior to 0.12, this should be only the tag passed to tf.Summary.
-                # By default, "training/hptuning/metric" will be used.
-            "maxParallelTrials": 42, # Optional. The number of training trials to run concurrently.
-                # You can reduce the time it takes to perform hyperparameter tuning by adding
-                # trials in parallel. However, each trail only benefits from the information
-                # gained in completed trials. That means that a trial does not get access to
-                # the results of trials running at the same time, which could reduce the
-                # quality of the overall optimization.
-                #
-                # Each trial will use the same scale tier and machine types.
-                #
-                # Defaults to one.
-          },
-          "args": [ # Optional. Command-line arguments passed to the training application when it
-              # starts. If your job uses a custom container, then the arguments are passed
-              # to the container's &lt;a class="external" target="_blank"
-              # href="https://docs.docker.com/engine/reference/builder/#entrypoint"&gt;
-              # `ENTRYPOINT`&lt;/a&gt; command.
-            "A String",
-          ],
-          "pythonModule": "A String", # Required. The Python module name to run after installing the packages.
-          "workerCount": "A String", # Optional. The number of worker replicas to use for the training job. Each
-              # replica in the cluster will be of the type specified in `worker_type`.
-              #
-              # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
-              # set this value, you must also set `worker_type`.
-              #
-              # The default value is zero.
-          "encryptionConfig": { # Represents a custom encryption key configuration that can be applied to # Optional. Options for using customer-managed encryption keys (CMEK) to
-              # protect resources created by a training job, instead of using Google's
-              # default encryption. If this is set, then all resources created by the
-              # training job will be encrypted with the customer-managed encryption key
-              # that you specify.
-              #
-              # [Learn how and when to use CMEK with AI Platform
-              # Training](/ai-platform/training/docs/cmek).
-              # a resource.
-            "kmsKeyName": "A String", # The Cloud KMS resource identifier of the customer-managed encryption key
-                # used to protect a resource, such as a training job. It has the following
-                # format:
-                # `projects/{PROJECT_ID}/locations/{REGION}/keyRings/{KEY_RING_NAME}/cryptoKeys/{KEY_NAME}`
-          },
-          "parameterServerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for parameter servers.
-              #
-              # You should only set `parameterServerConfig.acceleratorConfig` if
-              # `parameterServerType` is set to a Compute Engine machine type. [Learn
-              # about restrictions on accelerator configurations for
-              # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-              #
-              # Set `parameterServerConfig.imageUri` only if you build a custom image for
-              # your parameter server. If `parameterServerConfig.imageUri` has not been
-              # set, AI Platform uses the value of `masterConfig.imageUri`. Learn more about [configuring custom
-              # containers](/ai-platform/training/docs/distributed-training-containers).
-            "tpuTfVersion": "A String", # The AI Platform runtime version that includes a TensorFlow version matching
-                # the one used in the custom container. This field is required if the replica
-                # is a TPU worker that uses a custom container. Otherwise, do not specify
-                # this field. This must be a [runtime version that currently supports
-                # training with
-                # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
-                #
-                # Note that the version of TensorFlow included in a runtime version may
-                # differ from the numbering of the runtime version itself, because it may
-                # have a different [patch
-                # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
-                # In this field, you must specify the runtime version (TensorFlow minor
-                # version). For example, if your custom container runs TensorFlow `1.x.y`,
-                # specify `1.x`.
-            "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
-                # [Learn about restrictions on accelerator configurations for
-                # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-                # Note that the AcceleratorConfig can be used in both Jobs and Versions.
-                # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
-                # [accelerators for online
-                # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
-              "count": "A String", # The number of accelerators to attach to each machine running the job.
-              "type": "A String", # The type of accelerator to use.
+            &quot;hyperparameterMetricTag&quot;: &quot;A String&quot;, # The TensorFlow summary tag name used for optimizing hyperparameter tuning
+                # trials. See
+                # [`HyperparameterSpec.hyperparameterMetricTag`](#HyperparameterSpec.FIELDS.hyperparameter_metric_tag)
+                # for more information. Only set for hyperparameter tuning jobs.
+            &quot;completedTrialCount&quot;: &quot;A String&quot;, # The number of hyperparameter tuning trials that completed successfully.
+                # Only set for hyperparameter tuning jobs.
+            &quot;isHyperparameterTuningJob&quot;: True or False, # Whether this job is a hyperparameter tuning job.
+            &quot;consumedMLUnits&quot;: 3.14, # The amount of ML units consumed by the job.
+            &quot;isBuiltInAlgorithmJob&quot;: True or False, # Whether this job is a built-in Algorithm job.
+            &quot;builtInAlgorithmOutput&quot;: { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
+                # Only set for built-in algorithms jobs.
+              &quot;framework&quot;: &quot;A String&quot;, # Framework on which the built-in algorithm was trained.
+              &quot;modelPath&quot;: &quot;A String&quot;, # The Cloud Storage path to the `model/` directory where the training job
+                  # saves the trained model. Only set for successful jobs that don&#x27;t use
+                  # hyperparameter tuning.
+              &quot;pythonVersion&quot;: &quot;A String&quot;, # Python version on which the built-in algorithm was trained.
+              &quot;runtimeVersion&quot;: &quot;A String&quot;, # AI Platform runtime version on which the built-in algorithm was
+                  # trained.
             },
-            "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
-                # Registry. Learn more about [configuring custom
-                # containers](/ai-platform/training/docs/distributed-training-containers).
           },
-          "region": "A String", # Required. The region to run the training job in. See the [available
-              # regions](/ai-platform/training/docs/regions) for AI Platform Training.
-          "pythonVersion": "A String", # Optional. The version of Python used in training. You must either specify
-              # this field or specify `masterConfig.imageUri`.
-              #
-              # The following Python versions are available:
-              #
-              # * Python '3.7' is available when `runtime_version` is set to '1.15' or
-              #   later.
-              # * Python '3.5' is available when `runtime_version` is set to a version
-              #   from '1.4' to '1.14'.
-              # * Python '2.7' is available when `runtime_version` is set to '1.15' or
-              #   earlier.
-              #
-              # Read more about the Python versions available for [each runtime
-              # version](/ml-engine/docs/runtime-version-list).
-          "evaluatorType": "A String", # Optional. Specifies the type of virtual machine to use for your training
-              # job's evaluator nodes.
-              #
-              # The supported values are the same as those described in the entry for
-              # `masterType`.
-              #
-              # This value must be consistent with the category of machine type that
-              # `masterType` uses. In other words, both must be Compute Engine machine
-              # types or both must be legacy machine types.
-              #
-              # This value must be present when `scaleTier` is set to `CUSTOM` and
-              # `evaluatorCount` is greater than zero.
-          "parameterServerType": "A String", # Optional. Specifies the type of virtual machine to use for your training
-              # job's parameter server.
-              #
-              # The supported values are the same as those described in the entry for
-              # `master_type`.
-              #
-              # This value must be consistent with the category of machine type that
-              # `masterType` uses. In other words, both must be Compute Engine machine
-              # types or both must be legacy machine types.
-              #
-              # This value must be present when `scaleTier` is set to `CUSTOM` and
-              # `parameter_server_count` is greater than zero.
+          &quot;createTime&quot;: &quot;A String&quot;, # Output only. When the job was created.
+          &quot;labels&quot;: { # Optional. One or more labels that you can add, to organize your jobs.
+              # Each label is a key-value pair, where both the key and the value are
+              # arbitrary strings that you supply.
+              # For more information, see the documentation on
+              # &lt;a href=&quot;/ml-engine/docs/tensorflow/resource-labels&quot;&gt;using labels&lt;/a&gt;.
+            &quot;a_key&quot;: &quot;A String&quot;,
+          },
+          &quot;predictionInput&quot;: { # Represents input parameters for a prediction job. # Input parameters to create a prediction job.
+            &quot;outputPath&quot;: &quot;A String&quot;, # Required. The output Google Cloud Storage location.
+            &quot;outputDataFormat&quot;: &quot;A String&quot;, # Optional. Format of the output data files, defaults to JSON.
+            &quot;dataFormat&quot;: &quot;A String&quot;, # Required. The format of the input data files.
+            &quot;batchSize&quot;: &quot;A String&quot;, # Optional. Number of records per batch, defaults to 64.
+                # The service will buffer batch_size number of records in memory before
+                # invoking one Tensorflow prediction call internally. So take the record
+                # size and memory available into consideration when setting this parameter.
+            &quot;runtimeVersion&quot;: &quot;A String&quot;, # Optional. The AI Platform runtime version to use for this batch
+                # prediction. If not set, AI Platform will pick the runtime version used
+                # during the CreateVersion request for this model version, or choose the
+                # latest stable version when model version information is not available
+                # such as when the model is specified by uri.
+            &quot;inputPaths&quot;: [ # Required. The Cloud Storage location of the input data files. May contain
+                # &lt;a href=&quot;/storage/docs/gsutil/addlhelp/WildcardNames&quot;&gt;wildcards&lt;/a&gt;.
+              &quot;A String&quot;,
+            ],
+            &quot;region&quot;: &quot;A String&quot;, # Required. The Google Compute Engine region to run the prediction job in.
+                # See the &lt;a href=&quot;/ml-engine/docs/tensorflow/regions&quot;&gt;available regions&lt;/a&gt;
+                # for AI Platform services.
+            &quot;versionName&quot;: &quot;A String&quot;, # Use this field if you want to specify a version of the model to use. The
+                # string is formatted the same way as `model_version`, with the addition
+                # of the version information:
+                #
+                # `&quot;projects/YOUR_PROJECT/models/YOUR_MODEL/versions/YOUR_VERSION&quot;`
+            &quot;modelName&quot;: &quot;A String&quot;, # Use this field if you want to use the default version for the specified
+                # model. The string must use the following format:
+                #
+                # `&quot;projects/YOUR_PROJECT/models/YOUR_MODEL&quot;`
+            &quot;uri&quot;: &quot;A String&quot;, # Use this field if you want to specify a Google Cloud Storage path for
+                # the model to use.
+            &quot;maxWorkerCount&quot;: &quot;A String&quot;, # Optional. The maximum number of workers to be used for parallel processing.
+                # Defaults to 10 if not specified.
+            &quot;signatureName&quot;: &quot;A String&quot;, # Optional. The name of the signature defined in the SavedModel to use for
+                # this job. Please refer to
+                # [SavedModel](https://tensorflow.github.io/serving/serving_basic.html)
+                # for information about how to use signatures.
+                #
+                # Defaults to
+                # [DEFAULT_SERVING_SIGNATURE_DEF_KEY](https://www.tensorflow.org/api_docs/python/tf/saved_model/signature_constants)
+                # , which is &quot;serving_default&quot;.
+          },
+          &quot;errorMessage&quot;: &quot;A String&quot;, # Output only. The details of a failure or a cancellation.
         },
-        "endTime": "A String", # Output only. When the job processing was completed.
-        "predictionOutput": { # Represents results of a prediction job. # The current prediction job result.
-          "outputPath": "A String", # The output Google Cloud Storage location provided at the job creation time.
-          "nodeHours": 3.14, # Node hours used by the batch prediction job.
-          "predictionCount": "A String", # The number of generated predictions.
-          "errorCount": "A String", # The number of data instances which resulted in errors.
-        },
-        "createTime": "A String", # Output only. When the job was created.
-      },
     ],
+    &quot;nextPageToken&quot;: &quot;A String&quot;, # Optional. Pass this token as the `page_token` field of the request for a
+        # subsequent call.
   }</pre>
 </div>
 
@@ -2764,7 +3275,7 @@
   previous_response: The response from the request for the previous page. (required)
 
 Returns:
-  A request object that you can call 'execute()' on to request the next
+  A request object that you can call &#x27;execute()&#x27; on to request the next
   page. Returns None if there are no more items in the collection.
     </pre>
 </div>
@@ -2781,718 +3292,256 @@
     The object takes the form of:
 
 { # Represents a training or prediction job.
-  "errorMessage": "A String", # Output only. The details of a failure or a cancellation.
-  "trainingOutput": { # Represents results of a training job. Output only. # The current training job result.
-    "completedTrialCount": "A String", # The number of hyperparameter tuning trials that completed successfully.
-        # Only set for hyperparameter tuning jobs.
-    "trials": [ # Results for individual Hyperparameter trials.
-        # Only set for hyperparameter tuning jobs.
-      { # Represents the result of a single hyperparameter tuning trial from a
-          # training job. The TrainingOutput object that is returned on successful
-          # completion of a training job with hyperparameter tuning includes a list
-          # of HyperparameterOutput objects, one for each successful trial.
-        "startTime": "A String", # Output only. Start time for the trial.
-        "hyperparameters": { # The hyperparameters given to this trial.
-          "a_key": "A String",
-        },
-        "finalMetric": { # An observed value of a metric. # The final objective metric seen for this trial.
-          "trainingStep": "A String", # The global training step for this metric.
-          "objectiveValue": 3.14, # The objective value at this training step.
-        },
-        "state": "A String", # Output only. The detailed state of the trial.
-        "allMetrics": [ # All recorded object metrics for this trial. This field is not currently
-            # populated.
-          { # An observed value of a metric.
-            "trainingStep": "A String", # The global training step for this metric.
-            "objectiveValue": 3.14, # The objective value at this training step.
-          },
-        ],
-        "isTrialStoppedEarly": True or False, # True if the trial is stopped early.
-        "endTime": "A String", # Output only. End time for the trial.
-        "trialId": "A String", # The trial id for these results.
-        "builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
-            # Only set for trials of built-in algorithms jobs that have succeeded.
-          "framework": "A String", # Framework on which the built-in algorithm was trained.
-          "modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job
-              # saves the trained model. Only set for successful jobs that don't use
-              # hyperparameter tuning.
-          "runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was
-              # trained.
-          "pythonVersion": "A String", # Python version on which the built-in algorithm was trained.
-        },
-      },
-    ],
-    "isHyperparameterTuningJob": True or False, # Whether this job is a hyperparameter tuning job.
-    "isBuiltInAlgorithmJob": True or False, # Whether this job is a built-in Algorithm job.
-    "consumedMLUnits": 3.14, # The amount of ML units consumed by the job.
-    "hyperparameterMetricTag": "A String", # The TensorFlow summary tag name used for optimizing hyperparameter tuning
-        # trials. See
-        # [`HyperparameterSpec.hyperparameterMetricTag`](#HyperparameterSpec.FIELDS.hyperparameter_metric_tag)
-        # for more information. Only set for hyperparameter tuning jobs.
-    "builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
-        # Only set for built-in algorithms jobs.
-      "framework": "A String", # Framework on which the built-in algorithm was trained.
-      "modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job
-          # saves the trained model. Only set for successful jobs that don't use
-          # hyperparameter tuning.
-      "runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was
-          # trained.
-      "pythonVersion": "A String", # Python version on which the built-in algorithm was trained.
-    },
-  },
-  "predictionInput": { # Represents input parameters for a prediction job. # Input parameters to create a prediction job.
-    "modelName": "A String", # Use this field if you want to use the default version for the specified
-        # model. The string must use the following format:
-        #
-        # `"projects/YOUR_PROJECT/models/YOUR_MODEL"`
-    "signatureName": "A String", # Optional. The name of the signature defined in the SavedModel to use for
-        # this job. Please refer to
-        # [SavedModel](https://tensorflow.github.io/serving/serving_basic.html)
-        # for information about how to use signatures.
-        #
-        # Defaults to
-        # [DEFAULT_SERVING_SIGNATURE_DEF_KEY](https://www.tensorflow.org/api_docs/python/tf/saved_model/signature_constants)
-        # , which is "serving_default".
-    "runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for this batch
-        # prediction. If not set, AI Platform will pick the runtime version used
-        # during the CreateVersion request for this model version, or choose the
-        # latest stable version when model version information is not available
-        # such as when the model is specified by uri.
-    "batchSize": "A String", # Optional. Number of records per batch, defaults to 64.
-        # The service will buffer batch_size number of records in memory before
-        # invoking one Tensorflow prediction call internally. So take the record
-        # size and memory available into consideration when setting this parameter.
-    "maxWorkerCount": "A String", # Optional. The maximum number of workers to be used for parallel processing.
-        # Defaults to 10 if not specified.
-    "uri": "A String", # Use this field if you want to specify a Google Cloud Storage path for
-        # the model to use.
-    "outputPath": "A String", # Required. The output Google Cloud Storage location.
-    "dataFormat": "A String", # Required. The format of the input data files.
-    "versionName": "A String", # Use this field if you want to specify a version of the model to use. The
-        # string is formatted the same way as `model_version`, with the addition
-        # of the version information:
-        #
-        # `"projects/YOUR_PROJECT/models/YOUR_MODEL/versions/YOUR_VERSION"`
-    "region": "A String", # Required. The Google Compute Engine region to run the prediction job in.
-        # See the &lt;a href="/ml-engine/docs/tensorflow/regions"&gt;available regions&lt;/a&gt;
-        # for AI Platform services.
-    "inputPaths": [ # Required. The Cloud Storage location of the input data files. May contain
-        # &lt;a href="/storage/docs/gsutil/addlhelp/WildcardNames"&gt;wildcards&lt;/a&gt;.
-      "A String",
-    ],
-    "outputDataFormat": "A String", # Optional. Format of the output data files, defaults to JSON.
-  },
-  "labels": { # Optional. One or more labels that you can add, to organize your jobs.
-      # Each label is a key-value pair, where both the key and the value are
-      # arbitrary strings that you supply.
-      # For more information, see the documentation on
-      # &lt;a href="/ml-engine/docs/tensorflow/resource-labels"&gt;using labels&lt;/a&gt;.
-    "a_key": "A String",
-  },
-  "jobId": "A String", # Required. The user-specified id of the job.
-  "state": "A String", # Output only. The detailed state of a job.
-  "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help
-      # prevent simultaneous updates of a job from overwriting each other.
-      # It is strongly suggested that systems make use of the `etag` in the
-      # read-modify-write cycle to perform job updates in order to avoid race
-      # conditions: An `etag` is returned in the response to `GetJob`, and
-      # systems are expected to put that etag in the request to `UpdateJob` to
-      # ensure that their change will be applied to the same version of the job.
-  "startTime": "A String", # Output only. When the job processing was started.
-  "trainingInput": { # Represents input parameters for a training job. When using the gcloud command # Input parameters to create a training job.
-      # to submit your training job, you can specify the input parameters as
-      # command-line arguments and/or in a YAML configuration file referenced from
-      # the --config command-line argument. For details, see the guide to [submitting
-      # a training job](/ai-platform/training/docs/training-jobs).
-    "masterType": "A String", # Optional. Specifies the type of virtual machine to use for your training
-        # job's master worker. You must specify this field when `scaleTier` is set to
-        # `CUSTOM`.
-        #
-        # You can use certain Compute Engine machine types directly in this field.
-        # The following types are supported:
-        #
-        # - `n1-standard-4`
-        # - `n1-standard-8`
-        # - `n1-standard-16`
-        # - `n1-standard-32`
-        # - `n1-standard-64`
-        # - `n1-standard-96`
-        # - `n1-highmem-2`
-        # - `n1-highmem-4`
-        # - `n1-highmem-8`
-        # - `n1-highmem-16`
-        # - `n1-highmem-32`
-        # - `n1-highmem-64`
-        # - `n1-highmem-96`
-        # - `n1-highcpu-16`
-        # - `n1-highcpu-32`
-        # - `n1-highcpu-64`
-        # - `n1-highcpu-96`
-        #
-        # Learn more about [using Compute Engine machine
-        # types](/ml-engine/docs/machine-types#compute-engine-machine-types).
-        #
-        # Alternatively, you can use the following legacy machine types:
-        #
-        # - `standard`
-        # - `large_model`
-        # - `complex_model_s`
-        # - `complex_model_m`
-        # - `complex_model_l`
-        # - `standard_gpu`
-        # - `complex_model_m_gpu`
-        # - `complex_model_l_gpu`
-        # - `standard_p100`
-        # - `complex_model_m_p100`
-        # - `standard_v100`
-        # - `large_model_v100`
-        # - `complex_model_m_v100`
-        # - `complex_model_l_v100`
-        #
-        # Learn more about [using legacy machine
-        # types](/ml-engine/docs/machine-types#legacy-machine-types).
-        #
-        # Finally, if you want to use a TPU for training, specify `cloud_tpu` in this
-        # field. Learn more about the [special configuration options for training
-        # with
-        # TPUs](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine).
-    "jobDir": "A String", # Optional. A Google Cloud Storage path in which to store training outputs
-        # and other data needed for training. This path is passed to your TensorFlow
-        # program as the '--job-dir' command-line argument. The benefit of specifying
-        # this field is that Cloud ML validates the path for use in training.
-    "scheduling": { # All parameters related to scheduling of training jobs. # Optional. Scheduling options for a training job.
-      "maxRunningTime": "A String", # Optional. The maximum job running time, expressed in seconds. The field can
-          # contain up to nine fractional digits, terminated by `s`. By default there
-          # is no limit to the running time.
-          #
-          # If the training job is still running after this duration, AI Platform
-          # Training cancels it.
-          #
-          # For example, if you want to ensure your job runs for no more than 2 hours,
-          # set this field to `7200s` (2 hours * 60 minutes / hour * 60 seconds /
-          # minute).
-          #
-          # If you submit your training job using the `gcloud` tool, you can [provide
-          # this field in a `config.yaml`
-          # file](/ai-platform/training/docs/training-jobs#formatting_your_configuration_parameters).
-          # For example:
-          #
-          # ```yaml
-          # trainingInput:
-          #   ...
-          #   scheduling:
-          #     maxRunningTime: 7200s
-          #   ...
-          # ```
-    },
-    "parameterServerCount": "A String", # Optional. The number of parameter server replicas to use for the training
-        # job. Each replica in the cluster will be of the type specified in
-        # `parameter_server_type`.
-        #
-        # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
-        # set this value, you must also set `parameter_server_type`.
-        #
-        # The default value is zero.
-    "evaluatorCount": "A String", # Optional. The number of evaluator replicas to use for the training job.
-        # Each replica in the cluster will be of the type specified in
-        # `evaluator_type`.
-        #
-        # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
-        # set this value, you must also set `evaluator_type`.
-        #
-        # The default value is zero.
-    "workerType": "A String", # Optional. Specifies the type of virtual machine to use for your training
-        # job's worker nodes.
-        #
-        # The supported values are the same as those described in the entry for
-        # `masterType`.
-        #
-        # This value must be consistent with the category of machine type that
-        # `masterType` uses. In other words, both must be Compute Engine machine
-        # types or both must be legacy machine types.
-        #
-        # If you use `cloud_tpu` for this value, see special instructions for
-        # [configuring a custom TPU
-        # machine](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine).
-        #
-        # This value must be present when `scaleTier` is set to `CUSTOM` and
-        # `workerCount` is greater than zero.
-    "scaleTier": "A String", # Required. Specifies the machine types, the number of replicas for workers
-        # and parameter servers.
-    "packageUris": [ # Required. The Google Cloud Storage location of the packages with
-        # the training program and any additional dependencies.
-        # The maximum number of package URIs is 100.
-      "A String",
-    ],
-    "workerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for workers.
-        #
-        # You should only set `workerConfig.acceleratorConfig` if `workerType` is set
-        # to a Compute Engine machine type. [Learn about restrictions on accelerator
-        # configurations for
-        # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-        #
-        # Set `workerConfig.imageUri` only if you build a custom image for your
-        # worker. If `workerConfig.imageUri` has not been set, AI Platform uses
-        # the value of `masterConfig.imageUri`. Learn more about [configuring custom
-        # containers](/ai-platform/training/docs/distributed-training-containers).
-      "tpuTfVersion": "A String", # The AI Platform runtime version that includes a TensorFlow version matching
-          # the one used in the custom container. This field is required if the replica
-          # is a TPU worker that uses a custom container. Otherwise, do not specify
-          # this field. This must be a [runtime version that currently supports
-          # training with
-          # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
-          #
-          # Note that the version of TensorFlow included in a runtime version may
-          # differ from the numbering of the runtime version itself, because it may
-          # have a different [patch
-          # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
-          # In this field, you must specify the runtime version (TensorFlow minor
-          # version). For example, if your custom container runs TensorFlow `1.x.y`,
-          # specify `1.x`.
-      "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
-          # [Learn about restrictions on accelerator configurations for
-          # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-          # Note that the AcceleratorConfig can be used in both Jobs and Versions.
-          # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
-          # [accelerators for online
-          # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
-        "count": "A String", # The number of accelerators to attach to each machine running the job.
-        "type": "A String", # The type of accelerator to use.
-      },
-      "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
-          # Registry. Learn more about [configuring custom
-          # containers](/ai-platform/training/docs/distributed-training-containers).
-    },
-    "evaluatorConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for evaluators.
-        #
-        # You should only set `evaluatorConfig.acceleratorConfig` if
-        # `evaluatorType` is set to a Compute Engine machine type. [Learn
-        # about restrictions on accelerator configurations for
-        # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-        #
-        # Set `evaluatorConfig.imageUri` only if you build a custom image for
-        # your evaluator. If `evaluatorConfig.imageUri` has not been
-        # set, AI Platform uses the value of `masterConfig.imageUri`. Learn more about [configuring custom
-        # containers](/ai-platform/training/docs/distributed-training-containers).
-      "tpuTfVersion": "A String", # The AI Platform runtime version that includes a TensorFlow version matching
-          # the one used in the custom container. This field is required if the replica
-          # is a TPU worker that uses a custom container. Otherwise, do not specify
-          # this field. This must be a [runtime version that currently supports
-          # training with
-          # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
-          #
-          # Note that the version of TensorFlow included in a runtime version may
-          # differ from the numbering of the runtime version itself, because it may
-          # have a different [patch
-          # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
-          # In this field, you must specify the runtime version (TensorFlow minor
-          # version). For example, if your custom container runs TensorFlow `1.x.y`,
-          # specify `1.x`.
-      "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
-          # [Learn about restrictions on accelerator configurations for
-          # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-          # Note that the AcceleratorConfig can be used in both Jobs and Versions.
-          # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
-          # [accelerators for online
-          # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
-        "count": "A String", # The number of accelerators to attach to each machine running the job.
-        "type": "A String", # The type of accelerator to use.
-      },
-      "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
-          # Registry. Learn more about [configuring custom
-          # containers](/ai-platform/training/docs/distributed-training-containers).
-    },
-    "useChiefInTfConfig": True or False, # Optional. Use `chief` instead of `master` in the `TF_CONFIG` environment
-        # variable when training with a custom container. Defaults to `false`. [Learn
-        # more about this
-        # field.](/ai-platform/training/docs/distributed-training-details#chief-versus-master)
-        #
-        # This field has no effect for training jobs that don't use a custom
-        # container.
-    "masterConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for your master worker.
-        #
-        # You should only set `masterConfig.acceleratorConfig` if `masterType` is set
-        # to a Compute Engine machine type. Learn about [restrictions on accelerator
-        # configurations for
-        # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-        #
-        # Set `masterConfig.imageUri` only if you build a custom image. Only one of
-        # `masterConfig.imageUri` and `runtimeVersion` should be set. Learn more
-        # about [configuring custom
-        # containers](/ai-platform/training/docs/distributed-training-containers).
-      "tpuTfVersion": "A String", # The AI Platform runtime version that includes a TensorFlow version matching
-          # the one used in the custom container. This field is required if the replica
-          # is a TPU worker that uses a custom container. Otherwise, do not specify
-          # this field. This must be a [runtime version that currently supports
-          # training with
-          # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
-          #
-          # Note that the version of TensorFlow included in a runtime version may
-          # differ from the numbering of the runtime version itself, because it may
-          # have a different [patch
-          # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
-          # In this field, you must specify the runtime version (TensorFlow minor
-          # version). For example, if your custom container runs TensorFlow `1.x.y`,
-          # specify `1.x`.
-      "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
-          # [Learn about restrictions on accelerator configurations for
-          # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-          # Note that the AcceleratorConfig can be used in both Jobs and Versions.
-          # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
-          # [accelerators for online
-          # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
-        "count": "A String", # The number of accelerators to attach to each machine running the job.
-        "type": "A String", # The type of accelerator to use.
-      },
-      "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
-          # Registry. Learn more about [configuring custom
-          # containers](/ai-platform/training/docs/distributed-training-containers).
-    },
-    "runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for training. You must
-        # either specify this field or specify `masterConfig.imageUri`.
-        #
-        # For more information, see the [runtime version
-        # list](/ai-platform/training/docs/runtime-version-list) and learn [how to
-        # manage runtime versions](/ai-platform/training/docs/versioning).
-    "hyperparameters": { # Represents a set of hyperparameters to optimize. # Optional. The set of Hyperparameters to tune.
-      "maxTrials": 42, # Optional. How many training trials should be attempted to optimize
-          # the specified hyperparameters.
-          #
-          # Defaults to one.
-      "goal": "A String", # Required. The type of goal to use for tuning. Available types are
-          # `MAXIMIZE` and `MINIMIZE`.
-          #
-          # Defaults to `MAXIMIZE`.
-      "algorithm": "A String", # Optional. The search algorithm specified for the hyperparameter
-          # tuning job.
-          # Uses the default AI Platform hyperparameter tuning
-          # algorithm if unspecified.
-      "maxFailedTrials": 42, # Optional. The number of failed trials that need to be seen before failing
-          # the hyperparameter tuning job. You can specify this field to override the
-          # default failing criteria for AI Platform hyperparameter tuning jobs.
-          #
-          # Defaults to zero, which means the service decides when a hyperparameter
-          # job should fail.
-      "enableTrialEarlyStopping": True or False, # Optional. Indicates if the hyperparameter tuning job enables auto trial
-          # early stopping.
-      "resumePreviousJobId": "A String", # Optional. The prior hyperparameter tuning job id that users hope to
-          # continue with. The job id will be used to find the corresponding vizier
-          # study guid and resume the study.
-      "params": [ # Required. The set of parameters to tune.
-        { # Represents a single hyperparameter to optimize.
-          "maxValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
-              # should be unset if type is `CATEGORICAL`. This value should be integers if
-              # type is `INTEGER`.
-          "minValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
-              # should be unset if type is `CATEGORICAL`. This value should be integers if
-              # type is INTEGER.
-          "discreteValues": [ # Required if type is `DISCRETE`.
-              # A list of feasible points.
-              # The list should be in strictly increasing order. For instance, this
-              # parameter might have possible settings of 1.5, 2.5, and 4.0. This list
-              # should not contain more than 1,000 values.
-            3.14,
-          ],
-          "parameterName": "A String", # Required. The parameter name must be unique amongst all ParameterConfigs in
-              # a HyperparameterSpec message. E.g., "learning_rate".
-          "categoricalValues": [ # Required if type is `CATEGORICAL`. The list of possible categories.
-            "A String",
-          ],
-          "type": "A String", # Required. The type of the parameter.
-          "scaleType": "A String", # Optional. How the parameter should be scaled to the hypercube.
-              # Leave unset for categorical parameters.
-              # Some kind of scaling is strongly recommended for real or integral
-              # parameters (e.g., `UNIT_LINEAR_SCALE`).
-        },
-      ],
-      "hyperparameterMetricTag": "A String", # Optional. The TensorFlow summary tag name to use for optimizing trials. For
-          # current versions of TensorFlow, this tag name should exactly match what is
-          # shown in TensorBoard, including all scopes.  For versions of TensorFlow
-          # prior to 0.12, this should be only the tag passed to tf.Summary.
-          # By default, "training/hptuning/metric" will be used.
-      "maxParallelTrials": 42, # Optional. The number of training trials to run concurrently.
-          # You can reduce the time it takes to perform hyperparameter tuning by adding
-          # trials in parallel. However, each trail only benefits from the information
-          # gained in completed trials. That means that a trial does not get access to
-          # the results of trials running at the same time, which could reduce the
-          # quality of the overall optimization.
-          #
-          # Each trial will use the same scale tier and machine types.
-          #
-          # Defaults to one.
-    },
-    "args": [ # Optional. Command-line arguments passed to the training application when it
-        # starts. If your job uses a custom container, then the arguments are passed
-        # to the container's &lt;a class="external" target="_blank"
-        # href="https://docs.docker.com/engine/reference/builder/#entrypoint"&gt;
-        # `ENTRYPOINT`&lt;/a&gt; command.
-      "A String",
-    ],
-    "pythonModule": "A String", # Required. The Python module name to run after installing the packages.
-    "workerCount": "A String", # Optional. The number of worker replicas to use for the training job. Each
-        # replica in the cluster will be of the type specified in `worker_type`.
-        #
-        # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
-        # set this value, you must also set `worker_type`.
-        #
-        # The default value is zero.
-    "encryptionConfig": { # Represents a custom encryption key configuration that can be applied to # Optional. Options for using customer-managed encryption keys (CMEK) to
-        # protect resources created by a training job, instead of using Google's
-        # default encryption. If this is set, then all resources created by the
-        # training job will be encrypted with the customer-managed encryption key
-        # that you specify.
-        #
-        # [Learn how and when to use CMEK with AI Platform
-        # Training](/ai-platform/training/docs/cmek).
-        # a resource.
-      "kmsKeyName": "A String", # The Cloud KMS resource identifier of the customer-managed encryption key
-          # used to protect a resource, such as a training job. It has the following
-          # format:
-          # `projects/{PROJECT_ID}/locations/{REGION}/keyRings/{KEY_RING_NAME}/cryptoKeys/{KEY_NAME}`
-    },
-    "parameterServerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for parameter servers.
-        #
-        # You should only set `parameterServerConfig.acceleratorConfig` if
-        # `parameterServerType` is set to a Compute Engine machine type. [Learn
-        # about restrictions on accelerator configurations for
-        # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-        #
-        # Set `parameterServerConfig.imageUri` only if you build a custom image for
-        # your parameter server. If `parameterServerConfig.imageUri` has not been
-        # set, AI Platform uses the value of `masterConfig.imageUri`. Learn more about [configuring custom
-        # containers](/ai-platform/training/docs/distributed-training-containers).
-      "tpuTfVersion": "A String", # The AI Platform runtime version that includes a TensorFlow version matching
-          # the one used in the custom container. This field is required if the replica
-          # is a TPU worker that uses a custom container. Otherwise, do not specify
-          # this field. This must be a [runtime version that currently supports
-          # training with
-          # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
-          #
-          # Note that the version of TensorFlow included in a runtime version may
-          # differ from the numbering of the runtime version itself, because it may
-          # have a different [patch
-          # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
-          # In this field, you must specify the runtime version (TensorFlow minor
-          # version). For example, if your custom container runs TensorFlow `1.x.y`,
-          # specify `1.x`.
-      "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
-          # [Learn about restrictions on accelerator configurations for
-          # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-          # Note that the AcceleratorConfig can be used in both Jobs and Versions.
-          # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
-          # [accelerators for online
-          # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
-        "count": "A String", # The number of accelerators to attach to each machine running the job.
-        "type": "A String", # The type of accelerator to use.
-      },
-      "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
-          # Registry. Learn more about [configuring custom
-          # containers](/ai-platform/training/docs/distributed-training-containers).
-    },
-    "region": "A String", # Required. The region to run the training job in. See the [available
-        # regions](/ai-platform/training/docs/regions) for AI Platform Training.
-    "pythonVersion": "A String", # Optional. The version of Python used in training. You must either specify
-        # this field or specify `masterConfig.imageUri`.
-        #
-        # The following Python versions are available:
-        #
-        # * Python '3.7' is available when `runtime_version` is set to '1.15' or
-        #   later.
-        # * Python '3.5' is available when `runtime_version` is set to a version
-        #   from '1.4' to '1.14'.
-        # * Python '2.7' is available when `runtime_version` is set to '1.15' or
-        #   earlier.
-        #
-        # Read more about the Python versions available for [each runtime
-        # version](/ml-engine/docs/runtime-version-list).
-    "evaluatorType": "A String", # Optional. Specifies the type of virtual machine to use for your training
-        # job's evaluator nodes.
-        #
-        # The supported values are the same as those described in the entry for
-        # `masterType`.
-        #
-        # This value must be consistent with the category of machine type that
-        # `masterType` uses. In other words, both must be Compute Engine machine
-        # types or both must be legacy machine types.
-        #
-        # This value must be present when `scaleTier` is set to `CUSTOM` and
-        # `evaluatorCount` is greater than zero.
-    "parameterServerType": "A String", # Optional. Specifies the type of virtual machine to use for your training
-        # job's parameter server.
-        #
-        # The supported values are the same as those described in the entry for
-        # `master_type`.
-        #
-        # This value must be consistent with the category of machine type that
-        # `masterType` uses. In other words, both must be Compute Engine machine
-        # types or both must be legacy machine types.
-        #
-        # This value must be present when `scaleTier` is set to `CUSTOM` and
-        # `parameter_server_count` is greater than zero.
-  },
-  "endTime": "A String", # Output only. When the job processing was completed.
-  "predictionOutput": { # Represents results of a prediction job. # The current prediction job result.
-    "outputPath": "A String", # The output Google Cloud Storage location provided at the job creation time.
-    "nodeHours": 3.14, # Node hours used by the batch prediction job.
-    "predictionCount": "A String", # The number of generated predictions.
-    "errorCount": "A String", # The number of data instances which resulted in errors.
-  },
-  "createTime": "A String", # Output only. When the job was created.
-}
-
-  updateMask: string, Required. Specifies the path, relative to `Job`, of the field to update.
-To adopt etag mechanism, include `etag` field in the mask, and include the
-`etag` value in your job resource.
-
-For example, to change the labels of a job, the `update_mask` parameter
-would be specified as `labels`, `etag`, and the
-`PATCH` request body would specify the new value, as follows:
-    {
-      "labels": {
-         "owner": "Google",
-         "color": "Blue"
-      }
-      "etag": "33a64df551425fcc55e4d42a148795d9f25f89d4"
-    }
-If `etag` matches the one on the server, the labels of the job will be
-replaced with the given ones, and the server end `etag` will be
-recalculated.
-
-Currently the only supported update masks are `labels` and `etag`.
-  x__xgafv: string, V1 error format.
-    Allowed values
-      1 - v1 error format
-      2 - v2 error format
-
-Returns:
-  An object of the form:
-
-    { # Represents a training or prediction job.
-    "errorMessage": "A String", # Output only. The details of a failure or a cancellation.
-    "trainingOutput": { # Represents results of a training job. Output only. # The current training job result.
-      "completedTrialCount": "A String", # The number of hyperparameter tuning trials that completed successfully.
-          # Only set for hyperparameter tuning jobs.
-      "trials": [ # Results for individual Hyperparameter trials.
-          # Only set for hyperparameter tuning jobs.
-        { # Represents the result of a single hyperparameter tuning trial from a
-            # training job. The TrainingOutput object that is returned on successful
-            # completion of a training job with hyperparameter tuning includes a list
-            # of HyperparameterOutput objects, one for each successful trial.
-          "startTime": "A String", # Output only. Start time for the trial.
-          "hyperparameters": { # The hyperparameters given to this trial.
-            "a_key": "A String",
-          },
-          "finalMetric": { # An observed value of a metric. # The final objective metric seen for this trial.
-            "trainingStep": "A String", # The global training step for this metric.
-            "objectiveValue": 3.14, # The objective value at this training step.
-          },
-          "state": "A String", # Output only. The detailed state of the trial.
-          "allMetrics": [ # All recorded object metrics for this trial. This field is not currently
-              # populated.
-            { # An observed value of a metric.
-              "trainingStep": "A String", # The global training step for this metric.
-              "objectiveValue": 3.14, # The objective value at this training step.
-            },
-          ],
-          "isTrialStoppedEarly": True or False, # True if the trial is stopped early.
-          "endTime": "A String", # Output only. End time for the trial.
-          "trialId": "A String", # The trial id for these results.
-          "builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
-              # Only set for trials of built-in algorithms jobs that have succeeded.
-            "framework": "A String", # Framework on which the built-in algorithm was trained.
-            "modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job
-                # saves the trained model. Only set for successful jobs that don't use
-                # hyperparameter tuning.
-            "runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was
-                # trained.
-            "pythonVersion": "A String", # Python version on which the built-in algorithm was trained.
-          },
-        },
-      ],
-      "isHyperparameterTuningJob": True or False, # Whether this job is a hyperparameter tuning job.
-      "isBuiltInAlgorithmJob": True or False, # Whether this job is a built-in Algorithm job.
-      "consumedMLUnits": 3.14, # The amount of ML units consumed by the job.
-      "hyperparameterMetricTag": "A String", # The TensorFlow summary tag name used for optimizing hyperparameter tuning
-          # trials. See
-          # [`HyperparameterSpec.hyperparameterMetricTag`](#HyperparameterSpec.FIELDS.hyperparameter_metric_tag)
-          # for more information. Only set for hyperparameter tuning jobs.
-      "builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
-          # Only set for built-in algorithms jobs.
-        "framework": "A String", # Framework on which the built-in algorithm was trained.
-        "modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job
-            # saves the trained model. Only set for successful jobs that don't use
-            # hyperparameter tuning.
-        "runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was
-            # trained.
-        "pythonVersion": "A String", # Python version on which the built-in algorithm was trained.
-      },
-    },
-    "predictionInput": { # Represents input parameters for a prediction job. # Input parameters to create a prediction job.
-      "modelName": "A String", # Use this field if you want to use the default version for the specified
-          # model. The string must use the following format:
-          #
-          # `"projects/YOUR_PROJECT/models/YOUR_MODEL"`
-      "signatureName": "A String", # Optional. The name of the signature defined in the SavedModel to use for
-          # this job. Please refer to
-          # [SavedModel](https://tensorflow.github.io/serving/serving_basic.html)
-          # for information about how to use signatures.
-          #
-          # Defaults to
-          # [DEFAULT_SERVING_SIGNATURE_DEF_KEY](https://www.tensorflow.org/api_docs/python/tf/saved_model/signature_constants)
-          # , which is "serving_default".
-      "runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for this batch
-          # prediction. If not set, AI Platform will pick the runtime version used
-          # during the CreateVersion request for this model version, or choose the
-          # latest stable version when model version information is not available
-          # such as when the model is specified by uri.
-      "batchSize": "A String", # Optional. Number of records per batch, defaults to 64.
-          # The service will buffer batch_size number of records in memory before
-          # invoking one Tensorflow prediction call internally. So take the record
-          # size and memory available into consideration when setting this parameter.
-      "maxWorkerCount": "A String", # Optional. The maximum number of workers to be used for parallel processing.
-          # Defaults to 10 if not specified.
-      "uri": "A String", # Use this field if you want to specify a Google Cloud Storage path for
-          # the model to use.
-      "outputPath": "A String", # Required. The output Google Cloud Storage location.
-      "dataFormat": "A String", # Required. The format of the input data files.
-      "versionName": "A String", # Use this field if you want to specify a version of the model to use. The
-          # string is formatted the same way as `model_version`, with the addition
-          # of the version information:
-          #
-          # `"projects/YOUR_PROJECT/models/YOUR_MODEL/versions/YOUR_VERSION"`
-      "region": "A String", # Required. The Google Compute Engine region to run the prediction job in.
-          # See the &lt;a href="/ml-engine/docs/tensorflow/regions"&gt;available regions&lt;/a&gt;
-          # for AI Platform services.
-      "inputPaths": [ # Required. The Cloud Storage location of the input data files. May contain
-          # &lt;a href="/storage/docs/gsutil/addlhelp/WildcardNames"&gt;wildcards&lt;/a&gt;.
-        "A String",
-      ],
-      "outputDataFormat": "A String", # Optional. Format of the output data files, defaults to JSON.
-    },
-    "labels": { # Optional. One or more labels that you can add, to organize your jobs.
-        # Each label is a key-value pair, where both the key and the value are
-        # arbitrary strings that you supply.
-        # For more information, see the documentation on
-        # &lt;a href="/ml-engine/docs/tensorflow/resource-labels"&gt;using labels&lt;/a&gt;.
-      "a_key": "A String",
-    },
-    "jobId": "A String", # Required. The user-specified id of the job.
-    "state": "A String", # Output only. The detailed state of a job.
-    "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help
+    &quot;etag&quot;: &quot;A String&quot;, # `etag` is used for optimistic concurrency control as a way to help
         # prevent simultaneous updates of a job from overwriting each other.
         # It is strongly suggested that systems make use of the `etag` in the
         # read-modify-write cycle to perform job updates in order to avoid race
         # conditions: An `etag` is returned in the response to `GetJob`, and
         # systems are expected to put that etag in the request to `UpdateJob` to
         # ensure that their change will be applied to the same version of the job.
-    "startTime": "A String", # Output only. When the job processing was started.
-    "trainingInput": { # Represents input parameters for a training job. When using the gcloud command # Input parameters to create a training job.
+    &quot;trainingInput&quot;: { # Represents input parameters for a training job. When using the gcloud command # Input parameters to create a training job.
         # to submit your training job, you can specify the input parameters as
         # command-line arguments and/or in a YAML configuration file referenced from
         # the --config command-line argument. For details, see the guide to [submitting
         # a training job](/ai-platform/training/docs/training-jobs).
-      "masterType": "A String", # Optional. Specifies the type of virtual machine to use for your training
-          # job's master worker. You must specify this field when `scaleTier` is set to
+      &quot;parameterServerConfig&quot;: { # Represents the configuration for a replica in a cluster. # Optional. The configuration for parameter servers.
+          #
+          # You should only set `parameterServerConfig.acceleratorConfig` if
+          # `parameterServerType` is set to a Compute Engine machine type. [Learn
+          # about restrictions on accelerator configurations for
+          # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+          #
+          # Set `parameterServerConfig.imageUri` only if you build a custom image for
+          # your parameter server. If `parameterServerConfig.imageUri` has not been
+          # set, AI Platform uses the value of `masterConfig.imageUri`. Learn more about [configuring custom
+          # containers](/ai-platform/training/docs/distributed-training-containers).
+        &quot;tpuTfVersion&quot;: &quot;A String&quot;, # The AI Platform runtime version that includes a TensorFlow version matching
+            # the one used in the custom container. This field is required if the replica
+            # is a TPU worker that uses a custom container. Otherwise, do not specify
+            # this field. This must be a [runtime version that currently supports
+            # training with
+            # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
+            #
+            # Note that the version of TensorFlow included in a runtime version may
+            # differ from the numbering of the runtime version itself, because it may
+            # have a different [patch
+            # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
+            # In this field, you must specify the runtime version (TensorFlow minor
+            # version). For example, if your custom container runs TensorFlow `1.x.y`,
+            # specify `1.x`.
+        &quot;containerCommand&quot;: [ # The command with which the replica&#x27;s custom container is run.
+            # If provided, it will override default ENTRYPOINT of the docker image.
+            # If not provided, the docker image&#x27;s ENTRYPOINT is used.
+            # It cannot be set if custom container image is
+            # not provided.
+            # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+            # both cannot be set at the same time.
+          &quot;A String&quot;,
+        ],
+        &quot;imageUri&quot;: &quot;A String&quot;, # The Docker image to run on the replica. This image must be in Container
+            # Registry. Learn more about [configuring custom
+            # containers](/ai-platform/training/docs/distributed-training-containers).
+        &quot;containerArgs&quot;: [ # Arguments to the entrypoint command.
+            # The following rules apply for container_command and container_args:
+            # - If you do not supply command or args:
+            #   The defaults defined in the Docker image are used.
+            # - If you supply a command but no args:
+            #   The default EntryPoint and the default Cmd defined in the Docker image
+            #   are ignored. Your command is run without any arguments.
+            # - If you supply only args:
+            #   The default Entrypoint defined in the Docker image is run with the args
+            #   that you supplied.
+            # - If you supply a command and args:
+            #   The default Entrypoint and the default Cmd defined in the Docker image
+            #   are ignored. Your command is run with your args.
+            # It cannot be set if custom container image is
+            # not provided.
+            # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+            # both cannot be set at the same time.
+          &quot;A String&quot;,
+        ],
+        &quot;acceleratorConfig&quot;: { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
+            # [Learn about restrictions on accelerator configurations for
+            # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+            # Note that the AcceleratorConfig can be used in both Jobs and Versions.
+            # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
+            # [accelerators for online
+            # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
+          &quot;count&quot;: &quot;A String&quot;, # The number of accelerators to attach to each machine running the job.
+          &quot;type&quot;: &quot;A String&quot;, # The type of accelerator to use.
+        },
+      },
+      &quot;encryptionConfig&quot;: { # Represents a custom encryption key configuration that can be applied to # Optional. Options for using customer-managed encryption keys (CMEK) to
+          # protect resources created by a training job, instead of using Google&#x27;s
+          # default encryption. If this is set, then all resources created by the
+          # training job will be encrypted with the customer-managed encryption key
+          # that you specify.
+          #
+          # [Learn how and when to use CMEK with AI Platform
+          # Training](/ai-platform/training/docs/cmek).
+          # a resource.
+        &quot;kmsKeyName&quot;: &quot;A String&quot;, # The Cloud KMS resource identifier of the customer-managed encryption key
+            # used to protect a resource, such as a training job. It has the following
+            # format:
+            # `projects/{PROJECT_ID}/locations/{REGION}/keyRings/{KEY_RING_NAME}/cryptoKeys/{KEY_NAME}`
+      },
+      &quot;hyperparameters&quot;: { # Represents a set of hyperparameters to optimize. # Optional. The set of Hyperparameters to tune.
+        &quot;hyperparameterMetricTag&quot;: &quot;A String&quot;, # Optional. The TensorFlow summary tag name to use for optimizing trials. For
+            # current versions of TensorFlow, this tag name should exactly match what is
+            # shown in TensorBoard, including all scopes.  For versions of TensorFlow
+            # prior to 0.12, this should be only the tag passed to tf.Summary.
+            # By default, &quot;training/hptuning/metric&quot; will be used.
+        &quot;params&quot;: [ # Required. The set of parameters to tune.
+          { # Represents a single hyperparameter to optimize.
+            &quot;type&quot;: &quot;A String&quot;, # Required. The type of the parameter.
+            &quot;categoricalValues&quot;: [ # Required if type is `CATEGORICAL`. The list of possible categories.
+              &quot;A String&quot;,
+            ],
+            &quot;parameterName&quot;: &quot;A String&quot;, # Required. The parameter name must be unique amongst all ParameterConfigs in
+                # a HyperparameterSpec message. E.g., &quot;learning_rate&quot;.
+            &quot;minValue&quot;: 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
+                # should be unset if type is `CATEGORICAL`. This value should be integers if
+                # type is INTEGER.
+            &quot;discreteValues&quot;: [ # Required if type is `DISCRETE`.
+                # A list of feasible points.
+                # The list should be in strictly increasing order. For instance, this
+                # parameter might have possible settings of 1.5, 2.5, and 4.0. This list
+                # should not contain more than 1,000 values.
+              3.14,
+            ],
+            &quot;scaleType&quot;: &quot;A String&quot;, # Optional. How the parameter should be scaled to the hypercube.
+                # Leave unset for categorical parameters.
+                # Some kind of scaling is strongly recommended for real or integral
+                # parameters (e.g., `UNIT_LINEAR_SCALE`).
+            &quot;maxValue&quot;: 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
+                # should be unset if type is `CATEGORICAL`. This value should be integers if
+                # type is `INTEGER`.
+          },
+        ],
+        &quot;enableTrialEarlyStopping&quot;: True or False, # Optional. Indicates if the hyperparameter tuning job enables auto trial
+            # early stopping.
+        &quot;resumePreviousJobId&quot;: &quot;A String&quot;, # Optional. The prior hyperparameter tuning job id that users hope to
+            # continue with. The job id will be used to find the corresponding vizier
+            # study guid and resume the study.
+        &quot;maxParallelTrials&quot;: 42, # Optional. The number of training trials to run concurrently.
+            # You can reduce the time it takes to perform hyperparameter tuning by adding
+            # trials in parallel. However, each trail only benefits from the information
+            # gained in completed trials. That means that a trial does not get access to
+            # the results of trials running at the same time, which could reduce the
+            # quality of the overall optimization.
+            #
+            # Each trial will use the same scale tier and machine types.
+            #
+            # Defaults to one.
+        &quot;maxFailedTrials&quot;: 42, # Optional. The number of failed trials that need to be seen before failing
+            # the hyperparameter tuning job. You can specify this field to override the
+            # default failing criteria for AI Platform hyperparameter tuning jobs.
+            #
+            # Defaults to zero, which means the service decides when a hyperparameter
+            # job should fail.
+        &quot;goal&quot;: &quot;A String&quot;, # Required. The type of goal to use for tuning. Available types are
+            # `MAXIMIZE` and `MINIMIZE`.
+            #
+            # Defaults to `MAXIMIZE`.
+        &quot;maxTrials&quot;: 42, # Optional. How many training trials should be attempted to optimize
+            # the specified hyperparameters.
+            #
+            # Defaults to one.
+        &quot;algorithm&quot;: &quot;A String&quot;, # Optional. The search algorithm specified for the hyperparameter
+            # tuning job.
+            # Uses the default AI Platform hyperparameter tuning
+            # algorithm if unspecified.
+      },
+      &quot;workerConfig&quot;: { # Represents the configuration for a replica in a cluster. # Optional. The configuration for workers.
+          #
+          # You should only set `workerConfig.acceleratorConfig` if `workerType` is set
+          # to a Compute Engine machine type. [Learn about restrictions on accelerator
+          # configurations for
+          # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+          #
+          # Set `workerConfig.imageUri` only if you build a custom image for your
+          # worker. If `workerConfig.imageUri` has not been set, AI Platform uses
+          # the value of `masterConfig.imageUri`. Learn more about [configuring custom
+          # containers](/ai-platform/training/docs/distributed-training-containers).
+        &quot;tpuTfVersion&quot;: &quot;A String&quot;, # The AI Platform runtime version that includes a TensorFlow version matching
+            # the one used in the custom container. This field is required if the replica
+            # is a TPU worker that uses a custom container. Otherwise, do not specify
+            # this field. This must be a [runtime version that currently supports
+            # training with
+            # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
+            #
+            # Note that the version of TensorFlow included in a runtime version may
+            # differ from the numbering of the runtime version itself, because it may
+            # have a different [patch
+            # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
+            # In this field, you must specify the runtime version (TensorFlow minor
+            # version). For example, if your custom container runs TensorFlow `1.x.y`,
+            # specify `1.x`.
+        &quot;containerCommand&quot;: [ # The command with which the replica&#x27;s custom container is run.
+            # If provided, it will override default ENTRYPOINT of the docker image.
+            # If not provided, the docker image&#x27;s ENTRYPOINT is used.
+            # It cannot be set if custom container image is
+            # not provided.
+            # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+            # both cannot be set at the same time.
+          &quot;A String&quot;,
+        ],
+        &quot;imageUri&quot;: &quot;A String&quot;, # The Docker image to run on the replica. This image must be in Container
+            # Registry. Learn more about [configuring custom
+            # containers](/ai-platform/training/docs/distributed-training-containers).
+        &quot;containerArgs&quot;: [ # Arguments to the entrypoint command.
+            # The following rules apply for container_command and container_args:
+            # - If you do not supply command or args:
+            #   The defaults defined in the Docker image are used.
+            # - If you supply a command but no args:
+            #   The default EntryPoint and the default Cmd defined in the Docker image
+            #   are ignored. Your command is run without any arguments.
+            # - If you supply only args:
+            #   The default Entrypoint defined in the Docker image is run with the args
+            #   that you supplied.
+            # - If you supply a command and args:
+            #   The default Entrypoint and the default Cmd defined in the Docker image
+            #   are ignored. Your command is run with your args.
+            # It cannot be set if custom container image is
+            # not provided.
+            # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+            # both cannot be set at the same time.
+          &quot;A String&quot;,
+        ],
+        &quot;acceleratorConfig&quot;: { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
+            # [Learn about restrictions on accelerator configurations for
+            # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+            # Note that the AcceleratorConfig can be used in both Jobs and Versions.
+            # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
+            # [accelerators for online
+            # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
+          &quot;count&quot;: &quot;A String&quot;, # The number of accelerators to attach to each machine running the job.
+          &quot;type&quot;: &quot;A String&quot;, # The type of accelerator to use.
+        },
+      },
+      &quot;parameterServerCount&quot;: &quot;A String&quot;, # Optional. The number of parameter server replicas to use for the training
+          # job. Each replica in the cluster will be of the type specified in
+          # `parameter_server_type`.
+          #
+          # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
+          # set this value, you must also set `parameter_server_type`.
+          #
+          # The default value is zero.
+      &quot;packageUris&quot;: [ # Required. The Google Cloud Storage location of the packages with
+          # the training program and any additional dependencies.
+          # The maximum number of package URIs is 100.
+        &quot;A String&quot;,
+      ],
+      &quot;evaluatorCount&quot;: &quot;A String&quot;, # Optional. The number of evaluator replicas to use for the training job.
+          # Each replica in the cluster will be of the type specified in
+          # `evaluator_type`.
+          #
+          # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
+          # set this value, you must also set `evaluator_type`.
+          #
+          # The default value is zero.
+      &quot;masterType&quot;: &quot;A String&quot;, # Optional. Specifies the type of virtual machine to use for your training
+          # job&#x27;s master worker. You must specify this field when `scaleTier` is set to
           # `CUSTOM`.
           #
           # You can use certain Compute Engine machine types directly in this field.
@@ -3543,14 +3592,156 @@
           # field. Learn more about the [special configuration options for training
           # with
           # TPUs](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine).
-      "jobDir": "A String", # Optional. A Google Cloud Storage path in which to store training outputs
+      &quot;runtimeVersion&quot;: &quot;A String&quot;, # Optional. The AI Platform runtime version to use for training. You must
+          # either specify this field or specify `masterConfig.imageUri`.
+          #
+          # For more information, see the [runtime version
+          # list](/ai-platform/training/docs/runtime-version-list) and learn [how to
+          # manage runtime versions](/ai-platform/training/docs/versioning).
+      &quot;evaluatorType&quot;: &quot;A String&quot;, # Optional. Specifies the type of virtual machine to use for your training
+          # job&#x27;s evaluator nodes.
+          #
+          # The supported values are the same as those described in the entry for
+          # `masterType`.
+          #
+          # This value must be consistent with the category of machine type that
+          # `masterType` uses. In other words, both must be Compute Engine machine
+          # types or both must be legacy machine types.
+          #
+          # This value must be present when `scaleTier` is set to `CUSTOM` and
+          # `evaluatorCount` is greater than zero.
+      &quot;region&quot;: &quot;A String&quot;, # Required. The region to run the training job in. See the [available
+          # regions](/ai-platform/training/docs/regions) for AI Platform Training.
+      &quot;workerType&quot;: &quot;A String&quot;, # Optional. Specifies the type of virtual machine to use for your training
+          # job&#x27;s worker nodes.
+          #
+          # The supported values are the same as those described in the entry for
+          # `masterType`.
+          #
+          # This value must be consistent with the category of machine type that
+          # `masterType` uses. In other words, both must be Compute Engine machine
+          # types or both must be legacy machine types.
+          #
+          # If you use `cloud_tpu` for this value, see special instructions for
+          # [configuring a custom TPU
+          # machine](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine).
+          #
+          # This value must be present when `scaleTier` is set to `CUSTOM` and
+          # `workerCount` is greater than zero.
+      &quot;parameterServerType&quot;: &quot;A String&quot;, # Optional. Specifies the type of virtual machine to use for your training
+          # job&#x27;s parameter server.
+          #
+          # The supported values are the same as those described in the entry for
+          # `master_type`.
+          #
+          # This value must be consistent with the category of machine type that
+          # `masterType` uses. In other words, both must be Compute Engine machine
+          # types or both must be legacy machine types.
+          #
+          # This value must be present when `scaleTier` is set to `CUSTOM` and
+          # `parameter_server_count` is greater than zero.
+      &quot;masterConfig&quot;: { # Represents the configuration for a replica in a cluster. # Optional. The configuration for your master worker.
+          #
+          # You should only set `masterConfig.acceleratorConfig` if `masterType` is set
+          # to a Compute Engine machine type. Learn about [restrictions on accelerator
+          # configurations for
+          # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+          #
+          # Set `masterConfig.imageUri` only if you build a custom image. Only one of
+          # `masterConfig.imageUri` and `runtimeVersion` should be set. Learn more
+          # about [configuring custom
+          # containers](/ai-platform/training/docs/distributed-training-containers).
+        &quot;tpuTfVersion&quot;: &quot;A String&quot;, # The AI Platform runtime version that includes a TensorFlow version matching
+            # the one used in the custom container. This field is required if the replica
+            # is a TPU worker that uses a custom container. Otherwise, do not specify
+            # this field. This must be a [runtime version that currently supports
+            # training with
+            # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
+            #
+            # Note that the version of TensorFlow included in a runtime version may
+            # differ from the numbering of the runtime version itself, because it may
+            # have a different [patch
+            # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
+            # In this field, you must specify the runtime version (TensorFlow minor
+            # version). For example, if your custom container runs TensorFlow `1.x.y`,
+            # specify `1.x`.
+        &quot;containerCommand&quot;: [ # The command with which the replica&#x27;s custom container is run.
+            # If provided, it will override default ENTRYPOINT of the docker image.
+            # If not provided, the docker image&#x27;s ENTRYPOINT is used.
+            # It cannot be set if custom container image is
+            # not provided.
+            # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+            # both cannot be set at the same time.
+          &quot;A String&quot;,
+        ],
+        &quot;imageUri&quot;: &quot;A String&quot;, # The Docker image to run on the replica. This image must be in Container
+            # Registry. Learn more about [configuring custom
+            # containers](/ai-platform/training/docs/distributed-training-containers).
+        &quot;containerArgs&quot;: [ # Arguments to the entrypoint command.
+            # The following rules apply for container_command and container_args:
+            # - If you do not supply command or args:
+            #   The defaults defined in the Docker image are used.
+            # - If you supply a command but no args:
+            #   The default EntryPoint and the default Cmd defined in the Docker image
+            #   are ignored. Your command is run without any arguments.
+            # - If you supply only args:
+            #   The default Entrypoint defined in the Docker image is run with the args
+            #   that you supplied.
+            # - If you supply a command and args:
+            #   The default Entrypoint and the default Cmd defined in the Docker image
+            #   are ignored. Your command is run with your args.
+            # It cannot be set if custom container image is
+            # not provided.
+            # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+            # both cannot be set at the same time.
+          &quot;A String&quot;,
+        ],
+        &quot;acceleratorConfig&quot;: { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
+            # [Learn about restrictions on accelerator configurations for
+            # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+            # Note that the AcceleratorConfig can be used in both Jobs and Versions.
+            # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
+            # [accelerators for online
+            # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
+          &quot;count&quot;: &quot;A String&quot;, # The number of accelerators to attach to each machine running the job.
+          &quot;type&quot;: &quot;A String&quot;, # The type of accelerator to use.
+        },
+      },
+      &quot;scaleTier&quot;: &quot;A String&quot;, # Required. Specifies the machine types, the number of replicas for workers
+          # and parameter servers.
+      &quot;jobDir&quot;: &quot;A String&quot;, # Optional. A Google Cloud Storage path in which to store training outputs
           # and other data needed for training. This path is passed to your TensorFlow
-          # program as the '--job-dir' command-line argument. The benefit of specifying
+          # program as the &#x27;--job-dir&#x27; command-line argument. The benefit of specifying
           # this field is that Cloud ML validates the path for use in training.
-      "scheduling": { # All parameters related to scheduling of training jobs. # Optional. Scheduling options for a training job.
-        "maxRunningTime": "A String", # Optional. The maximum job running time, expressed in seconds. The field can
-            # contain up to nine fractional digits, terminated by `s`. By default there
-            # is no limit to the running time.
+      &quot;pythonVersion&quot;: &quot;A String&quot;, # Optional. The version of Python used in training. You must either specify
+          # this field or specify `masterConfig.imageUri`.
+          #
+          # The following Python versions are available:
+          #
+          # * Python &#x27;3.7&#x27; is available when `runtime_version` is set to &#x27;1.15&#x27; or
+          #   later.
+          # * Python &#x27;3.5&#x27; is available when `runtime_version` is set to a version
+          #   from &#x27;1.4&#x27; to &#x27;1.14&#x27;.
+          # * Python &#x27;2.7&#x27; is available when `runtime_version` is set to &#x27;1.15&#x27; or
+          #   earlier.
+          #
+          # Read more about the Python versions available for [each runtime
+          # version](/ml-engine/docs/runtime-version-list).
+      &quot;network&quot;: &quot;A String&quot;, # Optional. The full name of the Google Compute Engine
+          # [network](/compute/docs/networks-and-firewalls#networks) to which the Job
+          # is peered. For example, projects/12345/global/networks/myVPC. Format is of
+          # the form projects/{project}/global/networks/{network}. Where {project} is a
+          # project number, as in &#x27;12345&#x27;, and {network} is network name.&quot;.
+          #
+          # Private services access must already be configured for the network. If left
+          # unspecified, the Job is not peered with any network. Learn more -
+          # Connecting Job to user network over private
+          # IP.
+      &quot;scheduling&quot;: { # All parameters related to scheduling of training jobs. # Optional. Scheduling options for a training job.
+        &quot;maxWaitTime&quot;: &quot;A String&quot;,
+        &quot;maxRunningTime&quot;: &quot;A String&quot;, # Optional. The maximum job running time, expressed in seconds. The field can
+            # contain up to nine fractional digits, terminated by `s`. If not specified,
+            # this field defaults to `604800s` (seven days).
             #
             # If the training job is still running after this duration, AI Platform
             # Training cancels it.
@@ -3572,85 +3763,7 @@
             #   ...
             # ```
       },
-      "parameterServerCount": "A String", # Optional. The number of parameter server replicas to use for the training
-          # job. Each replica in the cluster will be of the type specified in
-          # `parameter_server_type`.
-          #
-          # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
-          # set this value, you must also set `parameter_server_type`.
-          #
-          # The default value is zero.
-      "evaluatorCount": "A String", # Optional. The number of evaluator replicas to use for the training job.
-          # Each replica in the cluster will be of the type specified in
-          # `evaluator_type`.
-          #
-          # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
-          # set this value, you must also set `evaluator_type`.
-          #
-          # The default value is zero.
-      "workerType": "A String", # Optional. Specifies the type of virtual machine to use for your training
-          # job's worker nodes.
-          #
-          # The supported values are the same as those described in the entry for
-          # `masterType`.
-          #
-          # This value must be consistent with the category of machine type that
-          # `masterType` uses. In other words, both must be Compute Engine machine
-          # types or both must be legacy machine types.
-          #
-          # If you use `cloud_tpu` for this value, see special instructions for
-          # [configuring a custom TPU
-          # machine](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine).
-          #
-          # This value must be present when `scaleTier` is set to `CUSTOM` and
-          # `workerCount` is greater than zero.
-      "scaleTier": "A String", # Required. Specifies the machine types, the number of replicas for workers
-          # and parameter servers.
-      "packageUris": [ # Required. The Google Cloud Storage location of the packages with
-          # the training program and any additional dependencies.
-          # The maximum number of package URIs is 100.
-        "A String",
-      ],
-      "workerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for workers.
-          #
-          # You should only set `workerConfig.acceleratorConfig` if `workerType` is set
-          # to a Compute Engine machine type. [Learn about restrictions on accelerator
-          # configurations for
-          # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-          #
-          # Set `workerConfig.imageUri` only if you build a custom image for your
-          # worker. If `workerConfig.imageUri` has not been set, AI Platform uses
-          # the value of `masterConfig.imageUri`. Learn more about [configuring custom
-          # containers](/ai-platform/training/docs/distributed-training-containers).
-        "tpuTfVersion": "A String", # The AI Platform runtime version that includes a TensorFlow version matching
-            # the one used in the custom container. This field is required if the replica
-            # is a TPU worker that uses a custom container. Otherwise, do not specify
-            # this field. This must be a [runtime version that currently supports
-            # training with
-            # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
-            #
-            # Note that the version of TensorFlow included in a runtime version may
-            # differ from the numbering of the runtime version itself, because it may
-            # have a different [patch
-            # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
-            # In this field, you must specify the runtime version (TensorFlow minor
-            # version). For example, if your custom container runs TensorFlow `1.x.y`,
-            # specify `1.x`.
-        "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
-            # [Learn about restrictions on accelerator configurations for
-            # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-            # Note that the AcceleratorConfig can be used in both Jobs and Versions.
-            # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
-            # [accelerators for online
-            # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
-          "count": "A String", # The number of accelerators to attach to each machine running the job.
-          "type": "A String", # The type of accelerator to use.
-        },
-        "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
-            # Registry. Learn more about [configuring custom
-            # containers](/ai-platform/training/docs/distributed-training-containers).
-      },
-      "evaluatorConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for evaluators.
+      &quot;evaluatorConfig&quot;: { # Represents the configuration for a replica in a cluster. # Optional. The configuration for evaluators.
           #
           # You should only set `evaluatorConfig.acceleratorConfig` if
           # `evaluatorType` is set to a Compute Engine machine type. [Learn
@@ -3661,7 +3774,7 @@
           # your evaluator. If `evaluatorConfig.imageUri` has not been
           # set, AI Platform uses the value of `masterConfig.imageUri`. Learn more about [configuring custom
           # containers](/ai-platform/training/docs/distributed-training-containers).
-        "tpuTfVersion": "A String", # The AI Platform runtime version that includes a TensorFlow version matching
+        &quot;tpuTfVersion&quot;: &quot;A String&quot;, # The AI Platform runtime version that includes a TensorFlow version matching
             # the one used in the custom container. This field is required if the replica
             # is a TPU worker that uses a custom container. Otherwise, do not specify
             # this field. This must be a [runtime version that currently supports
@@ -3675,257 +3788,901 @@
             # In this field, you must specify the runtime version (TensorFlow minor
             # version). For example, if your custom container runs TensorFlow `1.x.y`,
             # specify `1.x`.
-        "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
+        &quot;containerCommand&quot;: [ # The command with which the replica&#x27;s custom container is run.
+            # If provided, it will override default ENTRYPOINT of the docker image.
+            # If not provided, the docker image&#x27;s ENTRYPOINT is used.
+            # It cannot be set if custom container image is
+            # not provided.
+            # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+            # both cannot be set at the same time.
+          &quot;A String&quot;,
+        ],
+        &quot;imageUri&quot;: &quot;A String&quot;, # The Docker image to run on the replica. This image must be in Container
+            # Registry. Learn more about [configuring custom
+            # containers](/ai-platform/training/docs/distributed-training-containers).
+        &quot;containerArgs&quot;: [ # Arguments to the entrypoint command.
+            # The following rules apply for container_command and container_args:
+            # - If you do not supply command or args:
+            #   The defaults defined in the Docker image are used.
+            # - If you supply a command but no args:
+            #   The default EntryPoint and the default Cmd defined in the Docker image
+            #   are ignored. Your command is run without any arguments.
+            # - If you supply only args:
+            #   The default Entrypoint defined in the Docker image is run with the args
+            #   that you supplied.
+            # - If you supply a command and args:
+            #   The default Entrypoint and the default Cmd defined in the Docker image
+            #   are ignored. Your command is run with your args.
+            # It cannot be set if custom container image is
+            # not provided.
+            # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+            # both cannot be set at the same time.
+          &quot;A String&quot;,
+        ],
+        &quot;acceleratorConfig&quot;: { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
             # [Learn about restrictions on accelerator configurations for
             # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
             # Note that the AcceleratorConfig can be used in both Jobs and Versions.
             # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
             # [accelerators for online
             # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
-          "count": "A String", # The number of accelerators to attach to each machine running the job.
-          "type": "A String", # The type of accelerator to use.
+          &quot;count&quot;: &quot;A String&quot;, # The number of accelerators to attach to each machine running the job.
+          &quot;type&quot;: &quot;A String&quot;, # The type of accelerator to use.
         },
-        "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
-            # Registry. Learn more about [configuring custom
-            # containers](/ai-platform/training/docs/distributed-training-containers).
       },
-      "useChiefInTfConfig": True or False, # Optional. Use `chief` instead of `master` in the `TF_CONFIG` environment
+      &quot;useChiefInTfConfig&quot;: True or False, # Optional. Use `chief` instead of `master` in the `TF_CONFIG` environment
           # variable when training with a custom container. Defaults to `false`. [Learn
           # more about this
           # field.](/ai-platform/training/docs/distributed-training-details#chief-versus-master)
           #
-          # This field has no effect for training jobs that don't use a custom
+          # This field has no effect for training jobs that don&#x27;t use a custom
           # container.
-      "masterConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for your master worker.
-          #
-          # You should only set `masterConfig.acceleratorConfig` if `masterType` is set
-          # to a Compute Engine machine type. Learn about [restrictions on accelerator
-          # configurations for
-          # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-          #
-          # Set `masterConfig.imageUri` only if you build a custom image. Only one of
-          # `masterConfig.imageUri` and `runtimeVersion` should be set. Learn more
-          # about [configuring custom
-          # containers](/ai-platform/training/docs/distributed-training-containers).
-        "tpuTfVersion": "A String", # The AI Platform runtime version that includes a TensorFlow version matching
-            # the one used in the custom container. This field is required if the replica
-            # is a TPU worker that uses a custom container. Otherwise, do not specify
-            # this field. This must be a [runtime version that currently supports
-            # training with
-            # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
-            #
-            # Note that the version of TensorFlow included in a runtime version may
-            # differ from the numbering of the runtime version itself, because it may
-            # have a different [patch
-            # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
-            # In this field, you must specify the runtime version (TensorFlow minor
-            # version). For example, if your custom container runs TensorFlow `1.x.y`,
-            # specify `1.x`.
-        "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
-            # [Learn about restrictions on accelerator configurations for
-            # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-            # Note that the AcceleratorConfig can be used in both Jobs and Versions.
-            # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
-            # [accelerators for online
-            # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
-          "count": "A String", # The number of accelerators to attach to each machine running the job.
-          "type": "A String", # The type of accelerator to use.
-        },
-        "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
-            # Registry. Learn more about [configuring custom
-            # containers](/ai-platform/training/docs/distributed-training-containers).
-      },
-      "runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for training. You must
-          # either specify this field or specify `masterConfig.imageUri`.
-          #
-          # For more information, see the [runtime version
-          # list](/ai-platform/training/docs/runtime-version-list) and learn [how to
-          # manage runtime versions](/ai-platform/training/docs/versioning).
-      "hyperparameters": { # Represents a set of hyperparameters to optimize. # Optional. The set of Hyperparameters to tune.
-        "maxTrials": 42, # Optional. How many training trials should be attempted to optimize
-            # the specified hyperparameters.
-            #
-            # Defaults to one.
-        "goal": "A String", # Required. The type of goal to use for tuning. Available types are
-            # `MAXIMIZE` and `MINIMIZE`.
-            #
-            # Defaults to `MAXIMIZE`.
-        "algorithm": "A String", # Optional. The search algorithm specified for the hyperparameter
-            # tuning job.
-            # Uses the default AI Platform hyperparameter tuning
-            # algorithm if unspecified.
-        "maxFailedTrials": 42, # Optional. The number of failed trials that need to be seen before failing
-            # the hyperparameter tuning job. You can specify this field to override the
-            # default failing criteria for AI Platform hyperparameter tuning jobs.
-            #
-            # Defaults to zero, which means the service decides when a hyperparameter
-            # job should fail.
-        "enableTrialEarlyStopping": True or False, # Optional. Indicates if the hyperparameter tuning job enables auto trial
-            # early stopping.
-        "resumePreviousJobId": "A String", # Optional. The prior hyperparameter tuning job id that users hope to
-            # continue with. The job id will be used to find the corresponding vizier
-            # study guid and resume the study.
-        "params": [ # Required. The set of parameters to tune.
-          { # Represents a single hyperparameter to optimize.
-            "maxValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
-                # should be unset if type is `CATEGORICAL`. This value should be integers if
-                # type is `INTEGER`.
-            "minValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
-                # should be unset if type is `CATEGORICAL`. This value should be integers if
-                # type is INTEGER.
-            "discreteValues": [ # Required if type is `DISCRETE`.
-                # A list of feasible points.
-                # The list should be in strictly increasing order. For instance, this
-                # parameter might have possible settings of 1.5, 2.5, and 4.0. This list
-                # should not contain more than 1,000 values.
-              3.14,
-            ],
-            "parameterName": "A String", # Required. The parameter name must be unique amongst all ParameterConfigs in
-                # a HyperparameterSpec message. E.g., "learning_rate".
-            "categoricalValues": [ # Required if type is `CATEGORICAL`. The list of possible categories.
-              "A String",
-            ],
-            "type": "A String", # Required. The type of the parameter.
-            "scaleType": "A String", # Optional. How the parameter should be scaled to the hypercube.
-                # Leave unset for categorical parameters.
-                # Some kind of scaling is strongly recommended for real or integral
-                # parameters (e.g., `UNIT_LINEAR_SCALE`).
-          },
-        ],
-        "hyperparameterMetricTag": "A String", # Optional. The TensorFlow summary tag name to use for optimizing trials. For
-            # current versions of TensorFlow, this tag name should exactly match what is
-            # shown in TensorBoard, including all scopes.  For versions of TensorFlow
-            # prior to 0.12, this should be only the tag passed to tf.Summary.
-            # By default, "training/hptuning/metric" will be used.
-        "maxParallelTrials": 42, # Optional. The number of training trials to run concurrently.
-            # You can reduce the time it takes to perform hyperparameter tuning by adding
-            # trials in parallel. However, each trail only benefits from the information
-            # gained in completed trials. That means that a trial does not get access to
-            # the results of trials running at the same time, which could reduce the
-            # quality of the overall optimization.
-            #
-            # Each trial will use the same scale tier and machine types.
-            #
-            # Defaults to one.
-      },
-      "args": [ # Optional. Command-line arguments passed to the training application when it
-          # starts. If your job uses a custom container, then the arguments are passed
-          # to the container's &lt;a class="external" target="_blank"
-          # href="https://docs.docker.com/engine/reference/builder/#entrypoint"&gt;
-          # `ENTRYPOINT`&lt;/a&gt; command.
-        "A String",
-      ],
-      "pythonModule": "A String", # Required. The Python module name to run after installing the packages.
-      "workerCount": "A String", # Optional. The number of worker replicas to use for the training job. Each
+      &quot;workerCount&quot;: &quot;A String&quot;, # Optional. The number of worker replicas to use for the training job. Each
           # replica in the cluster will be of the type specified in `worker_type`.
           #
           # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
           # set this value, you must also set `worker_type`.
           #
           # The default value is zero.
-      "encryptionConfig": { # Represents a custom encryption key configuration that can be applied to # Optional. Options for using customer-managed encryption keys (CMEK) to
-          # protect resources created by a training job, instead of using Google's
-          # default encryption. If this is set, then all resources created by the
-          # training job will be encrypted with the customer-managed encryption key
-          # that you specify.
-          #
-          # [Learn how and when to use CMEK with AI Platform
-          # Training](/ai-platform/training/docs/cmek).
-          # a resource.
-        "kmsKeyName": "A String", # The Cloud KMS resource identifier of the customer-managed encryption key
-            # used to protect a resource, such as a training job. It has the following
-            # format:
-            # `projects/{PROJECT_ID}/locations/{REGION}/keyRings/{KEY_RING_NAME}/cryptoKeys/{KEY_NAME}`
-      },
-      "parameterServerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for parameter servers.
-          #
-          # You should only set `parameterServerConfig.acceleratorConfig` if
-          # `parameterServerType` is set to a Compute Engine machine type. [Learn
-          # about restrictions on accelerator configurations for
-          # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-          #
-          # Set `parameterServerConfig.imageUri` only if you build a custom image for
-          # your parameter server. If `parameterServerConfig.imageUri` has not been
-          # set, AI Platform uses the value of `masterConfig.imageUri`. Learn more about [configuring custom
-          # containers](/ai-platform/training/docs/distributed-training-containers).
-        "tpuTfVersion": "A String", # The AI Platform runtime version that includes a TensorFlow version matching
-            # the one used in the custom container. This field is required if the replica
-            # is a TPU worker that uses a custom container. Otherwise, do not specify
-            # this field. This must be a [runtime version that currently supports
-            # training with
-            # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
-            #
-            # Note that the version of TensorFlow included in a runtime version may
-            # differ from the numbering of the runtime version itself, because it may
-            # have a different [patch
-            # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
-            # In this field, you must specify the runtime version (TensorFlow minor
-            # version). For example, if your custom container runs TensorFlow `1.x.y`,
-            # specify `1.x`.
-        "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
-            # [Learn about restrictions on accelerator configurations for
-            # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
-            # Note that the AcceleratorConfig can be used in both Jobs and Versions.
-            # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
-            # [accelerators for online
-            # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
-          "count": "A String", # The number of accelerators to attach to each machine running the job.
-          "type": "A String", # The type of accelerator to use.
+      &quot;pythonModule&quot;: &quot;A String&quot;, # Required. The Python module name to run after installing the packages.
+      &quot;args&quot;: [ # Optional. Command-line arguments passed to the training application when it
+          # starts. If your job uses a custom container, then the arguments are passed
+          # to the container&#x27;s &lt;a class=&quot;external&quot; target=&quot;_blank&quot;
+          # href=&quot;https://docs.docker.com/engine/reference/builder/#entrypoint&quot;&gt;
+          # `ENTRYPOINT`&lt;/a&gt; command.
+        &quot;A String&quot;,
+      ],
+    },
+    &quot;state&quot;: &quot;A String&quot;, # Output only. The detailed state of a job.
+    &quot;jobId&quot;: &quot;A String&quot;, # Required. The user-specified id of the job.
+    &quot;endTime&quot;: &quot;A String&quot;, # Output only. When the job processing was completed.
+    &quot;startTime&quot;: &quot;A String&quot;, # Output only. When the job processing was started.
+    &quot;predictionOutput&quot;: { # Represents results of a prediction job. # The current prediction job result.
+      &quot;errorCount&quot;: &quot;A String&quot;, # The number of data instances which resulted in errors.
+      &quot;outputPath&quot;: &quot;A String&quot;, # The output Google Cloud Storage location provided at the job creation time.
+      &quot;nodeHours&quot;: 3.14, # Node hours used by the batch prediction job.
+      &quot;predictionCount&quot;: &quot;A String&quot;, # The number of generated predictions.
+    },
+    &quot;trainingOutput&quot;: { # Represents results of a training job. Output only. # The current training job result.
+      &quot;trials&quot;: [ # Results for individual Hyperparameter trials.
+          # Only set for hyperparameter tuning jobs.
+        { # Represents the result of a single hyperparameter tuning trial from a
+            # training job. The TrainingOutput object that is returned on successful
+            # completion of a training job with hyperparameter tuning includes a list
+            # of HyperparameterOutput objects, one for each successful trial.
+          &quot;allMetrics&quot;: [ # All recorded object metrics for this trial. This field is not currently
+              # populated.
+            { # An observed value of a metric.
+              &quot;objectiveValue&quot;: 3.14, # The objective value at this training step.
+              &quot;trainingStep&quot;: &quot;A String&quot;, # The global training step for this metric.
+            },
+          ],
+          &quot;hyperparameters&quot;: { # The hyperparameters given to this trial.
+            &quot;a_key&quot;: &quot;A String&quot;,
+          },
+          &quot;trialId&quot;: &quot;A String&quot;, # The trial id for these results.
+          &quot;endTime&quot;: &quot;A String&quot;, # Output only. End time for the trial.
+          &quot;isTrialStoppedEarly&quot;: True or False, # True if the trial is stopped early.
+          &quot;startTime&quot;: &quot;A String&quot;, # Output only. Start time for the trial.
+          &quot;finalMetric&quot;: { # An observed value of a metric. # The final objective metric seen for this trial.
+            &quot;objectiveValue&quot;: 3.14, # The objective value at this training step.
+            &quot;trainingStep&quot;: &quot;A String&quot;, # The global training step for this metric.
+          },
+          &quot;builtInAlgorithmOutput&quot;: { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
+              # Only set for trials of built-in algorithms jobs that have succeeded.
+            &quot;framework&quot;: &quot;A String&quot;, # Framework on which the built-in algorithm was trained.
+            &quot;modelPath&quot;: &quot;A String&quot;, # The Cloud Storage path to the `model/` directory where the training job
+                # saves the trained model. Only set for successful jobs that don&#x27;t use
+                # hyperparameter tuning.
+            &quot;pythonVersion&quot;: &quot;A String&quot;, # Python version on which the built-in algorithm was trained.
+            &quot;runtimeVersion&quot;: &quot;A String&quot;, # AI Platform runtime version on which the built-in algorithm was
+                # trained.
+          },
+          &quot;state&quot;: &quot;A String&quot;, # Output only. The detailed state of the trial.
         },
-        "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
-            # Registry. Learn more about [configuring custom
-            # containers](/ai-platform/training/docs/distributed-training-containers).
+      ],
+      &quot;hyperparameterMetricTag&quot;: &quot;A String&quot;, # The TensorFlow summary tag name used for optimizing hyperparameter tuning
+          # trials. See
+          # [`HyperparameterSpec.hyperparameterMetricTag`](#HyperparameterSpec.FIELDS.hyperparameter_metric_tag)
+          # for more information. Only set for hyperparameter tuning jobs.
+      &quot;completedTrialCount&quot;: &quot;A String&quot;, # The number of hyperparameter tuning trials that completed successfully.
+          # Only set for hyperparameter tuning jobs.
+      &quot;isHyperparameterTuningJob&quot;: True or False, # Whether this job is a hyperparameter tuning job.
+      &quot;consumedMLUnits&quot;: 3.14, # The amount of ML units consumed by the job.
+      &quot;isBuiltInAlgorithmJob&quot;: True or False, # Whether this job is a built-in Algorithm job.
+      &quot;builtInAlgorithmOutput&quot;: { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
+          # Only set for built-in algorithms jobs.
+        &quot;framework&quot;: &quot;A String&quot;, # Framework on which the built-in algorithm was trained.
+        &quot;modelPath&quot;: &quot;A String&quot;, # The Cloud Storage path to the `model/` directory where the training job
+            # saves the trained model. Only set for successful jobs that don&#x27;t use
+            # hyperparameter tuning.
+        &quot;pythonVersion&quot;: &quot;A String&quot;, # Python version on which the built-in algorithm was trained.
+        &quot;runtimeVersion&quot;: &quot;A String&quot;, # AI Platform runtime version on which the built-in algorithm was
+            # trained.
       },
-      "region": "A String", # Required. The region to run the training job in. See the [available
-          # regions](/ai-platform/training/docs/regions) for AI Platform Training.
-      "pythonVersion": "A String", # Optional. The version of Python used in training. You must either specify
-          # this field or specify `masterConfig.imageUri`.
-          #
-          # The following Python versions are available:
-          #
-          # * Python '3.7' is available when `runtime_version` is set to '1.15' or
-          #   later.
-          # * Python '3.5' is available when `runtime_version` is set to a version
-          #   from '1.4' to '1.14'.
-          # * Python '2.7' is available when `runtime_version` is set to '1.15' or
-          #   earlier.
-          #
-          # Read more about the Python versions available for [each runtime
-          # version](/ml-engine/docs/runtime-version-list).
-      "evaluatorType": "A String", # Optional. Specifies the type of virtual machine to use for your training
-          # job's evaluator nodes.
-          #
-          # The supported values are the same as those described in the entry for
-          # `masterType`.
-          #
-          # This value must be consistent with the category of machine type that
-          # `masterType` uses. In other words, both must be Compute Engine machine
-          # types or both must be legacy machine types.
-          #
-          # This value must be present when `scaleTier` is set to `CUSTOM` and
-          # `evaluatorCount` is greater than zero.
-      "parameterServerType": "A String", # Optional. Specifies the type of virtual machine to use for your training
-          # job's parameter server.
-          #
-          # The supported values are the same as those described in the entry for
-          # `master_type`.
-          #
-          # This value must be consistent with the category of machine type that
-          # `masterType` uses. In other words, both must be Compute Engine machine
-          # types or both must be legacy machine types.
-          #
-          # This value must be present when `scaleTier` is set to `CUSTOM` and
-          # `parameter_server_count` is greater than zero.
     },
-    "endTime": "A String", # Output only. When the job processing was completed.
-    "predictionOutput": { # Represents results of a prediction job. # The current prediction job result.
-      "outputPath": "A String", # The output Google Cloud Storage location provided at the job creation time.
-      "nodeHours": 3.14, # Node hours used by the batch prediction job.
-      "predictionCount": "A String", # The number of generated predictions.
-      "errorCount": "A String", # The number of data instances which resulted in errors.
+    &quot;createTime&quot;: &quot;A String&quot;, # Output only. When the job was created.
+    &quot;labels&quot;: { # Optional. One or more labels that you can add, to organize your jobs.
+        # Each label is a key-value pair, where both the key and the value are
+        # arbitrary strings that you supply.
+        # For more information, see the documentation on
+        # &lt;a href=&quot;/ml-engine/docs/tensorflow/resource-labels&quot;&gt;using labels&lt;/a&gt;.
+      &quot;a_key&quot;: &quot;A String&quot;,
     },
-    "createTime": "A String", # Output only. When the job was created.
-  }</pre>
+    &quot;predictionInput&quot;: { # Represents input parameters for a prediction job. # Input parameters to create a prediction job.
+      &quot;outputPath&quot;: &quot;A String&quot;, # Required. The output Google Cloud Storage location.
+      &quot;outputDataFormat&quot;: &quot;A String&quot;, # Optional. Format of the output data files, defaults to JSON.
+      &quot;dataFormat&quot;: &quot;A String&quot;, # Required. The format of the input data files.
+      &quot;batchSize&quot;: &quot;A String&quot;, # Optional. Number of records per batch, defaults to 64.
+          # The service will buffer batch_size number of records in memory before
+          # invoking one Tensorflow prediction call internally. So take the record
+          # size and memory available into consideration when setting this parameter.
+      &quot;runtimeVersion&quot;: &quot;A String&quot;, # Optional. The AI Platform runtime version to use for this batch
+          # prediction. If not set, AI Platform will pick the runtime version used
+          # during the CreateVersion request for this model version, or choose the
+          # latest stable version when model version information is not available
+          # such as when the model is specified by uri.
+      &quot;inputPaths&quot;: [ # Required. The Cloud Storage location of the input data files. May contain
+          # &lt;a href=&quot;/storage/docs/gsutil/addlhelp/WildcardNames&quot;&gt;wildcards&lt;/a&gt;.
+        &quot;A String&quot;,
+      ],
+      &quot;region&quot;: &quot;A String&quot;, # Required. The Google Compute Engine region to run the prediction job in.
+          # See the &lt;a href=&quot;/ml-engine/docs/tensorflow/regions&quot;&gt;available regions&lt;/a&gt;
+          # for AI Platform services.
+      &quot;versionName&quot;: &quot;A String&quot;, # Use this field if you want to specify a version of the model to use. The
+          # string is formatted the same way as `model_version`, with the addition
+          # of the version information:
+          #
+          # `&quot;projects/YOUR_PROJECT/models/YOUR_MODEL/versions/YOUR_VERSION&quot;`
+      &quot;modelName&quot;: &quot;A String&quot;, # Use this field if you want to use the default version for the specified
+          # model. The string must use the following format:
+          #
+          # `&quot;projects/YOUR_PROJECT/models/YOUR_MODEL&quot;`
+      &quot;uri&quot;: &quot;A String&quot;, # Use this field if you want to specify a Google Cloud Storage path for
+          # the model to use.
+      &quot;maxWorkerCount&quot;: &quot;A String&quot;, # Optional. The maximum number of workers to be used for parallel processing.
+          # Defaults to 10 if not specified.
+      &quot;signatureName&quot;: &quot;A String&quot;, # Optional. The name of the signature defined in the SavedModel to use for
+          # this job. Please refer to
+          # [SavedModel](https://tensorflow.github.io/serving/serving_basic.html)
+          # for information about how to use signatures.
+          #
+          # Defaults to
+          # [DEFAULT_SERVING_SIGNATURE_DEF_KEY](https://www.tensorflow.org/api_docs/python/tf/saved_model/signature_constants)
+          # , which is &quot;serving_default&quot;.
+    },
+    &quot;errorMessage&quot;: &quot;A String&quot;, # Output only. The details of a failure or a cancellation.
+  }
+
+  updateMask: string, Required. Specifies the path, relative to `Job`, of the field to update.
+To adopt etag mechanism, include `etag` field in the mask, and include the
+`etag` value in your job resource.
+
+For example, to change the labels of a job, the `update_mask` parameter
+would be specified as `labels`, `etag`, and the
+`PATCH` request body would specify the new value, as follows:
+    {
+      &quot;labels&quot;: {
+         &quot;owner&quot;: &quot;Google&quot;,
+         &quot;color&quot;: &quot;Blue&quot;
+      }
+      &quot;etag&quot;: &quot;33a64df551425fcc55e4d42a148795d9f25f89d4&quot;
+    }
+If `etag` matches the one on the server, the labels of the job will be
+replaced with the given ones, and the server end `etag` will be
+recalculated.
+
+Currently the only supported update masks are `labels` and `etag`.
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # Represents a training or prediction job.
+      &quot;etag&quot;: &quot;A String&quot;, # `etag` is used for optimistic concurrency control as a way to help
+          # prevent simultaneous updates of a job from overwriting each other.
+          # It is strongly suggested that systems make use of the `etag` in the
+          # read-modify-write cycle to perform job updates in order to avoid race
+          # conditions: An `etag` is returned in the response to `GetJob`, and
+          # systems are expected to put that etag in the request to `UpdateJob` to
+          # ensure that their change will be applied to the same version of the job.
+      &quot;trainingInput&quot;: { # Represents input parameters for a training job. When using the gcloud command # Input parameters to create a training job.
+          # to submit your training job, you can specify the input parameters as
+          # command-line arguments and/or in a YAML configuration file referenced from
+          # the --config command-line argument. For details, see the guide to [submitting
+          # a training job](/ai-platform/training/docs/training-jobs).
+        &quot;parameterServerConfig&quot;: { # Represents the configuration for a replica in a cluster. # Optional. The configuration for parameter servers.
+            #
+            # You should only set `parameterServerConfig.acceleratorConfig` if
+            # `parameterServerType` is set to a Compute Engine machine type. [Learn
+            # about restrictions on accelerator configurations for
+            # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+            #
+            # Set `parameterServerConfig.imageUri` only if you build a custom image for
+            # your parameter server. If `parameterServerConfig.imageUri` has not been
+            # set, AI Platform uses the value of `masterConfig.imageUri`. Learn more about [configuring custom
+            # containers](/ai-platform/training/docs/distributed-training-containers).
+          &quot;tpuTfVersion&quot;: &quot;A String&quot;, # The AI Platform runtime version that includes a TensorFlow version matching
+              # the one used in the custom container. This field is required if the replica
+              # is a TPU worker that uses a custom container. Otherwise, do not specify
+              # this field. This must be a [runtime version that currently supports
+              # training with
+              # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
+              #
+              # Note that the version of TensorFlow included in a runtime version may
+              # differ from the numbering of the runtime version itself, because it may
+              # have a different [patch
+              # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
+              # In this field, you must specify the runtime version (TensorFlow minor
+              # version). For example, if your custom container runs TensorFlow `1.x.y`,
+              # specify `1.x`.
+          &quot;containerCommand&quot;: [ # The command with which the replica&#x27;s custom container is run.
+              # If provided, it will override default ENTRYPOINT of the docker image.
+              # If not provided, the docker image&#x27;s ENTRYPOINT is used.
+              # It cannot be set if custom container image is
+              # not provided.
+              # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+              # both cannot be set at the same time.
+            &quot;A String&quot;,
+          ],
+          &quot;imageUri&quot;: &quot;A String&quot;, # The Docker image to run on the replica. This image must be in Container
+              # Registry. Learn more about [configuring custom
+              # containers](/ai-platform/training/docs/distributed-training-containers).
+          &quot;containerArgs&quot;: [ # Arguments to the entrypoint command.
+              # The following rules apply for container_command and container_args:
+              # - If you do not supply command or args:
+              #   The defaults defined in the Docker image are used.
+              # - If you supply a command but no args:
+              #   The default EntryPoint and the default Cmd defined in the Docker image
+              #   are ignored. Your command is run without any arguments.
+              # - If you supply only args:
+              #   The default Entrypoint defined in the Docker image is run with the args
+              #   that you supplied.
+              # - If you supply a command and args:
+              #   The default Entrypoint and the default Cmd defined in the Docker image
+              #   are ignored. Your command is run with your args.
+              # It cannot be set if custom container image is
+              # not provided.
+              # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+              # both cannot be set at the same time.
+            &quot;A String&quot;,
+          ],
+          &quot;acceleratorConfig&quot;: { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
+              # [Learn about restrictions on accelerator configurations for
+              # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+              # Note that the AcceleratorConfig can be used in both Jobs and Versions.
+              # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
+              # [accelerators for online
+              # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
+            &quot;count&quot;: &quot;A String&quot;, # The number of accelerators to attach to each machine running the job.
+            &quot;type&quot;: &quot;A String&quot;, # The type of accelerator to use.
+          },
+        },
+        &quot;encryptionConfig&quot;: { # Represents a custom encryption key configuration that can be applied to # Optional. Options for using customer-managed encryption keys (CMEK) to
+            # protect resources created by a training job, instead of using Google&#x27;s
+            # default encryption. If this is set, then all resources created by the
+            # training job will be encrypted with the customer-managed encryption key
+            # that you specify.
+            #
+            # [Learn how and when to use CMEK with AI Platform
+            # Training](/ai-platform/training/docs/cmek).
+            # a resource.
+          &quot;kmsKeyName&quot;: &quot;A String&quot;, # The Cloud KMS resource identifier of the customer-managed encryption key
+              # used to protect a resource, such as a training job. It has the following
+              # format:
+              # `projects/{PROJECT_ID}/locations/{REGION}/keyRings/{KEY_RING_NAME}/cryptoKeys/{KEY_NAME}`
+        },
+        &quot;hyperparameters&quot;: { # Represents a set of hyperparameters to optimize. # Optional. The set of Hyperparameters to tune.
+          &quot;hyperparameterMetricTag&quot;: &quot;A String&quot;, # Optional. The TensorFlow summary tag name to use for optimizing trials. For
+              # current versions of TensorFlow, this tag name should exactly match what is
+              # shown in TensorBoard, including all scopes.  For versions of TensorFlow
+              # prior to 0.12, this should be only the tag passed to tf.Summary.
+              # By default, &quot;training/hptuning/metric&quot; will be used.
+          &quot;params&quot;: [ # Required. The set of parameters to tune.
+            { # Represents a single hyperparameter to optimize.
+              &quot;type&quot;: &quot;A String&quot;, # Required. The type of the parameter.
+              &quot;categoricalValues&quot;: [ # Required if type is `CATEGORICAL`. The list of possible categories.
+                &quot;A String&quot;,
+              ],
+              &quot;parameterName&quot;: &quot;A String&quot;, # Required. The parameter name must be unique amongst all ParameterConfigs in
+                  # a HyperparameterSpec message. E.g., &quot;learning_rate&quot;.
+              &quot;minValue&quot;: 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
+                  # should be unset if type is `CATEGORICAL`. This value should be integers if
+                  # type is INTEGER.
+              &quot;discreteValues&quot;: [ # Required if type is `DISCRETE`.
+                  # A list of feasible points.
+                  # The list should be in strictly increasing order. For instance, this
+                  # parameter might have possible settings of 1.5, 2.5, and 4.0. This list
+                  # should not contain more than 1,000 values.
+                3.14,
+              ],
+              &quot;scaleType&quot;: &quot;A String&quot;, # Optional. How the parameter should be scaled to the hypercube.
+                  # Leave unset for categorical parameters.
+                  # Some kind of scaling is strongly recommended for real or integral
+                  # parameters (e.g., `UNIT_LINEAR_SCALE`).
+              &quot;maxValue&quot;: 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
+                  # should be unset if type is `CATEGORICAL`. This value should be integers if
+                  # type is `INTEGER`.
+            },
+          ],
+          &quot;enableTrialEarlyStopping&quot;: True or False, # Optional. Indicates if the hyperparameter tuning job enables auto trial
+              # early stopping.
+          &quot;resumePreviousJobId&quot;: &quot;A String&quot;, # Optional. The prior hyperparameter tuning job id that users hope to
+              # continue with. The job id will be used to find the corresponding vizier
+              # study guid and resume the study.
+          &quot;maxParallelTrials&quot;: 42, # Optional. The number of training trials to run concurrently.
+              # You can reduce the time it takes to perform hyperparameter tuning by adding
+              # trials in parallel. However, each trail only benefits from the information
+              # gained in completed trials. That means that a trial does not get access to
+              # the results of trials running at the same time, which could reduce the
+              # quality of the overall optimization.
+              #
+              # Each trial will use the same scale tier and machine types.
+              #
+              # Defaults to one.
+          &quot;maxFailedTrials&quot;: 42, # Optional. The number of failed trials that need to be seen before failing
+              # the hyperparameter tuning job. You can specify this field to override the
+              # default failing criteria for AI Platform hyperparameter tuning jobs.
+              #
+              # Defaults to zero, which means the service decides when a hyperparameter
+              # job should fail.
+          &quot;goal&quot;: &quot;A String&quot;, # Required. The type of goal to use for tuning. Available types are
+              # `MAXIMIZE` and `MINIMIZE`.
+              #
+              # Defaults to `MAXIMIZE`.
+          &quot;maxTrials&quot;: 42, # Optional. How many training trials should be attempted to optimize
+              # the specified hyperparameters.
+              #
+              # Defaults to one.
+          &quot;algorithm&quot;: &quot;A String&quot;, # Optional. The search algorithm specified for the hyperparameter
+              # tuning job.
+              # Uses the default AI Platform hyperparameter tuning
+              # algorithm if unspecified.
+        },
+        &quot;workerConfig&quot;: { # Represents the configuration for a replica in a cluster. # Optional. The configuration for workers.
+            #
+            # You should only set `workerConfig.acceleratorConfig` if `workerType` is set
+            # to a Compute Engine machine type. [Learn about restrictions on accelerator
+            # configurations for
+            # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+            #
+            # Set `workerConfig.imageUri` only if you build a custom image for your
+            # worker. If `workerConfig.imageUri` has not been set, AI Platform uses
+            # the value of `masterConfig.imageUri`. Learn more about [configuring custom
+            # containers](/ai-platform/training/docs/distributed-training-containers).
+          &quot;tpuTfVersion&quot;: &quot;A String&quot;, # The AI Platform runtime version that includes a TensorFlow version matching
+              # the one used in the custom container. This field is required if the replica
+              # is a TPU worker that uses a custom container. Otherwise, do not specify
+              # this field. This must be a [runtime version that currently supports
+              # training with
+              # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
+              #
+              # Note that the version of TensorFlow included in a runtime version may
+              # differ from the numbering of the runtime version itself, because it may
+              # have a different [patch
+              # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
+              # In this field, you must specify the runtime version (TensorFlow minor
+              # version). For example, if your custom container runs TensorFlow `1.x.y`,
+              # specify `1.x`.
+          &quot;containerCommand&quot;: [ # The command with which the replica&#x27;s custom container is run.
+              # If provided, it will override default ENTRYPOINT of the docker image.
+              # If not provided, the docker image&#x27;s ENTRYPOINT is used.
+              # It cannot be set if custom container image is
+              # not provided.
+              # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+              # both cannot be set at the same time.
+            &quot;A String&quot;,
+          ],
+          &quot;imageUri&quot;: &quot;A String&quot;, # The Docker image to run on the replica. This image must be in Container
+              # Registry. Learn more about [configuring custom
+              # containers](/ai-platform/training/docs/distributed-training-containers).
+          &quot;containerArgs&quot;: [ # Arguments to the entrypoint command.
+              # The following rules apply for container_command and container_args:
+              # - If you do not supply command or args:
+              #   The defaults defined in the Docker image are used.
+              # - If you supply a command but no args:
+              #   The default EntryPoint and the default Cmd defined in the Docker image
+              #   are ignored. Your command is run without any arguments.
+              # - If you supply only args:
+              #   The default Entrypoint defined in the Docker image is run with the args
+              #   that you supplied.
+              # - If you supply a command and args:
+              #   The default Entrypoint and the default Cmd defined in the Docker image
+              #   are ignored. Your command is run with your args.
+              # It cannot be set if custom container image is
+              # not provided.
+              # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+              # both cannot be set at the same time.
+            &quot;A String&quot;,
+          ],
+          &quot;acceleratorConfig&quot;: { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
+              # [Learn about restrictions on accelerator configurations for
+              # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+              # Note that the AcceleratorConfig can be used in both Jobs and Versions.
+              # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
+              # [accelerators for online
+              # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
+            &quot;count&quot;: &quot;A String&quot;, # The number of accelerators to attach to each machine running the job.
+            &quot;type&quot;: &quot;A String&quot;, # The type of accelerator to use.
+          },
+        },
+        &quot;parameterServerCount&quot;: &quot;A String&quot;, # Optional. The number of parameter server replicas to use for the training
+            # job. Each replica in the cluster will be of the type specified in
+            # `parameter_server_type`.
+            #
+            # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
+            # set this value, you must also set `parameter_server_type`.
+            #
+            # The default value is zero.
+        &quot;packageUris&quot;: [ # Required. The Google Cloud Storage location of the packages with
+            # the training program and any additional dependencies.
+            # The maximum number of package URIs is 100.
+          &quot;A String&quot;,
+        ],
+        &quot;evaluatorCount&quot;: &quot;A String&quot;, # Optional. The number of evaluator replicas to use for the training job.
+            # Each replica in the cluster will be of the type specified in
+            # `evaluator_type`.
+            #
+            # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
+            # set this value, you must also set `evaluator_type`.
+            #
+            # The default value is zero.
+        &quot;masterType&quot;: &quot;A String&quot;, # Optional. Specifies the type of virtual machine to use for your training
+            # job&#x27;s master worker. You must specify this field when `scaleTier` is set to
+            # `CUSTOM`.
+            #
+            # You can use certain Compute Engine machine types directly in this field.
+            # The following types are supported:
+            #
+            # - `n1-standard-4`
+            # - `n1-standard-8`
+            # - `n1-standard-16`
+            # - `n1-standard-32`
+            # - `n1-standard-64`
+            # - `n1-standard-96`
+            # - `n1-highmem-2`
+            # - `n1-highmem-4`
+            # - `n1-highmem-8`
+            # - `n1-highmem-16`
+            # - `n1-highmem-32`
+            # - `n1-highmem-64`
+            # - `n1-highmem-96`
+            # - `n1-highcpu-16`
+            # - `n1-highcpu-32`
+            # - `n1-highcpu-64`
+            # - `n1-highcpu-96`
+            #
+            # Learn more about [using Compute Engine machine
+            # types](/ml-engine/docs/machine-types#compute-engine-machine-types).
+            #
+            # Alternatively, you can use the following legacy machine types:
+            #
+            # - `standard`
+            # - `large_model`
+            # - `complex_model_s`
+            # - `complex_model_m`
+            # - `complex_model_l`
+            # - `standard_gpu`
+            # - `complex_model_m_gpu`
+            # - `complex_model_l_gpu`
+            # - `standard_p100`
+            # - `complex_model_m_p100`
+            # - `standard_v100`
+            # - `large_model_v100`
+            # - `complex_model_m_v100`
+            # - `complex_model_l_v100`
+            #
+            # Learn more about [using legacy machine
+            # types](/ml-engine/docs/machine-types#legacy-machine-types).
+            #
+            # Finally, if you want to use a TPU for training, specify `cloud_tpu` in this
+            # field. Learn more about the [special configuration options for training
+            # with
+            # TPUs](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine).
+        &quot;runtimeVersion&quot;: &quot;A String&quot;, # Optional. The AI Platform runtime version to use for training. You must
+            # either specify this field or specify `masterConfig.imageUri`.
+            #
+            # For more information, see the [runtime version
+            # list](/ai-platform/training/docs/runtime-version-list) and learn [how to
+            # manage runtime versions](/ai-platform/training/docs/versioning).
+        &quot;evaluatorType&quot;: &quot;A String&quot;, # Optional. Specifies the type of virtual machine to use for your training
+            # job&#x27;s evaluator nodes.
+            #
+            # The supported values are the same as those described in the entry for
+            # `masterType`.
+            #
+            # This value must be consistent with the category of machine type that
+            # `masterType` uses. In other words, both must be Compute Engine machine
+            # types or both must be legacy machine types.
+            #
+            # This value must be present when `scaleTier` is set to `CUSTOM` and
+            # `evaluatorCount` is greater than zero.
+        &quot;region&quot;: &quot;A String&quot;, # Required. The region to run the training job in. See the [available
+            # regions](/ai-platform/training/docs/regions) for AI Platform Training.
+        &quot;workerType&quot;: &quot;A String&quot;, # Optional. Specifies the type of virtual machine to use for your training
+            # job&#x27;s worker nodes.
+            #
+            # The supported values are the same as those described in the entry for
+            # `masterType`.
+            #
+            # This value must be consistent with the category of machine type that
+            # `masterType` uses. In other words, both must be Compute Engine machine
+            # types or both must be legacy machine types.
+            #
+            # If you use `cloud_tpu` for this value, see special instructions for
+            # [configuring a custom TPU
+            # machine](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine).
+            #
+            # This value must be present when `scaleTier` is set to `CUSTOM` and
+            # `workerCount` is greater than zero.
+        &quot;parameterServerType&quot;: &quot;A String&quot;, # Optional. Specifies the type of virtual machine to use for your training
+            # job&#x27;s parameter server.
+            #
+            # The supported values are the same as those described in the entry for
+            # `master_type`.
+            #
+            # This value must be consistent with the category of machine type that
+            # `masterType` uses. In other words, both must be Compute Engine machine
+            # types or both must be legacy machine types.
+            #
+            # This value must be present when `scaleTier` is set to `CUSTOM` and
+            # `parameter_server_count` is greater than zero.
+        &quot;masterConfig&quot;: { # Represents the configuration for a replica in a cluster. # Optional. The configuration for your master worker.
+            #
+            # You should only set `masterConfig.acceleratorConfig` if `masterType` is set
+            # to a Compute Engine machine type. Learn about [restrictions on accelerator
+            # configurations for
+            # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+            #
+            # Set `masterConfig.imageUri` only if you build a custom image. Only one of
+            # `masterConfig.imageUri` and `runtimeVersion` should be set. Learn more
+            # about [configuring custom
+            # containers](/ai-platform/training/docs/distributed-training-containers).
+          &quot;tpuTfVersion&quot;: &quot;A String&quot;, # The AI Platform runtime version that includes a TensorFlow version matching
+              # the one used in the custom container. This field is required if the replica
+              # is a TPU worker that uses a custom container. Otherwise, do not specify
+              # this field. This must be a [runtime version that currently supports
+              # training with
+              # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
+              #
+              # Note that the version of TensorFlow included in a runtime version may
+              # differ from the numbering of the runtime version itself, because it may
+              # have a different [patch
+              # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
+              # In this field, you must specify the runtime version (TensorFlow minor
+              # version). For example, if your custom container runs TensorFlow `1.x.y`,
+              # specify `1.x`.
+          &quot;containerCommand&quot;: [ # The command with which the replica&#x27;s custom container is run.
+              # If provided, it will override default ENTRYPOINT of the docker image.
+              # If not provided, the docker image&#x27;s ENTRYPOINT is used.
+              # It cannot be set if custom container image is
+              # not provided.
+              # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+              # both cannot be set at the same time.
+            &quot;A String&quot;,
+          ],
+          &quot;imageUri&quot;: &quot;A String&quot;, # The Docker image to run on the replica. This image must be in Container
+              # Registry. Learn more about [configuring custom
+              # containers](/ai-platform/training/docs/distributed-training-containers).
+          &quot;containerArgs&quot;: [ # Arguments to the entrypoint command.
+              # The following rules apply for container_command and container_args:
+              # - If you do not supply command or args:
+              #   The defaults defined in the Docker image are used.
+              # - If you supply a command but no args:
+              #   The default EntryPoint and the default Cmd defined in the Docker image
+              #   are ignored. Your command is run without any arguments.
+              # - If you supply only args:
+              #   The default Entrypoint defined in the Docker image is run with the args
+              #   that you supplied.
+              # - If you supply a command and args:
+              #   The default Entrypoint and the default Cmd defined in the Docker image
+              #   are ignored. Your command is run with your args.
+              # It cannot be set if custom container image is
+              # not provided.
+              # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+              # both cannot be set at the same time.
+            &quot;A String&quot;,
+          ],
+          &quot;acceleratorConfig&quot;: { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
+              # [Learn about restrictions on accelerator configurations for
+              # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+              # Note that the AcceleratorConfig can be used in both Jobs and Versions.
+              # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
+              # [accelerators for online
+              # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
+            &quot;count&quot;: &quot;A String&quot;, # The number of accelerators to attach to each machine running the job.
+            &quot;type&quot;: &quot;A String&quot;, # The type of accelerator to use.
+          },
+        },
+        &quot;scaleTier&quot;: &quot;A String&quot;, # Required. Specifies the machine types, the number of replicas for workers
+            # and parameter servers.
+        &quot;jobDir&quot;: &quot;A String&quot;, # Optional. A Google Cloud Storage path in which to store training outputs
+            # and other data needed for training. This path is passed to your TensorFlow
+            # program as the &#x27;--job-dir&#x27; command-line argument. The benefit of specifying
+            # this field is that Cloud ML validates the path for use in training.
+        &quot;pythonVersion&quot;: &quot;A String&quot;, # Optional. The version of Python used in training. You must either specify
+            # this field or specify `masterConfig.imageUri`.
+            #
+            # The following Python versions are available:
+            #
+            # * Python &#x27;3.7&#x27; is available when `runtime_version` is set to &#x27;1.15&#x27; or
+            #   later.
+            # * Python &#x27;3.5&#x27; is available when `runtime_version` is set to a version
+            #   from &#x27;1.4&#x27; to &#x27;1.14&#x27;.
+            # * Python &#x27;2.7&#x27; is available when `runtime_version` is set to &#x27;1.15&#x27; or
+            #   earlier.
+            #
+            # Read more about the Python versions available for [each runtime
+            # version](/ml-engine/docs/runtime-version-list).
+        &quot;network&quot;: &quot;A String&quot;, # Optional. The full name of the Google Compute Engine
+            # [network](/compute/docs/networks-and-firewalls#networks) to which the Job
+            # is peered. For example, projects/12345/global/networks/myVPC. Format is of
+            # the form projects/{project}/global/networks/{network}. Where {project} is a
+            # project number, as in &#x27;12345&#x27;, and {network} is network name.&quot;.
+            #
+            # Private services access must already be configured for the network. If left
+            # unspecified, the Job is not peered with any network. Learn more -
+            # Connecting Job to user network over private
+            # IP.
+        &quot;scheduling&quot;: { # All parameters related to scheduling of training jobs. # Optional. Scheduling options for a training job.
+          &quot;maxWaitTime&quot;: &quot;A String&quot;,
+          &quot;maxRunningTime&quot;: &quot;A String&quot;, # Optional. The maximum job running time, expressed in seconds. The field can
+              # contain up to nine fractional digits, terminated by `s`. If not specified,
+              # this field defaults to `604800s` (seven days).
+              #
+              # If the training job is still running after this duration, AI Platform
+              # Training cancels it.
+              #
+              # For example, if you want to ensure your job runs for no more than 2 hours,
+              # set this field to `7200s` (2 hours * 60 minutes / hour * 60 seconds /
+              # minute).
+              #
+              # If you submit your training job using the `gcloud` tool, you can [provide
+              # this field in a `config.yaml`
+              # file](/ai-platform/training/docs/training-jobs#formatting_your_configuration_parameters).
+              # For example:
+              #
+              # ```yaml
+              # trainingInput:
+              #   ...
+              #   scheduling:
+              #     maxRunningTime: 7200s
+              #   ...
+              # ```
+        },
+        &quot;evaluatorConfig&quot;: { # Represents the configuration for a replica in a cluster. # Optional. The configuration for evaluators.
+            #
+            # You should only set `evaluatorConfig.acceleratorConfig` if
+            # `evaluatorType` is set to a Compute Engine machine type. [Learn
+            # about restrictions on accelerator configurations for
+            # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+            #
+            # Set `evaluatorConfig.imageUri` only if you build a custom image for
+            # your evaluator. If `evaluatorConfig.imageUri` has not been
+            # set, AI Platform uses the value of `masterConfig.imageUri`. Learn more about [configuring custom
+            # containers](/ai-platform/training/docs/distributed-training-containers).
+          &quot;tpuTfVersion&quot;: &quot;A String&quot;, # The AI Platform runtime version that includes a TensorFlow version matching
+              # the one used in the custom container. This field is required if the replica
+              # is a TPU worker that uses a custom container. Otherwise, do not specify
+              # this field. This must be a [runtime version that currently supports
+              # training with
+              # TPUs](/ml-engine/docs/tensorflow/runtime-version-list#tpu-support).
+              #
+              # Note that the version of TensorFlow included in a runtime version may
+              # differ from the numbering of the runtime version itself, because it may
+              # have a different [patch
+              # version](https://www.tensorflow.org/guide/version_compat#semantic_versioning_20).
+              # In this field, you must specify the runtime version (TensorFlow minor
+              # version). For example, if your custom container runs TensorFlow `1.x.y`,
+              # specify `1.x`.
+          &quot;containerCommand&quot;: [ # The command with which the replica&#x27;s custom container is run.
+              # If provided, it will override default ENTRYPOINT of the docker image.
+              # If not provided, the docker image&#x27;s ENTRYPOINT is used.
+              # It cannot be set if custom container image is
+              # not provided.
+              # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+              # both cannot be set at the same time.
+            &quot;A String&quot;,
+          ],
+          &quot;imageUri&quot;: &quot;A String&quot;, # The Docker image to run on the replica. This image must be in Container
+              # Registry. Learn more about [configuring custom
+              # containers](/ai-platform/training/docs/distributed-training-containers).
+          &quot;containerArgs&quot;: [ # Arguments to the entrypoint command.
+              # The following rules apply for container_command and container_args:
+              # - If you do not supply command or args:
+              #   The defaults defined in the Docker image are used.
+              # - If you supply a command but no args:
+              #   The default EntryPoint and the default Cmd defined in the Docker image
+              #   are ignored. Your command is run without any arguments.
+              # - If you supply only args:
+              #   The default Entrypoint defined in the Docker image is run with the args
+              #   that you supplied.
+              # - If you supply a command and args:
+              #   The default Entrypoint and the default Cmd defined in the Docker image
+              #   are ignored. Your command is run with your args.
+              # It cannot be set if custom container image is
+              # not provided.
+              # Note that this field and [TrainingInput.args] are mutually exclusive, i.e.,
+              # both cannot be set at the same time.
+            &quot;A String&quot;,
+          ],
+          &quot;acceleratorConfig&quot;: { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
+              # [Learn about restrictions on accelerator configurations for
+              # training.](/ai-platform/training/docs/using-gpus#compute-engine-machine-types-with-gpu)
+              # Note that the AcceleratorConfig can be used in both Jobs and Versions.
+              # Learn more about [accelerators for training](/ml-engine/docs/using-gpus) and
+              # [accelerators for online
+              # prediction](/ml-engine/docs/machine-types-online-prediction#gpus).
+            &quot;count&quot;: &quot;A String&quot;, # The number of accelerators to attach to each machine running the job.
+            &quot;type&quot;: &quot;A String&quot;, # The type of accelerator to use.
+          },
+        },
+        &quot;useChiefInTfConfig&quot;: True or False, # Optional. Use `chief` instead of `master` in the `TF_CONFIG` environment
+            # variable when training with a custom container. Defaults to `false`. [Learn
+            # more about this
+            # field.](/ai-platform/training/docs/distributed-training-details#chief-versus-master)
+            #
+            # This field has no effect for training jobs that don&#x27;t use a custom
+            # container.
+        &quot;workerCount&quot;: &quot;A String&quot;, # Optional. The number of worker replicas to use for the training job. Each
+            # replica in the cluster will be of the type specified in `worker_type`.
+            #
+            # This value can only be used when `scale_tier` is set to `CUSTOM`. If you
+            # set this value, you must also set `worker_type`.
+            #
+            # The default value is zero.
+        &quot;pythonModule&quot;: &quot;A String&quot;, # Required. The Python module name to run after installing the packages.
+        &quot;args&quot;: [ # Optional. Command-line arguments passed to the training application when it
+            # starts. If your job uses a custom container, then the arguments are passed
+            # to the container&#x27;s &lt;a class=&quot;external&quot; target=&quot;_blank&quot;
+            # href=&quot;https://docs.docker.com/engine/reference/builder/#entrypoint&quot;&gt;
+            # `ENTRYPOINT`&lt;/a&gt; command.
+          &quot;A String&quot;,
+        ],
+      },
+      &quot;state&quot;: &quot;A String&quot;, # Output only. The detailed state of a job.
+      &quot;jobId&quot;: &quot;A String&quot;, # Required. The user-specified id of the job.
+      &quot;endTime&quot;: &quot;A String&quot;, # Output only. When the job processing was completed.
+      &quot;startTime&quot;: &quot;A String&quot;, # Output only. When the job processing was started.
+      &quot;predictionOutput&quot;: { # Represents results of a prediction job. # The current prediction job result.
+        &quot;errorCount&quot;: &quot;A String&quot;, # The number of data instances which resulted in errors.
+        &quot;outputPath&quot;: &quot;A String&quot;, # The output Google Cloud Storage location provided at the job creation time.
+        &quot;nodeHours&quot;: 3.14, # Node hours used by the batch prediction job.
+        &quot;predictionCount&quot;: &quot;A String&quot;, # The number of generated predictions.
+      },
+      &quot;trainingOutput&quot;: { # Represents results of a training job. Output only. # The current training job result.
+        &quot;trials&quot;: [ # Results for individual Hyperparameter trials.
+            # Only set for hyperparameter tuning jobs.
+          { # Represents the result of a single hyperparameter tuning trial from a
+              # training job. The TrainingOutput object that is returned on successful
+              # completion of a training job with hyperparameter tuning includes a list
+              # of HyperparameterOutput objects, one for each successful trial.
+            &quot;allMetrics&quot;: [ # All recorded object metrics for this trial. This field is not currently
+                # populated.
+              { # An observed value of a metric.
+                &quot;objectiveValue&quot;: 3.14, # The objective value at this training step.
+                &quot;trainingStep&quot;: &quot;A String&quot;, # The global training step for this metric.
+              },
+            ],
+            &quot;hyperparameters&quot;: { # The hyperparameters given to this trial.
+              &quot;a_key&quot;: &quot;A String&quot;,
+            },
+            &quot;trialId&quot;: &quot;A String&quot;, # The trial id for these results.
+            &quot;endTime&quot;: &quot;A String&quot;, # Output only. End time for the trial.
+            &quot;isTrialStoppedEarly&quot;: True or False, # True if the trial is stopped early.
+            &quot;startTime&quot;: &quot;A String&quot;, # Output only. Start time for the trial.
+            &quot;finalMetric&quot;: { # An observed value of a metric. # The final objective metric seen for this trial.
+              &quot;objectiveValue&quot;: 3.14, # The objective value at this training step.
+              &quot;trainingStep&quot;: &quot;A String&quot;, # The global training step for this metric.
+            },
+            &quot;builtInAlgorithmOutput&quot;: { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
+                # Only set for trials of built-in algorithms jobs that have succeeded.
+              &quot;framework&quot;: &quot;A String&quot;, # Framework on which the built-in algorithm was trained.
+              &quot;modelPath&quot;: &quot;A String&quot;, # The Cloud Storage path to the `model/` directory where the training job
+                  # saves the trained model. Only set for successful jobs that don&#x27;t use
+                  # hyperparameter tuning.
+              &quot;pythonVersion&quot;: &quot;A String&quot;, # Python version on which the built-in algorithm was trained.
+              &quot;runtimeVersion&quot;: &quot;A String&quot;, # AI Platform runtime version on which the built-in algorithm was
+                  # trained.
+            },
+            &quot;state&quot;: &quot;A String&quot;, # Output only. The detailed state of the trial.
+          },
+        ],
+        &quot;hyperparameterMetricTag&quot;: &quot;A String&quot;, # The TensorFlow summary tag name used for optimizing hyperparameter tuning
+            # trials. See
+            # [`HyperparameterSpec.hyperparameterMetricTag`](#HyperparameterSpec.FIELDS.hyperparameter_metric_tag)
+            # for more information. Only set for hyperparameter tuning jobs.
+        &quot;completedTrialCount&quot;: &quot;A String&quot;, # The number of hyperparameter tuning trials that completed successfully.
+            # Only set for hyperparameter tuning jobs.
+        &quot;isHyperparameterTuningJob&quot;: True or False, # Whether this job is a hyperparameter tuning job.
+        &quot;consumedMLUnits&quot;: 3.14, # The amount of ML units consumed by the job.
+        &quot;isBuiltInAlgorithmJob&quot;: True or False, # Whether this job is a built-in Algorithm job.
+        &quot;builtInAlgorithmOutput&quot;: { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
+            # Only set for built-in algorithms jobs.
+          &quot;framework&quot;: &quot;A String&quot;, # Framework on which the built-in algorithm was trained.
+          &quot;modelPath&quot;: &quot;A String&quot;, # The Cloud Storage path to the `model/` directory where the training job
+              # saves the trained model. Only set for successful jobs that don&#x27;t use
+              # hyperparameter tuning.
+          &quot;pythonVersion&quot;: &quot;A String&quot;, # Python version on which the built-in algorithm was trained.
+          &quot;runtimeVersion&quot;: &quot;A String&quot;, # AI Platform runtime version on which the built-in algorithm was
+              # trained.
+        },
+      },
+      &quot;createTime&quot;: &quot;A String&quot;, # Output only. When the job was created.
+      &quot;labels&quot;: { # Optional. One or more labels that you can add, to organize your jobs.
+          # Each label is a key-value pair, where both the key and the value are
+          # arbitrary strings that you supply.
+          # For more information, see the documentation on
+          # &lt;a href=&quot;/ml-engine/docs/tensorflow/resource-labels&quot;&gt;using labels&lt;/a&gt;.
+        &quot;a_key&quot;: &quot;A String&quot;,
+      },
+      &quot;predictionInput&quot;: { # Represents input parameters for a prediction job. # Input parameters to create a prediction job.
+        &quot;outputPath&quot;: &quot;A String&quot;, # Required. The output Google Cloud Storage location.
+        &quot;outputDataFormat&quot;: &quot;A String&quot;, # Optional. Format of the output data files, defaults to JSON.
+        &quot;dataFormat&quot;: &quot;A String&quot;, # Required. The format of the input data files.
+        &quot;batchSize&quot;: &quot;A String&quot;, # Optional. Number of records per batch, defaults to 64.
+            # The service will buffer batch_size number of records in memory before
+            # invoking one Tensorflow prediction call internally. So take the record
+            # size and memory available into consideration when setting this parameter.
+        &quot;runtimeVersion&quot;: &quot;A String&quot;, # Optional. The AI Platform runtime version to use for this batch
+            # prediction. If not set, AI Platform will pick the runtime version used
+            # during the CreateVersion request for this model version, or choose the
+            # latest stable version when model version information is not available
+            # such as when the model is specified by uri.
+        &quot;inputPaths&quot;: [ # Required. The Cloud Storage location of the input data files. May contain
+            # &lt;a href=&quot;/storage/docs/gsutil/addlhelp/WildcardNames&quot;&gt;wildcards&lt;/a&gt;.
+          &quot;A String&quot;,
+        ],
+        &quot;region&quot;: &quot;A String&quot;, # Required. The Google Compute Engine region to run the prediction job in.
+            # See the &lt;a href=&quot;/ml-engine/docs/tensorflow/regions&quot;&gt;available regions&lt;/a&gt;
+            # for AI Platform services.
+        &quot;versionName&quot;: &quot;A String&quot;, # Use this field if you want to specify a version of the model to use. The
+            # string is formatted the same way as `model_version`, with the addition
+            # of the version information:
+            #
+            # `&quot;projects/YOUR_PROJECT/models/YOUR_MODEL/versions/YOUR_VERSION&quot;`
+        &quot;modelName&quot;: &quot;A String&quot;, # Use this field if you want to use the default version for the specified
+            # model. The string must use the following format:
+            #
+            # `&quot;projects/YOUR_PROJECT/models/YOUR_MODEL&quot;`
+        &quot;uri&quot;: &quot;A String&quot;, # Use this field if you want to specify a Google Cloud Storage path for
+            # the model to use.
+        &quot;maxWorkerCount&quot;: &quot;A String&quot;, # Optional. The maximum number of workers to be used for parallel processing.
+            # Defaults to 10 if not specified.
+        &quot;signatureName&quot;: &quot;A String&quot;, # Optional. The name of the signature defined in the SavedModel to use for
+            # this job. Please refer to
+            # [SavedModel](https://tensorflow.github.io/serving/serving_basic.html)
+            # for information about how to use signatures.
+            #
+            # Defaults to
+            # [DEFAULT_SERVING_SIGNATURE_DEF_KEY](https://www.tensorflow.org/api_docs/python/tf/saved_model/signature_constants)
+            # , which is &quot;serving_default&quot;.
+      },
+      &quot;errorMessage&quot;: &quot;A String&quot;, # Output only. The details of a failure or a cancellation.
+    }</pre>
 </div>
 
 <div class="method">
@@ -3933,7 +4690,7 @@
   <pre>Sets the access control policy on the specified resource. Replaces any
 existing policy.
 
-Can return Public Errors: NOT_FOUND, INVALID_ARGUMENT and PERMISSION_DENIED
+Can return `NOT_FOUND`, `INVALID_ARGUMENT`, and `PERMISSION_DENIED` errors.
 
 Args:
   resource: string, REQUIRED: The resource for which the policy is being specified.
@@ -3942,7 +4699,7 @@
     The object takes the form of:
 
 { # Request message for `SetIamPolicy` method.
-    "policy": { # An Identity and Access Management (IAM) policy, which specifies access # REQUIRED: The complete policy to be applied to the `resource`. The size of
+    &quot;policy&quot;: { # An Identity and Access Management (IAM) policy, which specifies access # REQUIRED: The complete policy to be applied to the `resource`. The size of
         # the policy is limited to a few 10s of KB. An empty policy is a
         # valid policy but certain Cloud Platform services (such as Projects)
         # might reject them.
@@ -3955,36 +4712,40 @@
         # permissions; each `role` can be an IAM predefined role or a user-created
         # custom role.
         #
-        # Optionally, a `binding` can specify a `condition`, which is a logical
-        # expression that allows access to a resource only if the expression evaluates
-        # to `true`. A condition can add constraints based on attributes of the
-        # request, the resource, or both.
+        # For some types of Google Cloud resources, a `binding` can also specify a
+        # `condition`, which is a logical expression that allows access to a resource
+        # only if the expression evaluates to `true`. A condition can add constraints
+        # based on attributes of the request, the resource, or both. To learn which
+        # resources support conditions in their IAM policies, see the
+        # [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
         #
         # **JSON example:**
         #
         #     {
-        #       "bindings": [
+        #       &quot;bindings&quot;: [
         #         {
-        #           "role": "roles/resourcemanager.organizationAdmin",
-        #           "members": [
-        #             "user:mike@example.com",
-        #             "group:admins@example.com",
-        #             "domain:google.com",
-        #             "serviceAccount:my-project-id@appspot.gserviceaccount.com"
+        #           &quot;role&quot;: &quot;roles/resourcemanager.organizationAdmin&quot;,
+        #           &quot;members&quot;: [
+        #             &quot;user:mike@example.com&quot;,
+        #             &quot;group:admins@example.com&quot;,
+        #             &quot;domain:google.com&quot;,
+        #             &quot;serviceAccount:my-project-id@appspot.gserviceaccount.com&quot;
         #           ]
         #         },
         #         {
-        #           "role": "roles/resourcemanager.organizationViewer",
-        #           "members": ["user:eve@example.com"],
-        #           "condition": {
-        #             "title": "expirable access",
-        #             "description": "Does not grant access after Sep 2020",
-        #             "expression": "request.time &lt; timestamp('2020-10-01T00:00:00.000Z')",
+        #           &quot;role&quot;: &quot;roles/resourcemanager.organizationViewer&quot;,
+        #           &quot;members&quot;: [
+        #             &quot;user:eve@example.com&quot;
+        #           ],
+        #           &quot;condition&quot;: {
+        #             &quot;title&quot;: &quot;expirable access&quot;,
+        #             &quot;description&quot;: &quot;Does not grant access after Sep 2020&quot;,
+        #             &quot;expression&quot;: &quot;request.time &lt; timestamp(&#x27;2020-10-01T00:00:00.000Z&#x27;)&quot;,
         #           }
         #         }
         #       ],
-        #       "etag": "BwWWja0YfJA=",
-        #       "version": 3
+        #       &quot;etag&quot;: &quot;BwWWja0YfJA=&quot;,
+        #       &quot;version&quot;: 3
         #     }
         #
         # **YAML example:**
@@ -4002,19 +4763,190 @@
         #       condition:
         #         title: expirable access
         #         description: Does not grant access after Sep 2020
-        #         expression: request.time &lt; timestamp('2020-10-01T00:00:00.000Z')
+        #         expression: request.time &lt; timestamp(&#x27;2020-10-01T00:00:00.000Z&#x27;)
         #     - etag: BwWWja0YfJA=
         #     - version: 3
         #
         # For a description of IAM and its features, see the
         # [IAM documentation](https://cloud.google.com/iam/docs/).
-      "bindings": [ # Associates a list of `members` to a `role`. Optionally, may specify a
+      &quot;etag&quot;: &quot;A String&quot;, # `etag` is used for optimistic concurrency control as a way to help
+          # prevent simultaneous updates of a policy from overwriting each other.
+          # It is strongly suggested that systems make use of the `etag` in the
+          # read-modify-write cycle to perform policy updates in order to avoid race
+          # conditions: An `etag` is returned in the response to `getIamPolicy`, and
+          # systems are expected to put that etag in the request to `setIamPolicy` to
+          # ensure that their change will be applied to the same version of the policy.
+          #
+          # **Important:** If you use IAM Conditions, you must include the `etag` field
+          # whenever you call `setIamPolicy`. If you omit this field, then IAM allows
+          # you to overwrite a version `3` policy with a version `1` policy, and all of
+          # the conditions in the version `3` policy are lost.
+      &quot;version&quot;: 42, # Specifies the format of the policy.
+          #
+          # Valid values are `0`, `1`, and `3`. Requests that specify an invalid value
+          # are rejected.
+          #
+          # Any operation that affects conditional role bindings must specify version
+          # `3`. This requirement applies to the following operations:
+          #
+          # * Getting a policy that includes a conditional role binding
+          # * Adding a conditional role binding to a policy
+          # * Changing a conditional role binding in a policy
+          # * Removing any role binding, with or without a condition, from a policy
+          #   that includes conditions
+          #
+          # **Important:** If you use IAM Conditions, you must include the `etag` field
+          # whenever you call `setIamPolicy`. If you omit this field, then IAM allows
+          # you to overwrite a version `3` policy with a version `1` policy, and all of
+          # the conditions in the version `3` policy are lost.
+          #
+          # If a policy does not include any conditions, operations on that policy may
+          # specify any valid version or leave the field unset.
+          #
+          # To learn which resources support conditions in their IAM policies, see the
+          # [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+      &quot;auditConfigs&quot;: [ # Specifies cloud audit logging configuration for this policy.
+        { # Specifies the audit configuration for a service.
+            # The configuration determines which permission types are logged, and what
+            # identities, if any, are exempted from logging.
+            # An AuditConfig must have one or more AuditLogConfigs.
+            #
+            # If there are AuditConfigs for both `allServices` and a specific service,
+            # the union of the two AuditConfigs is used for that service: the log_types
+            # specified in each AuditConfig are enabled, and the exempted_members in each
+            # AuditLogConfig are exempted.
+            #
+            # Example Policy with multiple AuditConfigs:
+            #
+            #     {
+            #       &quot;audit_configs&quot;: [
+            #         {
+            #           &quot;service&quot;: &quot;allServices&quot;
+            #           &quot;audit_log_configs&quot;: [
+            #             {
+            #               &quot;log_type&quot;: &quot;DATA_READ&quot;,
+            #               &quot;exempted_members&quot;: [
+            #                 &quot;user:jose@example.com&quot;
+            #               ]
+            #             },
+            #             {
+            #               &quot;log_type&quot;: &quot;DATA_WRITE&quot;,
+            #             },
+            #             {
+            #               &quot;log_type&quot;: &quot;ADMIN_READ&quot;,
+            #             }
+            #           ]
+            #         },
+            #         {
+            #           &quot;service&quot;: &quot;sampleservice.googleapis.com&quot;
+            #           &quot;audit_log_configs&quot;: [
+            #             {
+            #               &quot;log_type&quot;: &quot;DATA_READ&quot;,
+            #             },
+            #             {
+            #               &quot;log_type&quot;: &quot;DATA_WRITE&quot;,
+            #               &quot;exempted_members&quot;: [
+            #                 &quot;user:aliya@example.com&quot;
+            #               ]
+            #             }
+            #           ]
+            #         }
+            #       ]
+            #     }
+            #
+            # For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ
+            # logging. It also exempts jose@example.com from DATA_READ logging, and
+            # aliya@example.com from DATA_WRITE logging.
+          &quot;service&quot;: &quot;A String&quot;, # Specifies a service that will be enabled for audit logging.
+              # For example, `storage.googleapis.com`, `cloudsql.googleapis.com`.
+              # `allServices` is a special value that covers all services.
+          &quot;auditLogConfigs&quot;: [ # The configuration for logging of each type of permission.
+            { # Provides the configuration for logging a type of permissions.
+                # Example:
+                #
+                #     {
+                #       &quot;audit_log_configs&quot;: [
+                #         {
+                #           &quot;log_type&quot;: &quot;DATA_READ&quot;,
+                #           &quot;exempted_members&quot;: [
+                #             &quot;user:jose@example.com&quot;
+                #           ]
+                #         },
+                #         {
+                #           &quot;log_type&quot;: &quot;DATA_WRITE&quot;,
+                #         }
+                #       ]
+                #     }
+                #
+                # This enables &#x27;DATA_READ&#x27; and &#x27;DATA_WRITE&#x27; logging, while exempting
+                # jose@example.com from DATA_READ logging.
+              &quot;exemptedMembers&quot;: [ # Specifies the identities that do not cause logging for this type of
+                  # permission.
+                  # Follows the same format of Binding.members.
+                &quot;A String&quot;,
+              ],
+              &quot;logType&quot;: &quot;A String&quot;, # The log type that this config enables.
+            },
+          ],
+        },
+      ],
+      &quot;bindings&quot;: [ # Associates a list of `members` to a `role`. Optionally, may specify a
           # `condition` that determines how and when the `bindings` are applied. Each
           # of the `bindings` must contain at least one member.
         { # Associates `members` with a `role`.
-          "role": "A String", # Role that is assigned to `members`.
-              # For example, `roles/viewer`, `roles/editor`, or `roles/owner`.
-          "members": [ # Specifies the identities requesting access for a Cloud Platform resource.
+          &quot;condition&quot;: { # Represents a textual expression in the Common Expression Language (CEL) # The condition that is associated with this binding.
+              #
+              # If the condition evaluates to `true`, then this binding applies to the
+              # current request.
+              #
+              # If the condition evaluates to `false`, then this binding does not apply to
+              # the current request. However, a different role binding might grant the same
+              # role to one or more of the members in this binding.
+              #
+              # To learn which resources support conditions in their IAM policies, see the
+              # [IAM
+              # documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+              # syntax. CEL is a C-like expression language. The syntax and semantics of CEL
+              # are documented at https://github.com/google/cel-spec.
+              #
+              # Example (Comparison):
+              #
+              #     title: &quot;Summary size limit&quot;
+              #     description: &quot;Determines if a summary is less than 100 chars&quot;
+              #     expression: &quot;document.summary.size() &lt; 100&quot;
+              #
+              # Example (Equality):
+              #
+              #     title: &quot;Requestor is owner&quot;
+              #     description: &quot;Determines if requestor is the document owner&quot;
+              #     expression: &quot;document.owner == request.auth.claims.email&quot;
+              #
+              # Example (Logic):
+              #
+              #     title: &quot;Public documents&quot;
+              #     description: &quot;Determine whether the document should be publicly visible&quot;
+              #     expression: &quot;document.type != &#x27;private&#x27; &amp;&amp; document.type != &#x27;internal&#x27;&quot;
+              #
+              # Example (Data Manipulation):
+              #
+              #     title: &quot;Notification string&quot;
+              #     description: &quot;Create a notification string with a timestamp.&quot;
+              #     expression: &quot;&#x27;New message received at &#x27; + string(document.create_time)&quot;
+              #
+              # The exact variables and functions that may be referenced within an expression
+              # are determined by the service that evaluates it. See the service
+              # documentation for additional information.
+            &quot;title&quot;: &quot;A String&quot;, # Optional. Title for the expression, i.e. a short string describing
+                # its purpose. This can be used e.g. in UIs which allow to enter the
+                # expression.
+            &quot;location&quot;: &quot;A String&quot;, # Optional. String indicating the location of the expression for error
+                # reporting, e.g. a file name and a position in the file.
+            &quot;description&quot;: &quot;A String&quot;, # Optional. Description of the expression. This is a longer text which
+                # describes the expression, e.g. when hovered over it in a UI.
+            &quot;expression&quot;: &quot;A String&quot;, # Textual representation of an expression in Common Expression Language
+                # syntax.
+          },
+          &quot;members&quot;: [ # Specifies the identities requesting access for a Cloud Platform resource.
               # `members` can have the following values:
               #
               # * `allUsers`: A special identifier that represents anyone who is
@@ -4057,178 +4989,18 @@
               # * `domain:{domain}`: The G Suite domain (primary) that represents all the
               #    users of that domain. For example, `google.com` or `example.com`.
               #
-            "A String",
+            &quot;A String&quot;,
           ],
-          "condition": { # Represents a textual expression in the Common Expression Language (CEL) # The condition that is associated with this binding.
-              # NOTE: An unsatisfied condition will not allow user access via current
-              # binding. Different bindings, including their conditions, are examined
-              # independently.
-              # syntax. CEL is a C-like expression language. The syntax and semantics of CEL
-              # are documented at https://github.com/google/cel-spec.
-              #
-              # Example (Comparison):
-              #
-              #     title: "Summary size limit"
-              #     description: "Determines if a summary is less than 100 chars"
-              #     expression: "document.summary.size() &lt; 100"
-              #
-              # Example (Equality):
-              #
-              #     title: "Requestor is owner"
-              #     description: "Determines if requestor is the document owner"
-              #     expression: "document.owner == request.auth.claims.email"
-              #
-              # Example (Logic):
-              #
-              #     title: "Public documents"
-              #     description: "Determine whether the document should be publicly visible"
-              #     expression: "document.type != 'private' &amp;&amp; document.type != 'internal'"
-              #
-              # Example (Data Manipulation):
-              #
-              #     title: "Notification string"
-              #     description: "Create a notification string with a timestamp."
-              #     expression: "'New message received at ' + string(document.create_time)"
-              #
-              # The exact variables and functions that may be referenced within an expression
-              # are determined by the service that evaluates it. See the service
-              # documentation for additional information.
-            "description": "A String", # Optional. Description of the expression. This is a longer text which
-                # describes the expression, e.g. when hovered over it in a UI.
-            "expression": "A String", # Textual representation of an expression in Common Expression Language
-                # syntax.
-            "location": "A String", # Optional. String indicating the location of the expression for error
-                # reporting, e.g. a file name and a position in the file.
-            "title": "A String", # Optional. Title for the expression, i.e. a short string describing
-                # its purpose. This can be used e.g. in UIs which allow to enter the
-                # expression.
-          },
+          &quot;role&quot;: &quot;A String&quot;, # Role that is assigned to `members`.
+              # For example, `roles/viewer`, `roles/editor`, or `roles/owner`.
         },
       ],
-      "auditConfigs": [ # Specifies cloud audit logging configuration for this policy.
-        { # Specifies the audit configuration for a service.
-            # The configuration determines which permission types are logged, and what
-            # identities, if any, are exempted from logging.
-            # An AuditConfig must have one or more AuditLogConfigs.
-            #
-            # If there are AuditConfigs for both `allServices` and a specific service,
-            # the union of the two AuditConfigs is used for that service: the log_types
-            # specified in each AuditConfig are enabled, and the exempted_members in each
-            # AuditLogConfig are exempted.
-            #
-            # Example Policy with multiple AuditConfigs:
-            #
-            #     {
-            #       "audit_configs": [
-            #         {
-            #           "service": "allServices"
-            #           "audit_log_configs": [
-            #             {
-            #               "log_type": "DATA_READ",
-            #               "exempted_members": [
-            #                 "user:jose@example.com"
-            #               ]
-            #             },
-            #             {
-            #               "log_type": "DATA_WRITE",
-            #             },
-            #             {
-            #               "log_type": "ADMIN_READ",
-            #             }
-            #           ]
-            #         },
-            #         {
-            #           "service": "sampleservice.googleapis.com"
-            #           "audit_log_configs": [
-            #             {
-            #               "log_type": "DATA_READ",
-            #             },
-            #             {
-            #               "log_type": "DATA_WRITE",
-            #               "exempted_members": [
-            #                 "user:aliya@example.com"
-            #               ]
-            #             }
-            #           ]
-            #         }
-            #       ]
-            #     }
-            #
-            # For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ
-            # logging. It also exempts jose@example.com from DATA_READ logging, and
-            # aliya@example.com from DATA_WRITE logging.
-          "auditLogConfigs": [ # The configuration for logging of each type of permission.
-            { # Provides the configuration for logging a type of permissions.
-                # Example:
-                #
-                #     {
-                #       "audit_log_configs": [
-                #         {
-                #           "log_type": "DATA_READ",
-                #           "exempted_members": [
-                #             "user:jose@example.com"
-                #           ]
-                #         },
-                #         {
-                #           "log_type": "DATA_WRITE",
-                #         }
-                #       ]
-                #     }
-                #
-                # This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting
-                # jose@example.com from DATA_READ logging.
-              "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of
-                  # permission.
-                  # Follows the same format of Binding.members.
-                "A String",
-              ],
-              "logType": "A String", # The log type that this config enables.
-            },
-          ],
-          "service": "A String", # Specifies a service that will be enabled for audit logging.
-              # For example, `storage.googleapis.com`, `cloudsql.googleapis.com`.
-              # `allServices` is a special value that covers all services.
-        },
-      ],
-      "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help
-          # prevent simultaneous updates of a policy from overwriting each other.
-          # It is strongly suggested that systems make use of the `etag` in the
-          # read-modify-write cycle to perform policy updates in order to avoid race
-          # conditions: An `etag` is returned in the response to `getIamPolicy`, and
-          # systems are expected to put that etag in the request to `setIamPolicy` to
-          # ensure that their change will be applied to the same version of the policy.
-          #
-          # **Important:** If you use IAM Conditions, you must include the `etag` field
-          # whenever you call `setIamPolicy`. If you omit this field, then IAM allows
-          # you to overwrite a version `3` policy with a version `1` policy, and all of
-          # the conditions in the version `3` policy are lost.
-      "version": 42, # Specifies the format of the policy.
-          #
-          # Valid values are `0`, `1`, and `3`. Requests that specify an invalid value
-          # are rejected.
-          #
-          # Any operation that affects conditional role bindings must specify version
-          # `3`. This requirement applies to the following operations:
-          #
-          # * Getting a policy that includes a conditional role binding
-          # * Adding a conditional role binding to a policy
-          # * Changing a conditional role binding in a policy
-          # * Removing any role binding, with or without a condition, from a policy
-          #   that includes conditions
-          #
-          # **Important:** If you use IAM Conditions, you must include the `etag` field
-          # whenever you call `setIamPolicy`. If you omit this field, then IAM allows
-          # you to overwrite a version `3` policy with a version `1` policy, and all of
-          # the conditions in the version `3` policy are lost.
-          #
-          # If a policy does not include any conditions, operations on that policy may
-          # specify any valid version or leave the field unset.
     },
-    "updateMask": "A String", # OPTIONAL: A FieldMask specifying which fields of the policy to modify. Only
+    &quot;updateMask&quot;: &quot;A String&quot;, # OPTIONAL: A FieldMask specifying which fields of the policy to modify. Only
         # the fields in the mask will be modified. If no mask is provided, the
         # following default mask is used:
-        # paths: "bindings, etag"
-        # This field is only used by Cloud IAM.
+        # 
+        # `paths: &quot;bindings, etag&quot;`
   }
 
   x__xgafv: string, V1 error format.
@@ -4249,36 +5021,40 @@
       # permissions; each `role` can be an IAM predefined role or a user-created
       # custom role.
       #
-      # Optionally, a `binding` can specify a `condition`, which is a logical
-      # expression that allows access to a resource only if the expression evaluates
-      # to `true`. A condition can add constraints based on attributes of the
-      # request, the resource, or both.
+      # For some types of Google Cloud resources, a `binding` can also specify a
+      # `condition`, which is a logical expression that allows access to a resource
+      # only if the expression evaluates to `true`. A condition can add constraints
+      # based on attributes of the request, the resource, or both. To learn which
+      # resources support conditions in their IAM policies, see the
+      # [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
       #
       # **JSON example:**
       #
       #     {
-      #       "bindings": [
+      #       &quot;bindings&quot;: [
       #         {
-      #           "role": "roles/resourcemanager.organizationAdmin",
-      #           "members": [
-      #             "user:mike@example.com",
-      #             "group:admins@example.com",
-      #             "domain:google.com",
-      #             "serviceAccount:my-project-id@appspot.gserviceaccount.com"
+      #           &quot;role&quot;: &quot;roles/resourcemanager.organizationAdmin&quot;,
+      #           &quot;members&quot;: [
+      #             &quot;user:mike@example.com&quot;,
+      #             &quot;group:admins@example.com&quot;,
+      #             &quot;domain:google.com&quot;,
+      #             &quot;serviceAccount:my-project-id@appspot.gserviceaccount.com&quot;
       #           ]
       #         },
       #         {
-      #           "role": "roles/resourcemanager.organizationViewer",
-      #           "members": ["user:eve@example.com"],
-      #           "condition": {
-      #             "title": "expirable access",
-      #             "description": "Does not grant access after Sep 2020",
-      #             "expression": "request.time &lt; timestamp('2020-10-01T00:00:00.000Z')",
+      #           &quot;role&quot;: &quot;roles/resourcemanager.organizationViewer&quot;,
+      #           &quot;members&quot;: [
+      #             &quot;user:eve@example.com&quot;
+      #           ],
+      #           &quot;condition&quot;: {
+      #             &quot;title&quot;: &quot;expirable access&quot;,
+      #             &quot;description&quot;: &quot;Does not grant access after Sep 2020&quot;,
+      #             &quot;expression&quot;: &quot;request.time &lt; timestamp(&#x27;2020-10-01T00:00:00.000Z&#x27;)&quot;,
       #           }
       #         }
       #       ],
-      #       "etag": "BwWWja0YfJA=",
-      #       "version": 3
+      #       &quot;etag&quot;: &quot;BwWWja0YfJA=&quot;,
+      #       &quot;version&quot;: 3
       #     }
       #
       # **YAML example:**
@@ -4296,19 +5072,190 @@
       #       condition:
       #         title: expirable access
       #         description: Does not grant access after Sep 2020
-      #         expression: request.time &lt; timestamp('2020-10-01T00:00:00.000Z')
+      #         expression: request.time &lt; timestamp(&#x27;2020-10-01T00:00:00.000Z&#x27;)
       #     - etag: BwWWja0YfJA=
       #     - version: 3
       #
       # For a description of IAM and its features, see the
       # [IAM documentation](https://cloud.google.com/iam/docs/).
-    "bindings": [ # Associates a list of `members` to a `role`. Optionally, may specify a
+    &quot;etag&quot;: &quot;A String&quot;, # `etag` is used for optimistic concurrency control as a way to help
+        # prevent simultaneous updates of a policy from overwriting each other.
+        # It is strongly suggested that systems make use of the `etag` in the
+        # read-modify-write cycle to perform policy updates in order to avoid race
+        # conditions: An `etag` is returned in the response to `getIamPolicy`, and
+        # systems are expected to put that etag in the request to `setIamPolicy` to
+        # ensure that their change will be applied to the same version of the policy.
+        #
+        # **Important:** If you use IAM Conditions, you must include the `etag` field
+        # whenever you call `setIamPolicy`. If you omit this field, then IAM allows
+        # you to overwrite a version `3` policy with a version `1` policy, and all of
+        # the conditions in the version `3` policy are lost.
+    &quot;version&quot;: 42, # Specifies the format of the policy.
+        #
+        # Valid values are `0`, `1`, and `3`. Requests that specify an invalid value
+        # are rejected.
+        #
+        # Any operation that affects conditional role bindings must specify version
+        # `3`. This requirement applies to the following operations:
+        #
+        # * Getting a policy that includes a conditional role binding
+        # * Adding a conditional role binding to a policy
+        # * Changing a conditional role binding in a policy
+        # * Removing any role binding, with or without a condition, from a policy
+        #   that includes conditions
+        #
+        # **Important:** If you use IAM Conditions, you must include the `etag` field
+        # whenever you call `setIamPolicy`. If you omit this field, then IAM allows
+        # you to overwrite a version `3` policy with a version `1` policy, and all of
+        # the conditions in the version `3` policy are lost.
+        #
+        # If a policy does not include any conditions, operations on that policy may
+        # specify any valid version or leave the field unset.
+        #
+        # To learn which resources support conditions in their IAM policies, see the
+        # [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+    &quot;auditConfigs&quot;: [ # Specifies cloud audit logging configuration for this policy.
+      { # Specifies the audit configuration for a service.
+          # The configuration determines which permission types are logged, and what
+          # identities, if any, are exempted from logging.
+          # An AuditConfig must have one or more AuditLogConfigs.
+          #
+          # If there are AuditConfigs for both `allServices` and a specific service,
+          # the union of the two AuditConfigs is used for that service: the log_types
+          # specified in each AuditConfig are enabled, and the exempted_members in each
+          # AuditLogConfig are exempted.
+          #
+          # Example Policy with multiple AuditConfigs:
+          #
+          #     {
+          #       &quot;audit_configs&quot;: [
+          #         {
+          #           &quot;service&quot;: &quot;allServices&quot;
+          #           &quot;audit_log_configs&quot;: [
+          #             {
+          #               &quot;log_type&quot;: &quot;DATA_READ&quot;,
+          #               &quot;exempted_members&quot;: [
+          #                 &quot;user:jose@example.com&quot;
+          #               ]
+          #             },
+          #             {
+          #               &quot;log_type&quot;: &quot;DATA_WRITE&quot;,
+          #             },
+          #             {
+          #               &quot;log_type&quot;: &quot;ADMIN_READ&quot;,
+          #             }
+          #           ]
+          #         },
+          #         {
+          #           &quot;service&quot;: &quot;sampleservice.googleapis.com&quot;
+          #           &quot;audit_log_configs&quot;: [
+          #             {
+          #               &quot;log_type&quot;: &quot;DATA_READ&quot;,
+          #             },
+          #             {
+          #               &quot;log_type&quot;: &quot;DATA_WRITE&quot;,
+          #               &quot;exempted_members&quot;: [
+          #                 &quot;user:aliya@example.com&quot;
+          #               ]
+          #             }
+          #           ]
+          #         }
+          #       ]
+          #     }
+          #
+          # For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ
+          # logging. It also exempts jose@example.com from DATA_READ logging, and
+          # aliya@example.com from DATA_WRITE logging.
+        &quot;service&quot;: &quot;A String&quot;, # Specifies a service that will be enabled for audit logging.
+            # For example, `storage.googleapis.com`, `cloudsql.googleapis.com`.
+            # `allServices` is a special value that covers all services.
+        &quot;auditLogConfigs&quot;: [ # The configuration for logging of each type of permission.
+          { # Provides the configuration for logging a type of permissions.
+              # Example:
+              #
+              #     {
+              #       &quot;audit_log_configs&quot;: [
+              #         {
+              #           &quot;log_type&quot;: &quot;DATA_READ&quot;,
+              #           &quot;exempted_members&quot;: [
+              #             &quot;user:jose@example.com&quot;
+              #           ]
+              #         },
+              #         {
+              #           &quot;log_type&quot;: &quot;DATA_WRITE&quot;,
+              #         }
+              #       ]
+              #     }
+              #
+              # This enables &#x27;DATA_READ&#x27; and &#x27;DATA_WRITE&#x27; logging, while exempting
+              # jose@example.com from DATA_READ logging.
+            &quot;exemptedMembers&quot;: [ # Specifies the identities that do not cause logging for this type of
+                # permission.
+                # Follows the same format of Binding.members.
+              &quot;A String&quot;,
+            ],
+            &quot;logType&quot;: &quot;A String&quot;, # The log type that this config enables.
+          },
+        ],
+      },
+    ],
+    &quot;bindings&quot;: [ # Associates a list of `members` to a `role`. Optionally, may specify a
         # `condition` that determines how and when the `bindings` are applied. Each
         # of the `bindings` must contain at least one member.
       { # Associates `members` with a `role`.
-        "role": "A String", # Role that is assigned to `members`.
-            # For example, `roles/viewer`, `roles/editor`, or `roles/owner`.
-        "members": [ # Specifies the identities requesting access for a Cloud Platform resource.
+        &quot;condition&quot;: { # Represents a textual expression in the Common Expression Language (CEL) # The condition that is associated with this binding.
+            #
+            # If the condition evaluates to `true`, then this binding applies to the
+            # current request.
+            #
+            # If the condition evaluates to `false`, then this binding does not apply to
+            # the current request. However, a different role binding might grant the same
+            # role to one or more of the members in this binding.
+            #
+            # To learn which resources support conditions in their IAM policies, see the
+            # [IAM
+            # documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+            # syntax. CEL is a C-like expression language. The syntax and semantics of CEL
+            # are documented at https://github.com/google/cel-spec.
+            #
+            # Example (Comparison):
+            #
+            #     title: &quot;Summary size limit&quot;
+            #     description: &quot;Determines if a summary is less than 100 chars&quot;
+            #     expression: &quot;document.summary.size() &lt; 100&quot;
+            #
+            # Example (Equality):
+            #
+            #     title: &quot;Requestor is owner&quot;
+            #     description: &quot;Determines if requestor is the document owner&quot;
+            #     expression: &quot;document.owner == request.auth.claims.email&quot;
+            #
+            # Example (Logic):
+            #
+            #     title: &quot;Public documents&quot;
+            #     description: &quot;Determine whether the document should be publicly visible&quot;
+            #     expression: &quot;document.type != &#x27;private&#x27; &amp;&amp; document.type != &#x27;internal&#x27;&quot;
+            #
+            # Example (Data Manipulation):
+            #
+            #     title: &quot;Notification string&quot;
+            #     description: &quot;Create a notification string with a timestamp.&quot;
+            #     expression: &quot;&#x27;New message received at &#x27; + string(document.create_time)&quot;
+            #
+            # The exact variables and functions that may be referenced within an expression
+            # are determined by the service that evaluates it. See the service
+            # documentation for additional information.
+          &quot;title&quot;: &quot;A String&quot;, # Optional. Title for the expression, i.e. a short string describing
+              # its purpose. This can be used e.g. in UIs which allow to enter the
+              # expression.
+          &quot;location&quot;: &quot;A String&quot;, # Optional. String indicating the location of the expression for error
+              # reporting, e.g. a file name and a position in the file.
+          &quot;description&quot;: &quot;A String&quot;, # Optional. Description of the expression. This is a longer text which
+              # describes the expression, e.g. when hovered over it in a UI.
+          &quot;expression&quot;: &quot;A String&quot;, # Textual representation of an expression in Common Expression Language
+              # syntax.
+        },
+        &quot;members&quot;: [ # Specifies the identities requesting access for a Cloud Platform resource.
             # `members` can have the following values:
             #
             # * `allUsers`: A special identifier that represents anyone who is
@@ -4351,172 +5298,12 @@
             # * `domain:{domain}`: The G Suite domain (primary) that represents all the
             #    users of that domain. For example, `google.com` or `example.com`.
             #
-          "A String",
+          &quot;A String&quot;,
         ],
-        "condition": { # Represents a textual expression in the Common Expression Language (CEL) # The condition that is associated with this binding.
-            # NOTE: An unsatisfied condition will not allow user access via current
-            # binding. Different bindings, including their conditions, are examined
-            # independently.
-            # syntax. CEL is a C-like expression language. The syntax and semantics of CEL
-            # are documented at https://github.com/google/cel-spec.
-            #
-            # Example (Comparison):
-            #
-            #     title: "Summary size limit"
-            #     description: "Determines if a summary is less than 100 chars"
-            #     expression: "document.summary.size() &lt; 100"
-            #
-            # Example (Equality):
-            #
-            #     title: "Requestor is owner"
-            #     description: "Determines if requestor is the document owner"
-            #     expression: "document.owner == request.auth.claims.email"
-            #
-            # Example (Logic):
-            #
-            #     title: "Public documents"
-            #     description: "Determine whether the document should be publicly visible"
-            #     expression: "document.type != 'private' &amp;&amp; document.type != 'internal'"
-            #
-            # Example (Data Manipulation):
-            #
-            #     title: "Notification string"
-            #     description: "Create a notification string with a timestamp."
-            #     expression: "'New message received at ' + string(document.create_time)"
-            #
-            # The exact variables and functions that may be referenced within an expression
-            # are determined by the service that evaluates it. See the service
-            # documentation for additional information.
-          "description": "A String", # Optional. Description of the expression. This is a longer text which
-              # describes the expression, e.g. when hovered over it in a UI.
-          "expression": "A String", # Textual representation of an expression in Common Expression Language
-              # syntax.
-          "location": "A String", # Optional. String indicating the location of the expression for error
-              # reporting, e.g. a file name and a position in the file.
-          "title": "A String", # Optional. Title for the expression, i.e. a short string describing
-              # its purpose. This can be used e.g. in UIs which allow to enter the
-              # expression.
-        },
+        &quot;role&quot;: &quot;A String&quot;, # Role that is assigned to `members`.
+            # For example, `roles/viewer`, `roles/editor`, or `roles/owner`.
       },
     ],
-    "auditConfigs": [ # Specifies cloud audit logging configuration for this policy.
-      { # Specifies the audit configuration for a service.
-          # The configuration determines which permission types are logged, and what
-          # identities, if any, are exempted from logging.
-          # An AuditConfig must have one or more AuditLogConfigs.
-          #
-          # If there are AuditConfigs for both `allServices` and a specific service,
-          # the union of the two AuditConfigs is used for that service: the log_types
-          # specified in each AuditConfig are enabled, and the exempted_members in each
-          # AuditLogConfig are exempted.
-          #
-          # Example Policy with multiple AuditConfigs:
-          #
-          #     {
-          #       "audit_configs": [
-          #         {
-          #           "service": "allServices"
-          #           "audit_log_configs": [
-          #             {
-          #               "log_type": "DATA_READ",
-          #               "exempted_members": [
-          #                 "user:jose@example.com"
-          #               ]
-          #             },
-          #             {
-          #               "log_type": "DATA_WRITE",
-          #             },
-          #             {
-          #               "log_type": "ADMIN_READ",
-          #             }
-          #           ]
-          #         },
-          #         {
-          #           "service": "sampleservice.googleapis.com"
-          #           "audit_log_configs": [
-          #             {
-          #               "log_type": "DATA_READ",
-          #             },
-          #             {
-          #               "log_type": "DATA_WRITE",
-          #               "exempted_members": [
-          #                 "user:aliya@example.com"
-          #               ]
-          #             }
-          #           ]
-          #         }
-          #       ]
-          #     }
-          #
-          # For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ
-          # logging. It also exempts jose@example.com from DATA_READ logging, and
-          # aliya@example.com from DATA_WRITE logging.
-        "auditLogConfigs": [ # The configuration for logging of each type of permission.
-          { # Provides the configuration for logging a type of permissions.
-              # Example:
-              #
-              #     {
-              #       "audit_log_configs": [
-              #         {
-              #           "log_type": "DATA_READ",
-              #           "exempted_members": [
-              #             "user:jose@example.com"
-              #           ]
-              #         },
-              #         {
-              #           "log_type": "DATA_WRITE",
-              #         }
-              #       ]
-              #     }
-              #
-              # This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting
-              # jose@example.com from DATA_READ logging.
-            "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of
-                # permission.
-                # Follows the same format of Binding.members.
-              "A String",
-            ],
-            "logType": "A String", # The log type that this config enables.
-          },
-        ],
-        "service": "A String", # Specifies a service that will be enabled for audit logging.
-            # For example, `storage.googleapis.com`, `cloudsql.googleapis.com`.
-            # `allServices` is a special value that covers all services.
-      },
-    ],
-    "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help
-        # prevent simultaneous updates of a policy from overwriting each other.
-        # It is strongly suggested that systems make use of the `etag` in the
-        # read-modify-write cycle to perform policy updates in order to avoid race
-        # conditions: An `etag` is returned in the response to `getIamPolicy`, and
-        # systems are expected to put that etag in the request to `setIamPolicy` to
-        # ensure that their change will be applied to the same version of the policy.
-        #
-        # **Important:** If you use IAM Conditions, you must include the `etag` field
-        # whenever you call `setIamPolicy`. If you omit this field, then IAM allows
-        # you to overwrite a version `3` policy with a version `1` policy, and all of
-        # the conditions in the version `3` policy are lost.
-    "version": 42, # Specifies the format of the policy.
-        #
-        # Valid values are `0`, `1`, and `3`. Requests that specify an invalid value
-        # are rejected.
-        #
-        # Any operation that affects conditional role bindings must specify version
-        # `3`. This requirement applies to the following operations:
-        #
-        # * Getting a policy that includes a conditional role binding
-        # * Adding a conditional role binding to a policy
-        # * Changing a conditional role binding in a policy
-        # * Removing any role binding, with or without a condition, from a policy
-        #   that includes conditions
-        #
-        # **Important:** If you use IAM Conditions, you must include the `etag` field
-        # whenever you call `setIamPolicy`. If you omit this field, then IAM allows
-        # you to overwrite a version `3` policy with a version `1` policy, and all of
-        # the conditions in the version `3` policy are lost.
-        #
-        # If a policy does not include any conditions, operations on that policy may
-        # specify any valid version or leave the field unset.
   }</pre>
 </div>
 
@@ -4524,11 +5311,11 @@
     <code class="details" id="testIamPermissions">testIamPermissions(resource, body=None, x__xgafv=None)</code>
   <pre>Returns permissions that a caller has on the specified resource.
 If the resource does not exist, this will return an empty set of
-permissions, not a NOT_FOUND error.
+permissions, not a `NOT_FOUND` error.
 
 Note: This operation is designed to be used for building permission-aware
 UIs and command-line tools, not for authorization checking. This operation
-may "fail open" without warning.
+may &quot;fail open&quot; without warning.
 
 Args:
   resource: string, REQUIRED: The resource for which the policy detail is being requested.
@@ -4537,11 +5324,11 @@
     The object takes the form of:
 
 { # Request message for `TestIamPermissions` method.
-    "permissions": [ # The set of permissions to check for the `resource`. Permissions with
-        # wildcards (such as '*' or 'storage.*') are not allowed. For more
+    &quot;permissions&quot;: [ # The set of permissions to check for the `resource`. Permissions with
+        # wildcards (such as &#x27;*&#x27; or &#x27;storage.*&#x27;) are not allowed. For more
         # information see
         # [IAM Overview](https://cloud.google.com/iam/docs/overview#permissions).
-      "A String",
+      &quot;A String&quot;,
     ],
   }
 
@@ -4554,9 +5341,9 @@
   An object of the form:
 
     { # Response message for `TestIamPermissions` method.
-    "permissions": [ # A subset of `TestPermissionsRequest.permissions` that the caller is
+    &quot;permissions&quot;: [ # A subset of `TestPermissionsRequest.permissions` that the caller is
         # allowed.
-      "A String",
+      &quot;A String&quot;,
     ],
   }</pre>
 </div>