Patch

Updates a specific job resource

104 variables
102 variables

Updates a specific job resource.

Currently the only supported fields to update are labels

Authorization

To use this building block you will have to grant access to at least one of the following scopes:

  • View and manage your data across Google Cloud Platform services

Input

This building block consumes 104 input parameters

  = Parameter name
  = Format

name STRING Required

Required. The job name

updateMask ANY

Required. Specifies the path, relative to Job, of the field to update. To adopt etag mechanism, include etag field in the mask, and include the etag value in your job resource.

For example, to change the labels of a job, the update_mask parameter would be specified as labels, etag, and the PATCH request body would specify the new value, as follows: { "labels": { "owner": "Google", "color": "Blue" } "etag": "33a64df551425fcc55e4d42a148795d9f25f89d4" } If etag matches the one on the server, the labels of the job will be replaced with the given ones, and the server end etag will be recalculated.

Currently the only supported update masks are labels and etag

trainingOutput OBJECT

Represents results of a training job. Output only

trainingOutput.isBuiltInAlgorithmJob BOOLEAN

Whether this job is a built-in Algorithm job

trainingOutput.builtInAlgorithmOutput OBJECT

Represents output related to a built-in algorithm Job

trainingOutput.builtInAlgorithmOutput.pythonVersion STRING

Python version on which the built-in algorithm was trained

trainingOutput.builtInAlgorithmOutput.runtimeVersion STRING

AI Platform runtime version on which the built-in algorithm was trained

trainingOutput.builtInAlgorithmOutput.framework STRING

Framework on which the built-in algorithm was trained

trainingOutput.builtInAlgorithmOutput.modelPath STRING

The Cloud Storage path to the model/ directory where the training job saves the trained model. Only set for successful jobs that don't use hyperparameter tuning

trainingOutput.trials[] OBJECT

Represents the result of a single hyperparameter tuning trial from a training job. The TrainingOutput object that is returned on successful completion of a training job with hyperparameter tuning includes a list of HyperparameterOutput objects, one for each successful trial

trainingOutput.trials[].hyperparameters OBJECT

The hyperparameters given to this trial

trainingOutput.trials[].hyperparameters.customKey.value STRING Required

The hyperparameters given to this trial

trainingOutput.trials[].trialId STRING

The trial id for these results

trainingOutput.trials[].endTime ANY

Output only. End time for the trial

trainingOutput.trials[].isTrialStoppedEarly BOOLEAN

True if the trial is stopped early

trainingOutput.trials[].startTime ANY

Output only. Start time for the trial

trainingOutput.trials[].finalMetric OBJECT

An observed value of a metric

trainingOutput.trials[].finalMetric.trainingStep INTEGER

The global training step for this metric

trainingOutput.trials[].finalMetric.objectiveValue NUMBER

The objective value at this training step

trainingOutput.trials[].builtInAlgorithmOutput OBJECT

Represents output related to a built-in algorithm Job

trainingOutput.trials[].builtInAlgorithmOutput.pythonVersion STRING

Python version on which the built-in algorithm was trained

trainingOutput.trials[].builtInAlgorithmOutput.runtimeVersion STRING

AI Platform runtime version on which the built-in algorithm was trained

trainingOutput.trials[].builtInAlgorithmOutput.framework STRING

Framework on which the built-in algorithm was trained

trainingOutput.trials[].builtInAlgorithmOutput.modelPath STRING

The Cloud Storage path to the model/ directory where the training job saves the trained model. Only set for successful jobs that don't use hyperparameter tuning

trainingOutput.trials[].state ENUMERATION

Output only. The detailed state of the trial

trainingOutput.trials[].allMetrics[] OBJECT

An observed value of a metric

trainingOutput.hyperparameterMetricTag STRING

The TensorFlow summary tag name used for optimizing hyperparameter tuning trials. See HyperparameterSpec.hyperparameterMetricTag for more information. Only set for hyperparameter tuning jobs

trainingOutput.completedTrialCount INTEGER

The number of hyperparameter tuning trials that completed successfully. Only set for hyperparameter tuning jobs

trainingOutput.isHyperparameterTuningJob BOOLEAN

Whether this job is a hyperparameter tuning job

trainingOutput.consumedMLUnits NUMBER

The amount of ML units consumed by the job

createTime ANY

Output only. When the job was created

labels OBJECT

Optional. One or more labels that you can add, to organize your jobs. Each label is a key-value pair, where both the key and the value are arbitrary strings that you supply. For more information, see the documentation on using labels

labels.customKey.value STRING Required

Optional. One or more labels that you can add, to organize your jobs. Each label is a key-value pair, where both the key and the value are arbitrary strings that you supply. For more information, see the documentation on using labels

predictionInput OBJECT

Represents input parameters for a prediction job

predictionInput.outputPath STRING

Required. The output Google Cloud Storage location

predictionInput.outputDataFormat ENUMERATION

Optional. Format of the output data files, defaults to JSON

predictionInput.dataFormat ENUMERATION

Required. The format of the input data files

predictionInput.batchSize INTEGER

Optional. Number of records per batch, defaults to 64. The service will buffer batch_size number of records in memory before invoking one Tensorflow prediction call internally. So take the record size and memory available into consideration when setting this parameter

predictionInput.runtimeVersion STRING

Optional. The AI Platform runtime version to use for this batch prediction. If not set, AI Platform will pick the runtime version used during the CreateVersion request for this model version, or choose the latest stable version when model version information is not available such as when the model is specified by uri

predictionInput.inputPaths[] STRING

predictionInput.region STRING

Required. The Google Compute Engine region to run the prediction job in. See the available regions for AI Platform services

predictionInput.versionName STRING

Use this field if you want to specify a version of the model to use. The string is formatted the same way as model_version, with the addition of the version information:

"projects/YOUR_PROJECT/models/YOUR_MODEL/versions/YOUR_VERSION"

predictionInput.modelName STRING

Use this field if you want to use the default version for the specified model. The string must use the following format:

"projects/YOUR_PROJECT/models/YOUR_MODEL"

predictionInput.uri STRING

Use this field if you want to specify a Google Cloud Storage path for the model to use

predictionInput.maxWorkerCount INTEGER

Optional. The maximum number of workers to be used for parallel processing. Defaults to 10 if not specified

predictionInput.signatureName STRING

Optional. The name of the signature defined in the SavedModel to use for this job. Please refer to SavedModel for information about how to use signatures.

Defaults to DEFAULT_SERVING_SIGNATURE_DEF_KEY , which is "serving_default"

errorMessage STRING

Output only. The details of a failure or a cancellation

etag BINARY

etag is used for optimistic concurrency control as a way to help prevent simultaneous updates of a job from overwriting each other. It is strongly suggested that systems make use of the etag in the read-modify-write cycle to perform job updates in order to avoid race conditions: An etag is returned in the response to GetJob, and systems are expected to put that etag in the request to UpdateJob to ensure that their change will be applied to the same version of the job

trainingInput OBJECT

Represents input parameters for a training job. When using the gcloud command to submit your training job, you can specify the input parameters as command-line arguments and/or in a YAML configuration file referenced from the --config command-line argument. For details, see the guide to submitting a training job

trainingInput.workerCount INTEGER

Optional. The number of worker replicas to use for the training job. Each replica in the cluster will be of the type specified in worker_type.

This value can only be used when scale_tier is set to CUSTOM. If you set this value, you must also set worker_type.

The default value is zero

trainingInput.masterType STRING

Optional. Specifies the type of virtual machine to use for your training job's master worker.

The following types are supported:

<dl> <dt>standard</dt> <dd> A basic machine configuration suitable for training simple models with small to moderate datasets. </dd> <dt>large_model</dt> <dd> A machine with a lot of memory, specially suited for parameter servers when your model is large (having many hidden layers or layers with very large numbers of nodes). </dd> <dt>complex_model_s</dt> <dd> A machine suitable for the master and workers of the cluster when your model requires more computation than the standard machine can handle satisfactorily. </dd> <dt>complex_model_m</dt> <dd> A machine with roughly twice the number of cores and roughly double the memory of <i>complex_model_s</i>. </dd> <dt>complex_model_l</dt> <dd> A machine with roughly twice the number of cores and roughly double the memory of <i>complex_model_m</i>. </dd> <dt>standard_gpu</dt> <dd> A machine equivalent to <i>standard</i> that also includes a single NVIDIA Tesla K80 GPU. See more about <a href="/ml-engine/docs/tensorflow/using-gpus">using GPUs to train your model</a>. </dd> <dt>complex_model_m_gpu</dt> <dd> A machine equivalent to <i>complex_model_m</i> that also includes four NVIDIA Tesla K80 GPUs. </dd> <dt>complex_model_l_gpu</dt> <dd> A machine equivalent to <i>complex_model_l</i> that also includes eight NVIDIA Tesla K80 GPUs. </dd> <dt>standard_p100</dt> <dd> A machine equivalent to <i>standard</i> that also includes a single NVIDIA Tesla P100 GPU. </dd> <dt>complex_model_m_p100</dt> <dd> A machine equivalent to <i>complex_model_m</i> that also includes four NVIDIA Tesla P100 GPUs. </dd> <dt>standard_v100</dt> <dd> A machine equivalent to <i>standard</i> that also includes a single NVIDIA Tesla V100 GPU. </dd> <dt>large_model_v100</dt> <dd> A machine equivalent to <i>large_model</i> that also includes a single NVIDIA Tesla V100 GPU. </dd> <dt>complex_model_m_v100</dt> <dd> A machine equivalent to <i>complex_model_m</i> that also includes four NVIDIA Tesla V100 GPUs. </dd> <dt>complex_model_l_v100</dt> <dd> A machine equivalent to <i>complex_model_l</i> that also includes eight NVIDIA Tesla V100 GPUs. </dd> <dt>cloud_tpu</dt> <dd> A TPU VM including one Cloud TPU. See more about <a href="/ml-engine/docs/tensorflow/using-tpus">using TPUs to train your model</a>. </dd> </dl>

You may also use certain Compute Engine machine types directly in this field. The following types are supported:

  • n1-standard-4
  • n1-standard-8
  • n1-standard-16
  • n1-standard-32
  • n1-standard-64
  • n1-standard-96
  • n1-highmem-2
  • n1-highmem-4
  • n1-highmem-8
  • n1-highmem-16
  • n1-highmem-32
  • n1-highmem-64
  • n1-highmem-96
  • n1-highcpu-16
  • n1-highcpu-32
  • n1-highcpu-64
  • n1-highcpu-96

See more about using Compute Engine machine types.

You must set this value when scaleTier is set to CUSTOM

trainingInput.maxRunningTime ANY

Optional. The maximum job running time. The default is 7 days

trainingInput.runtimeVersion STRING

Optional. The AI Platform runtime version to use for training. If not set, AI Platform uses the default stable version, 1.0. For more information, see the runtime version list and how to manage runtime versions

trainingInput.pythonModule STRING

Required. The Python module name to run after installing the packages

trainingInput.args[] STRING

trainingInput.region STRING

Required. The Google Compute Engine region to run the training job in. See the available regions for AI Platform services

trainingInput.workerType STRING

Optional. Specifies the type of virtual machine to use for your training job's worker nodes.

The supported values are the same as those described in the entry for masterType.

This value must be consistent with the category of machine type that masterType uses. In other words, both must be AI Platform machine types or both must be Compute Engine machine types.

If you use cloud_tpu for this value, see special instructions for configuring a custom TPU machine.

This value must be present when scaleTier is set to CUSTOM and workerCount is greater than zero

trainingInput.parameterServerType STRING

Optional. Specifies the type of virtual machine to use for your training job's parameter server.

The supported values are the same as those described in the entry for master_type.

This value must be consistent with the category of machine type that masterType uses. In other words, both must be AI Platform machine types or both must be Compute Engine machine types.

This value must be present when scaleTier is set to CUSTOM and parameter_server_count is greater than zero

trainingInput.parameterServerConfig OBJECT

Represents the configuration for a replica in a cluster

trainingInput.parameterServerConfig.acceleratorConfig OBJECT

Represents a hardware accelerator request config

trainingInput.parameterServerConfig.acceleratorConfig.type ENUMERATION

The type of accelerator to use

trainingInput.parameterServerConfig.acceleratorConfig.count INTEGER

The number of accelerators to attach to each machine running the job

trainingInput.parameterServerConfig.imageUri STRING

The Docker image to run on the replica. This image must be in Container Registry. Learn more about configuring custom containers

trainingInput.masterConfig OBJECT

Represents the configuration for a replica in a cluster

trainingInput.masterConfig.acceleratorConfig OBJECT

Represents a hardware accelerator request config

trainingInput.masterConfig.acceleratorConfig.type ENUMERATION

The type of accelerator to use

trainingInput.masterConfig.acceleratorConfig.count INTEGER

The number of accelerators to attach to each machine running the job

trainingInput.masterConfig.imageUri STRING

The Docker image to run on the replica. This image must be in Container Registry. Learn more about configuring custom containers

trainingInput.scaleTier ENUMERATION

Required. Specifies the machine types, the number of replicas for workers and parameter servers

trainingInput.jobDir STRING

Optional. A Google Cloud Storage path in which to store training outputs and other data needed for training. This path is passed to your TensorFlow program as the '--job-dir' command-line argument. The benefit of specifying this field is that Cloud ML validates the path for use in training

trainingInput.hyperparameters OBJECT

Represents a set of hyperparameters to optimize

trainingInput.hyperparameters.algorithm ENUMERATION

Optional. The search algorithm specified for the hyperparameter tuning job. Uses the default AI Platform hyperparameter tuning algorithm if unspecified

trainingInput.hyperparameters.hyperparameterMetricTag STRING

Optional. The TensorFlow summary tag name to use for optimizing trials. For current versions of TensorFlow, this tag name should exactly match what is shown in TensorBoard, including all scopes. For versions of TensorFlow prior to 0.12, this should be only the tag passed to tf.Summary. By default, "training/hptuning/metric" will be used

trainingInput.hyperparameters.params[] OBJECT

Represents a single hyperparameter to optimize

trainingInput.hyperparameters.params[].minValue NUMBER

Required if type is DOUBLE or INTEGER. This field should be unset if type is CATEGORICAL. This value should be integers if type is INTEGER

trainingInput.hyperparameters.params[].discreteValues[] NUMBER

trainingInput.hyperparameters.params[].scaleType ENUMERATION

Optional. How the parameter should be scaled to the hypercube. Leave unset for categorical parameters. Some kind of scaling is strongly recommended for real or integral parameters (e.g., UNIT_LINEAR_SCALE)

trainingInput.hyperparameters.params[].maxValue NUMBER

Required if type is DOUBLE or INTEGER. This field should be unset if type is CATEGORICAL. This value should be integers if type is INTEGER

trainingInput.hyperparameters.params[].type ENUMERATION

Required. The type of the parameter

trainingInput.hyperparameters.params[].categoricalValues[] STRING

trainingInput.hyperparameters.params[].parameterName STRING

Required. The parameter name must be unique amongst all ParameterConfigs in a HyperparameterSpec message. E.g., "learning_rate"

trainingInput.hyperparameters.enableTrialEarlyStopping BOOLEAN

Optional. Indicates if the hyperparameter tuning job enables auto trial early stopping

trainingInput.hyperparameters.resumePreviousJobId STRING

Optional. The prior hyperparameter tuning job id that users hope to continue with. The job id will be used to find the corresponding vizier study guid and resume the study

trainingInput.hyperparameters.maxParallelTrials INTEGER

Optional. The number of training trials to run concurrently. You can reduce the time it takes to perform hyperparameter tuning by adding trials in parallel. However, each trail only benefits from the information gained in completed trials. That means that a trial does not get access to the results of trials running at the same time, which could reduce the quality of the overall optimization.

Each trial will use the same scale tier and machine types.

Defaults to one

trainingInput.hyperparameters.maxFailedTrials INTEGER

Optional. The number of failed trials that need to be seen before failing the hyperparameter tuning job. You can specify this field to override the default failing criteria for AI Platform hyperparameter tuning jobs.

Defaults to zero, which means the service decides when a hyperparameter job should fail

trainingInput.hyperparameters.goal ENUMERATION

Required. The type of goal to use for tuning. Available types are MAXIMIZE and MINIMIZE.

Defaults to MAXIMIZE

trainingInput.hyperparameters.maxTrials INTEGER

Optional. How many training trials should be attempted to optimize the specified hyperparameters.

Defaults to one

trainingInput.pythonVersion STRING

Optional. The version of Python used in training. If not set, the default version is '2.7'. Python '3.5' is available when runtime_version is set to '1.4' and above. Python '2.7' works with all supported runtime versions

trainingInput.workerConfig OBJECT

Represents the configuration for a replica in a cluster

trainingInput.workerConfig.acceleratorConfig OBJECT

Represents a hardware accelerator request config

trainingInput.workerConfig.acceleratorConfig.type ENUMERATION

The type of accelerator to use

trainingInput.workerConfig.acceleratorConfig.count INTEGER

The number of accelerators to attach to each machine running the job

trainingInput.workerConfig.imageUri STRING

The Docker image to run on the replica. This image must be in Container Registry. Learn more about configuring custom containers

trainingInput.parameterServerCount INTEGER

Optional. The number of parameter server replicas to use for the training job. Each replica in the cluster will be of the type specified in parameter_server_type.

This value can only be used when scale_tier is set to CUSTOM.If you set this value, you must also set parameter_server_type.

The default value is zero

trainingInput.packageUris[] STRING

state ENUMERATION

Output only. The detailed state of a job

jobId STRING

Required. The user-specified id of the job

endTime ANY

Output only. When the job processing was completed

startTime ANY

Output only. When the job processing was started

predictionOutput OBJECT

Represents results of a prediction job

predictionOutput.errorCount INTEGER

The number of data instances which resulted in errors

predictionOutput.outputPath STRING

The output Google Cloud Storage location provided at the job creation time

predictionOutput.nodeHours NUMBER

Node hours used by the batch prediction job

predictionOutput.predictionCount INTEGER

The number of generated predictions

Output

This building block provides 102 output parameters

  = Parameter name
  = Format

trainingOutput OBJECT

Represents results of a training job. Output only

trainingOutput.isBuiltInAlgorithmJob BOOLEAN

Whether this job is a built-in Algorithm job

trainingOutput.builtInAlgorithmOutput OBJECT

Represents output related to a built-in algorithm Job

trainingOutput.builtInAlgorithmOutput.pythonVersion STRING

Python version on which the built-in algorithm was trained

trainingOutput.builtInAlgorithmOutput.runtimeVersion STRING

AI Platform runtime version on which the built-in algorithm was trained

trainingOutput.builtInAlgorithmOutput.framework STRING

Framework on which the built-in algorithm was trained

trainingOutput.builtInAlgorithmOutput.modelPath STRING

The Cloud Storage path to the model/ directory where the training job saves the trained model. Only set for successful jobs that don't use hyperparameter tuning

trainingOutput.trials[] OBJECT

Represents the result of a single hyperparameter tuning trial from a training job. The TrainingOutput object that is returned on successful completion of a training job with hyperparameter tuning includes a list of HyperparameterOutput objects, one for each successful trial

trainingOutput.trials[].hyperparameters OBJECT

The hyperparameters given to this trial

trainingOutput.trials[].hyperparameters.customKey.value STRING

The hyperparameters given to this trial

trainingOutput.trials[].trialId STRING

The trial id for these results

trainingOutput.trials[].endTime ANY

Output only. End time for the trial

trainingOutput.trials[].isTrialStoppedEarly BOOLEAN

True if the trial is stopped early

trainingOutput.trials[].startTime ANY

Output only. Start time for the trial

trainingOutput.trials[].finalMetric OBJECT

An observed value of a metric

trainingOutput.trials[].finalMetric.trainingStep INTEGER

The global training step for this metric

trainingOutput.trials[].finalMetric.objectiveValue NUMBER

The objective value at this training step

trainingOutput.trials[].builtInAlgorithmOutput OBJECT

Represents output related to a built-in algorithm Job

trainingOutput.trials[].builtInAlgorithmOutput.pythonVersion STRING

Python version on which the built-in algorithm was trained

trainingOutput.trials[].builtInAlgorithmOutput.runtimeVersion STRING

AI Platform runtime version on which the built-in algorithm was trained

trainingOutput.trials[].builtInAlgorithmOutput.framework STRING

Framework on which the built-in algorithm was trained

trainingOutput.trials[].builtInAlgorithmOutput.modelPath STRING

The Cloud Storage path to the model/ directory where the training job saves the trained model. Only set for successful jobs that don't use hyperparameter tuning

trainingOutput.trials[].state ENUMERATION

Output only. The detailed state of the trial

trainingOutput.trials[].allMetrics[] OBJECT

An observed value of a metric

trainingOutput.hyperparameterMetricTag STRING

The TensorFlow summary tag name used for optimizing hyperparameter tuning trials. See HyperparameterSpec.hyperparameterMetricTag for more information. Only set for hyperparameter tuning jobs

trainingOutput.completedTrialCount INTEGER

The number of hyperparameter tuning trials that completed successfully. Only set for hyperparameter tuning jobs

trainingOutput.isHyperparameterTuningJob BOOLEAN

Whether this job is a hyperparameter tuning job

trainingOutput.consumedMLUnits NUMBER

The amount of ML units consumed by the job

createTime ANY

Output only. When the job was created

labels OBJECT

Optional. One or more labels that you can add, to organize your jobs. Each label is a key-value pair, where both the key and the value are arbitrary strings that you supply. For more information, see the documentation on using labels

labels.customKey.value STRING

Optional. One or more labels that you can add, to organize your jobs. Each label is a key-value pair, where both the key and the value are arbitrary strings that you supply. For more information, see the documentation on using labels

predictionInput OBJECT

Represents input parameters for a prediction job

predictionInput.outputPath STRING

Required. The output Google Cloud Storage location

predictionInput.outputDataFormat ENUMERATION

Optional. Format of the output data files, defaults to JSON

predictionInput.dataFormat ENUMERATION

Required. The format of the input data files

predictionInput.batchSize INTEGER

Optional. Number of records per batch, defaults to 64. The service will buffer batch_size number of records in memory before invoking one Tensorflow prediction call internally. So take the record size and memory available into consideration when setting this parameter

predictionInput.runtimeVersion STRING

Optional. The AI Platform runtime version to use for this batch prediction. If not set, AI Platform will pick the runtime version used during the CreateVersion request for this model version, or choose the latest stable version when model version information is not available such as when the model is specified by uri

predictionInput.inputPaths[] STRING

predictionInput.region STRING

Required. The Google Compute Engine region to run the prediction job in. See the available regions for AI Platform services

predictionInput.versionName STRING

Use this field if you want to specify a version of the model to use. The string is formatted the same way as model_version, with the addition of the version information:

"projects/YOUR_PROJECT/models/YOUR_MODEL/versions/YOUR_VERSION"

predictionInput.modelName STRING

Use this field if you want to use the default version for the specified model. The string must use the following format:

"projects/YOUR_PROJECT/models/YOUR_MODEL"

predictionInput.uri STRING

Use this field if you want to specify a Google Cloud Storage path for the model to use

predictionInput.maxWorkerCount INTEGER

Optional. The maximum number of workers to be used for parallel processing. Defaults to 10 if not specified

predictionInput.signatureName STRING

Optional. The name of the signature defined in the SavedModel to use for this job. Please refer to SavedModel for information about how to use signatures.

Defaults to DEFAULT_SERVING_SIGNATURE_DEF_KEY , which is "serving_default"

errorMessage STRING

Output only. The details of a failure or a cancellation

etag BINARY

etag is used for optimistic concurrency control as a way to help prevent simultaneous updates of a job from overwriting each other. It is strongly suggested that systems make use of the etag in the read-modify-write cycle to perform job updates in order to avoid race conditions: An etag is returned in the response to GetJob, and systems are expected to put that etag in the request to UpdateJob to ensure that their change will be applied to the same version of the job

trainingInput OBJECT

Represents input parameters for a training job. When using the gcloud command to submit your training job, you can specify the input parameters as command-line arguments and/or in a YAML configuration file referenced from the --config command-line argument. For details, see the guide to submitting a training job

trainingInput.workerCount INTEGER

Optional. The number of worker replicas to use for the training job. Each replica in the cluster will be of the type specified in worker_type.

This value can only be used when scale_tier is set to CUSTOM. If you set this value, you must also set worker_type.

The default value is zero

trainingInput.masterType STRING

Optional. Specifies the type of virtual machine to use for your training job's master worker.

The following types are supported:

<dl> <dt>standard</dt> <dd> A basic machine configuration suitable for training simple models with small to moderate datasets. </dd> <dt>large_model</dt> <dd> A machine with a lot of memory, specially suited for parameter servers when your model is large (having many hidden layers or layers with very large numbers of nodes). </dd> <dt>complex_model_s</dt> <dd> A machine suitable for the master and workers of the cluster when your model requires more computation than the standard machine can handle satisfactorily. </dd> <dt>complex_model_m</dt> <dd> A machine with roughly twice the number of cores and roughly double the memory of <i>complex_model_s</i>. </dd> <dt>complex_model_l</dt> <dd> A machine with roughly twice the number of cores and roughly double the memory of <i>complex_model_m</i>. </dd> <dt>standard_gpu</dt> <dd> A machine equivalent to <i>standard</i> that also includes a single NVIDIA Tesla K80 GPU. See more about <a href="/ml-engine/docs/tensorflow/using-gpus">using GPUs to train your model</a>. </dd> <dt>complex_model_m_gpu</dt> <dd> A machine equivalent to <i>complex_model_m</i> that also includes four NVIDIA Tesla K80 GPUs. </dd> <dt>complex_model_l_gpu</dt> <dd> A machine equivalent to <i>complex_model_l</i> that also includes eight NVIDIA Tesla K80 GPUs. </dd> <dt>standard_p100</dt> <dd> A machine equivalent to <i>standard</i> that also includes a single NVIDIA Tesla P100 GPU. </dd> <dt>complex_model_m_p100</dt> <dd> A machine equivalent to <i>complex_model_m</i> that also includes four NVIDIA Tesla P100 GPUs. </dd> <dt>standard_v100</dt> <dd> A machine equivalent to <i>standard</i> that also includes a single NVIDIA Tesla V100 GPU. </dd> <dt>large_model_v100</dt> <dd> A machine equivalent to <i>large_model</i> that also includes a single NVIDIA Tesla V100 GPU. </dd> <dt>complex_model_m_v100</dt> <dd> A machine equivalent to <i>complex_model_m</i> that also includes four NVIDIA Tesla V100 GPUs. </dd> <dt>complex_model_l_v100</dt> <dd> A machine equivalent to <i>complex_model_l</i> that also includes eight NVIDIA Tesla V100 GPUs. </dd> <dt>cloud_tpu</dt> <dd> A TPU VM including one Cloud TPU. See more about <a href="/ml-engine/docs/tensorflow/using-tpus">using TPUs to train your model</a>. </dd> </dl>

You may also use certain Compute Engine machine types directly in this field. The following types are supported:

  • n1-standard-4
  • n1-standard-8
  • n1-standard-16
  • n1-standard-32
  • n1-standard-64
  • n1-standard-96
  • n1-highmem-2
  • n1-highmem-4
  • n1-highmem-8
  • n1-highmem-16
  • n1-highmem-32
  • n1-highmem-64
  • n1-highmem-96
  • n1-highcpu-16
  • n1-highcpu-32
  • n1-highcpu-64
  • n1-highcpu-96

See more about using Compute Engine machine types.

You must set this value when scaleTier is set to CUSTOM

trainingInput.maxRunningTime ANY

Optional. The maximum job running time. The default is 7 days

trainingInput.runtimeVersion STRING

Optional. The AI Platform runtime version to use for training. If not set, AI Platform uses the default stable version, 1.0. For more information, see the runtime version list and how to manage runtime versions

trainingInput.pythonModule STRING

Required. The Python module name to run after installing the packages

trainingInput.args[] STRING

trainingInput.region STRING

Required. The Google Compute Engine region to run the training job in. See the available regions for AI Platform services

trainingInput.workerType STRING

Optional. Specifies the type of virtual machine to use for your training job's worker nodes.

The supported values are the same as those described in the entry for masterType.

This value must be consistent with the category of machine type that masterType uses. In other words, both must be AI Platform machine types or both must be Compute Engine machine types.

If you use cloud_tpu for this value, see special instructions for configuring a custom TPU machine.

This value must be present when scaleTier is set to CUSTOM and workerCount is greater than zero

trainingInput.parameterServerType STRING

Optional. Specifies the type of virtual machine to use for your training job's parameter server.

The supported values are the same as those described in the entry for master_type.

This value must be consistent with the category of machine type that masterType uses. In other words, both must be AI Platform machine types or both must be Compute Engine machine types.

This value must be present when scaleTier is set to CUSTOM and parameter_server_count is greater than zero

trainingInput.parameterServerConfig OBJECT

Represents the configuration for a replica in a cluster

trainingInput.parameterServerConfig.acceleratorConfig OBJECT

Represents a hardware accelerator request config

trainingInput.parameterServerConfig.acceleratorConfig.type ENUMERATION

The type of accelerator to use

trainingInput.parameterServerConfig.acceleratorConfig.count INTEGER

The number of accelerators to attach to each machine running the job

trainingInput.parameterServerConfig.imageUri STRING

The Docker image to run on the replica. This image must be in Container Registry. Learn more about configuring custom containers

trainingInput.masterConfig OBJECT

Represents the configuration for a replica in a cluster

trainingInput.masterConfig.acceleratorConfig OBJECT

Represents a hardware accelerator request config

trainingInput.masterConfig.acceleratorConfig.type ENUMERATION

The type of accelerator to use

trainingInput.masterConfig.acceleratorConfig.count INTEGER

The number of accelerators to attach to each machine running the job

trainingInput.masterConfig.imageUri STRING

The Docker image to run on the replica. This image must be in Container Registry. Learn more about configuring custom containers

trainingInput.scaleTier ENUMERATION

Required. Specifies the machine types, the number of replicas for workers and parameter servers

trainingInput.jobDir STRING

Optional. A Google Cloud Storage path in which to store training outputs and other data needed for training. This path is passed to your TensorFlow program as the '--job-dir' command-line argument. The benefit of specifying this field is that Cloud ML validates the path for use in training

trainingInput.hyperparameters OBJECT

Represents a set of hyperparameters to optimize

trainingInput.hyperparameters.algorithm ENUMERATION

Optional. The search algorithm specified for the hyperparameter tuning job. Uses the default AI Platform hyperparameter tuning algorithm if unspecified

trainingInput.hyperparameters.hyperparameterMetricTag STRING

Optional. The TensorFlow summary tag name to use for optimizing trials. For current versions of TensorFlow, this tag name should exactly match what is shown in TensorBoard, including all scopes. For versions of TensorFlow prior to 0.12, this should be only the tag passed to tf.Summary. By default, "training/hptuning/metric" will be used

trainingInput.hyperparameters.params[] OBJECT

Represents a single hyperparameter to optimize

trainingInput.hyperparameters.params[].minValue NUMBER

Required if type is DOUBLE or INTEGER. This field should be unset if type is CATEGORICAL. This value should be integers if type is INTEGER

trainingInput.hyperparameters.params[].discreteValues[] NUMBER

trainingInput.hyperparameters.params[].scaleType ENUMERATION

Optional. How the parameter should be scaled to the hypercube. Leave unset for categorical parameters. Some kind of scaling is strongly recommended for real or integral parameters (e.g., UNIT_LINEAR_SCALE)

trainingInput.hyperparameters.params[].maxValue NUMBER

Required if type is DOUBLE or INTEGER. This field should be unset if type is CATEGORICAL. This value should be integers if type is INTEGER

trainingInput.hyperparameters.params[].type ENUMERATION

Required. The type of the parameter

trainingInput.hyperparameters.params[].categoricalValues[] STRING

trainingInput.hyperparameters.params[].parameterName STRING

Required. The parameter name must be unique amongst all ParameterConfigs in a HyperparameterSpec message. E.g., "learning_rate"

trainingInput.hyperparameters.enableTrialEarlyStopping BOOLEAN

Optional. Indicates if the hyperparameter tuning job enables auto trial early stopping

trainingInput.hyperparameters.resumePreviousJobId STRING

Optional. The prior hyperparameter tuning job id that users hope to continue with. The job id will be used to find the corresponding vizier study guid and resume the study

trainingInput.hyperparameters.maxParallelTrials INTEGER

Optional. The number of training trials to run concurrently. You can reduce the time it takes to perform hyperparameter tuning by adding trials in parallel. However, each trail only benefits from the information gained in completed trials. That means that a trial does not get access to the results of trials running at the same time, which could reduce the quality of the overall optimization.

Each trial will use the same scale tier and machine types.

Defaults to one

trainingInput.hyperparameters.maxFailedTrials INTEGER

Optional. The number of failed trials that need to be seen before failing the hyperparameter tuning job. You can specify this field to override the default failing criteria for AI Platform hyperparameter tuning jobs.

Defaults to zero, which means the service decides when a hyperparameter job should fail

trainingInput.hyperparameters.goal ENUMERATION

Required. The type of goal to use for tuning. Available types are MAXIMIZE and MINIMIZE.

Defaults to MAXIMIZE

trainingInput.hyperparameters.maxTrials INTEGER

Optional. How many training trials should be attempted to optimize the specified hyperparameters.

Defaults to one

trainingInput.pythonVersion STRING

Optional. The version of Python used in training. If not set, the default version is '2.7'. Python '3.5' is available when runtime_version is set to '1.4' and above. Python '2.7' works with all supported runtime versions

trainingInput.workerConfig OBJECT

Represents the configuration for a replica in a cluster

trainingInput.workerConfig.acceleratorConfig OBJECT

Represents a hardware accelerator request config

trainingInput.workerConfig.acceleratorConfig.type ENUMERATION

The type of accelerator to use

trainingInput.workerConfig.acceleratorConfig.count INTEGER

The number of accelerators to attach to each machine running the job

trainingInput.workerConfig.imageUri STRING

The Docker image to run on the replica. This image must be in Container Registry. Learn more about configuring custom containers

trainingInput.parameterServerCount INTEGER

Optional. The number of parameter server replicas to use for the training job. Each replica in the cluster will be of the type specified in parameter_server_type.

This value can only be used when scale_tier is set to CUSTOM.If you set this value, you must also set parameter_server_type.

The default value is zero

trainingInput.packageUris[] STRING

state ENUMERATION

Output only. The detailed state of a job

jobId STRING

Required. The user-specified id of the job

endTime ANY

Output only. When the job processing was completed

startTime ANY

Output only. When the job processing was started

predictionOutput OBJECT

Represents results of a prediction job

predictionOutput.errorCount INTEGER

The number of data instances which resulted in errors

predictionOutput.outputPath STRING

The output Google Cloud Storage location provided at the job creation time

predictionOutput.nodeHours NUMBER

Node hours used by the batch prediction job

predictionOutput.predictionCount INTEGER

The number of generated predictions