Update
|
|||||
|
|
||||
Updates (replaces) workflow template. The updated template must contain version that matches the current server version
Authorization
To use this building block you will have to grant access to at least one of the following scopes:
- View and manage your data across Google Cloud Platform services
Input
This building block consumes 113 input parameters
| Name | Format | Description |
|---|---|---|
name Required |
STRING |
Output only. The "resource name" of the template, as described in https://cloud.google.com/apis/design/resource_names of the form projects/{project_id}/regions/{region}/workflowTemplates/{template_id} |
placement |
OBJECT |
Specifies workflow execution target.Either managed_cluster or cluster_selector is required |
placement.clusterSelector |
OBJECT |
A selector that chooses target cluster for jobs based on metadata |
placement.clusterSelector.clusterLabels |
OBJECT |
Required. The cluster labels. Cluster must have all labels to match |
placement.clusterSelector.clusterLabels.customKey.value Required |
STRING |
Required. The cluster labels. Cluster must have all labels to match |
placement.clusterSelector.zone |
STRING |
Optional. The zone where workflow process executes. This parameter does not affect the selection of the cluster.If unspecified, the zone of the first cluster matching the selector is used |
placement.managedCluster |
OBJECT |
Cluster that is managed by the workflow |
placement.managedCluster.labels |
OBJECT |
Optional. The labels to associate with this cluster.Label keys must be between 1 and 63 characters long, and must conform to the following PCRE regular expression: \p{Ll}\p{Lo}{0,62}Label values must be between 1 and 63 characters long, and must conform to the following PCRE regular expression: \p{Ll}\p{Lo}\p{N}_-{0,63}No more than 32 labels can be associated with a given cluster |
placement.managedCluster.labels.customKey.value Required |
STRING |
Optional. The labels to associate with this cluster.Label keys must be between 1 and 63 characters long, and must conform to the following PCRE regular expression: \p{Ll}\p{Lo}{0,62}Label values must be between 1 and 63 characters long, and must conform to the following PCRE regular expression: \p{Ll}\p{Lo}\p{N}_-{0,63}No more than 32 labels can be associated with a given cluster |
placement.managedCluster.config |
OBJECT |
The cluster config |
placement.managedCluster.config.workerConfig |
OBJECT |
Optional. The config settings for Compute Engine resources in an instance group, such as a master or worker group |
placement.managedCluster.config.gceClusterConfig |
OBJECT |
Common config settings for resources of Compute Engine cluster instances, applicable to all instances in the cluster |
placement.managedCluster.config.softwareConfig |
OBJECT |
Specifies the selection and config of software inside the cluster |
placement.managedCluster.config.masterConfig |
OBJECT |
Optional. The config settings for Compute Engine resources in an instance group, such as a master or worker group |
placement.managedCluster.config.secondaryWorkerConfig |
OBJECT |
Optional. The config settings for Compute Engine resources in an instance group, such as a master or worker group |
placement.managedCluster.config.encryptionConfig |
OBJECT |
Encryption settings for the cluster |
placement.managedCluster.config.securityConfig |
OBJECT |
Security related configuration, including Kerberos |
placement.managedCluster.config.configBucket |
STRING |
Optional. A Google Cloud Storage bucket used to stage job dependencies, config files, and job driver console output. If you do not specify a staging bucket, Cloud Dataproc will determine a Cloud Storage location (US, ASIA, or EU) for your cluster's staging bucket according to the Google Compute Engine zone where your cluster is deployed, and then create and manage this project-level, per-location bucket (see Cloud Dataproc staging bucket) |
placement.managedCluster.clusterName |
STRING |
Required. The cluster name prefix. A unique cluster name will be formed by appending a random suffix.The name must contain only lower-case letters (a-z), numbers (0-9), and hyphens (-). Must begin with a letter. Cannot begin or end with hyphen. Must consist of between 2 and 35 characters |
updateTime |
ANY |
Output only. The time template was last updated |
parameters[] |
OBJECT |
A configurable parameter that replaces one or more fields in the template. Parameterizable fields: - Labels - File uris - Job properties - Job arguments - Script variables - Main class (in HadoopJob and SparkJob) - Zone (in ClusterSelector) |
parameters[].validation |
OBJECT |
Configuration for parameter validation |
parameters[].validation.values |
OBJECT |
Validation based on a list of allowed values |
parameters[].validation.values.values[] |
STRING |
|
parameters[].validation.regex |
OBJECT |
Validation based on regular expressions |
parameters[].validation.regex.regexes[] |
STRING |
|
parameters[].fields[] |
STRING |
|
parameters[].name |
STRING |
Required. Parameter name. The parameter name is used as the key, and paired with the parameter value, which are passed to the template when the template is instantiated. The name must contain only capital letters (A-Z), numbers (0-9), and underscores (_), and must not start with a number. The maximum length is 40 characters |
parameters[].description |
STRING |
Optional. Brief description of the parameter. Must not exceed 1024 characters |
name |
STRING |
Output only. The "resource name" of the template, as described in https://cloud.google.com/apis/design/resource_names of the form projects/{project_id}/regions/{region}/workflowTemplates/{template_id} |
version |
INTEGER |
Optional. Used to perform a consistent read-modify-write.This field should be left blank for a CreateWorkflowTemplate request. It is required for an UpdateWorkflowTemplate request, and must match the current server version. A typical update template flow would fetch the current template with a GetWorkflowTemplate request, which will return the current template with the version field filled in with the current server version. The user updates other fields in the template, then returns it as part of the UpdateWorkflowTemplate request |
id |
STRING |
Required. The template id.The id must contain only letters (a-z, A-Z), numbers (0-9), underscores (_), and hyphens (-). Cannot begin or end with underscore or hyphen. Must consist of between 3 and 50 characters |
jobs[] |
OBJECT |
A job executed by the workflow |
jobs[].hadoopJob |
OBJECT |
A Cloud Dataproc job for running Apache Hadoop MapReduce (https://hadoop.apache.org/docs/current/hadoop-mapreduce-client/hadoop-mapreduce-client-core/MapReduceTutorial.html) jobs on Apache Hadoop YARN (https://hadoop.apache.org/docs/r2.7.1/hadoop-yarn/hadoop-yarn-site/YARN.html) |
jobs[].hadoopJob.args[] |
STRING |
|
jobs[].hadoopJob.fileUris[] |
STRING |
|
jobs[].hadoopJob.mainClass |
STRING |
The name of the driver's main class. The jar file containing the class must be in the default CLASSPATH or specified in jar_file_uris |
jobs[].hadoopJob.archiveUris[] |
STRING |
|
jobs[].hadoopJob.mainJarFileUri |
STRING |
The HCFS URI of the jar file containing the main class. Examples: 'gs://foo-bucket/analytics-binaries/extract-useful-metrics-mr.jar' 'hdfs:/tmp/test-samples/custom-wordcount.jar' 'file:///home/usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples.jar' |
jobs[].hadoopJob.jarFileUris[] |
STRING |
|
jobs[].hadoopJob.loggingConfig |
OBJECT |
The runtime logging config of the job |
jobs[].hadoopJob.loggingConfig.driverLogLevels |
OBJECT |
The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
jobs[].hadoopJob.loggingConfig.driverLogLevels.customKey.value Required |
ENUMERATION |
The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
jobs[].hadoopJob.properties |
OBJECT |
Optional. A mapping of property names to values, used to configure Hadoop. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site and classes in user code |
jobs[].hadoopJob.properties.customKey.value Required |
STRING |
Optional. A mapping of property names to values, used to configure Hadoop. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site and classes in user code |
jobs[].hiveJob |
OBJECT |
A Cloud Dataproc job for running Apache Hive (https://hive.apache.org/) queries on YARN |
jobs[].hiveJob.continueOnFailure |
BOOLEAN |
Optional. Whether to continue executing queries if a query fails. The default value is false. Setting to true can be useful when executing independent parallel queries |
jobs[].hiveJob.queryFileUri |
STRING |
The HCFS URI of the script that contains Hive queries |
jobs[].hiveJob.queryList |
OBJECT |
A list of queries to run on a cluster |
jobs[].hiveJob.queryList.queries[] |
STRING |
|
jobs[].hiveJob.jarFileUris[] |
STRING |
|
jobs[].hiveJob.scriptVariables |
OBJECT |
Optional. Mapping of query variable names to values (equivalent to the Hive command: SET name="value";) |
jobs[].hiveJob.scriptVariables.customKey.value Required |
STRING |
Optional. Mapping of query variable names to values (equivalent to the Hive command: SET name="value";) |
jobs[].hiveJob.properties |
OBJECT |
Optional. A mapping of property names and values, used to configure Hive. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, /etc/hive/conf/hive-site.xml, and classes in user code |
jobs[].hiveJob.properties.customKey.value Required |
STRING |
Optional. A mapping of property names and values, used to configure Hive. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, /etc/hive/conf/hive-site.xml, and classes in user code |
jobs[].prerequisiteStepIds[] |
STRING |
|
jobs[].labels |
OBJECT |
Optional. The labels to associate with this job.Label keys must be between 1 and 63 characters long, and must conform to the following regular expression: \p{Ll}\p{Lo}{0,62}Label values must be between 1 and 63 characters long, and must conform to the following regular expression: \p{Ll}\p{Lo}\p{N}_-{0,63}No more than 32 labels can be associated with a given job |
jobs[].labels.customKey.value Required |
STRING |
Optional. The labels to associate with this job.Label keys must be between 1 and 63 characters long, and must conform to the following regular expression: \p{Ll}\p{Lo}{0,62}Label values must be between 1 and 63 characters long, and must conform to the following regular expression: \p{Ll}\p{Lo}\p{N}_-{0,63}No more than 32 labels can be associated with a given job |
jobs[].sparkJob |
OBJECT |
A Cloud Dataproc job for running Apache Spark (http://spark.apache.org/) applications on YARN |
jobs[].sparkJob.args[] |
STRING |
|
jobs[].sparkJob.fileUris[] |
STRING |
|
jobs[].sparkJob.mainClass |
STRING |
The name of the driver's main class. The jar file that contains the class must be in the default CLASSPATH or specified in jar_file_uris |
jobs[].sparkJob.archiveUris[] |
STRING |
|
jobs[].sparkJob.mainJarFileUri |
STRING |
The HCFS URI of the jar file that contains the main class |
jobs[].sparkJob.jarFileUris[] |
STRING |
|
jobs[].sparkJob.loggingConfig |
OBJECT |
The runtime logging config of the job |
jobs[].sparkJob.loggingConfig.driverLogLevels |
OBJECT |
The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
jobs[].sparkJob.loggingConfig.driverLogLevels.customKey.value Required |
ENUMERATION |
The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
jobs[].sparkJob.properties |
OBJECT |
Optional. A mapping of property names to values, used to configure Spark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code |
jobs[].sparkJob.properties.customKey.value Required |
STRING |
Optional. A mapping of property names to values, used to configure Spark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code |
jobs[].sparkSqlJob |
OBJECT |
A Cloud Dataproc job for running Apache Spark SQL (http://spark.apache.org/sql/) queries |
jobs[].sparkSqlJob.queryList |
OBJECT |
A list of queries to run on a cluster |
jobs[].sparkSqlJob.queryList.queries[] |
STRING |
|
jobs[].sparkSqlJob.queryFileUri |
STRING |
The HCFS URI of the script that contains SQL queries |
jobs[].sparkSqlJob.scriptVariables |
OBJECT |
Optional. Mapping of query variable names to values (equivalent to the Spark SQL command: SET name="value";) |
jobs[].sparkSqlJob.scriptVariables.customKey.value Required |
STRING |
Optional. Mapping of query variable names to values (equivalent to the Spark SQL command: SET name="value";) |
jobs[].sparkSqlJob.jarFileUris[] |
STRING |
|
jobs[].sparkSqlJob.loggingConfig |
OBJECT |
The runtime logging config of the job |
jobs[].sparkSqlJob.loggingConfig.driverLogLevels |
OBJECT |
The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
jobs[].sparkSqlJob.loggingConfig.driverLogLevels.customKey.value Required |
ENUMERATION |
The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
jobs[].sparkSqlJob.properties |
OBJECT |
Optional. A mapping of property names to values, used to configure Spark SQL's SparkConf. Properties that conflict with values set by the Cloud Dataproc API may be overwritten |
jobs[].sparkSqlJob.properties.customKey.value Required |
STRING |
Optional. A mapping of property names to values, used to configure Spark SQL's SparkConf. Properties that conflict with values set by the Cloud Dataproc API may be overwritten |
jobs[].pysparkJob |
OBJECT |
A Cloud Dataproc job for running Apache PySpark (https://spark.apache.org/docs/0.9.0/python-programming-guide.html) applications on YARN |
jobs[].pysparkJob.jarFileUris[] |
STRING |
|
jobs[].pysparkJob.loggingConfig |
OBJECT |
The runtime logging config of the job |
jobs[].pysparkJob.loggingConfig.driverLogLevels |
OBJECT |
The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
jobs[].pysparkJob.loggingConfig.driverLogLevels.customKey.value Required |
ENUMERATION |
The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
jobs[].pysparkJob.properties |
OBJECT |
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code |
jobs[].pysparkJob.properties.customKey.value Required |
STRING |
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code |
jobs[].pysparkJob.args[] |
STRING |
|
jobs[].pysparkJob.fileUris[] |
STRING |
|
jobs[].pysparkJob.pythonFileUris[] |
STRING |
|
jobs[].pysparkJob.mainPythonFileUri |
STRING |
Required. The HCFS URI of the main Python file to use as the driver. Must be a .py file |
jobs[].pysparkJob.archiveUris[] |
STRING |
|
jobs[].scheduling |
OBJECT |
Job scheduling options |
jobs[].scheduling.maxFailuresPerHour |
INTEGER |
Optional. Maximum number of times per hour a driver may be restarted as a result of driver terminating with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10 |
jobs[].pigJob |
OBJECT |
A Cloud Dataproc job for running Apache Pig (https://pig.apache.org/) queries on YARN |
jobs[].pigJob.loggingConfig |
OBJECT |
The runtime logging config of the job |
jobs[].pigJob.loggingConfig.driverLogLevels |
OBJECT |
The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
jobs[].pigJob.loggingConfig.driverLogLevels.customKey.value Required |
ENUMERATION |
The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
jobs[].pigJob.properties |
OBJECT |
Optional. A mapping of property names to values, used to configure Pig. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, /etc/pig/conf/pig.properties, and classes in user code |
jobs[].pigJob.properties.customKey.value Required |
STRING |
Optional. A mapping of property names to values, used to configure Pig. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, /etc/pig/conf/pig.properties, and classes in user code |
jobs[].pigJob.continueOnFailure |
BOOLEAN |
Optional. Whether to continue executing queries if a query fails. The default value is false. Setting to true can be useful when executing independent parallel queries |
jobs[].pigJob.queryFileUri |
STRING |
The HCFS URI of the script that contains the Pig queries |
jobs[].pigJob.queryList |
OBJECT |
A list of queries to run on a cluster |
jobs[].pigJob.queryList.queries[] |
STRING |
|
jobs[].pigJob.jarFileUris[] |
STRING |
|
jobs[].pigJob.scriptVariables |
OBJECT |
Optional. Mapping of query variable names to values (equivalent to the Pig command: name=[value]) |
jobs[].pigJob.scriptVariables.customKey.value Required |
STRING |
Optional. Mapping of query variable names to values (equivalent to the Pig command: name=[value]) |
jobs[].stepId |
STRING |
Required. The step id. The id must be unique among all jobs within the template.The step id is used as prefix for job id, as job goog-dataproc-workflow-step-id label, and in prerequisiteStepIds field from other steps.The id must contain only letters (a-z, A-Z), numbers (0-9), underscores (_), and hyphens (-). Cannot begin or end with underscore or hyphen. Must consist of between 3 and 50 characters |
createTime |
ANY |
Output only. The time template was created |
labels |
OBJECT |
Optional. The labels to associate with this template. These labels will be propagated to all jobs and clusters created by the workflow instance.Label keys must contain 1 to 63 characters, and must conform to RFC 1035 (https://www.ietf.org/rfc/rfc1035.txt).Label values may be empty, but, if present, must contain 1 to 63 characters, and must conform to RFC 1035 (https://www.ietf.org/rfc/rfc1035.txt).No more than 32 labels can be associated with a template |
labels.customKey.value Required |
STRING |
Optional. The labels to associate with this template. These labels will be propagated to all jobs and clusters created by the workflow instance.Label keys must contain 1 to 63 characters, and must conform to RFC 1035 (https://www.ietf.org/rfc/rfc1035.txt).Label values may be empty, but, if present, must contain 1 to 63 characters, and must conform to RFC 1035 (https://www.ietf.org/rfc/rfc1035.txt).No more than 32 labels can be associated with a template |
= Parameter name
= Format
|
name STRING Required Output only. The "resource name" of the template, as described in https://cloud.google.com/apis/design/resource_names of the form projects/{project_id}/regions/{region}/workflowTemplates/{template_id} |
|
placement OBJECT Specifies workflow execution target.Either managed_cluster or cluster_selector is required |
|
placement.clusterSelector OBJECT A selector that chooses target cluster for jobs based on metadata |
|
placement.clusterSelector.clusterLabels OBJECT Required. The cluster labels. Cluster must have all labels to match |
|
placement.clusterSelector.clusterLabels.customKey.value STRING Required Required. The cluster labels. Cluster must have all labels to match |
|
placement.clusterSelector.zone STRING Optional. The zone where workflow process executes. This parameter does not affect the selection of the cluster.If unspecified, the zone of the first cluster matching the selector is used |
|
placement.managedCluster OBJECT Cluster that is managed by the workflow |
|
placement.managedCluster.labels OBJECT Optional. The labels to associate with this cluster.Label keys must be between 1 and 63 characters long, and must conform to the following PCRE regular expression: \p{Ll}\p{Lo}{0,62}Label values must be between 1 and 63 characters long, and must conform to the following PCRE regular expression: \p{Ll}\p{Lo}\p{N}_-{0,63}No more than 32 labels can be associated with a given cluster |
|
placement.managedCluster.labels.customKey.value STRING Required Optional. The labels to associate with this cluster.Label keys must be between 1 and 63 characters long, and must conform to the following PCRE regular expression: \p{Ll}\p{Lo}{0,62}Label values must be between 1 and 63 characters long, and must conform to the following PCRE regular expression: \p{Ll}\p{Lo}\p{N}_-{0,63}No more than 32 labels can be associated with a given cluster |
|
placement.managedCluster.config OBJECT The cluster config |
|
placement.managedCluster.config.workerConfig OBJECT Optional. The config settings for Compute Engine resources in an instance group, such as a master or worker group |
|
placement.managedCluster.config.gceClusterConfig OBJECT Common config settings for resources of Compute Engine cluster instances, applicable to all instances in the cluster |
|
placement.managedCluster.config.softwareConfig OBJECT Specifies the selection and config of software inside the cluster |
|
placement.managedCluster.config.masterConfig OBJECT Optional. The config settings for Compute Engine resources in an instance group, such as a master or worker group |
|
placement.managedCluster.config.secondaryWorkerConfig OBJECT Optional. The config settings for Compute Engine resources in an instance group, such as a master or worker group |
|
placement.managedCluster.config.encryptionConfig OBJECT Encryption settings for the cluster |
|
placement.managedCluster.config.securityConfig OBJECT Security related configuration, including Kerberos |
|
placement.managedCluster.config.configBucket STRING Optional. A Google Cloud Storage bucket used to stage job dependencies, config files, and job driver console output. If you do not specify a staging bucket, Cloud Dataproc will determine a Cloud Storage location (US, ASIA, or EU) for your cluster's staging bucket according to the Google Compute Engine zone where your cluster is deployed, and then create and manage this project-level, per-location bucket (see Cloud Dataproc staging bucket) |
|
placement.managedCluster.clusterName STRING Required. The cluster name prefix. A unique cluster name will be formed by appending a random suffix.The name must contain only lower-case letters (a-z), numbers (0-9), and hyphens (-). Must begin with a letter. Cannot begin or end with hyphen. Must consist of between 2 and 35 characters |
|
updateTime ANY Output only. The time template was last updated |
|
parameters[] OBJECT A configurable parameter that replaces one or more fields in the template. Parameterizable fields: - Labels - File uris - Job properties - Job arguments - Script variables - Main class (in HadoopJob and SparkJob) - Zone (in ClusterSelector) |
|
parameters[].validation OBJECT Configuration for parameter validation |
|
parameters[].validation.values OBJECT Validation based on a list of allowed values |
|
parameters[].validation.values.values[] STRING |
|
parameters[].validation.regex OBJECT Validation based on regular expressions |
|
parameters[].validation.regex.regexes[] STRING |
|
parameters[].fields[] STRING |
|
parameters[].name STRING Required. Parameter name. The parameter name is used as the key, and paired with the parameter value, which are passed to the template when the template is instantiated. The name must contain only capital letters (A-Z), numbers (0-9), and underscores (_), and must not start with a number. The maximum length is 40 characters |
|
parameters[].description STRING Optional. Brief description of the parameter. Must not exceed 1024 characters |
|
name STRING Output only. The "resource name" of the template, as described in https://cloud.google.com/apis/design/resource_names of the form projects/{project_id}/regions/{region}/workflowTemplates/{template_id} |
|
version INTEGER Optional. Used to perform a consistent read-modify-write.This field should be left blank for a CreateWorkflowTemplate request. It is required for an UpdateWorkflowTemplate request, and must match the current server version. A typical update template flow would fetch the current template with a GetWorkflowTemplate request, which will return the current template with the version field filled in with the current server version. The user updates other fields in the template, then returns it as part of the UpdateWorkflowTemplate request |
|
id STRING Required. The template id.The id must contain only letters (a-z, A-Z), numbers (0-9), underscores (_), and hyphens (-). Cannot begin or end with underscore or hyphen. Must consist of between 3 and 50 characters |
|
jobs[] OBJECT A job executed by the workflow |
|
jobs[].hadoopJob OBJECT A Cloud Dataproc job for running Apache Hadoop MapReduce (https://hadoop.apache.org/docs/current/hadoop-mapreduce-client/hadoop-mapreduce-client-core/MapReduceTutorial.html) jobs on Apache Hadoop YARN (https://hadoop.apache.org/docs/r2.7.1/hadoop-yarn/hadoop-yarn-site/YARN.html) |
|
jobs[].hadoopJob.args[] STRING |
|
jobs[].hadoopJob.fileUris[] STRING |
|
jobs[].hadoopJob.mainClass STRING The name of the driver's main class. The jar file containing the class must be in the default CLASSPATH or specified in jar_file_uris |
|
jobs[].hadoopJob.archiveUris[] STRING |
|
jobs[].hadoopJob.mainJarFileUri STRING The HCFS URI of the jar file containing the main class. Examples: 'gs://foo-bucket/analytics-binaries/extract-useful-metrics-mr.jar' 'hdfs:/tmp/test-samples/custom-wordcount.jar' 'file:///home/usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples.jar' |
|
jobs[].hadoopJob.jarFileUris[] STRING |
|
jobs[].hadoopJob.loggingConfig OBJECT The runtime logging config of the job |
|
jobs[].hadoopJob.loggingConfig.driverLogLevels OBJECT The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
|
jobs[].hadoopJob.loggingConfig.driverLogLevels.customKey.value ENUMERATION Required The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
|
jobs[].hadoopJob.properties OBJECT Optional. A mapping of property names to values, used to configure Hadoop. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site and classes in user code |
|
jobs[].hadoopJob.properties.customKey.value STRING Required Optional. A mapping of property names to values, used to configure Hadoop. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site and classes in user code |
|
jobs[].hiveJob OBJECT A Cloud Dataproc job for running Apache Hive (https://hive.apache.org/) queries on YARN |
|
jobs[].hiveJob.continueOnFailure BOOLEAN Optional. Whether to continue executing queries if a query fails. The default value is false. Setting to true can be useful when executing independent parallel queries |
|
jobs[].hiveJob.queryFileUri STRING The HCFS URI of the script that contains Hive queries |
|
jobs[].hiveJob.queryList OBJECT A list of queries to run on a cluster |
|
jobs[].hiveJob.queryList.queries[] STRING |
|
jobs[].hiveJob.jarFileUris[] STRING |
|
jobs[].hiveJob.scriptVariables OBJECT Optional. Mapping of query variable names to values (equivalent to the Hive command: SET name="value";) |
|
jobs[].hiveJob.scriptVariables.customKey.value STRING Required Optional. Mapping of query variable names to values (equivalent to the Hive command: SET name="value";) |
|
jobs[].hiveJob.properties OBJECT Optional. A mapping of property names and values, used to configure Hive. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, /etc/hive/conf/hive-site.xml, and classes in user code |
|
jobs[].hiveJob.properties.customKey.value STRING Required Optional. A mapping of property names and values, used to configure Hive. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, /etc/hive/conf/hive-site.xml, and classes in user code |
|
jobs[].prerequisiteStepIds[] STRING |
|
jobs[].labels OBJECT Optional. The labels to associate with this job.Label keys must be between 1 and 63 characters long, and must conform to the following regular expression: \p{Ll}\p{Lo}{0,62}Label values must be between 1 and 63 characters long, and must conform to the following regular expression: \p{Ll}\p{Lo}\p{N}_-{0,63}No more than 32 labels can be associated with a given job |
|
jobs[].labels.customKey.value STRING Required Optional. The labels to associate with this job.Label keys must be between 1 and 63 characters long, and must conform to the following regular expression: \p{Ll}\p{Lo}{0,62}Label values must be between 1 and 63 characters long, and must conform to the following regular expression: \p{Ll}\p{Lo}\p{N}_-{0,63}No more than 32 labels can be associated with a given job |
|
jobs[].sparkJob OBJECT A Cloud Dataproc job for running Apache Spark (http://spark.apache.org/) applications on YARN |
|
jobs[].sparkJob.args[] STRING |
|
jobs[].sparkJob.fileUris[] STRING |
|
jobs[].sparkJob.mainClass STRING The name of the driver's main class. The jar file that contains the class must be in the default CLASSPATH or specified in jar_file_uris |
|
jobs[].sparkJob.archiveUris[] STRING |
|
jobs[].sparkJob.mainJarFileUri STRING The HCFS URI of the jar file that contains the main class |
|
jobs[].sparkJob.jarFileUris[] STRING |
|
jobs[].sparkJob.loggingConfig OBJECT The runtime logging config of the job |
|
jobs[].sparkJob.loggingConfig.driverLogLevels OBJECT The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
|
jobs[].sparkJob.loggingConfig.driverLogLevels.customKey.value ENUMERATION Required The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
|
jobs[].sparkJob.properties OBJECT Optional. A mapping of property names to values, used to configure Spark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code |
|
jobs[].sparkJob.properties.customKey.value STRING Required Optional. A mapping of property names to values, used to configure Spark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code |
|
jobs[].sparkSqlJob OBJECT A Cloud Dataproc job for running Apache Spark SQL (http://spark.apache.org/sql/) queries |
|
jobs[].sparkSqlJob.queryList OBJECT A list of queries to run on a cluster |
|
jobs[].sparkSqlJob.queryList.queries[] STRING |
|
jobs[].sparkSqlJob.queryFileUri STRING The HCFS URI of the script that contains SQL queries |
|
jobs[].sparkSqlJob.scriptVariables OBJECT Optional. Mapping of query variable names to values (equivalent to the Spark SQL command: SET name="value";) |
|
jobs[].sparkSqlJob.scriptVariables.customKey.value STRING Required Optional. Mapping of query variable names to values (equivalent to the Spark SQL command: SET name="value";) |
|
jobs[].sparkSqlJob.jarFileUris[] STRING |
|
jobs[].sparkSqlJob.loggingConfig OBJECT The runtime logging config of the job |
|
jobs[].sparkSqlJob.loggingConfig.driverLogLevels OBJECT The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
|
jobs[].sparkSqlJob.loggingConfig.driverLogLevels.customKey.value ENUMERATION Required The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
|
jobs[].sparkSqlJob.properties OBJECT Optional. A mapping of property names to values, used to configure Spark SQL's SparkConf. Properties that conflict with values set by the Cloud Dataproc API may be overwritten |
|
jobs[].sparkSqlJob.properties.customKey.value STRING Required Optional. A mapping of property names to values, used to configure Spark SQL's SparkConf. Properties that conflict with values set by the Cloud Dataproc API may be overwritten |
|
jobs[].pysparkJob OBJECT A Cloud Dataproc job for running Apache PySpark (https://spark.apache.org/docs/0.9.0/python-programming-guide.html) applications on YARN |
|
jobs[].pysparkJob.jarFileUris[] STRING |
|
jobs[].pysparkJob.loggingConfig OBJECT The runtime logging config of the job |
|
jobs[].pysparkJob.loggingConfig.driverLogLevels OBJECT The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
|
jobs[].pysparkJob.loggingConfig.driverLogLevels.customKey.value ENUMERATION Required The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
|
jobs[].pysparkJob.properties OBJECT Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code |
|
jobs[].pysparkJob.properties.customKey.value STRING Required Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code |
|
jobs[].pysparkJob.args[] STRING |
|
jobs[].pysparkJob.fileUris[] STRING |
|
jobs[].pysparkJob.pythonFileUris[] STRING |
|
jobs[].pysparkJob.mainPythonFileUri STRING Required. The HCFS URI of the main Python file to use as the driver. Must be a .py file |
|
jobs[].pysparkJob.archiveUris[] STRING |
|
jobs[].scheduling OBJECT Job scheduling options |
|
jobs[].scheduling.maxFailuresPerHour INTEGER Optional. Maximum number of times per hour a driver may be restarted as a result of driver terminating with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10 |
|
jobs[].pigJob OBJECT A Cloud Dataproc job for running Apache Pig (https://pig.apache.org/) queries on YARN |
|
jobs[].pigJob.loggingConfig OBJECT The runtime logging config of the job |
|
jobs[].pigJob.loggingConfig.driverLogLevels OBJECT The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
|
jobs[].pigJob.loggingConfig.driverLogLevels.customKey.value ENUMERATION Required The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
|
jobs[].pigJob.properties OBJECT Optional. A mapping of property names to values, used to configure Pig. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, /etc/pig/conf/pig.properties, and classes in user code |
|
jobs[].pigJob.properties.customKey.value STRING Required Optional. A mapping of property names to values, used to configure Pig. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, /etc/pig/conf/pig.properties, and classes in user code |
|
jobs[].pigJob.continueOnFailure BOOLEAN Optional. Whether to continue executing queries if a query fails. The default value is false. Setting to true can be useful when executing independent parallel queries |
|
jobs[].pigJob.queryFileUri STRING The HCFS URI of the script that contains the Pig queries |
|
jobs[].pigJob.queryList OBJECT A list of queries to run on a cluster |
|
jobs[].pigJob.queryList.queries[] STRING |
|
jobs[].pigJob.jarFileUris[] STRING |
|
jobs[].pigJob.scriptVariables OBJECT Optional. Mapping of query variable names to values (equivalent to the Pig command: name=[value]) |
|
jobs[].pigJob.scriptVariables.customKey.value STRING Required Optional. Mapping of query variable names to values (equivalent to the Pig command: name=[value]) |
|
jobs[].stepId STRING Required. The step id. The id must be unique among all jobs within the template.The step id is used as prefix for job id, as job goog-dataproc-workflow-step-id label, and in prerequisiteStepIds field from other steps.The id must contain only letters (a-z, A-Z), numbers (0-9), underscores (_), and hyphens (-). Cannot begin or end with underscore or hyphen. Must consist of between 3 and 50 characters |
|
createTime ANY Output only. The time template was created |
|
labels OBJECT Optional. The labels to associate with this template. These labels will be propagated to all jobs and clusters created by the workflow instance.Label keys must contain 1 to 63 characters, and must conform to RFC 1035 (https://www.ietf.org/rfc/rfc1035.txt).Label values may be empty, but, if present, must contain 1 to 63 characters, and must conform to RFC 1035 (https://www.ietf.org/rfc/rfc1035.txt).No more than 32 labels can be associated with a template |
|
labels.customKey.value STRING Required Optional. The labels to associate with this template. These labels will be propagated to all jobs and clusters created by the workflow instance.Label keys must contain 1 to 63 characters, and must conform to RFC 1035 (https://www.ietf.org/rfc/rfc1035.txt).Label values may be empty, but, if present, must contain 1 to 63 characters, and must conform to RFC 1035 (https://www.ietf.org/rfc/rfc1035.txt).No more than 32 labels can be associated with a template |
Output
This building block provides 112 output parameters
| Name | Format | Description |
|---|---|---|
placement |
OBJECT |
Specifies workflow execution target.Either managed_cluster or cluster_selector is required |
placement.clusterSelector |
OBJECT |
A selector that chooses target cluster for jobs based on metadata |
placement.clusterSelector.clusterLabels |
OBJECT |
Required. The cluster labels. Cluster must have all labels to match |
placement.clusterSelector.clusterLabels.customKey.value |
STRING |
Required. The cluster labels. Cluster must have all labels to match |
placement.clusterSelector.zone |
STRING |
Optional. The zone where workflow process executes. This parameter does not affect the selection of the cluster.If unspecified, the zone of the first cluster matching the selector is used |
placement.managedCluster |
OBJECT |
Cluster that is managed by the workflow |
placement.managedCluster.labels |
OBJECT |
Optional. The labels to associate with this cluster.Label keys must be between 1 and 63 characters long, and must conform to the following PCRE regular expression: \p{Ll}\p{Lo}{0,62}Label values must be between 1 and 63 characters long, and must conform to the following PCRE regular expression: \p{Ll}\p{Lo}\p{N}_-{0,63}No more than 32 labels can be associated with a given cluster |
placement.managedCluster.labels.customKey.value |
STRING |
Optional. The labels to associate with this cluster.Label keys must be between 1 and 63 characters long, and must conform to the following PCRE regular expression: \p{Ll}\p{Lo}{0,62}Label values must be between 1 and 63 characters long, and must conform to the following PCRE regular expression: \p{Ll}\p{Lo}\p{N}_-{0,63}No more than 32 labels can be associated with a given cluster |
placement.managedCluster.config |
OBJECT |
The cluster config |
placement.managedCluster.config.workerConfig |
OBJECT |
Optional. The config settings for Compute Engine resources in an instance group, such as a master or worker group |
placement.managedCluster.config.gceClusterConfig |
OBJECT |
Common config settings for resources of Compute Engine cluster instances, applicable to all instances in the cluster |
placement.managedCluster.config.softwareConfig |
OBJECT |
Specifies the selection and config of software inside the cluster |
placement.managedCluster.config.masterConfig |
OBJECT |
Optional. The config settings for Compute Engine resources in an instance group, such as a master or worker group |
placement.managedCluster.config.secondaryWorkerConfig |
OBJECT |
Optional. The config settings for Compute Engine resources in an instance group, such as a master or worker group |
placement.managedCluster.config.encryptionConfig |
OBJECT |
Encryption settings for the cluster |
placement.managedCluster.config.securityConfig |
OBJECT |
Security related configuration, including Kerberos |
placement.managedCluster.config.configBucket |
STRING |
Optional. A Google Cloud Storage bucket used to stage job dependencies, config files, and job driver console output. If you do not specify a staging bucket, Cloud Dataproc will determine a Cloud Storage location (US, ASIA, or EU) for your cluster's staging bucket according to the Google Compute Engine zone where your cluster is deployed, and then create and manage this project-level, per-location bucket (see Cloud Dataproc staging bucket) |
placement.managedCluster.clusterName |
STRING |
Required. The cluster name prefix. A unique cluster name will be formed by appending a random suffix.The name must contain only lower-case letters (a-z), numbers (0-9), and hyphens (-). Must begin with a letter. Cannot begin or end with hyphen. Must consist of between 2 and 35 characters |
updateTime |
ANY |
Output only. The time template was last updated |
parameters[] |
OBJECT |
A configurable parameter that replaces one or more fields in the template. Parameterizable fields: - Labels - File uris - Job properties - Job arguments - Script variables - Main class (in HadoopJob and SparkJob) - Zone (in ClusterSelector) |
parameters[].validation |
OBJECT |
Configuration for parameter validation |
parameters[].validation.values |
OBJECT |
Validation based on a list of allowed values |
parameters[].validation.values.values[] |
STRING |
|
parameters[].validation.regex |
OBJECT |
Validation based on regular expressions |
parameters[].validation.regex.regexes[] |
STRING |
|
parameters[].fields[] |
STRING |
|
parameters[].name |
STRING |
Required. Parameter name. The parameter name is used as the key, and paired with the parameter value, which are passed to the template when the template is instantiated. The name must contain only capital letters (A-Z), numbers (0-9), and underscores (_), and must not start with a number. The maximum length is 40 characters |
parameters[].description |
STRING |
Optional. Brief description of the parameter. Must not exceed 1024 characters |
name |
STRING |
Output only. The "resource name" of the template, as described in https://cloud.google.com/apis/design/resource_names of the form projects/{project_id}/regions/{region}/workflowTemplates/{template_id} |
version |
INTEGER |
Optional. Used to perform a consistent read-modify-write.This field should be left blank for a CreateWorkflowTemplate request. It is required for an UpdateWorkflowTemplate request, and must match the current server version. A typical update template flow would fetch the current template with a GetWorkflowTemplate request, which will return the current template with the version field filled in with the current server version. The user updates other fields in the template, then returns it as part of the UpdateWorkflowTemplate request |
id |
STRING |
Required. The template id.The id must contain only letters (a-z, A-Z), numbers (0-9), underscores (_), and hyphens (-). Cannot begin or end with underscore or hyphen. Must consist of between 3 and 50 characters |
jobs[] |
OBJECT |
A job executed by the workflow |
jobs[].hadoopJob |
OBJECT |
A Cloud Dataproc job for running Apache Hadoop MapReduce (https://hadoop.apache.org/docs/current/hadoop-mapreduce-client/hadoop-mapreduce-client-core/MapReduceTutorial.html) jobs on Apache Hadoop YARN (https://hadoop.apache.org/docs/r2.7.1/hadoop-yarn/hadoop-yarn-site/YARN.html) |
jobs[].hadoopJob.args[] |
STRING |
|
jobs[].hadoopJob.fileUris[] |
STRING |
|
jobs[].hadoopJob.mainClass |
STRING |
The name of the driver's main class. The jar file containing the class must be in the default CLASSPATH or specified in jar_file_uris |
jobs[].hadoopJob.archiveUris[] |
STRING |
|
jobs[].hadoopJob.mainJarFileUri |
STRING |
The HCFS URI of the jar file containing the main class. Examples: 'gs://foo-bucket/analytics-binaries/extract-useful-metrics-mr.jar' 'hdfs:/tmp/test-samples/custom-wordcount.jar' 'file:///home/usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples.jar' |
jobs[].hadoopJob.jarFileUris[] |
STRING |
|
jobs[].hadoopJob.loggingConfig |
OBJECT |
The runtime logging config of the job |
jobs[].hadoopJob.loggingConfig.driverLogLevels |
OBJECT |
The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
jobs[].hadoopJob.loggingConfig.driverLogLevels.customKey.value |
ENUMERATION |
The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
jobs[].hadoopJob.properties |
OBJECT |
Optional. A mapping of property names to values, used to configure Hadoop. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site and classes in user code |
jobs[].hadoopJob.properties.customKey.value |
STRING |
Optional. A mapping of property names to values, used to configure Hadoop. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site and classes in user code |
jobs[].hiveJob |
OBJECT |
A Cloud Dataproc job for running Apache Hive (https://hive.apache.org/) queries on YARN |
jobs[].hiveJob.continueOnFailure |
BOOLEAN |
Optional. Whether to continue executing queries if a query fails. The default value is false. Setting to true can be useful when executing independent parallel queries |
jobs[].hiveJob.queryFileUri |
STRING |
The HCFS URI of the script that contains Hive queries |
jobs[].hiveJob.queryList |
OBJECT |
A list of queries to run on a cluster |
jobs[].hiveJob.queryList.queries[] |
STRING |
|
jobs[].hiveJob.jarFileUris[] |
STRING |
|
jobs[].hiveJob.scriptVariables |
OBJECT |
Optional. Mapping of query variable names to values (equivalent to the Hive command: SET name="value";) |
jobs[].hiveJob.scriptVariables.customKey.value |
STRING |
Optional. Mapping of query variable names to values (equivalent to the Hive command: SET name="value";) |
jobs[].hiveJob.properties |
OBJECT |
Optional. A mapping of property names and values, used to configure Hive. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, /etc/hive/conf/hive-site.xml, and classes in user code |
jobs[].hiveJob.properties.customKey.value |
STRING |
Optional. A mapping of property names and values, used to configure Hive. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, /etc/hive/conf/hive-site.xml, and classes in user code |
jobs[].prerequisiteStepIds[] |
STRING |
|
jobs[].labels |
OBJECT |
Optional. The labels to associate with this job.Label keys must be between 1 and 63 characters long, and must conform to the following regular expression: \p{Ll}\p{Lo}{0,62}Label values must be between 1 and 63 characters long, and must conform to the following regular expression: \p{Ll}\p{Lo}\p{N}_-{0,63}No more than 32 labels can be associated with a given job |
jobs[].labels.customKey.value |
STRING |
Optional. The labels to associate with this job.Label keys must be between 1 and 63 characters long, and must conform to the following regular expression: \p{Ll}\p{Lo}{0,62}Label values must be between 1 and 63 characters long, and must conform to the following regular expression: \p{Ll}\p{Lo}\p{N}_-{0,63}No more than 32 labels can be associated with a given job |
jobs[].sparkJob |
OBJECT |
A Cloud Dataproc job for running Apache Spark (http://spark.apache.org/) applications on YARN |
jobs[].sparkJob.args[] |
STRING |
|
jobs[].sparkJob.fileUris[] |
STRING |
|
jobs[].sparkJob.mainClass |
STRING |
The name of the driver's main class. The jar file that contains the class must be in the default CLASSPATH or specified in jar_file_uris |
jobs[].sparkJob.archiveUris[] |
STRING |
|
jobs[].sparkJob.mainJarFileUri |
STRING |
The HCFS URI of the jar file that contains the main class |
jobs[].sparkJob.jarFileUris[] |
STRING |
|
jobs[].sparkJob.loggingConfig |
OBJECT |
The runtime logging config of the job |
jobs[].sparkJob.loggingConfig.driverLogLevels |
OBJECT |
The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
jobs[].sparkJob.loggingConfig.driverLogLevels.customKey.value |
ENUMERATION |
The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
jobs[].sparkJob.properties |
OBJECT |
Optional. A mapping of property names to values, used to configure Spark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code |
jobs[].sparkJob.properties.customKey.value |
STRING |
Optional. A mapping of property names to values, used to configure Spark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code |
jobs[].sparkSqlJob |
OBJECT |
A Cloud Dataproc job for running Apache Spark SQL (http://spark.apache.org/sql/) queries |
jobs[].sparkSqlJob.queryList |
OBJECT |
A list of queries to run on a cluster |
jobs[].sparkSqlJob.queryList.queries[] |
STRING |
|
jobs[].sparkSqlJob.queryFileUri |
STRING |
The HCFS URI of the script that contains SQL queries |
jobs[].sparkSqlJob.scriptVariables |
OBJECT |
Optional. Mapping of query variable names to values (equivalent to the Spark SQL command: SET name="value";) |
jobs[].sparkSqlJob.scriptVariables.customKey.value |
STRING |
Optional. Mapping of query variable names to values (equivalent to the Spark SQL command: SET name="value";) |
jobs[].sparkSqlJob.jarFileUris[] |
STRING |
|
jobs[].sparkSqlJob.loggingConfig |
OBJECT |
The runtime logging config of the job |
jobs[].sparkSqlJob.loggingConfig.driverLogLevels |
OBJECT |
The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
jobs[].sparkSqlJob.loggingConfig.driverLogLevels.customKey.value |
ENUMERATION |
The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
jobs[].sparkSqlJob.properties |
OBJECT |
Optional. A mapping of property names to values, used to configure Spark SQL's SparkConf. Properties that conflict with values set by the Cloud Dataproc API may be overwritten |
jobs[].sparkSqlJob.properties.customKey.value |
STRING |
Optional. A mapping of property names to values, used to configure Spark SQL's SparkConf. Properties that conflict with values set by the Cloud Dataproc API may be overwritten |
jobs[].pysparkJob |
OBJECT |
A Cloud Dataproc job for running Apache PySpark (https://spark.apache.org/docs/0.9.0/python-programming-guide.html) applications on YARN |
jobs[].pysparkJob.jarFileUris[] |
STRING |
|
jobs[].pysparkJob.loggingConfig |
OBJECT |
The runtime logging config of the job |
jobs[].pysparkJob.loggingConfig.driverLogLevels |
OBJECT |
The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
jobs[].pysparkJob.loggingConfig.driverLogLevels.customKey.value |
ENUMERATION |
The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
jobs[].pysparkJob.properties |
OBJECT |
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code |
jobs[].pysparkJob.properties.customKey.value |
STRING |
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code |
jobs[].pysparkJob.args[] |
STRING |
|
jobs[].pysparkJob.fileUris[] |
STRING |
|
jobs[].pysparkJob.pythonFileUris[] |
STRING |
|
jobs[].pysparkJob.mainPythonFileUri |
STRING |
Required. The HCFS URI of the main Python file to use as the driver. Must be a .py file |
jobs[].pysparkJob.archiveUris[] |
STRING |
|
jobs[].scheduling |
OBJECT |
Job scheduling options |
jobs[].scheduling.maxFailuresPerHour |
INTEGER |
Optional. Maximum number of times per hour a driver may be restarted as a result of driver terminating with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10 |
jobs[].pigJob |
OBJECT |
A Cloud Dataproc job for running Apache Pig (https://pig.apache.org/) queries on YARN |
jobs[].pigJob.loggingConfig |
OBJECT |
The runtime logging config of the job |
jobs[].pigJob.loggingConfig.driverLogLevels |
OBJECT |
The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
jobs[].pigJob.loggingConfig.driverLogLevels.customKey.value |
ENUMERATION |
The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
jobs[].pigJob.properties |
OBJECT |
Optional. A mapping of property names to values, used to configure Pig. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, /etc/pig/conf/pig.properties, and classes in user code |
jobs[].pigJob.properties.customKey.value |
STRING |
Optional. A mapping of property names to values, used to configure Pig. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, /etc/pig/conf/pig.properties, and classes in user code |
jobs[].pigJob.continueOnFailure |
BOOLEAN |
Optional. Whether to continue executing queries if a query fails. The default value is false. Setting to true can be useful when executing independent parallel queries |
jobs[].pigJob.queryFileUri |
STRING |
The HCFS URI of the script that contains the Pig queries |
jobs[].pigJob.queryList |
OBJECT |
A list of queries to run on a cluster |
jobs[].pigJob.queryList.queries[] |
STRING |
|
jobs[].pigJob.jarFileUris[] |
STRING |
|
jobs[].pigJob.scriptVariables |
OBJECT |
Optional. Mapping of query variable names to values (equivalent to the Pig command: name=[value]) |
jobs[].pigJob.scriptVariables.customKey.value |
STRING |
Optional. Mapping of query variable names to values (equivalent to the Pig command: name=[value]) |
jobs[].stepId |
STRING |
Required. The step id. The id must be unique among all jobs within the template.The step id is used as prefix for job id, as job goog-dataproc-workflow-step-id label, and in prerequisiteStepIds field from other steps.The id must contain only letters (a-z, A-Z), numbers (0-9), underscores (_), and hyphens (-). Cannot begin or end with underscore or hyphen. Must consist of between 3 and 50 characters |
createTime |
ANY |
Output only. The time template was created |
labels |
OBJECT |
Optional. The labels to associate with this template. These labels will be propagated to all jobs and clusters created by the workflow instance.Label keys must contain 1 to 63 characters, and must conform to RFC 1035 (https://www.ietf.org/rfc/rfc1035.txt).Label values may be empty, but, if present, must contain 1 to 63 characters, and must conform to RFC 1035 (https://www.ietf.org/rfc/rfc1035.txt).No more than 32 labels can be associated with a template |
labels.customKey.value |
STRING |
Optional. The labels to associate with this template. These labels will be propagated to all jobs and clusters created by the workflow instance.Label keys must contain 1 to 63 characters, and must conform to RFC 1035 (https://www.ietf.org/rfc/rfc1035.txt).Label values may be empty, but, if present, must contain 1 to 63 characters, and must conform to RFC 1035 (https://www.ietf.org/rfc/rfc1035.txt).No more than 32 labels can be associated with a template |
= Parameter name
= Format
|
placement OBJECT Specifies workflow execution target.Either managed_cluster or cluster_selector is required |
|
placement.clusterSelector OBJECT A selector that chooses target cluster for jobs based on metadata |
|
placement.clusterSelector.clusterLabels OBJECT Required. The cluster labels. Cluster must have all labels to match |
|
placement.clusterSelector.clusterLabels.customKey.value STRING Required. The cluster labels. Cluster must have all labels to match |
|
placement.clusterSelector.zone STRING Optional. The zone where workflow process executes. This parameter does not affect the selection of the cluster.If unspecified, the zone of the first cluster matching the selector is used |
|
placement.managedCluster OBJECT Cluster that is managed by the workflow |
|
placement.managedCluster.labels OBJECT Optional. The labels to associate with this cluster.Label keys must be between 1 and 63 characters long, and must conform to the following PCRE regular expression: \p{Ll}\p{Lo}{0,62}Label values must be between 1 and 63 characters long, and must conform to the following PCRE regular expression: \p{Ll}\p{Lo}\p{N}_-{0,63}No more than 32 labels can be associated with a given cluster |
|
placement.managedCluster.labels.customKey.value STRING Optional. The labels to associate with this cluster.Label keys must be between 1 and 63 characters long, and must conform to the following PCRE regular expression: \p{Ll}\p{Lo}{0,62}Label values must be between 1 and 63 characters long, and must conform to the following PCRE regular expression: \p{Ll}\p{Lo}\p{N}_-{0,63}No more than 32 labels can be associated with a given cluster |
|
placement.managedCluster.config OBJECT The cluster config |
|
placement.managedCluster.config.workerConfig OBJECT Optional. The config settings for Compute Engine resources in an instance group, such as a master or worker group |
|
placement.managedCluster.config.gceClusterConfig OBJECT Common config settings for resources of Compute Engine cluster instances, applicable to all instances in the cluster |
|
placement.managedCluster.config.softwareConfig OBJECT Specifies the selection and config of software inside the cluster |
|
placement.managedCluster.config.masterConfig OBJECT Optional. The config settings for Compute Engine resources in an instance group, such as a master or worker group |
|
placement.managedCluster.config.secondaryWorkerConfig OBJECT Optional. The config settings for Compute Engine resources in an instance group, such as a master or worker group |
|
placement.managedCluster.config.encryptionConfig OBJECT Encryption settings for the cluster |
|
placement.managedCluster.config.securityConfig OBJECT Security related configuration, including Kerberos |
|
placement.managedCluster.config.configBucket STRING Optional. A Google Cloud Storage bucket used to stage job dependencies, config files, and job driver console output. If you do not specify a staging bucket, Cloud Dataproc will determine a Cloud Storage location (US, ASIA, or EU) for your cluster's staging bucket according to the Google Compute Engine zone where your cluster is deployed, and then create and manage this project-level, per-location bucket (see Cloud Dataproc staging bucket) |
|
placement.managedCluster.clusterName STRING Required. The cluster name prefix. A unique cluster name will be formed by appending a random suffix.The name must contain only lower-case letters (a-z), numbers (0-9), and hyphens (-). Must begin with a letter. Cannot begin or end with hyphen. Must consist of between 2 and 35 characters |
|
updateTime ANY Output only. The time template was last updated |
|
parameters[] OBJECT A configurable parameter that replaces one or more fields in the template. Parameterizable fields: - Labels - File uris - Job properties - Job arguments - Script variables - Main class (in HadoopJob and SparkJob) - Zone (in ClusterSelector) |
|
parameters[].validation OBJECT Configuration for parameter validation |
|
parameters[].validation.values OBJECT Validation based on a list of allowed values |
|
parameters[].validation.values.values[] STRING |
|
parameters[].validation.regex OBJECT Validation based on regular expressions |
|
parameters[].validation.regex.regexes[] STRING |
|
parameters[].fields[] STRING |
|
parameters[].name STRING Required. Parameter name. The parameter name is used as the key, and paired with the parameter value, which are passed to the template when the template is instantiated. The name must contain only capital letters (A-Z), numbers (0-9), and underscores (_), and must not start with a number. The maximum length is 40 characters |
|
parameters[].description STRING Optional. Brief description of the parameter. Must not exceed 1024 characters |
|
name STRING Output only. The "resource name" of the template, as described in https://cloud.google.com/apis/design/resource_names of the form projects/{project_id}/regions/{region}/workflowTemplates/{template_id} |
|
version INTEGER Optional. Used to perform a consistent read-modify-write.This field should be left blank for a CreateWorkflowTemplate request. It is required for an UpdateWorkflowTemplate request, and must match the current server version. A typical update template flow would fetch the current template with a GetWorkflowTemplate request, which will return the current template with the version field filled in with the current server version. The user updates other fields in the template, then returns it as part of the UpdateWorkflowTemplate request |
|
id STRING Required. The template id.The id must contain only letters (a-z, A-Z), numbers (0-9), underscores (_), and hyphens (-). Cannot begin or end with underscore or hyphen. Must consist of between 3 and 50 characters |
|
jobs[] OBJECT A job executed by the workflow |
|
jobs[].hadoopJob OBJECT A Cloud Dataproc job for running Apache Hadoop MapReduce (https://hadoop.apache.org/docs/current/hadoop-mapreduce-client/hadoop-mapreduce-client-core/MapReduceTutorial.html) jobs on Apache Hadoop YARN (https://hadoop.apache.org/docs/r2.7.1/hadoop-yarn/hadoop-yarn-site/YARN.html) |
|
jobs[].hadoopJob.args[] STRING |
|
jobs[].hadoopJob.fileUris[] STRING |
|
jobs[].hadoopJob.mainClass STRING The name of the driver's main class. The jar file containing the class must be in the default CLASSPATH or specified in jar_file_uris |
|
jobs[].hadoopJob.archiveUris[] STRING |
|
jobs[].hadoopJob.mainJarFileUri STRING The HCFS URI of the jar file containing the main class. Examples: 'gs://foo-bucket/analytics-binaries/extract-useful-metrics-mr.jar' 'hdfs:/tmp/test-samples/custom-wordcount.jar' 'file:///home/usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples.jar' |
|
jobs[].hadoopJob.jarFileUris[] STRING |
|
jobs[].hadoopJob.loggingConfig OBJECT The runtime logging config of the job |
|
jobs[].hadoopJob.loggingConfig.driverLogLevels OBJECT The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
|
jobs[].hadoopJob.loggingConfig.driverLogLevels.customKey.value ENUMERATION The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
|
jobs[].hadoopJob.properties OBJECT Optional. A mapping of property names to values, used to configure Hadoop. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site and classes in user code |
|
jobs[].hadoopJob.properties.customKey.value STRING Optional. A mapping of property names to values, used to configure Hadoop. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site and classes in user code |
|
jobs[].hiveJob OBJECT A Cloud Dataproc job for running Apache Hive (https://hive.apache.org/) queries on YARN |
|
jobs[].hiveJob.continueOnFailure BOOLEAN Optional. Whether to continue executing queries if a query fails. The default value is false. Setting to true can be useful when executing independent parallel queries |
|
jobs[].hiveJob.queryFileUri STRING The HCFS URI of the script that contains Hive queries |
|
jobs[].hiveJob.queryList OBJECT A list of queries to run on a cluster |
|
jobs[].hiveJob.queryList.queries[] STRING |
|
jobs[].hiveJob.jarFileUris[] STRING |
|
jobs[].hiveJob.scriptVariables OBJECT Optional. Mapping of query variable names to values (equivalent to the Hive command: SET name="value";) |
|
jobs[].hiveJob.scriptVariables.customKey.value STRING Optional. Mapping of query variable names to values (equivalent to the Hive command: SET name="value";) |
|
jobs[].hiveJob.properties OBJECT Optional. A mapping of property names and values, used to configure Hive. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, /etc/hive/conf/hive-site.xml, and classes in user code |
|
jobs[].hiveJob.properties.customKey.value STRING Optional. A mapping of property names and values, used to configure Hive. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, /etc/hive/conf/hive-site.xml, and classes in user code |
|
jobs[].prerequisiteStepIds[] STRING |
|
jobs[].labels OBJECT Optional. The labels to associate with this job.Label keys must be between 1 and 63 characters long, and must conform to the following regular expression: \p{Ll}\p{Lo}{0,62}Label values must be between 1 and 63 characters long, and must conform to the following regular expression: \p{Ll}\p{Lo}\p{N}_-{0,63}No more than 32 labels can be associated with a given job |
|
jobs[].labels.customKey.value STRING Optional. The labels to associate with this job.Label keys must be between 1 and 63 characters long, and must conform to the following regular expression: \p{Ll}\p{Lo}{0,62}Label values must be between 1 and 63 characters long, and must conform to the following regular expression: \p{Ll}\p{Lo}\p{N}_-{0,63}No more than 32 labels can be associated with a given job |
|
jobs[].sparkJob OBJECT A Cloud Dataproc job for running Apache Spark (http://spark.apache.org/) applications on YARN |
|
jobs[].sparkJob.args[] STRING |
|
jobs[].sparkJob.fileUris[] STRING |
|
jobs[].sparkJob.mainClass STRING The name of the driver's main class. The jar file that contains the class must be in the default CLASSPATH or specified in jar_file_uris |
|
jobs[].sparkJob.archiveUris[] STRING |
|
jobs[].sparkJob.mainJarFileUri STRING The HCFS URI of the jar file that contains the main class |
|
jobs[].sparkJob.jarFileUris[] STRING |
|
jobs[].sparkJob.loggingConfig OBJECT The runtime logging config of the job |
|
jobs[].sparkJob.loggingConfig.driverLogLevels OBJECT The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
|
jobs[].sparkJob.loggingConfig.driverLogLevels.customKey.value ENUMERATION The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
|
jobs[].sparkJob.properties OBJECT Optional. A mapping of property names to values, used to configure Spark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code |
|
jobs[].sparkJob.properties.customKey.value STRING Optional. A mapping of property names to values, used to configure Spark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code |
|
jobs[].sparkSqlJob OBJECT A Cloud Dataproc job for running Apache Spark SQL (http://spark.apache.org/sql/) queries |
|
jobs[].sparkSqlJob.queryList OBJECT A list of queries to run on a cluster |
|
jobs[].sparkSqlJob.queryList.queries[] STRING |
|
jobs[].sparkSqlJob.queryFileUri STRING The HCFS URI of the script that contains SQL queries |
|
jobs[].sparkSqlJob.scriptVariables OBJECT Optional. Mapping of query variable names to values (equivalent to the Spark SQL command: SET name="value";) |
|
jobs[].sparkSqlJob.scriptVariables.customKey.value STRING Optional. Mapping of query variable names to values (equivalent to the Spark SQL command: SET name="value";) |
|
jobs[].sparkSqlJob.jarFileUris[] STRING |
|
jobs[].sparkSqlJob.loggingConfig OBJECT The runtime logging config of the job |
|
jobs[].sparkSqlJob.loggingConfig.driverLogLevels OBJECT The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
|
jobs[].sparkSqlJob.loggingConfig.driverLogLevels.customKey.value ENUMERATION The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
|
jobs[].sparkSqlJob.properties OBJECT Optional. A mapping of property names to values, used to configure Spark SQL's SparkConf. Properties that conflict with values set by the Cloud Dataproc API may be overwritten |
|
jobs[].sparkSqlJob.properties.customKey.value STRING Optional. A mapping of property names to values, used to configure Spark SQL's SparkConf. Properties that conflict with values set by the Cloud Dataproc API may be overwritten |
|
jobs[].pysparkJob OBJECT A Cloud Dataproc job for running Apache PySpark (https://spark.apache.org/docs/0.9.0/python-programming-guide.html) applications on YARN |
|
jobs[].pysparkJob.jarFileUris[] STRING |
|
jobs[].pysparkJob.loggingConfig OBJECT The runtime logging config of the job |
|
jobs[].pysparkJob.loggingConfig.driverLogLevels OBJECT The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
|
jobs[].pysparkJob.loggingConfig.driverLogLevels.customKey.value ENUMERATION The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
|
jobs[].pysparkJob.properties OBJECT Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code |
|
jobs[].pysparkJob.properties.customKey.value STRING Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code |
|
jobs[].pysparkJob.args[] STRING |
|
jobs[].pysparkJob.fileUris[] STRING |
|
jobs[].pysparkJob.pythonFileUris[] STRING |
|
jobs[].pysparkJob.mainPythonFileUri STRING Required. The HCFS URI of the main Python file to use as the driver. Must be a .py file |
|
jobs[].pysparkJob.archiveUris[] STRING |
|
jobs[].scheduling OBJECT Job scheduling options |
|
jobs[].scheduling.maxFailuresPerHour INTEGER Optional. Maximum number of times per hour a driver may be restarted as a result of driver terminating with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10 |
|
jobs[].pigJob OBJECT A Cloud Dataproc job for running Apache Pig (https://pig.apache.org/) queries on YARN |
|
jobs[].pigJob.loggingConfig OBJECT The runtime logging config of the job |
|
jobs[].pigJob.loggingConfig.driverLogLevels OBJECT The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
|
jobs[].pigJob.loggingConfig.driverLogLevels.customKey.value ENUMERATION The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' |
|
jobs[].pigJob.properties OBJECT Optional. A mapping of property names to values, used to configure Pig. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, /etc/pig/conf/pig.properties, and classes in user code |
|
jobs[].pigJob.properties.customKey.value STRING Optional. A mapping of property names to values, used to configure Pig. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, /etc/pig/conf/pig.properties, and classes in user code |
|
jobs[].pigJob.continueOnFailure BOOLEAN Optional. Whether to continue executing queries if a query fails. The default value is false. Setting to true can be useful when executing independent parallel queries |
|
jobs[].pigJob.queryFileUri STRING The HCFS URI of the script that contains the Pig queries |
|
jobs[].pigJob.queryList OBJECT A list of queries to run on a cluster |
|
jobs[].pigJob.queryList.queries[] STRING |
|
jobs[].pigJob.jarFileUris[] STRING |
|
jobs[].pigJob.scriptVariables OBJECT Optional. Mapping of query variable names to values (equivalent to the Pig command: name=[value]) |
|
jobs[].pigJob.scriptVariables.customKey.value STRING Optional. Mapping of query variable names to values (equivalent to the Pig command: name=[value]) |
|
jobs[].stepId STRING Required. The step id. The id must be unique among all jobs within the template.The step id is used as prefix for job id, as job goog-dataproc-workflow-step-id label, and in prerequisiteStepIds field from other steps.The id must contain only letters (a-z, A-Z), numbers (0-9), underscores (_), and hyphens (-). Cannot begin or end with underscore or hyphen. Must consist of between 3 and 50 characters |
|
createTime ANY Output only. The time template was created |
|
labels OBJECT Optional. The labels to associate with this template. These labels will be propagated to all jobs and clusters created by the workflow instance.Label keys must contain 1 to 63 characters, and must conform to RFC 1035 (https://www.ietf.org/rfc/rfc1035.txt).Label values may be empty, but, if present, must contain 1 to 63 characters, and must conform to RFC 1035 (https://www.ietf.org/rfc/rfc1035.txt).No more than 32 labels can be associated with a template |
|
labels.customKey.value STRING Optional. The labels to associate with this template. These labels will be propagated to all jobs and clusters created by the workflow instance.Label keys must contain 1 to 63 characters, and must conform to RFC 1035 (https://www.ietf.org/rfc/rfc1035.txt).Label values may be empty, but, if present, must contain 1 to 63 characters, and must conform to RFC 1035 (https://www.ietf.org/rfc/rfc1035.txt).No more than 32 labels can be associated with a template |