gcp.dataproc.Job
Explore with Pulumi AI
Manages a job resource within a Dataproc cluster within GCE. For more information see the official dataproc documentation.
!> Note: This resource does not support ‘update’ and changing any attributes will cause the resource to be recreated.
Example Usage
using System.Collections.Generic;
using System.Linq;
using Pulumi;
using Gcp = Pulumi.Gcp;
return await Deployment.RunAsync(() =>
{
var mycluster = new Gcp.Dataproc.Cluster("mycluster", new()
{
Region = "us-central1",
});
// Submit an example spark job to a dataproc cluster
var spark = new Gcp.Dataproc.Job("spark", new()
{
Region = mycluster.Region,
ForceDelete = true,
Placement = new Gcp.Dataproc.Inputs.JobPlacementArgs
{
ClusterName = mycluster.Name,
},
SparkConfig = new Gcp.Dataproc.Inputs.JobSparkConfigArgs
{
MainClass = "org.apache.spark.examples.SparkPi",
JarFileUris = new[]
{
"file:///usr/lib/spark/examples/jars/spark-examples.jar",
},
Args = new[]
{
"1000",
},
Properties =
{
{ "spark.logConf", "true" },
},
LoggingConfig = new Gcp.Dataproc.Inputs.JobSparkConfigLoggingConfigArgs
{
DriverLogLevels =
{
{ "root", "INFO" },
},
},
},
});
// Submit an example pyspark job to a dataproc cluster
var pyspark = new Gcp.Dataproc.Job("pyspark", new()
{
Region = mycluster.Region,
ForceDelete = true,
Placement = new Gcp.Dataproc.Inputs.JobPlacementArgs
{
ClusterName = mycluster.Name,
},
PysparkConfig = new Gcp.Dataproc.Inputs.JobPysparkConfigArgs
{
MainPythonFileUri = "gs://dataproc-examples-2f10d78d114f6aaec76462e3c310f31f/src/pyspark/hello-world/hello-world.py",
Properties =
{
{ "spark.logConf", "true" },
},
},
});
return new Dictionary<string, object?>
{
["sparkStatus"] = spark.Statuses.Apply(statuses => statuses[0].State),
["pysparkStatus"] = pyspark.Statuses.Apply(statuses => statuses[0].State),
};
});
package main
import (
"github.com/pulumi/pulumi-gcp/sdk/v6/go/gcp/dataproc"
"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
)
func main() {
pulumi.Run(func(ctx *pulumi.Context) error {
mycluster, err := dataproc.NewCluster(ctx, "mycluster", &dataproc.ClusterArgs{
Region: pulumi.String("us-central1"),
})
if err != nil {
return err
}
spark, err := dataproc.NewJob(ctx, "spark", &dataproc.JobArgs{
Region: mycluster.Region,
ForceDelete: pulumi.Bool(true),
Placement: &dataproc.JobPlacementArgs{
ClusterName: mycluster.Name,
},
SparkConfig: &dataproc.JobSparkConfigArgs{
MainClass: pulumi.String("org.apache.spark.examples.SparkPi"),
JarFileUris: pulumi.StringArray{
pulumi.String("file:///usr/lib/spark/examples/jars/spark-examples.jar"),
},
Args: pulumi.StringArray{
pulumi.String("1000"),
},
Properties: pulumi.StringMap{
"spark.logConf": pulumi.String("true"),
},
LoggingConfig: &dataproc.JobSparkConfigLoggingConfigArgs{
DriverLogLevels: pulumi.StringMap{
"root": pulumi.String("INFO"),
},
},
},
})
if err != nil {
return err
}
pyspark, err := dataproc.NewJob(ctx, "pyspark", &dataproc.JobArgs{
Region: mycluster.Region,
ForceDelete: pulumi.Bool(true),
Placement: &dataproc.JobPlacementArgs{
ClusterName: mycluster.Name,
},
PysparkConfig: &dataproc.JobPysparkConfigArgs{
MainPythonFileUri: pulumi.String("gs://dataproc-examples-2f10d78d114f6aaec76462e3c310f31f/src/pyspark/hello-world/hello-world.py"),
Properties: pulumi.StringMap{
"spark.logConf": pulumi.String("true"),
},
},
})
if err != nil {
return err
}
ctx.Export("sparkStatus", spark.Statuses.ApplyT(func(statuses []dataproc.JobStatus) (*string, error) {
return &statuses[0].State, nil
}).(pulumi.StringPtrOutput))
ctx.Export("pysparkStatus", pyspark.Statuses.ApplyT(func(statuses []dataproc.JobStatus) (*string, error) {
return &statuses[0].State, nil
}).(pulumi.StringPtrOutput))
return nil
})
}
package generated_program;
import com.pulumi.Context;
import com.pulumi.Pulumi;
import com.pulumi.core.Output;
import com.pulumi.gcp.dataproc.Cluster;
import com.pulumi.gcp.dataproc.ClusterArgs;
import com.pulumi.gcp.dataproc.Job;
import com.pulumi.gcp.dataproc.JobArgs;
import com.pulumi.gcp.dataproc.inputs.JobPlacementArgs;
import com.pulumi.gcp.dataproc.inputs.JobSparkConfigArgs;
import com.pulumi.gcp.dataproc.inputs.JobSparkConfigLoggingConfigArgs;
import com.pulumi.gcp.dataproc.inputs.JobPysparkConfigArgs;
import java.util.List;
import java.util.ArrayList;
import java.util.Map;
import java.io.File;
import java.nio.file.Files;
import java.nio.file.Paths;
public class App {
public static void main(String[] args) {
Pulumi.run(App::stack);
}
public static void stack(Context ctx) {
var mycluster = new Cluster("mycluster", ClusterArgs.builder()
.region("us-central1")
.build());
var spark = new Job("spark", JobArgs.builder()
.region(mycluster.region())
.forceDelete(true)
.placement(JobPlacementArgs.builder()
.clusterName(mycluster.name())
.build())
.sparkConfig(JobSparkConfigArgs.builder()
.mainClass("org.apache.spark.examples.SparkPi")
.jarFileUris("file:///usr/lib/spark/examples/jars/spark-examples.jar")
.args("1000")
.properties(Map.of("spark.logConf", "true"))
.loggingConfig(JobSparkConfigLoggingConfigArgs.builder()
.driverLogLevels(Map.of("root", "INFO"))
.build())
.build())
.build());
var pyspark = new Job("pyspark", JobArgs.builder()
.region(mycluster.region())
.forceDelete(true)
.placement(JobPlacementArgs.builder()
.clusterName(mycluster.name())
.build())
.pysparkConfig(JobPysparkConfigArgs.builder()
.mainPythonFileUri("gs://dataproc-examples-2f10d78d114f6aaec76462e3c310f31f/src/pyspark/hello-world/hello-world.py")
.properties(Map.of("spark.logConf", "true"))
.build())
.build());
ctx.export("sparkStatus", spark.statuses().applyValue(statuses -> statuses[0].state()));
ctx.export("pysparkStatus", pyspark.statuses().applyValue(statuses -> statuses[0].state()));
}
}
import pulumi
import pulumi_gcp as gcp
mycluster = gcp.dataproc.Cluster("mycluster", region="us-central1")
# Submit an example spark job to a dataproc cluster
spark = gcp.dataproc.Job("spark",
region=mycluster.region,
force_delete=True,
placement=gcp.dataproc.JobPlacementArgs(
cluster_name=mycluster.name,
),
spark_config=gcp.dataproc.JobSparkConfigArgs(
main_class="org.apache.spark.examples.SparkPi",
jar_file_uris=["file:///usr/lib/spark/examples/jars/spark-examples.jar"],
args=["1000"],
properties={
"spark.logConf": "true",
},
logging_config=gcp.dataproc.JobSparkConfigLoggingConfigArgs(
driver_log_levels={
"root": "INFO",
},
),
))
# Submit an example pyspark job to a dataproc cluster
pyspark = gcp.dataproc.Job("pyspark",
region=mycluster.region,
force_delete=True,
placement=gcp.dataproc.JobPlacementArgs(
cluster_name=mycluster.name,
),
pyspark_config=gcp.dataproc.JobPysparkConfigArgs(
main_python_file_uri="gs://dataproc-examples-2f10d78d114f6aaec76462e3c310f31f/src/pyspark/hello-world/hello-world.py",
properties={
"spark.logConf": "true",
},
))
pulumi.export("sparkStatus", spark.statuses[0].state)
pulumi.export("pysparkStatus", pyspark.statuses[0].state)
import * as pulumi from "@pulumi/pulumi";
import * as gcp from "@pulumi/gcp";
const mycluster = new gcp.dataproc.Cluster("mycluster", {region: "us-central1"});
// Submit an example spark job to a dataproc cluster
const spark = new gcp.dataproc.Job("spark", {
region: mycluster.region,
forceDelete: true,
placement: {
clusterName: mycluster.name,
},
sparkConfig: {
mainClass: "org.apache.spark.examples.SparkPi",
jarFileUris: ["file:///usr/lib/spark/examples/jars/spark-examples.jar"],
args: ["1000"],
properties: {
"spark.logConf": "true",
},
loggingConfig: {
driverLogLevels: {
root: "INFO",
},
},
},
});
// Submit an example pyspark job to a dataproc cluster
const pyspark = new gcp.dataproc.Job("pyspark", {
region: mycluster.region,
forceDelete: true,
placement: {
clusterName: mycluster.name,
},
pysparkConfig: {
mainPythonFileUri: "gs://dataproc-examples-2f10d78d114f6aaec76462e3c310f31f/src/pyspark/hello-world/hello-world.py",
properties: {
"spark.logConf": "true",
},
},
});
export const sparkStatus = spark.statuses.apply(statuses => statuses[0].state);
export const pysparkStatus = pyspark.statuses.apply(statuses => statuses[0].state);
resources:
mycluster:
type: gcp:dataproc:Cluster
properties:
region: us-central1
# Submit an example spark job to a dataproc cluster
spark:
type: gcp:dataproc:Job
properties:
region: ${mycluster.region}
forceDelete: true
placement:
clusterName: ${mycluster.name}
sparkConfig:
mainClass: org.apache.spark.examples.SparkPi
jarFileUris:
- file:///usr/lib/spark/examples/jars/spark-examples.jar
args:
- '1000'
properties:
spark.logConf: 'true'
loggingConfig:
driverLogLevels:
root: INFO
# Submit an example pyspark job to a dataproc cluster
pyspark:
type: gcp:dataproc:Job
properties:
region: ${mycluster.region}
forceDelete: true
placement:
clusterName: ${mycluster.name}
pysparkConfig:
mainPythonFileUri: gs://dataproc-examples-2f10d78d114f6aaec76462e3c310f31f/src/pyspark/hello-world/hello-world.py
properties:
spark.logConf: 'true'
outputs:
# Check out current state of the jobs
sparkStatus: ${spark.statuses[0].state}
pysparkStatus: ${pyspark.statuses[0].state}
Create Job Resource
new Job(name: string, args: JobArgs, opts?: CustomResourceOptions);
@overload
def Job(resource_name: str,
opts: Optional[ResourceOptions] = None,
force_delete: Optional[bool] = None,
hadoop_config: Optional[JobHadoopConfigArgs] = None,
hive_config: Optional[JobHiveConfigArgs] = None,
labels: Optional[Mapping[str, str]] = None,
pig_config: Optional[JobPigConfigArgs] = None,
placement: Optional[JobPlacementArgs] = None,
presto_config: Optional[JobPrestoConfigArgs] = None,
project: Optional[str] = None,
pyspark_config: Optional[JobPysparkConfigArgs] = None,
reference: Optional[JobReferenceArgs] = None,
region: Optional[str] = None,
scheduling: Optional[JobSchedulingArgs] = None,
spark_config: Optional[JobSparkConfigArgs] = None,
sparksql_config: Optional[JobSparksqlConfigArgs] = None)
@overload
def Job(resource_name: str,
args: JobArgs,
opts: Optional[ResourceOptions] = None)
func NewJob(ctx *Context, name string, args JobArgs, opts ...ResourceOption) (*Job, error)
public Job(string name, JobArgs args, CustomResourceOptions? opts = null)
type: gcp:dataproc:Job
properties: # The arguments to resource properties.
options: # Bag of options to control resource's behavior.
- name string
- The unique name of the resource.
- args JobArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- resource_name str
- The unique name of the resource.
- args JobArgs
- The arguments to resource properties.
- opts ResourceOptions
- Bag of options to control resource's behavior.
- ctx Context
- Context object for the current deployment.
- name string
- The unique name of the resource.
- args JobArgs
- The arguments to resource properties.
- opts ResourceOption
- Bag of options to control resource's behavior.
- name string
- The unique name of the resource.
- args JobArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- name String
- The unique name of the resource.
- args JobArgs
- The arguments to resource properties.
- options CustomResourceOptions
- Bag of options to control resource's behavior.
Job Resource Properties
To learn more about resource properties and how to use them, see Inputs and Outputs in the Architecture and Concepts docs.
Inputs
The Job resource accepts the following input properties:
- Placement
Job
Placement The config of job placement.
- Force
Delete bool By default, you can only delete inactive jobs within Dataproc. Setting this to true, and calling destroy, will ensure that the job is first cancelled before issuing the delete.
- Hadoop
Config JobHadoop Config The config of Hadoop job
- Hive
Config JobHive Config The config of hive job
- Labels Dictionary<string, string>
The list of labels (key/value pairs) to add to the job.
scheduling.max_failures_per_hour
- (Required) Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.scheduling.max_failures_total
- (Required) Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.
- Pig
Config JobPig Config The config of pag job.
- Presto
Config JobPresto Config The config of presto job
- Project string
The project in which the
cluster
can be found and jobs subsequently run against. If it is not provided, the provider project is used.- Pyspark
Config JobPyspark Config The config of pySpark job.
- Reference
Job
Reference The reference of the job
- Region string
The Cloud Dataproc region. This essentially determines which clusters are available for this job to be submitted to. If not specified, defaults to
global
.- Scheduling
Job
Scheduling Optional. Job scheduling configuration.
- Spark
Config JobSpark Config The config of the Spark job.
- Sparksql
Config JobSparksql Config The config of SparkSql job
- Placement
Job
Placement Args The config of job placement.
- Force
Delete bool By default, you can only delete inactive jobs within Dataproc. Setting this to true, and calling destroy, will ensure that the job is first cancelled before issuing the delete.
- Hadoop
Config JobHadoop Config Args The config of Hadoop job
- Hive
Config JobHive Config Args The config of hive job
- Labels map[string]string
The list of labels (key/value pairs) to add to the job.
scheduling.max_failures_per_hour
- (Required) Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.scheduling.max_failures_total
- (Required) Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.
- Pig
Config JobPig Config Args The config of pag job.
- Presto
Config JobPresto Config Args The config of presto job
- Project string
The project in which the
cluster
can be found and jobs subsequently run against. If it is not provided, the provider project is used.- Pyspark
Config JobPyspark Config Args The config of pySpark job.
- Reference
Job
Reference Args The reference of the job
- Region string
The Cloud Dataproc region. This essentially determines which clusters are available for this job to be submitted to. If not specified, defaults to
global
.- Scheduling
Job
Scheduling Args Optional. Job scheduling configuration.
- Spark
Config JobSpark Config Args The config of the Spark job.
- Sparksql
Config JobSparksql Config Args The config of SparkSql job
- placement
Job
Placement The config of job placement.
- force
Delete Boolean By default, you can only delete inactive jobs within Dataproc. Setting this to true, and calling destroy, will ensure that the job is first cancelled before issuing the delete.
- hadoop
Config JobHadoop Config The config of Hadoop job
- hive
Config JobHive Config The config of hive job
- labels Map<String,String>
The list of labels (key/value pairs) to add to the job.
scheduling.max_failures_per_hour
- (Required) Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.scheduling.max_failures_total
- (Required) Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.
- pig
Config JobPig Config The config of pag job.
- presto
Config JobPresto Config The config of presto job
- project String
The project in which the
cluster
can be found and jobs subsequently run against. If it is not provided, the provider project is used.- pyspark
Config JobPyspark Config The config of pySpark job.
- reference
Job
Reference The reference of the job
- region String
The Cloud Dataproc region. This essentially determines which clusters are available for this job to be submitted to. If not specified, defaults to
global
.- scheduling
Job
Scheduling Optional. Job scheduling configuration.
- spark
Config JobSpark Config The config of the Spark job.
- sparksql
Config JobSparksql Config The config of SparkSql job
- placement
Job
Placement The config of job placement.
- force
Delete boolean By default, you can only delete inactive jobs within Dataproc. Setting this to true, and calling destroy, will ensure that the job is first cancelled before issuing the delete.
- hadoop
Config JobHadoop Config The config of Hadoop job
- hive
Config JobHive Config The config of hive job
- labels {[key: string]: string}
The list of labels (key/value pairs) to add to the job.
scheduling.max_failures_per_hour
- (Required) Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.scheduling.max_failures_total
- (Required) Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.
- pig
Config JobPig Config The config of pag job.
- presto
Config JobPresto Config The config of presto job
- project string
The project in which the
cluster
can be found and jobs subsequently run against. If it is not provided, the provider project is used.- pyspark
Config JobPyspark Config The config of pySpark job.
- reference
Job
Reference The reference of the job
- region string
The Cloud Dataproc region. This essentially determines which clusters are available for this job to be submitted to. If not specified, defaults to
global
.- scheduling
Job
Scheduling Optional. Job scheduling configuration.
- spark
Config JobSpark Config The config of the Spark job.
- sparksql
Config JobSparksql Config The config of SparkSql job
- placement
Job
Placement Args The config of job placement.
- force_
delete bool By default, you can only delete inactive jobs within Dataproc. Setting this to true, and calling destroy, will ensure that the job is first cancelled before issuing the delete.
- hadoop_
config JobHadoop Config Args The config of Hadoop job
- hive_
config JobHive Config Args The config of hive job
- labels Mapping[str, str]
The list of labels (key/value pairs) to add to the job.
scheduling.max_failures_per_hour
- (Required) Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.scheduling.max_failures_total
- (Required) Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.
- pig_
config JobPig Config Args The config of pag job.
- presto_
config JobPresto Config Args The config of presto job
- project str
The project in which the
cluster
can be found and jobs subsequently run against. If it is not provided, the provider project is used.- pyspark_
config JobPyspark Config Args The config of pySpark job.
- reference
Job
Reference Args The reference of the job
- region str
The Cloud Dataproc region. This essentially determines which clusters are available for this job to be submitted to. If not specified, defaults to
global
.- scheduling
Job
Scheduling Args Optional. Job scheduling configuration.
- spark_
config JobSpark Config Args The config of the Spark job.
- sparksql_
config JobSparksql Config Args The config of SparkSql job
- placement Property Map
The config of job placement.
- force
Delete Boolean By default, you can only delete inactive jobs within Dataproc. Setting this to true, and calling destroy, will ensure that the job is first cancelled before issuing the delete.
- hadoop
Config Property Map The config of Hadoop job
- hive
Config Property Map The config of hive job
- labels Map<String>
The list of labels (key/value pairs) to add to the job.
scheduling.max_failures_per_hour
- (Required) Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.scheduling.max_failures_total
- (Required) Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.
- pig
Config Property Map The config of pag job.
- presto
Config Property Map The config of presto job
- project String
The project in which the
cluster
can be found and jobs subsequently run against. If it is not provided, the provider project is used.- pyspark
Config Property Map The config of pySpark job.
- reference Property Map
The reference of the job
- region String
The Cloud Dataproc region. This essentially determines which clusters are available for this job to be submitted to. If not specified, defaults to
global
.- scheduling Property Map
Optional. Job scheduling configuration.
- spark
Config Property Map The config of the Spark job.
- sparksql
Config Property Map The config of SparkSql job
Outputs
All input properties are implicitly available as output properties. Additionally, the Job resource produces the following output properties:
- Driver
Controls stringFiles Uri If present, the location of miscellaneous control files which may be used as part of job setup and handling. If not present, control files may be placed in the same location as driver_output_uri.
- Driver
Output stringResource Uri A URI pointing to the location of the stdout of the job's driver program.
- Id string
The provider-assigned unique ID for this managed resource.
- Statuses
List<Job
Status> The status of the job.
- Driver
Controls stringFiles Uri If present, the location of miscellaneous control files which may be used as part of job setup and handling. If not present, control files may be placed in the same location as driver_output_uri.
- Driver
Output stringResource Uri A URI pointing to the location of the stdout of the job's driver program.
- Id string
The provider-assigned unique ID for this managed resource.
- Statuses
[]Job
Status The status of the job.
- driver
Controls StringFiles Uri If present, the location of miscellaneous control files which may be used as part of job setup and handling. If not present, control files may be placed in the same location as driver_output_uri.
- driver
Output StringResource Uri A URI pointing to the location of the stdout of the job's driver program.
- id String
The provider-assigned unique ID for this managed resource.
- statuses
List<Job
Status> The status of the job.
- driver
Controls stringFiles Uri If present, the location of miscellaneous control files which may be used as part of job setup and handling. If not present, control files may be placed in the same location as driver_output_uri.
- driver
Output stringResource Uri A URI pointing to the location of the stdout of the job's driver program.
- id string
The provider-assigned unique ID for this managed resource.
- statuses
Job
Status[] The status of the job.
- driver_
controls_ strfiles_ uri If present, the location of miscellaneous control files which may be used as part of job setup and handling. If not present, control files may be placed in the same location as driver_output_uri.
- driver_
output_ strresource_ uri A URI pointing to the location of the stdout of the job's driver program.
- id str
The provider-assigned unique ID for this managed resource.
- statuses
Sequence[Job
Status] The status of the job.
- driver
Controls StringFiles Uri If present, the location of miscellaneous control files which may be used as part of job setup and handling. If not present, control files may be placed in the same location as driver_output_uri.
- driver
Output StringResource Uri A URI pointing to the location of the stdout of the job's driver program.
- id String
The provider-assigned unique ID for this managed resource.
- statuses List<Property Map>
The status of the job.
Look up Existing Job Resource
Get an existing Job resource’s state with the given name, ID, and optional extra properties used to qualify the lookup.
public static get(name: string, id: Input<ID>, state?: JobState, opts?: CustomResourceOptions): Job
@staticmethod
def get(resource_name: str,
id: str,
opts: Optional[ResourceOptions] = None,
driver_controls_files_uri: Optional[str] = None,
driver_output_resource_uri: Optional[str] = None,
force_delete: Optional[bool] = None,
hadoop_config: Optional[JobHadoopConfigArgs] = None,
hive_config: Optional[JobHiveConfigArgs] = None,
labels: Optional[Mapping[str, str]] = None,
pig_config: Optional[JobPigConfigArgs] = None,
placement: Optional[JobPlacementArgs] = None,
presto_config: Optional[JobPrestoConfigArgs] = None,
project: Optional[str] = None,
pyspark_config: Optional[JobPysparkConfigArgs] = None,
reference: Optional[JobReferenceArgs] = None,
region: Optional[str] = None,
scheduling: Optional[JobSchedulingArgs] = None,
spark_config: Optional[JobSparkConfigArgs] = None,
sparksql_config: Optional[JobSparksqlConfigArgs] = None,
statuses: Optional[Sequence[JobStatusArgs]] = None) -> Job
func GetJob(ctx *Context, name string, id IDInput, state *JobState, opts ...ResourceOption) (*Job, error)
public static Job Get(string name, Input<string> id, JobState? state, CustomResourceOptions? opts = null)
public static Job get(String name, Output<String> id, JobState state, CustomResourceOptions options)
Resource lookup is not supported in YAML
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- resource_name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- Driver
Controls stringFiles Uri If present, the location of miscellaneous control files which may be used as part of job setup and handling. If not present, control files may be placed in the same location as driver_output_uri.
- Driver
Output stringResource Uri A URI pointing to the location of the stdout of the job's driver program.
- Force
Delete bool By default, you can only delete inactive jobs within Dataproc. Setting this to true, and calling destroy, will ensure that the job is first cancelled before issuing the delete.
- Hadoop
Config JobHadoop Config The config of Hadoop job
- Hive
Config JobHive Config The config of hive job
- Labels Dictionary<string, string>
The list of labels (key/value pairs) to add to the job.
scheduling.max_failures_per_hour
- (Required) Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.scheduling.max_failures_total
- (Required) Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.
- Pig
Config JobPig Config The config of pag job.
- Placement
Job
Placement The config of job placement.
- Presto
Config JobPresto Config The config of presto job
- Project string
The project in which the
cluster
can be found and jobs subsequently run against. If it is not provided, the provider project is used.- Pyspark
Config JobPyspark Config The config of pySpark job.
- Reference
Job
Reference The reference of the job
- Region string
The Cloud Dataproc region. This essentially determines which clusters are available for this job to be submitted to. If not specified, defaults to
global
.- Scheduling
Job
Scheduling Optional. Job scheduling configuration.
- Spark
Config JobSpark Config The config of the Spark job.
- Sparksql
Config JobSparksql Config The config of SparkSql job
- Statuses
List<Job
Status> The status of the job.
- Driver
Controls stringFiles Uri If present, the location of miscellaneous control files which may be used as part of job setup and handling. If not present, control files may be placed in the same location as driver_output_uri.
- Driver
Output stringResource Uri A URI pointing to the location of the stdout of the job's driver program.
- Force
Delete bool By default, you can only delete inactive jobs within Dataproc. Setting this to true, and calling destroy, will ensure that the job is first cancelled before issuing the delete.
- Hadoop
Config JobHadoop Config Args The config of Hadoop job
- Hive
Config JobHive Config Args The config of hive job
- Labels map[string]string
The list of labels (key/value pairs) to add to the job.
scheduling.max_failures_per_hour
- (Required) Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.scheduling.max_failures_total
- (Required) Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.
- Pig
Config JobPig Config Args The config of pag job.
- Placement
Job
Placement Args The config of job placement.
- Presto
Config JobPresto Config Args The config of presto job
- Project string
The project in which the
cluster
can be found and jobs subsequently run against. If it is not provided, the provider project is used.- Pyspark
Config JobPyspark Config Args The config of pySpark job.
- Reference
Job
Reference Args The reference of the job
- Region string
The Cloud Dataproc region. This essentially determines which clusters are available for this job to be submitted to. If not specified, defaults to
global
.- Scheduling
Job
Scheduling Args Optional. Job scheduling configuration.
- Spark
Config JobSpark Config Args The config of the Spark job.
- Sparksql
Config JobSparksql Config Args The config of SparkSql job
- Statuses
[]Job
Status Args The status of the job.
- driver
Controls StringFiles Uri If present, the location of miscellaneous control files which may be used as part of job setup and handling. If not present, control files may be placed in the same location as driver_output_uri.
- driver
Output StringResource Uri A URI pointing to the location of the stdout of the job's driver program.
- force
Delete Boolean By default, you can only delete inactive jobs within Dataproc. Setting this to true, and calling destroy, will ensure that the job is first cancelled before issuing the delete.
- hadoop
Config JobHadoop Config The config of Hadoop job
- hive
Config JobHive Config The config of hive job
- labels Map<String,String>
The list of labels (key/value pairs) to add to the job.
scheduling.max_failures_per_hour
- (Required) Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.scheduling.max_failures_total
- (Required) Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.
- pig
Config JobPig Config The config of pag job.
- placement
Job
Placement The config of job placement.
- presto
Config JobPresto Config The config of presto job
- project String
The project in which the
cluster
can be found and jobs subsequently run against. If it is not provided, the provider project is used.- pyspark
Config JobPyspark Config The config of pySpark job.
- reference
Job
Reference The reference of the job
- region String
The Cloud Dataproc region. This essentially determines which clusters are available for this job to be submitted to. If not specified, defaults to
global
.- scheduling
Job
Scheduling Optional. Job scheduling configuration.
- spark
Config JobSpark Config The config of the Spark job.
- sparksql
Config JobSparksql Config The config of SparkSql job
- statuses
List<Job
Status> The status of the job.
- driver
Controls stringFiles Uri If present, the location of miscellaneous control files which may be used as part of job setup and handling. If not present, control files may be placed in the same location as driver_output_uri.
- driver
Output stringResource Uri A URI pointing to the location of the stdout of the job's driver program.
- force
Delete boolean By default, you can only delete inactive jobs within Dataproc. Setting this to true, and calling destroy, will ensure that the job is first cancelled before issuing the delete.
- hadoop
Config JobHadoop Config The config of Hadoop job
- hive
Config JobHive Config The config of hive job
- labels {[key: string]: string}
The list of labels (key/value pairs) to add to the job.
scheduling.max_failures_per_hour
- (Required) Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.scheduling.max_failures_total
- (Required) Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.
- pig
Config JobPig Config The config of pag job.
- placement
Job
Placement The config of job placement.
- presto
Config JobPresto Config The config of presto job
- project string
The project in which the
cluster
can be found and jobs subsequently run against. If it is not provided, the provider project is used.- pyspark
Config JobPyspark Config The config of pySpark job.
- reference
Job
Reference The reference of the job
- region string
The Cloud Dataproc region. This essentially determines which clusters are available for this job to be submitted to. If not specified, defaults to
global
.- scheduling
Job
Scheduling Optional. Job scheduling configuration.
- spark
Config JobSpark Config The config of the Spark job.
- sparksql
Config JobSparksql Config The config of SparkSql job
- statuses
Job
Status[] The status of the job.
- driver_
controls_ strfiles_ uri If present, the location of miscellaneous control files which may be used as part of job setup and handling. If not present, control files may be placed in the same location as driver_output_uri.
- driver_
output_ strresource_ uri A URI pointing to the location of the stdout of the job's driver program.
- force_
delete bool By default, you can only delete inactive jobs within Dataproc. Setting this to true, and calling destroy, will ensure that the job is first cancelled before issuing the delete.
- hadoop_
config JobHadoop Config Args The config of Hadoop job
- hive_
config JobHive Config Args The config of hive job
- labels Mapping[str, str]
The list of labels (key/value pairs) to add to the job.
scheduling.max_failures_per_hour
- (Required) Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.scheduling.max_failures_total
- (Required) Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.
- pig_
config JobPig Config Args The config of pag job.
- placement
Job
Placement Args The config of job placement.
- presto_
config JobPresto Config Args The config of presto job
- project str
The project in which the
cluster
can be found and jobs subsequently run against. If it is not provided, the provider project is used.- pyspark_
config JobPyspark Config Args The config of pySpark job.
- reference
Job
Reference Args The reference of the job
- region str
The Cloud Dataproc region. This essentially determines which clusters are available for this job to be submitted to. If not specified, defaults to
global
.- scheduling
Job
Scheduling Args Optional. Job scheduling configuration.
- spark_
config JobSpark Config Args The config of the Spark job.
- sparksql_
config JobSparksql Config Args The config of SparkSql job
- statuses
Sequence[Job
Status Args] The status of the job.
- driver
Controls StringFiles Uri If present, the location of miscellaneous control files which may be used as part of job setup and handling. If not present, control files may be placed in the same location as driver_output_uri.
- driver
Output StringResource Uri A URI pointing to the location of the stdout of the job's driver program.
- force
Delete Boolean By default, you can only delete inactive jobs within Dataproc. Setting this to true, and calling destroy, will ensure that the job is first cancelled before issuing the delete.
- hadoop
Config Property Map The config of Hadoop job
- hive
Config Property Map The config of hive job
- labels Map<String>
The list of labels (key/value pairs) to add to the job.
scheduling.max_failures_per_hour
- (Required) Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.scheduling.max_failures_total
- (Required) Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.
- pig
Config Property Map The config of pag job.
- placement Property Map
The config of job placement.
- presto
Config Property Map The config of presto job
- project String
The project in which the
cluster
can be found and jobs subsequently run against. If it is not provided, the provider project is used.- pyspark
Config Property Map The config of pySpark job.
- reference Property Map
The reference of the job
- region String
The Cloud Dataproc region. This essentially determines which clusters are available for this job to be submitted to. If not specified, defaults to
global
.- scheduling Property Map
Optional. Job scheduling configuration.
- spark
Config Property Map The config of the Spark job.
- sparksql
Config Property Map The config of SparkSql job
- statuses List<Property Map>
The status of the job.
Supporting Types
JobHadoopConfig, JobHadoopConfigArgs
- Archive
Uris List<string> HCFS URIs of archives to be extracted in the working directory of .jar, .tar, .tar.gz, .tgz, and .zip.
- Args List<string>
The arguments to pass to the driver. Do not include arguments, such as -libjars or -Dfoo=bar, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
- File
Uris List<string> HCFS URIs of files to be copied to the working directory of Hadoop drivers and distributed tasks. Useful for naively parallel tasks.
- Jar
File List<string>Uris HCFS URIs of jar files to add to the CLASSPATHs of the Spark driver and tasks.
- Logging
Config JobHadoop Config Logging Config - Main
Class string The name of the driver's main class. The jar file containing the class must be in the default CLASSPATH or specified in
jar_file_uris
. Conflicts withmain_jar_file_uri
- Main
Jar stringFile Uri The HCFS URI of the jar file containing the main class. Examples: 'gs://foo-bucket/analytics-binaries/extract-useful-metrics-mr.jar' 'hdfs:/tmp/test-samples/custom-wordcount.jar' 'file:///home/usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples.jar'. Conflicts with
main_class
- Properties Dictionary<string, string>
A mapping of property names to values, used to configure Hadoop. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in
/etc/hadoop/conf/*-site
and classes in user code..logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- Archive
Uris []string HCFS URIs of archives to be extracted in the working directory of .jar, .tar, .tar.gz, .tgz, and .zip.
- Args []string
The arguments to pass to the driver. Do not include arguments, such as -libjars or -Dfoo=bar, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
- File
Uris []string HCFS URIs of files to be copied to the working directory of Hadoop drivers and distributed tasks. Useful for naively parallel tasks.
- Jar
File []stringUris HCFS URIs of jar files to add to the CLASSPATHs of the Spark driver and tasks.
- Logging
Config JobHadoop Config Logging Config - Main
Class string The name of the driver's main class. The jar file containing the class must be in the default CLASSPATH or specified in
jar_file_uris
. Conflicts withmain_jar_file_uri
- Main
Jar stringFile Uri The HCFS URI of the jar file containing the main class. Examples: 'gs://foo-bucket/analytics-binaries/extract-useful-metrics-mr.jar' 'hdfs:/tmp/test-samples/custom-wordcount.jar' 'file:///home/usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples.jar'. Conflicts with
main_class
- Properties map[string]string
A mapping of property names to values, used to configure Hadoop. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in
/etc/hadoop/conf/*-site
and classes in user code..logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- archive
Uris List<String> HCFS URIs of archives to be extracted in the working directory of .jar, .tar, .tar.gz, .tgz, and .zip.
- args List<String>
The arguments to pass to the driver. Do not include arguments, such as -libjars or -Dfoo=bar, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
- file
Uris List<String> HCFS URIs of files to be copied to the working directory of Hadoop drivers and distributed tasks. Useful for naively parallel tasks.
- jar
File List<String>Uris HCFS URIs of jar files to add to the CLASSPATHs of the Spark driver and tasks.
- logging
Config JobHadoop Config Logging Config - main
Class String The name of the driver's main class. The jar file containing the class must be in the default CLASSPATH or specified in
jar_file_uris
. Conflicts withmain_jar_file_uri
- main
Jar StringFile Uri The HCFS URI of the jar file containing the main class. Examples: 'gs://foo-bucket/analytics-binaries/extract-useful-metrics-mr.jar' 'hdfs:/tmp/test-samples/custom-wordcount.jar' 'file:///home/usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples.jar'. Conflicts with
main_class
- properties Map<String,String>
A mapping of property names to values, used to configure Hadoop. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in
/etc/hadoop/conf/*-site
and classes in user code..logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- archive
Uris string[] HCFS URIs of archives to be extracted in the working directory of .jar, .tar, .tar.gz, .tgz, and .zip.
- args string[]
The arguments to pass to the driver. Do not include arguments, such as -libjars or -Dfoo=bar, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
- file
Uris string[] HCFS URIs of files to be copied to the working directory of Hadoop drivers and distributed tasks. Useful for naively parallel tasks.
- jar
File string[]Uris HCFS URIs of jar files to add to the CLASSPATHs of the Spark driver and tasks.
- logging
Config JobHadoop Config Logging Config - main
Class string The name of the driver's main class. The jar file containing the class must be in the default CLASSPATH or specified in
jar_file_uris
. Conflicts withmain_jar_file_uri
- main
Jar stringFile Uri The HCFS URI of the jar file containing the main class. Examples: 'gs://foo-bucket/analytics-binaries/extract-useful-metrics-mr.jar' 'hdfs:/tmp/test-samples/custom-wordcount.jar' 'file:///home/usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples.jar'. Conflicts with
main_class
- properties {[key: string]: string}
A mapping of property names to values, used to configure Hadoop. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in
/etc/hadoop/conf/*-site
and classes in user code..logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- archive_
uris Sequence[str] HCFS URIs of archives to be extracted in the working directory of .jar, .tar, .tar.gz, .tgz, and .zip.
- args Sequence[str]
The arguments to pass to the driver. Do not include arguments, such as -libjars or -Dfoo=bar, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
- file_
uris Sequence[str] HCFS URIs of files to be copied to the working directory of Hadoop drivers and distributed tasks. Useful for naively parallel tasks.
- jar_
file_ Sequence[str]uris HCFS URIs of jar files to add to the CLASSPATHs of the Spark driver and tasks.
- logging_
config JobHadoop Config Logging Config - main_
class str The name of the driver's main class. The jar file containing the class must be in the default CLASSPATH or specified in
jar_file_uris
. Conflicts withmain_jar_file_uri
- main_
jar_ strfile_ uri The HCFS URI of the jar file containing the main class. Examples: 'gs://foo-bucket/analytics-binaries/extract-useful-metrics-mr.jar' 'hdfs:/tmp/test-samples/custom-wordcount.jar' 'file:///home/usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples.jar'. Conflicts with
main_class
- properties Mapping[str, str]
A mapping of property names to values, used to configure Hadoop. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in
/etc/hadoop/conf/*-site
and classes in user code..logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- archive
Uris List<String> HCFS URIs of archives to be extracted in the working directory of .jar, .tar, .tar.gz, .tgz, and .zip.
- args List<String>
The arguments to pass to the driver. Do not include arguments, such as -libjars or -Dfoo=bar, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
- file
Uris List<String> HCFS URIs of files to be copied to the working directory of Hadoop drivers and distributed tasks. Useful for naively parallel tasks.
- jar
File List<String>Uris HCFS URIs of jar files to add to the CLASSPATHs of the Spark driver and tasks.
- logging
Config Property Map - main
Class String The name of the driver's main class. The jar file containing the class must be in the default CLASSPATH or specified in
jar_file_uris
. Conflicts withmain_jar_file_uri
- main
Jar StringFile Uri The HCFS URI of the jar file containing the main class. Examples: 'gs://foo-bucket/analytics-binaries/extract-useful-metrics-mr.jar' 'hdfs:/tmp/test-samples/custom-wordcount.jar' 'file:///home/usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples.jar'. Conflicts with
main_class
- properties Map<String>
A mapping of property names to values, used to configure Hadoop. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in
/etc/hadoop/conf/*-site
and classes in user code..logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
JobHadoopConfigLoggingConfig, JobHadoopConfigLoggingConfigArgs
- Driver
Log Dictionary<string, string>Levels
- Driver
Log map[string]stringLevels
- driver
Log Map<String,String>Levels
- driver
Log {[key: string]: string}Levels
- driver_
log_ Mapping[str, str]levels
- driver
Log Map<String>Levels
JobHiveConfig, JobHiveConfigArgs
- Continue
On boolFailure Whether to continue executing queries if a query fails. The default value is false. Setting to true can be useful when executing independent parallel queries. Defaults to false.
- Jar
File List<string>Uris HCFS URIs of jar files to add to the CLASSPATH of the Hive server and Hadoop MapReduce (MR) tasks. Can contain Hive SerDes and UDFs.
- Properties Dictionary<string, string>
A mapping of property names and values, used to configure Hive. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in
/etc/hadoop/conf/*-site.xml
,/etc/hive/conf/hive-site.xml
, and classes in user code..- Query
File stringUri HCFS URI of file containing Hive script to execute as the job. Conflicts with
query_list
- Query
Lists List<string> The list of Hive queries or statements to execute as part of the job. Conflicts with
query_file_uri
- Script
Variables Dictionary<string, string> Mapping of query variable names to values (equivalent to the Hive command:
SET name="value";
).
- Continue
On boolFailure Whether to continue executing queries if a query fails. The default value is false. Setting to true can be useful when executing independent parallel queries. Defaults to false.
- Jar
File []stringUris HCFS URIs of jar files to add to the CLASSPATH of the Hive server and Hadoop MapReduce (MR) tasks. Can contain Hive SerDes and UDFs.
- Properties map[string]string
A mapping of property names and values, used to configure Hive. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in
/etc/hadoop/conf/*-site.xml
,/etc/hive/conf/hive-site.xml
, and classes in user code..- Query
File stringUri HCFS URI of file containing Hive script to execute as the job. Conflicts with
query_list
- Query
Lists []string The list of Hive queries or statements to execute as part of the job. Conflicts with
query_file_uri
- Script
Variables map[string]string Mapping of query variable names to values (equivalent to the Hive command:
SET name="value";
).
- continue
On BooleanFailure Whether to continue executing queries if a query fails. The default value is false. Setting to true can be useful when executing independent parallel queries. Defaults to false.
- jar
File List<String>Uris HCFS URIs of jar files to add to the CLASSPATH of the Hive server and Hadoop MapReduce (MR) tasks. Can contain Hive SerDes and UDFs.
- properties Map<String,String>
A mapping of property names and values, used to configure Hive. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in
/etc/hadoop/conf/*-site.xml
,/etc/hive/conf/hive-site.xml
, and classes in user code..- query
File StringUri HCFS URI of file containing Hive script to execute as the job. Conflicts with
query_list
- query
Lists List<String> The list of Hive queries or statements to execute as part of the job. Conflicts with
query_file_uri
- script
Variables Map<String,String> Mapping of query variable names to values (equivalent to the Hive command:
SET name="value";
).
- continue
On booleanFailure Whether to continue executing queries if a query fails. The default value is false. Setting to true can be useful when executing independent parallel queries. Defaults to false.
- jar
File string[]Uris HCFS URIs of jar files to add to the CLASSPATH of the Hive server and Hadoop MapReduce (MR) tasks. Can contain Hive SerDes and UDFs.
- properties {[key: string]: string}
A mapping of property names and values, used to configure Hive. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in
/etc/hadoop/conf/*-site.xml
,/etc/hive/conf/hive-site.xml
, and classes in user code..- query
File stringUri HCFS URI of file containing Hive script to execute as the job. Conflicts with
query_list
- query
Lists string[] The list of Hive queries or statements to execute as part of the job. Conflicts with
query_file_uri
- script
Variables {[key: string]: string} Mapping of query variable names to values (equivalent to the Hive command:
SET name="value";
).
- continue_
on_ boolfailure Whether to continue executing queries if a query fails. The default value is false. Setting to true can be useful when executing independent parallel queries. Defaults to false.
- jar_
file_ Sequence[str]uris HCFS URIs of jar files to add to the CLASSPATH of the Hive server and Hadoop MapReduce (MR) tasks. Can contain Hive SerDes and UDFs.
- properties Mapping[str, str]
A mapping of property names and values, used to configure Hive. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in
/etc/hadoop/conf/*-site.xml
,/etc/hive/conf/hive-site.xml
, and classes in user code..- query_
file_ struri HCFS URI of file containing Hive script to execute as the job. Conflicts with
query_list
- query_
lists Sequence[str] The list of Hive queries or statements to execute as part of the job. Conflicts with
query_file_uri
- script_
variables Mapping[str, str] Mapping of query variable names to values (equivalent to the Hive command:
SET name="value";
).
- continue
On BooleanFailure Whether to continue executing queries if a query fails. The default value is false. Setting to true can be useful when executing independent parallel queries. Defaults to false.
- jar
File List<String>Uris HCFS URIs of jar files to add to the CLASSPATH of the Hive server and Hadoop MapReduce (MR) tasks. Can contain Hive SerDes and UDFs.
- properties Map<String>
A mapping of property names and values, used to configure Hive. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in
/etc/hadoop/conf/*-site.xml
,/etc/hive/conf/hive-site.xml
, and classes in user code..- query
File StringUri HCFS URI of file containing Hive script to execute as the job. Conflicts with
query_list
- query
Lists List<String> The list of Hive queries or statements to execute as part of the job. Conflicts with
query_file_uri
- script
Variables Map<String> Mapping of query variable names to values (equivalent to the Hive command:
SET name="value";
).
JobPigConfig, JobPigConfigArgs
- Continue
On boolFailure Whether to continue executing queries if a query fails. The default value is false. Setting to true can be useful when executing independent parallel queries. Defaults to false.
- Jar
File List<string>Uris HCFS URIs of jar files to add to the CLASSPATH of the Pig Client and Hadoop MapReduce (MR) tasks. Can contain Pig UDFs.
logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- Logging
Config JobPig Config Logging Config - Properties Dictionary<string, string>
A mapping of property names to values, used to configure Pig. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in
/etc/hadoop/conf/*-site.xml
,/etc/pig/conf/pig.properties
, and classes in user code.- Query
File stringUri HCFS URI of file containing Hive script to execute as the job. Conflicts with
query_list
- Query
Lists List<string> The list of Hive queries or statements to execute as part of the job. Conflicts with
query_file_uri
- Script
Variables Dictionary<string, string> Mapping of query variable names to values (equivalent to the Pig command:
name=[value]
).
- Continue
On boolFailure Whether to continue executing queries if a query fails. The default value is false. Setting to true can be useful when executing independent parallel queries. Defaults to false.
- Jar
File []stringUris HCFS URIs of jar files to add to the CLASSPATH of the Pig Client and Hadoop MapReduce (MR) tasks. Can contain Pig UDFs.
logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- Logging
Config JobPig Config Logging Config - Properties map[string]string
A mapping of property names to values, used to configure Pig. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in
/etc/hadoop/conf/*-site.xml
,/etc/pig/conf/pig.properties
, and classes in user code.- Query
File stringUri HCFS URI of file containing Hive script to execute as the job. Conflicts with
query_list
- Query
Lists []string The list of Hive queries or statements to execute as part of the job. Conflicts with
query_file_uri
- Script
Variables map[string]string Mapping of query variable names to values (equivalent to the Pig command:
name=[value]
).
- continue
On BooleanFailure Whether to continue executing queries if a query fails. The default value is false. Setting to true can be useful when executing independent parallel queries. Defaults to false.
- jar
File List<String>Uris HCFS URIs of jar files to add to the CLASSPATH of the Pig Client and Hadoop MapReduce (MR) tasks. Can contain Pig UDFs.
logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- logging
Config JobPig Config Logging Config - properties Map<String,String>
A mapping of property names to values, used to configure Pig. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in
/etc/hadoop/conf/*-site.xml
,/etc/pig/conf/pig.properties
, and classes in user code.- query
File StringUri HCFS URI of file containing Hive script to execute as the job. Conflicts with
query_list
- query
Lists List<String> The list of Hive queries or statements to execute as part of the job. Conflicts with
query_file_uri
- script
Variables Map<String,String> Mapping of query variable names to values (equivalent to the Pig command:
name=[value]
).
- continue
On booleanFailure Whether to continue executing queries if a query fails. The default value is false. Setting to true can be useful when executing independent parallel queries. Defaults to false.
- jar
File string[]Uris HCFS URIs of jar files to add to the CLASSPATH of the Pig Client and Hadoop MapReduce (MR) tasks. Can contain Pig UDFs.
logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- logging
Config JobPig Config Logging Config - properties {[key: string]: string}
A mapping of property names to values, used to configure Pig. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in
/etc/hadoop/conf/*-site.xml
,/etc/pig/conf/pig.properties
, and classes in user code.- query
File stringUri HCFS URI of file containing Hive script to execute as the job. Conflicts with
query_list
- query
Lists string[] The list of Hive queries or statements to execute as part of the job. Conflicts with
query_file_uri
- script
Variables {[key: string]: string} Mapping of query variable names to values (equivalent to the Pig command:
name=[value]
).
- continue_
on_ boolfailure Whether to continue executing queries if a query fails. The default value is false. Setting to true can be useful when executing independent parallel queries. Defaults to false.
- jar_
file_ Sequence[str]uris HCFS URIs of jar files to add to the CLASSPATH of the Pig Client and Hadoop MapReduce (MR) tasks. Can contain Pig UDFs.
logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- logging_
config JobPig Config Logging Config - properties Mapping[str, str]
A mapping of property names to values, used to configure Pig. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in
/etc/hadoop/conf/*-site.xml
,/etc/pig/conf/pig.properties
, and classes in user code.- query_
file_ struri HCFS URI of file containing Hive script to execute as the job. Conflicts with
query_list
- query_
lists Sequence[str] The list of Hive queries or statements to execute as part of the job. Conflicts with
query_file_uri
- script_
variables Mapping[str, str] Mapping of query variable names to values (equivalent to the Pig command:
name=[value]
).
- continue
On BooleanFailure Whether to continue executing queries if a query fails. The default value is false. Setting to true can be useful when executing independent parallel queries. Defaults to false.
- jar
File List<String>Uris HCFS URIs of jar files to add to the CLASSPATH of the Pig Client and Hadoop MapReduce (MR) tasks. Can contain Pig UDFs.
logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- logging
Config Property Map - properties Map<String>
A mapping of property names to values, used to configure Pig. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in
/etc/hadoop/conf/*-site.xml
,/etc/pig/conf/pig.properties
, and classes in user code.- query
File StringUri HCFS URI of file containing Hive script to execute as the job. Conflicts with
query_list
- query
Lists List<String> The list of Hive queries or statements to execute as part of the job. Conflicts with
query_file_uri
- script
Variables Map<String> Mapping of query variable names to values (equivalent to the Pig command:
name=[value]
).
JobPigConfigLoggingConfig, JobPigConfigLoggingConfigArgs
- Driver
Log Dictionary<string, string>Levels
- Driver
Log map[string]stringLevels
- driver
Log Map<String,String>Levels
- driver
Log {[key: string]: string}Levels
- driver_
log_ Mapping[str, str]levels
- driver
Log Map<String>Levels
JobPlacement, JobPlacementArgs
- Cluster
Name string - Cluster
Uuid string
- Cluster
Name string - Cluster
Uuid string
- cluster
Name String - cluster
Uuid String
- cluster
Name string - cluster
Uuid string
- cluster_
name str - cluster_
uuid str
- cluster
Name String - cluster
Uuid String
JobPrestoConfig, JobPrestoConfigArgs
- List<string>
Presto client tags to attach to this query.
- Continue
On boolFailure Whether to continue executing queries if a query fails. Setting to true can be useful when executing independent parallel queries. Defaults to false.
- Logging
Config JobPresto Config Logging Config - Output
Format string The format in which query output will be displayed. See the Presto documentation for supported output formats.
logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- Properties Dictionary<string, string>
A mapping of property names to values. Used to set Presto session properties Equivalent to using the --session flag in the Presto CLI.
- Query
File stringUri The HCFS URI of the script that contains SQL queries. Conflicts with
query_list
- Query
Lists List<string> The list of SQL queries or statements to execute as part of the job. Conflicts with
query_file_uri
- []string
Presto client tags to attach to this query.
- Continue
On boolFailure Whether to continue executing queries if a query fails. Setting to true can be useful when executing independent parallel queries. Defaults to false.
- Logging
Config JobPresto Config Logging Config - Output
Format string The format in which query output will be displayed. See the Presto documentation for supported output formats.
logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- Properties map[string]string
A mapping of property names to values. Used to set Presto session properties Equivalent to using the --session flag in the Presto CLI.
- Query
File stringUri The HCFS URI of the script that contains SQL queries. Conflicts with
query_list
- Query
Lists []string The list of SQL queries or statements to execute as part of the job. Conflicts with
query_file_uri
- List<String>
Presto client tags to attach to this query.
- continue
On BooleanFailure Whether to continue executing queries if a query fails. Setting to true can be useful when executing independent parallel queries. Defaults to false.
- logging
Config JobPresto Config Logging Config - output
Format String The format in which query output will be displayed. See the Presto documentation for supported output formats.
logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- properties Map<String,String>
A mapping of property names to values. Used to set Presto session properties Equivalent to using the --session flag in the Presto CLI.
- query
File StringUri The HCFS URI of the script that contains SQL queries. Conflicts with
query_list
- query
Lists List<String> The list of SQL queries or statements to execute as part of the job. Conflicts with
query_file_uri
- string[]
Presto client tags to attach to this query.
- continue
On booleanFailure Whether to continue executing queries if a query fails. Setting to true can be useful when executing independent parallel queries. Defaults to false.
- logging
Config JobPresto Config Logging Config - output
Format string The format in which query output will be displayed. See the Presto documentation for supported output formats.
logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- properties {[key: string]: string}
A mapping of property names to values. Used to set Presto session properties Equivalent to using the --session flag in the Presto CLI.
- query
File stringUri The HCFS URI of the script that contains SQL queries. Conflicts with
query_list
- query
Lists string[] The list of SQL queries or statements to execute as part of the job. Conflicts with
query_file_uri
- Sequence[str]
Presto client tags to attach to this query.
- continue_
on_ boolfailure Whether to continue executing queries if a query fails. Setting to true can be useful when executing independent parallel queries. Defaults to false.
- logging_
config JobPresto Config Logging Config - output_
format str The format in which query output will be displayed. See the Presto documentation for supported output formats.
logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- properties Mapping[str, str]
A mapping of property names to values. Used to set Presto session properties Equivalent to using the --session flag in the Presto CLI.
- query_
file_ struri The HCFS URI of the script that contains SQL queries. Conflicts with
query_list
- query_
lists Sequence[str] The list of SQL queries or statements to execute as part of the job. Conflicts with
query_file_uri
- List<String>
Presto client tags to attach to this query.
- continue
On BooleanFailure Whether to continue executing queries if a query fails. Setting to true can be useful when executing independent parallel queries. Defaults to false.
- logging
Config Property Map - output
Format String The format in which query output will be displayed. See the Presto documentation for supported output formats.
logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- properties Map<String>
A mapping of property names to values. Used to set Presto session properties Equivalent to using the --session flag in the Presto CLI.
- query
File StringUri The HCFS URI of the script that contains SQL queries. Conflicts with
query_list
- query
Lists List<String> The list of SQL queries or statements to execute as part of the job. Conflicts with
query_file_uri
JobPrestoConfigLoggingConfig, JobPrestoConfigLoggingConfigArgs
- Driver
Log Dictionary<string, string>Levels
- Driver
Log map[string]stringLevels
- driver
Log Map<String,String>Levels
- driver
Log {[key: string]: string}Levels
- driver_
log_ Mapping[str, str]levels
- driver
Log Map<String>Levels
JobPysparkConfig, JobPysparkConfigArgs
- Main
Python stringFile Uri The HCFS URI of the main Python file to use as the driver. Must be a .py file.
- Archive
Uris List<string> HCFS URIs of archives to be extracted in the working directory of .jar, .tar, .tar.gz, .tgz, and .zip.
- Args List<string>
The arguments to pass to the driver.
- File
Uris List<string> HCFS URIs of files to be copied to the working directory of Python drivers and distributed tasks. Useful for naively parallel tasks.
- Jar
File List<string>Uris HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
- Logging
Config JobPyspark Config Logging Config - Properties Dictionary<string, string>
A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in
/etc/spark/conf/spark-defaults.conf
and classes in user code.logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- Python
File List<string>Uris HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
- Main
Python stringFile Uri The HCFS URI of the main Python file to use as the driver. Must be a .py file.
- Archive
Uris []string HCFS URIs of archives to be extracted in the working directory of .jar, .tar, .tar.gz, .tgz, and .zip.
- Args []string
The arguments to pass to the driver.
- File
Uris []string HCFS URIs of files to be copied to the working directory of Python drivers and distributed tasks. Useful for naively parallel tasks.
- Jar
File []stringUris HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
- Logging
Config JobPyspark Config Logging Config - Properties map[string]string
A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in
/etc/spark/conf/spark-defaults.conf
and classes in user code.logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- Python
File []stringUris HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
- main
Python StringFile Uri The HCFS URI of the main Python file to use as the driver. Must be a .py file.
- archive
Uris List<String> HCFS URIs of archives to be extracted in the working directory of .jar, .tar, .tar.gz, .tgz, and .zip.
- args List<String>
The arguments to pass to the driver.
- file
Uris List<String> HCFS URIs of files to be copied to the working directory of Python drivers and distributed tasks. Useful for naively parallel tasks.
- jar
File List<String>Uris HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
- logging
Config JobPyspark Config Logging Config - properties Map<String,String>
A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in
/etc/spark/conf/spark-defaults.conf
and classes in user code.logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- python
File List<String>Uris HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
- main
Python stringFile Uri The HCFS URI of the main Python file to use as the driver. Must be a .py file.
- archive
Uris string[] HCFS URIs of archives to be extracted in the working directory of .jar, .tar, .tar.gz, .tgz, and .zip.
- args string[]
The arguments to pass to the driver.
- file
Uris string[] HCFS URIs of files to be copied to the working directory of Python drivers and distributed tasks. Useful for naively parallel tasks.
- jar
File string[]Uris HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
- logging
Config JobPyspark Config Logging Config - properties {[key: string]: string}
A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in
/etc/spark/conf/spark-defaults.conf
and classes in user code.logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- python
File string[]Uris HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
- main_
python_ strfile_ uri The HCFS URI of the main Python file to use as the driver. Must be a .py file.
- archive_
uris Sequence[str] HCFS URIs of archives to be extracted in the working directory of .jar, .tar, .tar.gz, .tgz, and .zip.
- args Sequence[str]
The arguments to pass to the driver.
- file_
uris Sequence[str] HCFS URIs of files to be copied to the working directory of Python drivers and distributed tasks. Useful for naively parallel tasks.
- jar_
file_ Sequence[str]uris HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
- logging_
config JobPyspark Config Logging Config - properties Mapping[str, str]
A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in
/etc/spark/conf/spark-defaults.conf
and classes in user code.logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- python_
file_ Sequence[str]uris HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
- main
Python StringFile Uri The HCFS URI of the main Python file to use as the driver. Must be a .py file.
- archive
Uris List<String> HCFS URIs of archives to be extracted in the working directory of .jar, .tar, .tar.gz, .tgz, and .zip.
- args List<String>
The arguments to pass to the driver.
- file
Uris List<String> HCFS URIs of files to be copied to the working directory of Python drivers and distributed tasks. Useful for naively parallel tasks.
- jar
File List<String>Uris HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
- logging
Config Property Map - properties Map<String>
A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in
/etc/spark/conf/spark-defaults.conf
and classes in user code.logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- python
File List<String>Uris HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
JobPysparkConfigLoggingConfig, JobPysparkConfigLoggingConfigArgs
- Driver
Log Dictionary<string, string>Levels
- Driver
Log map[string]stringLevels
- driver
Log Map<String,String>Levels
- driver
Log {[key: string]: string}Levels
- driver_
log_ Mapping[str, str]levels
- driver
Log Map<String>Levels
JobReference, JobReferenceArgs
- Job
Id string
- Job
Id string
- job
Id String
- job
Id string
- job_
id str
- job
Id String
JobScheduling, JobSchedulingArgs
- max
Failures IntegerPer Hour - max
Failures IntegerTotal
- max
Failures numberPer Hour - max
Failures numberTotal
- max
Failures NumberPer Hour - max
Failures NumberTotal
JobSparkConfig, JobSparkConfigArgs
- Archive
Uris List<string> HCFS URIs of archives to be extracted in the working directory of .jar, .tar, .tar.gz, .tgz, and .zip.
- Args List<string>
The arguments to pass to the driver.
- File
Uris List<string> HCFS URIs of files to be copied to the working directory of Spark drivers and distributed tasks. Useful for naively parallel tasks.
- Jar
File List<string>Uris HCFS URIs of jar files to add to the CLASSPATHs of the Spark driver and tasks.
- Logging
Config JobSpark Config Logging Config - Main
Class string The class containing the main method of the driver. Must be in a provided jar or jar that is already on the classpath. Conflicts with
main_jar_file_uri
- Main
Jar stringFile Uri The HCFS URI of jar file containing the driver jar. Conflicts with
main_class
- Properties Dictionary<string, string>
A mapping of property names to values, used to configure Spark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in
/etc/spark/conf/spark-defaults.conf
and classes in user code.logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- Archive
Uris []string HCFS URIs of archives to be extracted in the working directory of .jar, .tar, .tar.gz, .tgz, and .zip.
- Args []string
The arguments to pass to the driver.
- File
Uris []string HCFS URIs of files to be copied to the working directory of Spark drivers and distributed tasks. Useful for naively parallel tasks.
- Jar
File []stringUris HCFS URIs of jar files to add to the CLASSPATHs of the Spark driver and tasks.
- Logging
Config JobSpark Config Logging Config - Main
Class string The class containing the main method of the driver. Must be in a provided jar or jar that is already on the classpath. Conflicts with
main_jar_file_uri
- Main
Jar stringFile Uri The HCFS URI of jar file containing the driver jar. Conflicts with
main_class
- Properties map[string]string
A mapping of property names to values, used to configure Spark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in
/etc/spark/conf/spark-defaults.conf
and classes in user code.logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- archive
Uris List<String> HCFS URIs of archives to be extracted in the working directory of .jar, .tar, .tar.gz, .tgz, and .zip.
- args List<String>
The arguments to pass to the driver.
- file
Uris List<String> HCFS URIs of files to be copied to the working directory of Spark drivers and distributed tasks. Useful for naively parallel tasks.
- jar
File List<String>Uris HCFS URIs of jar files to add to the CLASSPATHs of the Spark driver and tasks.
- logging
Config JobSpark Config Logging Config - main
Class String The class containing the main method of the driver. Must be in a provided jar or jar that is already on the classpath. Conflicts with
main_jar_file_uri
- main
Jar StringFile Uri The HCFS URI of jar file containing the driver jar. Conflicts with
main_class
- properties Map<String,String>
A mapping of property names to values, used to configure Spark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in
/etc/spark/conf/spark-defaults.conf
and classes in user code.logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- archive
Uris string[] HCFS URIs of archives to be extracted in the working directory of .jar, .tar, .tar.gz, .tgz, and .zip.
- args string[]
The arguments to pass to the driver.
- file
Uris string[] HCFS URIs of files to be copied to the working directory of Spark drivers and distributed tasks. Useful for naively parallel tasks.
- jar
File string[]Uris HCFS URIs of jar files to add to the CLASSPATHs of the Spark driver and tasks.
- logging
Config JobSpark Config Logging Config - main
Class string The class containing the main method of the driver. Must be in a provided jar or jar that is already on the classpath. Conflicts with
main_jar_file_uri
- main
Jar stringFile Uri The HCFS URI of jar file containing the driver jar. Conflicts with
main_class
- properties {[key: string]: string}
A mapping of property names to values, used to configure Spark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in
/etc/spark/conf/spark-defaults.conf
and classes in user code.logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- archive_
uris Sequence[str] HCFS URIs of archives to be extracted in the working directory of .jar, .tar, .tar.gz, .tgz, and .zip.
- args Sequence[str]
The arguments to pass to the driver.
- file_
uris Sequence[str] HCFS URIs of files to be copied to the working directory of Spark drivers and distributed tasks. Useful for naively parallel tasks.
- jar_
file_ Sequence[str]uris HCFS URIs of jar files to add to the CLASSPATHs of the Spark driver and tasks.
- logging_
config JobSpark Config Logging Config - main_
class str The class containing the main method of the driver. Must be in a provided jar or jar that is already on the classpath. Conflicts with
main_jar_file_uri
- main_
jar_ strfile_ uri The HCFS URI of jar file containing the driver jar. Conflicts with
main_class
- properties Mapping[str, str]
A mapping of property names to values, used to configure Spark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in
/etc/spark/conf/spark-defaults.conf
and classes in user code.logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- archive
Uris List<String> HCFS URIs of archives to be extracted in the working directory of .jar, .tar, .tar.gz, .tgz, and .zip.
- args List<String>
The arguments to pass to the driver.
- file
Uris List<String> HCFS URIs of files to be copied to the working directory of Spark drivers and distributed tasks. Useful for naively parallel tasks.
- jar
File List<String>Uris HCFS URIs of jar files to add to the CLASSPATHs of the Spark driver and tasks.
- logging
Config Property Map - main
Class String The class containing the main method of the driver. Must be in a provided jar or jar that is already on the classpath. Conflicts with
main_jar_file_uri
- main
Jar StringFile Uri The HCFS URI of jar file containing the driver jar. Conflicts with
main_class
- properties Map<String>
A mapping of property names to values, used to configure Spark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in
/etc/spark/conf/spark-defaults.conf
and classes in user code.logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
JobSparkConfigLoggingConfig, JobSparkConfigLoggingConfigArgs
- Driver
Log Dictionary<string, string>Levels
- Driver
Log map[string]stringLevels
- driver
Log Map<String,String>Levels
- driver
Log {[key: string]: string}Levels
- driver_
log_ Mapping[str, str]levels
- driver
Log Map<String>Levels
JobSparksqlConfig, JobSparksqlConfigArgs
- Jar
File List<string>Uris HCFS URIs of jar files to be added to the Spark CLASSPATH.
logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- Logging
Config JobSparksql Config Logging Config - Properties Dictionary<string, string>
A mapping of property names to values, used to configure Spark SQL's SparkConf. Properties that conflict with values set by the Cloud Dataproc API may be overwritten.
- Query
File stringUri The HCFS URI of the script that contains SQL queries. Conflicts with
query_list
- Query
Lists List<string> The list of SQL queries or statements to execute as part of the job. Conflicts with
query_file_uri
- Script
Variables Dictionary<string, string> Mapping of query variable names to values (equivalent to the Spark SQL command:
SET name="value";
).
- Jar
File []stringUris HCFS URIs of jar files to be added to the Spark CLASSPATH.
logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- Logging
Config JobSparksql Config Logging Config - Properties map[string]string
A mapping of property names to values, used to configure Spark SQL's SparkConf. Properties that conflict with values set by the Cloud Dataproc API may be overwritten.
- Query
File stringUri The HCFS URI of the script that contains SQL queries. Conflicts with
query_list
- Query
Lists []string The list of SQL queries or statements to execute as part of the job. Conflicts with
query_file_uri
- Script
Variables map[string]string Mapping of query variable names to values (equivalent to the Spark SQL command:
SET name="value";
).
- jar
File List<String>Uris HCFS URIs of jar files to be added to the Spark CLASSPATH.
logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- logging
Config JobSparksql Config Logging Config - properties Map<String,String>
A mapping of property names to values, used to configure Spark SQL's SparkConf. Properties that conflict with values set by the Cloud Dataproc API may be overwritten.
- query
File StringUri The HCFS URI of the script that contains SQL queries. Conflicts with
query_list
- query
Lists List<String> The list of SQL queries or statements to execute as part of the job. Conflicts with
query_file_uri
- script
Variables Map<String,String> Mapping of query variable names to values (equivalent to the Spark SQL command:
SET name="value";
).
- jar
File string[]Uris HCFS URIs of jar files to be added to the Spark CLASSPATH.
logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- logging
Config JobSparksql Config Logging Config - properties {[key: string]: string}
A mapping of property names to values, used to configure Spark SQL's SparkConf. Properties that conflict with values set by the Cloud Dataproc API may be overwritten.
- query
File stringUri The HCFS URI of the script that contains SQL queries. Conflicts with
query_list
- query
Lists string[] The list of SQL queries or statements to execute as part of the job. Conflicts with
query_file_uri
- script
Variables {[key: string]: string} Mapping of query variable names to values (equivalent to the Spark SQL command:
SET name="value";
).
- jar_
file_ Sequence[str]uris HCFS URIs of jar files to be added to the Spark CLASSPATH.
logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- logging_
config JobSparksql Config Logging Config - properties Mapping[str, str]
A mapping of property names to values, used to configure Spark SQL's SparkConf. Properties that conflict with values set by the Cloud Dataproc API may be overwritten.
- query_
file_ struri The HCFS URI of the script that contains SQL queries. Conflicts with
query_list
- query_
lists Sequence[str] The list of SQL queries or statements to execute as part of the job. Conflicts with
query_file_uri
- script_
variables Mapping[str, str] Mapping of query variable names to values (equivalent to the Spark SQL command:
SET name="value";
).
- jar
File List<String>Uris HCFS URIs of jar files to be added to the Spark CLASSPATH.
logging_config.driver_log_levels
- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
- logging
Config Property Map - properties Map<String>
A mapping of property names to values, used to configure Spark SQL's SparkConf. Properties that conflict with values set by the Cloud Dataproc API may be overwritten.
- query
File StringUri The HCFS URI of the script that contains SQL queries. Conflicts with
query_list
- query
Lists List<String> The list of SQL queries or statements to execute as part of the job. Conflicts with
query_file_uri
- script
Variables Map<String> Mapping of query variable names to values (equivalent to the Spark SQL command:
SET name="value";
).
JobSparksqlConfigLoggingConfig, JobSparksqlConfigLoggingConfigArgs
- Driver
Log Dictionary<string, string>Levels
- Driver
Log map[string]stringLevels
- driver
Log Map<String,String>Levels
- driver
Log {[key: string]: string}Levels
- driver_
log_ Mapping[str, str]levels
- driver
Log Map<String>Levels
JobStatus, JobStatusArgs
- Details string
- State string
- State
Start stringTime - Substate string
- Details string
- State string
- State
Start stringTime - Substate string
- details String
- state String
- state
Start StringTime - substate String
- details string
- state string
- state
Start stringTime - substate string
- details str
- state str
- state_
start_ strtime - substate str
- details String
- state String
- state
Start StringTime - substate String
Import
This resource does not support import.
Package Details
- Repository
- Google Cloud (GCP) Classic pulumi/pulumi-gcp
- License
- Apache-2.0
- Notes
This Pulumi package is based on the
google-beta
Terraform Provider.