published on Monday, Mar 9, 2026 by Pulumi
published on Monday, Mar 9, 2026 by Pulumi
Import
The resource cluster can be imported using cluster id. bash
$ pulumi import databricks:index/cluster:Cluster this <cluster-id>
Create Cluster Resource
Resources are created with functions called constructors. To learn more about declaring and configuring resources, see Resources.
Constructor syntax
new Cluster(name: string, args: ClusterArgs, opts?: CustomResourceOptions);@overload
def Cluster(resource_name: str,
args: ClusterArgs,
opts: Optional[ResourceOptions] = None)
@overload
def Cluster(resource_name: str,
opts: Optional[ResourceOptions] = None,
spark_version: Optional[str] = None,
gcp_attributes: Optional[ClusterGcpAttributesArgs] = None,
spark_env_vars: Optional[Mapping[str, Any]] = None,
azure_attributes: Optional[ClusterAzureAttributesArgs] = None,
cluster_id: Optional[str] = None,
cluster_log_conf: Optional[ClusterClusterLogConfArgs] = None,
cluster_name: Optional[str] = None,
custom_tags: Optional[Mapping[str, Any]] = None,
data_security_mode: Optional[str] = None,
docker_image: Optional[ClusterDockerImageArgs] = None,
driver_instance_pool_id: Optional[str] = None,
driver_node_type_id: Optional[str] = None,
idempotency_token: Optional[str] = None,
aws_attributes: Optional[ClusterAwsAttributesArgs] = None,
enable_local_disk_encryption: Optional[bool] = None,
enable_elastic_disk: Optional[bool] = None,
init_scripts: Optional[Sequence[ClusterInitScriptArgs]] = None,
instance_pool_id: Optional[str] = None,
is_pinned: Optional[bool] = None,
libraries: Optional[Sequence[ClusterLibraryArgs]] = None,
node_type_id: Optional[str] = None,
num_workers: Optional[int] = None,
policy_id: Optional[str] = None,
single_user_name: Optional[str] = None,
spark_conf: Optional[Mapping[str, Any]] = None,
autoscale: Optional[ClusterAutoscaleArgs] = None,
autotermination_minutes: Optional[int] = None,
ssh_public_keys: Optional[Sequence[str]] = None)func NewCluster(ctx *Context, name string, args ClusterArgs, opts ...ResourceOption) (*Cluster, error)public Cluster(string name, ClusterArgs args, CustomResourceOptions? opts = null)
public Cluster(String name, ClusterArgs args)
public Cluster(String name, ClusterArgs args, CustomResourceOptions options)
type: databricks:Cluster
properties: # The arguments to resource properties.
options: # Bag of options to control resource's behavior.
Parameters
- name string
- The unique name of the resource.
- args ClusterArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- resource_name str
- The unique name of the resource.
- args ClusterArgs
- The arguments to resource properties.
- opts ResourceOptions
- Bag of options to control resource's behavior.
- ctx Context
- Context object for the current deployment.
- name string
- The unique name of the resource.
- args ClusterArgs
- The arguments to resource properties.
- opts ResourceOption
- Bag of options to control resource's behavior.
- name string
- The unique name of the resource.
- args ClusterArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- name String
- The unique name of the resource.
- args ClusterArgs
- The arguments to resource properties.
- options CustomResourceOptions
- Bag of options to control resource's behavior.
Constructor example
The following reference example uses placeholder values for all input properties.
var clusterResource = new Databricks.Cluster("clusterResource", new()
{
SparkVersion = "string",
GcpAttributes = new Databricks.Inputs.ClusterGcpAttributesArgs
{
Availability = "string",
BootDiskSize = 0,
GoogleServiceAccount = "string",
UsePreemptibleExecutors = false,
ZoneId = "string",
},
SparkEnvVars =
{
{ "string", "any" },
},
AzureAttributes = new Databricks.Inputs.ClusterAzureAttributesArgs
{
Availability = "string",
FirstOnDemand = 0,
SpotBidMaxPrice = 0,
},
ClusterId = "string",
ClusterLogConf = new Databricks.Inputs.ClusterClusterLogConfArgs
{
Dbfs = new Databricks.Inputs.ClusterClusterLogConfDbfsArgs
{
Destination = "string",
},
S3 = new Databricks.Inputs.ClusterClusterLogConfS3Args
{
Destination = "string",
CannedAcl = "string",
EnableEncryption = false,
EncryptionType = "string",
Endpoint = "string",
KmsKey = "string",
Region = "string",
},
},
ClusterName = "string",
CustomTags =
{
{ "string", "any" },
},
DataSecurityMode = "string",
DockerImage = new Databricks.Inputs.ClusterDockerImageArgs
{
Url = "string",
BasicAuth = new Databricks.Inputs.ClusterDockerImageBasicAuthArgs
{
Password = "string",
Username = "string",
},
},
DriverInstancePoolId = "string",
DriverNodeTypeId = "string",
IdempotencyToken = "string",
AwsAttributes = new Databricks.Inputs.ClusterAwsAttributesArgs
{
Availability = "string",
EbsVolumeCount = 0,
EbsVolumeSize = 0,
EbsVolumeType = "string",
FirstOnDemand = 0,
InstanceProfileArn = "string",
SpotBidPricePercent = 0,
ZoneId = "string",
},
EnableLocalDiskEncryption = false,
EnableElasticDisk = false,
InitScripts = new[]
{
new Databricks.Inputs.ClusterInitScriptArgs
{
Dbfs = new Databricks.Inputs.ClusterInitScriptDbfsArgs
{
Destination = "string",
},
File = new Databricks.Inputs.ClusterInitScriptFileArgs
{
Destination = "string",
},
S3 = new Databricks.Inputs.ClusterInitScriptS3Args
{
Destination = "string",
CannedAcl = "string",
EnableEncryption = false,
EncryptionType = "string",
Endpoint = "string",
KmsKey = "string",
Region = "string",
},
},
},
InstancePoolId = "string",
IsPinned = false,
Libraries = new[]
{
new Databricks.Inputs.ClusterLibraryArgs
{
Cran = new Databricks.Inputs.ClusterLibraryCranArgs
{
Package = "string",
Repo = "string",
},
Egg = "string",
Jar = "string",
Maven = new Databricks.Inputs.ClusterLibraryMavenArgs
{
Coordinates = "string",
Exclusions = new[]
{
"string",
},
Repo = "string",
},
Pypi = new Databricks.Inputs.ClusterLibraryPypiArgs
{
Package = "string",
Repo = "string",
},
Whl = "string",
},
},
NodeTypeId = "string",
NumWorkers = 0,
PolicyId = "string",
SingleUserName = "string",
SparkConf =
{
{ "string", "any" },
},
Autoscale = new Databricks.Inputs.ClusterAutoscaleArgs
{
MaxWorkers = 0,
MinWorkers = 0,
},
AutoterminationMinutes = 0,
SshPublicKeys = new[]
{
"string",
},
});
example, err := databricks.NewCluster(ctx, "clusterResource", &databricks.ClusterArgs{
SparkVersion: pulumi.String("string"),
GcpAttributes: &databricks.ClusterGcpAttributesArgs{
Availability: pulumi.String("string"),
BootDiskSize: pulumi.Int(0),
GoogleServiceAccount: pulumi.String("string"),
UsePreemptibleExecutors: pulumi.Bool(false),
ZoneId: pulumi.String("string"),
},
SparkEnvVars: pulumi.Map{
"string": pulumi.Any("any"),
},
AzureAttributes: &databricks.ClusterAzureAttributesArgs{
Availability: pulumi.String("string"),
FirstOnDemand: pulumi.Int(0),
SpotBidMaxPrice: pulumi.Float64(0),
},
ClusterId: pulumi.String("string"),
ClusterLogConf: &databricks.ClusterClusterLogConfArgs{
Dbfs: &databricks.ClusterClusterLogConfDbfsArgs{
Destination: pulumi.String("string"),
},
S3: &databricks.ClusterClusterLogConfS3Args{
Destination: pulumi.String("string"),
CannedAcl: pulumi.String("string"),
EnableEncryption: pulumi.Bool(false),
EncryptionType: pulumi.String("string"),
Endpoint: pulumi.String("string"),
KmsKey: pulumi.String("string"),
Region: pulumi.String("string"),
},
},
ClusterName: pulumi.String("string"),
CustomTags: pulumi.Map{
"string": pulumi.Any("any"),
},
DataSecurityMode: pulumi.String("string"),
DockerImage: &databricks.ClusterDockerImageArgs{
Url: pulumi.String("string"),
BasicAuth: &databricks.ClusterDockerImageBasicAuthArgs{
Password: pulumi.String("string"),
Username: pulumi.String("string"),
},
},
DriverInstancePoolId: pulumi.String("string"),
DriverNodeTypeId: pulumi.String("string"),
IdempotencyToken: pulumi.String("string"),
AwsAttributes: &databricks.ClusterAwsAttributesArgs{
Availability: pulumi.String("string"),
EbsVolumeCount: pulumi.Int(0),
EbsVolumeSize: pulumi.Int(0),
EbsVolumeType: pulumi.String("string"),
FirstOnDemand: pulumi.Int(0),
InstanceProfileArn: pulumi.String("string"),
SpotBidPricePercent: pulumi.Int(0),
ZoneId: pulumi.String("string"),
},
EnableLocalDiskEncryption: pulumi.Bool(false),
EnableElasticDisk: pulumi.Bool(false),
InitScripts: databricks.ClusterInitScriptArray{
&databricks.ClusterInitScriptArgs{
Dbfs: &databricks.ClusterInitScriptDbfsArgs{
Destination: pulumi.String("string"),
},
File: &databricks.ClusterInitScriptFileArgs{
Destination: pulumi.String("string"),
},
S3: &databricks.ClusterInitScriptS3Args{
Destination: pulumi.String("string"),
CannedAcl: pulumi.String("string"),
EnableEncryption: pulumi.Bool(false),
EncryptionType: pulumi.String("string"),
Endpoint: pulumi.String("string"),
KmsKey: pulumi.String("string"),
Region: pulumi.String("string"),
},
},
},
InstancePoolId: pulumi.String("string"),
IsPinned: pulumi.Bool(false),
Libraries: databricks.ClusterLibraryArray{
&databricks.ClusterLibraryArgs{
Cran: &databricks.ClusterLibraryCranArgs{
Package: pulumi.String("string"),
Repo: pulumi.String("string"),
},
Egg: pulumi.String("string"),
Jar: pulumi.String("string"),
Maven: &databricks.ClusterLibraryMavenArgs{
Coordinates: pulumi.String("string"),
Exclusions: pulumi.StringArray{
pulumi.String("string"),
},
Repo: pulumi.String("string"),
},
Pypi: &databricks.ClusterLibraryPypiArgs{
Package: pulumi.String("string"),
Repo: pulumi.String("string"),
},
Whl: pulumi.String("string"),
},
},
NodeTypeId: pulumi.String("string"),
NumWorkers: pulumi.Int(0),
PolicyId: pulumi.String("string"),
SingleUserName: pulumi.String("string"),
SparkConf: pulumi.Map{
"string": pulumi.Any("any"),
},
Autoscale: &databricks.ClusterAutoscaleArgs{
MaxWorkers: pulumi.Int(0),
MinWorkers: pulumi.Int(0),
},
AutoterminationMinutes: pulumi.Int(0),
SshPublicKeys: pulumi.StringArray{
pulumi.String("string"),
},
})
var clusterResource = new Cluster("clusterResource", ClusterArgs.builder()
.sparkVersion("string")
.gcpAttributes(ClusterGcpAttributesArgs.builder()
.availability("string")
.bootDiskSize(0)
.googleServiceAccount("string")
.usePreemptibleExecutors(false)
.zoneId("string")
.build())
.sparkEnvVars(Map.of("string", "any"))
.azureAttributes(ClusterAzureAttributesArgs.builder()
.availability("string")
.firstOnDemand(0)
.spotBidMaxPrice(0.0)
.build())
.clusterId("string")
.clusterLogConf(ClusterClusterLogConfArgs.builder()
.dbfs(ClusterClusterLogConfDbfsArgs.builder()
.destination("string")
.build())
.s3(ClusterClusterLogConfS3Args.builder()
.destination("string")
.cannedAcl("string")
.enableEncryption(false)
.encryptionType("string")
.endpoint("string")
.kmsKey("string")
.region("string")
.build())
.build())
.clusterName("string")
.customTags(Map.of("string", "any"))
.dataSecurityMode("string")
.dockerImage(ClusterDockerImageArgs.builder()
.url("string")
.basicAuth(ClusterDockerImageBasicAuthArgs.builder()
.password("string")
.username("string")
.build())
.build())
.driverInstancePoolId("string")
.driverNodeTypeId("string")
.idempotencyToken("string")
.awsAttributes(ClusterAwsAttributesArgs.builder()
.availability("string")
.ebsVolumeCount(0)
.ebsVolumeSize(0)
.ebsVolumeType("string")
.firstOnDemand(0)
.instanceProfileArn("string")
.spotBidPricePercent(0)
.zoneId("string")
.build())
.enableLocalDiskEncryption(false)
.enableElasticDisk(false)
.initScripts(ClusterInitScriptArgs.builder()
.dbfs(ClusterInitScriptDbfsArgs.builder()
.destination("string")
.build())
.file(ClusterInitScriptFileArgs.builder()
.destination("string")
.build())
.s3(ClusterInitScriptS3Args.builder()
.destination("string")
.cannedAcl("string")
.enableEncryption(false)
.encryptionType("string")
.endpoint("string")
.kmsKey("string")
.region("string")
.build())
.build())
.instancePoolId("string")
.isPinned(false)
.libraries(ClusterLibraryArgs.builder()
.cran(ClusterLibraryCranArgs.builder()
.package_("string")
.repo("string")
.build())
.egg("string")
.jar("string")
.maven(ClusterLibraryMavenArgs.builder()
.coordinates("string")
.exclusions("string")
.repo("string")
.build())
.pypi(ClusterLibraryPypiArgs.builder()
.package_("string")
.repo("string")
.build())
.whl("string")
.build())
.nodeTypeId("string")
.numWorkers(0)
.policyId("string")
.singleUserName("string")
.sparkConf(Map.of("string", "any"))
.autoscale(ClusterAutoscaleArgs.builder()
.maxWorkers(0)
.minWorkers(0)
.build())
.autoterminationMinutes(0)
.sshPublicKeys("string")
.build());
cluster_resource = databricks.Cluster("clusterResource",
spark_version="string",
gcp_attributes={
"availability": "string",
"boot_disk_size": 0,
"google_service_account": "string",
"use_preemptible_executors": False,
"zone_id": "string",
},
spark_env_vars={
"string": "any",
},
azure_attributes={
"availability": "string",
"first_on_demand": 0,
"spot_bid_max_price": 0,
},
cluster_id="string",
cluster_log_conf={
"dbfs": {
"destination": "string",
},
"s3": {
"destination": "string",
"canned_acl": "string",
"enable_encryption": False,
"encryption_type": "string",
"endpoint": "string",
"kms_key": "string",
"region": "string",
},
},
cluster_name="string",
custom_tags={
"string": "any",
},
data_security_mode="string",
docker_image={
"url": "string",
"basic_auth": {
"password": "string",
"username": "string",
},
},
driver_instance_pool_id="string",
driver_node_type_id="string",
idempotency_token="string",
aws_attributes={
"availability": "string",
"ebs_volume_count": 0,
"ebs_volume_size": 0,
"ebs_volume_type": "string",
"first_on_demand": 0,
"instance_profile_arn": "string",
"spot_bid_price_percent": 0,
"zone_id": "string",
},
enable_local_disk_encryption=False,
enable_elastic_disk=False,
init_scripts=[{
"dbfs": {
"destination": "string",
},
"file": {
"destination": "string",
},
"s3": {
"destination": "string",
"canned_acl": "string",
"enable_encryption": False,
"encryption_type": "string",
"endpoint": "string",
"kms_key": "string",
"region": "string",
},
}],
instance_pool_id="string",
is_pinned=False,
libraries=[{
"cran": {
"package": "string",
"repo": "string",
},
"egg": "string",
"jar": "string",
"maven": {
"coordinates": "string",
"exclusions": ["string"],
"repo": "string",
},
"pypi": {
"package": "string",
"repo": "string",
},
"whl": "string",
}],
node_type_id="string",
num_workers=0,
policy_id="string",
single_user_name="string",
spark_conf={
"string": "any",
},
autoscale={
"max_workers": 0,
"min_workers": 0,
},
autotermination_minutes=0,
ssh_public_keys=["string"])
const clusterResource = new databricks.Cluster("clusterResource", {
sparkVersion: "string",
gcpAttributes: {
availability: "string",
bootDiskSize: 0,
googleServiceAccount: "string",
usePreemptibleExecutors: false,
zoneId: "string",
},
sparkEnvVars: {
string: "any",
},
azureAttributes: {
availability: "string",
firstOnDemand: 0,
spotBidMaxPrice: 0,
},
clusterId: "string",
clusterLogConf: {
dbfs: {
destination: "string",
},
s3: {
destination: "string",
cannedAcl: "string",
enableEncryption: false,
encryptionType: "string",
endpoint: "string",
kmsKey: "string",
region: "string",
},
},
clusterName: "string",
customTags: {
string: "any",
},
dataSecurityMode: "string",
dockerImage: {
url: "string",
basicAuth: {
password: "string",
username: "string",
},
},
driverInstancePoolId: "string",
driverNodeTypeId: "string",
idempotencyToken: "string",
awsAttributes: {
availability: "string",
ebsVolumeCount: 0,
ebsVolumeSize: 0,
ebsVolumeType: "string",
firstOnDemand: 0,
instanceProfileArn: "string",
spotBidPricePercent: 0,
zoneId: "string",
},
enableLocalDiskEncryption: false,
enableElasticDisk: false,
initScripts: [{
dbfs: {
destination: "string",
},
file: {
destination: "string",
},
s3: {
destination: "string",
cannedAcl: "string",
enableEncryption: false,
encryptionType: "string",
endpoint: "string",
kmsKey: "string",
region: "string",
},
}],
instancePoolId: "string",
isPinned: false,
libraries: [{
cran: {
"package": "string",
repo: "string",
},
egg: "string",
jar: "string",
maven: {
coordinates: "string",
exclusions: ["string"],
repo: "string",
},
pypi: {
"package": "string",
repo: "string",
},
whl: "string",
}],
nodeTypeId: "string",
numWorkers: 0,
policyId: "string",
singleUserName: "string",
sparkConf: {
string: "any",
},
autoscale: {
maxWorkers: 0,
minWorkers: 0,
},
autoterminationMinutes: 0,
sshPublicKeys: ["string"],
});
type: databricks:Cluster
properties:
autoscale:
maxWorkers: 0
minWorkers: 0
autoterminationMinutes: 0
awsAttributes:
availability: string
ebsVolumeCount: 0
ebsVolumeSize: 0
ebsVolumeType: string
firstOnDemand: 0
instanceProfileArn: string
spotBidPricePercent: 0
zoneId: string
azureAttributes:
availability: string
firstOnDemand: 0
spotBidMaxPrice: 0
clusterId: string
clusterLogConf:
dbfs:
destination: string
s3:
cannedAcl: string
destination: string
enableEncryption: false
encryptionType: string
endpoint: string
kmsKey: string
region: string
clusterName: string
customTags:
string: any
dataSecurityMode: string
dockerImage:
basicAuth:
password: string
username: string
url: string
driverInstancePoolId: string
driverNodeTypeId: string
enableElasticDisk: false
enableLocalDiskEncryption: false
gcpAttributes:
availability: string
bootDiskSize: 0
googleServiceAccount: string
usePreemptibleExecutors: false
zoneId: string
idempotencyToken: string
initScripts:
- dbfs:
destination: string
file:
destination: string
s3:
cannedAcl: string
destination: string
enableEncryption: false
encryptionType: string
endpoint: string
kmsKey: string
region: string
instancePoolId: string
isPinned: false
libraries:
- cran:
package: string
repo: string
egg: string
jar: string
maven:
coordinates: string
exclusions:
- string
repo: string
pypi:
package: string
repo: string
whl: string
nodeTypeId: string
numWorkers: 0
policyId: string
singleUserName: string
sparkConf:
string: any
sparkEnvVars:
string: any
sparkVersion: string
sshPublicKeys:
- string
Cluster Resource Properties
To learn more about resource properties and how to use them, see Inputs and Outputs in the Architecture and Concepts docs.
Inputs
In Python, inputs that are objects can be passed either as argument classes or as dictionary literals.
The Cluster resource accepts the following input properties:
- Spark
Version string - Runtime version of the cluster. Any supported databricks.getSparkVersion id. We advise using Cluster Policies to restrict the list of versions for simplicity while maintaining enough control.
- Autoscale
Cluster
Autoscale - Autotermination
Minutes int - Automatically terminate the cluster after being inactive for this time in minutes. If not set, Databricks won't automatically terminate an inactive cluster. If specified, the threshold must be between 10 and 10000 minutes. You can also set this value to 0 to explicitly disable automatic termination. We highly recommend having this setting present for Interactive/BI clusters.
- Aws
Attributes ClusterAws Attributes - Azure
Attributes ClusterAzure Attributes - Cluster
Id string - Cluster
Log ClusterConf Cluster Log Conf - Cluster
Name string - Cluster name, which doesn’t have to be unique. If not specified at creation, the cluster name will be an empty string.
- Dictionary<string, object>
- Additional tags for cluster resources. Databricks will tag all cluster resources (e.g., AWS EC2 instances and EBS volumes) with these tags in addition to
default_tags. - Data
Security stringMode - Select the security features of the cluster. Unity Catalog requires
SINGLE_USERorUSER_ISOLATIONmode.LEGACY_PASSTHROUGHfor passthrough cluster andLEGACY_TABLE_ACLfor Table ACL cluster. Default toNONE, i.e. no security feature enabled. - Docker
Image ClusterDocker Image - Driver
Instance stringPool Id - similar to
instance_pool_id, but for driver node. If omitted, andinstance_pool_idis specified, then driver will be allocated from that pool. - Driver
Node stringType Id - The node type of the Spark driver. This field is optional; if unset, API will set the driver node type to the same value as
node_type_iddefined above. - Enable
Elastic boolDisk - If you don’t want to allocate a fixed number of EBS volumes at cluster creation time, use autoscaling local storage. With autoscaling local storage, Databricks monitors the amount of free disk space available on your cluster’s Spark workers. If a worker begins to run too low on disk, Databricks automatically attaches a new EBS volume to the worker before it runs out of disk space. EBS volumes are attached up to a limit of 5 TB of total disk space per instance (including the instance’s local storage). To scale down EBS usage, make sure you have
autotermination_minutesandautoscaleattributes set. More documentation available at cluster configuration page. - Enable
Local boolDisk Encryption - Some instance types you use to run clusters may have locally attached disks. Databricks may store shuffle data or temporary data on these locally attached disks. To ensure that all data at rest is encrypted for all storage types, including shuffle data stored temporarily on your cluster’s local disks, you can enable local disk encryption. When local disk encryption is enabled, Databricks generates an encryption key locally unique to each cluster node and encrypting all data stored on local disks. The scope of the key is local to each cluster node and is destroyed along with the cluster node itself. During its lifetime, the key resides in memory for encryption and decryption and is stored encrypted on the disk. Your workloads may run more slowly because of the performance impact of reading and writing encrypted data to and from local volumes. This feature is not available for all Azure Databricks subscriptions. Contact your Microsoft or Databricks account representative to request access.
- Gcp
Attributes ClusterGcp Attributes - Idempotency
Token string - An optional token to guarantee the idempotency of cluster creation requests. If an active cluster with the provided token already exists, the request will not create a new cluster, but it will return the existing running cluster's ID instead. If you specify the idempotency token, upon failure, you can retry until the request succeeds. Databricks platform guarantees to launch exactly one cluster with that idempotency token. This token should have at most 64 characters.
- Init
Scripts List<ClusterInit Script> - Instance
Pool stringId - To reduce cluster start time, you can attach a cluster to a predefined pool of idle instances. When attached to a pool, a cluster allocates its driver and worker nodes from the pool. If the pool does not have sufficient idle resources to accommodate the cluster’s request, it expands by allocating new instances from the instance provider. When an attached cluster changes its state to
TERMINATED, the instances it used are returned to the pool and reused by a different cluster.
- To reduce cluster start time, you can attach a cluster to a predefined pool of idle instances. When attached to a pool, a cluster allocates its driver and worker nodes from the pool. If the pool does not have sufficient idle resources to accommodate the cluster’s request, it expands by allocating new instances from the instance provider. When an attached cluster changes its state to
- Is
Pinned bool - boolean value specifying if cluster is pinned (not pinned by default). You must be a Databricks administrator to use this. The pinned clusters' maximum number is limited to 70, so
applymay fail if you have more than that. - Libraries
List<Cluster
Library> - Node
Type stringId - Any supported databricks.getNodeType id. If
instance_pool_idis specified, this field is not needed. - Num
Workers int - Policy
Id string - Identifier of Cluster Policy to validate cluster and preset certain defaults. The primary use for cluster policies is to allow users to create policy-scoped clusters via UI rather than sharing configuration for API-created clusters. For example, when you specify
policy_idof external metastore policy, you still have to fill in relevant keys forspark_conf. - Single
User stringName - The optional user name of the user to assign to an interactive cluster. This field is required when using standard AAD Passthrough for Azure Data Lake Storage (ADLS) with a single-user cluster (i.e., not high-concurrency clusters).
- Spark
Conf Dictionary<string, object> - Map with key-value pairs to fine-tune Spark clusters, where you can provide custom Spark configuration properties in a cluster configuration.
- Spark
Env Dictionary<string, object>Vars - Map with environment variable key-value pairs to fine-tune Spark clusters. Key-value pairs of the form (X,Y) are exported (i.e., X='Y') while launching the driver and workers.
- Ssh
Public List<string>Keys - SSH public key contents that will be added to each Spark node in this cluster. The corresponding private keys can be used to login with the user name ubuntu on port 2200. You can specify up to 10 keys.
- Spark
Version string - Runtime version of the cluster. Any supported databricks.getSparkVersion id. We advise using Cluster Policies to restrict the list of versions for simplicity while maintaining enough control.
- Autoscale
Cluster
Autoscale Args - Autotermination
Minutes int - Automatically terminate the cluster after being inactive for this time in minutes. If not set, Databricks won't automatically terminate an inactive cluster. If specified, the threshold must be between 10 and 10000 minutes. You can also set this value to 0 to explicitly disable automatic termination. We highly recommend having this setting present for Interactive/BI clusters.
- Aws
Attributes ClusterAws Attributes Args - Azure
Attributes ClusterAzure Attributes Args - Cluster
Id string - Cluster
Log ClusterConf Cluster Log Conf Args - Cluster
Name string - Cluster name, which doesn’t have to be unique. If not specified at creation, the cluster name will be an empty string.
- map[string]interface{}
- Additional tags for cluster resources. Databricks will tag all cluster resources (e.g., AWS EC2 instances and EBS volumes) with these tags in addition to
default_tags. - Data
Security stringMode - Select the security features of the cluster. Unity Catalog requires
SINGLE_USERorUSER_ISOLATIONmode.LEGACY_PASSTHROUGHfor passthrough cluster andLEGACY_TABLE_ACLfor Table ACL cluster. Default toNONE, i.e. no security feature enabled. - Docker
Image ClusterDocker Image Args - Driver
Instance stringPool Id - similar to
instance_pool_id, but for driver node. If omitted, andinstance_pool_idis specified, then driver will be allocated from that pool. - Driver
Node stringType Id - The node type of the Spark driver. This field is optional; if unset, API will set the driver node type to the same value as
node_type_iddefined above. - Enable
Elastic boolDisk - If you don’t want to allocate a fixed number of EBS volumes at cluster creation time, use autoscaling local storage. With autoscaling local storage, Databricks monitors the amount of free disk space available on your cluster’s Spark workers. If a worker begins to run too low on disk, Databricks automatically attaches a new EBS volume to the worker before it runs out of disk space. EBS volumes are attached up to a limit of 5 TB of total disk space per instance (including the instance’s local storage). To scale down EBS usage, make sure you have
autotermination_minutesandautoscaleattributes set. More documentation available at cluster configuration page. - Enable
Local boolDisk Encryption - Some instance types you use to run clusters may have locally attached disks. Databricks may store shuffle data or temporary data on these locally attached disks. To ensure that all data at rest is encrypted for all storage types, including shuffle data stored temporarily on your cluster’s local disks, you can enable local disk encryption. When local disk encryption is enabled, Databricks generates an encryption key locally unique to each cluster node and encrypting all data stored on local disks. The scope of the key is local to each cluster node and is destroyed along with the cluster node itself. During its lifetime, the key resides in memory for encryption and decryption and is stored encrypted on the disk. Your workloads may run more slowly because of the performance impact of reading and writing encrypted data to and from local volumes. This feature is not available for all Azure Databricks subscriptions. Contact your Microsoft or Databricks account representative to request access.
- Gcp
Attributes ClusterGcp Attributes Args - Idempotency
Token string - An optional token to guarantee the idempotency of cluster creation requests. If an active cluster with the provided token already exists, the request will not create a new cluster, but it will return the existing running cluster's ID instead. If you specify the idempotency token, upon failure, you can retry until the request succeeds. Databricks platform guarantees to launch exactly one cluster with that idempotency token. This token should have at most 64 characters.
- Init
Scripts []ClusterInit Script Args - Instance
Pool stringId - To reduce cluster start time, you can attach a cluster to a predefined pool of idle instances. When attached to a pool, a cluster allocates its driver and worker nodes from the pool. If the pool does not have sufficient idle resources to accommodate the cluster’s request, it expands by allocating new instances from the instance provider. When an attached cluster changes its state to
TERMINATED, the instances it used are returned to the pool and reused by a different cluster.
- To reduce cluster start time, you can attach a cluster to a predefined pool of idle instances. When attached to a pool, a cluster allocates its driver and worker nodes from the pool. If the pool does not have sufficient idle resources to accommodate the cluster’s request, it expands by allocating new instances from the instance provider. When an attached cluster changes its state to
- Is
Pinned bool - boolean value specifying if cluster is pinned (not pinned by default). You must be a Databricks administrator to use this. The pinned clusters' maximum number is limited to 70, so
applymay fail if you have more than that. - Libraries
[]Cluster
Library Args - Node
Type stringId - Any supported databricks.getNodeType id. If
instance_pool_idis specified, this field is not needed. - Num
Workers int - Policy
Id string - Identifier of Cluster Policy to validate cluster and preset certain defaults. The primary use for cluster policies is to allow users to create policy-scoped clusters via UI rather than sharing configuration for API-created clusters. For example, when you specify
policy_idof external metastore policy, you still have to fill in relevant keys forspark_conf. - Single
User stringName - The optional user name of the user to assign to an interactive cluster. This field is required when using standard AAD Passthrough for Azure Data Lake Storage (ADLS) with a single-user cluster (i.e., not high-concurrency clusters).
- Spark
Conf map[string]interface{} - Map with key-value pairs to fine-tune Spark clusters, where you can provide custom Spark configuration properties in a cluster configuration.
- Spark
Env map[string]interface{}Vars - Map with environment variable key-value pairs to fine-tune Spark clusters. Key-value pairs of the form (X,Y) are exported (i.e., X='Y') while launching the driver and workers.
- Ssh
Public []stringKeys - SSH public key contents that will be added to each Spark node in this cluster. The corresponding private keys can be used to login with the user name ubuntu on port 2200. You can specify up to 10 keys.
- spark
Version String - Runtime version of the cluster. Any supported databricks.getSparkVersion id. We advise using Cluster Policies to restrict the list of versions for simplicity while maintaining enough control.
- autoscale
Cluster
Autoscale - autotermination
Minutes Integer - Automatically terminate the cluster after being inactive for this time in minutes. If not set, Databricks won't automatically terminate an inactive cluster. If specified, the threshold must be between 10 and 10000 minutes. You can also set this value to 0 to explicitly disable automatic termination. We highly recommend having this setting present for Interactive/BI clusters.
- aws
Attributes ClusterAws Attributes - azure
Attributes ClusterAzure Attributes - cluster
Id String - cluster
Log ClusterConf Cluster Log Conf - cluster
Name String - Cluster name, which doesn’t have to be unique. If not specified at creation, the cluster name will be an empty string.
- Map<String,Object>
- Additional tags for cluster resources. Databricks will tag all cluster resources (e.g., AWS EC2 instances and EBS volumes) with these tags in addition to
default_tags. - data
Security StringMode - Select the security features of the cluster. Unity Catalog requires
SINGLE_USERorUSER_ISOLATIONmode.LEGACY_PASSTHROUGHfor passthrough cluster andLEGACY_TABLE_ACLfor Table ACL cluster. Default toNONE, i.e. no security feature enabled. - docker
Image ClusterDocker Image - driver
Instance StringPool Id - similar to
instance_pool_id, but for driver node. If omitted, andinstance_pool_idis specified, then driver will be allocated from that pool. - driver
Node StringType Id - The node type of the Spark driver. This field is optional; if unset, API will set the driver node type to the same value as
node_type_iddefined above. - enable
Elastic BooleanDisk - If you don’t want to allocate a fixed number of EBS volumes at cluster creation time, use autoscaling local storage. With autoscaling local storage, Databricks monitors the amount of free disk space available on your cluster’s Spark workers. If a worker begins to run too low on disk, Databricks automatically attaches a new EBS volume to the worker before it runs out of disk space. EBS volumes are attached up to a limit of 5 TB of total disk space per instance (including the instance’s local storage). To scale down EBS usage, make sure you have
autotermination_minutesandautoscaleattributes set. More documentation available at cluster configuration page. - enable
Local BooleanDisk Encryption - Some instance types you use to run clusters may have locally attached disks. Databricks may store shuffle data or temporary data on these locally attached disks. To ensure that all data at rest is encrypted for all storage types, including shuffle data stored temporarily on your cluster’s local disks, you can enable local disk encryption. When local disk encryption is enabled, Databricks generates an encryption key locally unique to each cluster node and encrypting all data stored on local disks. The scope of the key is local to each cluster node and is destroyed along with the cluster node itself. During its lifetime, the key resides in memory for encryption and decryption and is stored encrypted on the disk. Your workloads may run more slowly because of the performance impact of reading and writing encrypted data to and from local volumes. This feature is not available for all Azure Databricks subscriptions. Contact your Microsoft or Databricks account representative to request access.
- gcp
Attributes ClusterGcp Attributes - idempotency
Token String - An optional token to guarantee the idempotency of cluster creation requests. If an active cluster with the provided token already exists, the request will not create a new cluster, but it will return the existing running cluster's ID instead. If you specify the idempotency token, upon failure, you can retry until the request succeeds. Databricks platform guarantees to launch exactly one cluster with that idempotency token. This token should have at most 64 characters.
- init
Scripts List<ClusterInit Script> - instance
Pool StringId - To reduce cluster start time, you can attach a cluster to a predefined pool of idle instances. When attached to a pool, a cluster allocates its driver and worker nodes from the pool. If the pool does not have sufficient idle resources to accommodate the cluster’s request, it expands by allocating new instances from the instance provider. When an attached cluster changes its state to
TERMINATED, the instances it used are returned to the pool and reused by a different cluster.
- To reduce cluster start time, you can attach a cluster to a predefined pool of idle instances. When attached to a pool, a cluster allocates its driver and worker nodes from the pool. If the pool does not have sufficient idle resources to accommodate the cluster’s request, it expands by allocating new instances from the instance provider. When an attached cluster changes its state to
- is
Pinned Boolean - boolean value specifying if cluster is pinned (not pinned by default). You must be a Databricks administrator to use this. The pinned clusters' maximum number is limited to 70, so
applymay fail if you have more than that. - libraries
List<Cluster
Library> - node
Type StringId - Any supported databricks.getNodeType id. If
instance_pool_idis specified, this field is not needed. - num
Workers Integer - policy
Id String - Identifier of Cluster Policy to validate cluster and preset certain defaults. The primary use for cluster policies is to allow users to create policy-scoped clusters via UI rather than sharing configuration for API-created clusters. For example, when you specify
policy_idof external metastore policy, you still have to fill in relevant keys forspark_conf. - single
User StringName - The optional user name of the user to assign to an interactive cluster. This field is required when using standard AAD Passthrough for Azure Data Lake Storage (ADLS) with a single-user cluster (i.e., not high-concurrency clusters).
- spark
Conf Map<String,Object> - Map with key-value pairs to fine-tune Spark clusters, where you can provide custom Spark configuration properties in a cluster configuration.
- spark
Env Map<String,Object>Vars - Map with environment variable key-value pairs to fine-tune Spark clusters. Key-value pairs of the form (X,Y) are exported (i.e., X='Y') while launching the driver and workers.
- ssh
Public List<String>Keys - SSH public key contents that will be added to each Spark node in this cluster. The corresponding private keys can be used to login with the user name ubuntu on port 2200. You can specify up to 10 keys.
- spark
Version string - Runtime version of the cluster. Any supported databricks.getSparkVersion id. We advise using Cluster Policies to restrict the list of versions for simplicity while maintaining enough control.
- autoscale
Cluster
Autoscale - autotermination
Minutes number - Automatically terminate the cluster after being inactive for this time in minutes. If not set, Databricks won't automatically terminate an inactive cluster. If specified, the threshold must be between 10 and 10000 minutes. You can also set this value to 0 to explicitly disable automatic termination. We highly recommend having this setting present for Interactive/BI clusters.
- aws
Attributes ClusterAws Attributes - azure
Attributes ClusterAzure Attributes - cluster
Id string - cluster
Log ClusterConf Cluster Log Conf - cluster
Name string - Cluster name, which doesn’t have to be unique. If not specified at creation, the cluster name will be an empty string.
- {[key: string]: any}
- Additional tags for cluster resources. Databricks will tag all cluster resources (e.g., AWS EC2 instances and EBS volumes) with these tags in addition to
default_tags. - data
Security stringMode - Select the security features of the cluster. Unity Catalog requires
SINGLE_USERorUSER_ISOLATIONmode.LEGACY_PASSTHROUGHfor passthrough cluster andLEGACY_TABLE_ACLfor Table ACL cluster. Default toNONE, i.e. no security feature enabled. - docker
Image ClusterDocker Image - driver
Instance stringPool Id - similar to
instance_pool_id, but for driver node. If omitted, andinstance_pool_idis specified, then driver will be allocated from that pool. - driver
Node stringType Id - The node type of the Spark driver. This field is optional; if unset, API will set the driver node type to the same value as
node_type_iddefined above. - enable
Elastic booleanDisk - If you don’t want to allocate a fixed number of EBS volumes at cluster creation time, use autoscaling local storage. With autoscaling local storage, Databricks monitors the amount of free disk space available on your cluster’s Spark workers. If a worker begins to run too low on disk, Databricks automatically attaches a new EBS volume to the worker before it runs out of disk space. EBS volumes are attached up to a limit of 5 TB of total disk space per instance (including the instance’s local storage). To scale down EBS usage, make sure you have
autotermination_minutesandautoscaleattributes set. More documentation available at cluster configuration page. - enable
Local booleanDisk Encryption - Some instance types you use to run clusters may have locally attached disks. Databricks may store shuffle data or temporary data on these locally attached disks. To ensure that all data at rest is encrypted for all storage types, including shuffle data stored temporarily on your cluster’s local disks, you can enable local disk encryption. When local disk encryption is enabled, Databricks generates an encryption key locally unique to each cluster node and encrypting all data stored on local disks. The scope of the key is local to each cluster node and is destroyed along with the cluster node itself. During its lifetime, the key resides in memory for encryption and decryption and is stored encrypted on the disk. Your workloads may run more slowly because of the performance impact of reading and writing encrypted data to and from local volumes. This feature is not available for all Azure Databricks subscriptions. Contact your Microsoft or Databricks account representative to request access.
- gcp
Attributes ClusterGcp Attributes - idempotency
Token string - An optional token to guarantee the idempotency of cluster creation requests. If an active cluster with the provided token already exists, the request will not create a new cluster, but it will return the existing running cluster's ID instead. If you specify the idempotency token, upon failure, you can retry until the request succeeds. Databricks platform guarantees to launch exactly one cluster with that idempotency token. This token should have at most 64 characters.
- init
Scripts ClusterInit Script[] - instance
Pool stringId - To reduce cluster start time, you can attach a cluster to a predefined pool of idle instances. When attached to a pool, a cluster allocates its driver and worker nodes from the pool. If the pool does not have sufficient idle resources to accommodate the cluster’s request, it expands by allocating new instances from the instance provider. When an attached cluster changes its state to
TERMINATED, the instances it used are returned to the pool and reused by a different cluster.
- To reduce cluster start time, you can attach a cluster to a predefined pool of idle instances. When attached to a pool, a cluster allocates its driver and worker nodes from the pool. If the pool does not have sufficient idle resources to accommodate the cluster’s request, it expands by allocating new instances from the instance provider. When an attached cluster changes its state to
- is
Pinned boolean - boolean value specifying if cluster is pinned (not pinned by default). You must be a Databricks administrator to use this. The pinned clusters' maximum number is limited to 70, so
applymay fail if you have more than that. - libraries
Cluster
Library[] - node
Type stringId - Any supported databricks.getNodeType id. If
instance_pool_idis specified, this field is not needed. - num
Workers number - policy
Id string - Identifier of Cluster Policy to validate cluster and preset certain defaults. The primary use for cluster policies is to allow users to create policy-scoped clusters via UI rather than sharing configuration for API-created clusters. For example, when you specify
policy_idof external metastore policy, you still have to fill in relevant keys forspark_conf. - single
User stringName - The optional user name of the user to assign to an interactive cluster. This field is required when using standard AAD Passthrough for Azure Data Lake Storage (ADLS) with a single-user cluster (i.e., not high-concurrency clusters).
- spark
Conf {[key: string]: any} - Map with key-value pairs to fine-tune Spark clusters, where you can provide custom Spark configuration properties in a cluster configuration.
- spark
Env {[key: string]: any}Vars - Map with environment variable key-value pairs to fine-tune Spark clusters. Key-value pairs of the form (X,Y) are exported (i.e., X='Y') while launching the driver and workers.
- ssh
Public string[]Keys - SSH public key contents that will be added to each Spark node in this cluster. The corresponding private keys can be used to login with the user name ubuntu on port 2200. You can specify up to 10 keys.
- spark_
version str - Runtime version of the cluster. Any supported databricks.getSparkVersion id. We advise using Cluster Policies to restrict the list of versions for simplicity while maintaining enough control.
- autoscale
Cluster
Autoscale Args - autotermination_
minutes int - Automatically terminate the cluster after being inactive for this time in minutes. If not set, Databricks won't automatically terminate an inactive cluster. If specified, the threshold must be between 10 and 10000 minutes. You can also set this value to 0 to explicitly disable automatic termination. We highly recommend having this setting present for Interactive/BI clusters.
- aws_
attributes ClusterAws Attributes Args - azure_
attributes ClusterAzure Attributes Args - cluster_
id str - cluster_
log_ Clusterconf Cluster Log Conf Args - cluster_
name str - Cluster name, which doesn’t have to be unique. If not specified at creation, the cluster name will be an empty string.
- Mapping[str, Any]
- Additional tags for cluster resources. Databricks will tag all cluster resources (e.g., AWS EC2 instances and EBS volumes) with these tags in addition to
default_tags. - data_
security_ strmode - Select the security features of the cluster. Unity Catalog requires
SINGLE_USERorUSER_ISOLATIONmode.LEGACY_PASSTHROUGHfor passthrough cluster andLEGACY_TABLE_ACLfor Table ACL cluster. Default toNONE, i.e. no security feature enabled. - docker_
image ClusterDocker Image Args - driver_
instance_ strpool_ id - similar to
instance_pool_id, but for driver node. If omitted, andinstance_pool_idis specified, then driver will be allocated from that pool. - driver_
node_ strtype_ id - The node type of the Spark driver. This field is optional; if unset, API will set the driver node type to the same value as
node_type_iddefined above. - enable_
elastic_ booldisk - If you don’t want to allocate a fixed number of EBS volumes at cluster creation time, use autoscaling local storage. With autoscaling local storage, Databricks monitors the amount of free disk space available on your cluster’s Spark workers. If a worker begins to run too low on disk, Databricks automatically attaches a new EBS volume to the worker before it runs out of disk space. EBS volumes are attached up to a limit of 5 TB of total disk space per instance (including the instance’s local storage). To scale down EBS usage, make sure you have
autotermination_minutesandautoscaleattributes set. More documentation available at cluster configuration page. - enable_
local_ booldisk_ encryption - Some instance types you use to run clusters may have locally attached disks. Databricks may store shuffle data or temporary data on these locally attached disks. To ensure that all data at rest is encrypted for all storage types, including shuffle data stored temporarily on your cluster’s local disks, you can enable local disk encryption. When local disk encryption is enabled, Databricks generates an encryption key locally unique to each cluster node and encrypting all data stored on local disks. The scope of the key is local to each cluster node and is destroyed along with the cluster node itself. During its lifetime, the key resides in memory for encryption and decryption and is stored encrypted on the disk. Your workloads may run more slowly because of the performance impact of reading and writing encrypted data to and from local volumes. This feature is not available for all Azure Databricks subscriptions. Contact your Microsoft or Databricks account representative to request access.
- gcp_
attributes ClusterGcp Attributes Args - idempotency_
token str - An optional token to guarantee the idempotency of cluster creation requests. If an active cluster with the provided token already exists, the request will not create a new cluster, but it will return the existing running cluster's ID instead. If you specify the idempotency token, upon failure, you can retry until the request succeeds. Databricks platform guarantees to launch exactly one cluster with that idempotency token. This token should have at most 64 characters.
- init_
scripts Sequence[ClusterInit Script Args] - instance_
pool_ strid - To reduce cluster start time, you can attach a cluster to a predefined pool of idle instances. When attached to a pool, a cluster allocates its driver and worker nodes from the pool. If the pool does not have sufficient idle resources to accommodate the cluster’s request, it expands by allocating new instances from the instance provider. When an attached cluster changes its state to
TERMINATED, the instances it used are returned to the pool and reused by a different cluster.
- To reduce cluster start time, you can attach a cluster to a predefined pool of idle instances. When attached to a pool, a cluster allocates its driver and worker nodes from the pool. If the pool does not have sufficient idle resources to accommodate the cluster’s request, it expands by allocating new instances from the instance provider. When an attached cluster changes its state to
- is_
pinned bool - boolean value specifying if cluster is pinned (not pinned by default). You must be a Databricks administrator to use this. The pinned clusters' maximum number is limited to 70, so
applymay fail if you have more than that. - libraries
Sequence[Cluster
Library Args] - node_
type_ strid - Any supported databricks.getNodeType id. If
instance_pool_idis specified, this field is not needed. - num_
workers int - policy_
id str - Identifier of Cluster Policy to validate cluster and preset certain defaults. The primary use for cluster policies is to allow users to create policy-scoped clusters via UI rather than sharing configuration for API-created clusters. For example, when you specify
policy_idof external metastore policy, you still have to fill in relevant keys forspark_conf. - single_
user_ strname - The optional user name of the user to assign to an interactive cluster. This field is required when using standard AAD Passthrough for Azure Data Lake Storage (ADLS) with a single-user cluster (i.e., not high-concurrency clusters).
- spark_
conf Mapping[str, Any] - Map with key-value pairs to fine-tune Spark clusters, where you can provide custom Spark configuration properties in a cluster configuration.
- spark_
env_ Mapping[str, Any]vars - Map with environment variable key-value pairs to fine-tune Spark clusters. Key-value pairs of the form (X,Y) are exported (i.e., X='Y') while launching the driver and workers.
- ssh_
public_ Sequence[str]keys - SSH public key contents that will be added to each Spark node in this cluster. The corresponding private keys can be used to login with the user name ubuntu on port 2200. You can specify up to 10 keys.
- spark
Version String - Runtime version of the cluster. Any supported databricks.getSparkVersion id. We advise using Cluster Policies to restrict the list of versions for simplicity while maintaining enough control.
- autoscale Property Map
- autotermination
Minutes Number - Automatically terminate the cluster after being inactive for this time in minutes. If not set, Databricks won't automatically terminate an inactive cluster. If specified, the threshold must be between 10 and 10000 minutes. You can also set this value to 0 to explicitly disable automatic termination. We highly recommend having this setting present for Interactive/BI clusters.
- aws
Attributes Property Map - azure
Attributes Property Map - cluster
Id String - cluster
Log Property MapConf - cluster
Name String - Cluster name, which doesn’t have to be unique. If not specified at creation, the cluster name will be an empty string.
- Map<Any>
- Additional tags for cluster resources. Databricks will tag all cluster resources (e.g., AWS EC2 instances and EBS volumes) with these tags in addition to
default_tags. - data
Security StringMode - Select the security features of the cluster. Unity Catalog requires
SINGLE_USERorUSER_ISOLATIONmode.LEGACY_PASSTHROUGHfor passthrough cluster andLEGACY_TABLE_ACLfor Table ACL cluster. Default toNONE, i.e. no security feature enabled. - docker
Image Property Map - driver
Instance StringPool Id - similar to
instance_pool_id, but for driver node. If omitted, andinstance_pool_idis specified, then driver will be allocated from that pool. - driver
Node StringType Id - The node type of the Spark driver. This field is optional; if unset, API will set the driver node type to the same value as
node_type_iddefined above. - enable
Elastic BooleanDisk - If you don’t want to allocate a fixed number of EBS volumes at cluster creation time, use autoscaling local storage. With autoscaling local storage, Databricks monitors the amount of free disk space available on your cluster’s Spark workers. If a worker begins to run too low on disk, Databricks automatically attaches a new EBS volume to the worker before it runs out of disk space. EBS volumes are attached up to a limit of 5 TB of total disk space per instance (including the instance’s local storage). To scale down EBS usage, make sure you have
autotermination_minutesandautoscaleattributes set. More documentation available at cluster configuration page. - enable
Local BooleanDisk Encryption - Some instance types you use to run clusters may have locally attached disks. Databricks may store shuffle data or temporary data on these locally attached disks. To ensure that all data at rest is encrypted for all storage types, including shuffle data stored temporarily on your cluster’s local disks, you can enable local disk encryption. When local disk encryption is enabled, Databricks generates an encryption key locally unique to each cluster node and encrypting all data stored on local disks. The scope of the key is local to each cluster node and is destroyed along with the cluster node itself. During its lifetime, the key resides in memory for encryption and decryption and is stored encrypted on the disk. Your workloads may run more slowly because of the performance impact of reading and writing encrypted data to and from local volumes. This feature is not available for all Azure Databricks subscriptions. Contact your Microsoft or Databricks account representative to request access.
- gcp
Attributes Property Map - idempotency
Token String - An optional token to guarantee the idempotency of cluster creation requests. If an active cluster with the provided token already exists, the request will not create a new cluster, but it will return the existing running cluster's ID instead. If you specify the idempotency token, upon failure, you can retry until the request succeeds. Databricks platform guarantees to launch exactly one cluster with that idempotency token. This token should have at most 64 characters.
- init
Scripts List<Property Map> - instance
Pool StringId - To reduce cluster start time, you can attach a cluster to a predefined pool of idle instances. When attached to a pool, a cluster allocates its driver and worker nodes from the pool. If the pool does not have sufficient idle resources to accommodate the cluster’s request, it expands by allocating new instances from the instance provider. When an attached cluster changes its state to
TERMINATED, the instances it used are returned to the pool and reused by a different cluster.
- To reduce cluster start time, you can attach a cluster to a predefined pool of idle instances. When attached to a pool, a cluster allocates its driver and worker nodes from the pool. If the pool does not have sufficient idle resources to accommodate the cluster’s request, it expands by allocating new instances from the instance provider. When an attached cluster changes its state to
- is
Pinned Boolean - boolean value specifying if cluster is pinned (not pinned by default). You must be a Databricks administrator to use this. The pinned clusters' maximum number is limited to 70, so
applymay fail if you have more than that. - libraries List<Property Map>
- node
Type StringId - Any supported databricks.getNodeType id. If
instance_pool_idis specified, this field is not needed. - num
Workers Number - policy
Id String - Identifier of Cluster Policy to validate cluster and preset certain defaults. The primary use for cluster policies is to allow users to create policy-scoped clusters via UI rather than sharing configuration for API-created clusters. For example, when you specify
policy_idof external metastore policy, you still have to fill in relevant keys forspark_conf. - single
User StringName - The optional user name of the user to assign to an interactive cluster. This field is required when using standard AAD Passthrough for Azure Data Lake Storage (ADLS) with a single-user cluster (i.e., not high-concurrency clusters).
- spark
Conf Map<Any> - Map with key-value pairs to fine-tune Spark clusters, where you can provide custom Spark configuration properties in a cluster configuration.
- spark
Env Map<Any>Vars - Map with environment variable key-value pairs to fine-tune Spark clusters. Key-value pairs of the form (X,Y) are exported (i.e., X='Y') while launching the driver and workers.
- ssh
Public List<String>Keys - SSH public key contents that will be added to each Spark node in this cluster. The corresponding private keys can be used to login with the user name ubuntu on port 2200. You can specify up to 10 keys.
Outputs
All input properties are implicitly available as output properties. Additionally, the Cluster resource produces the following output properties:
- Dictionary<string, object>
- (map) Tags that are added by Databricks by default, regardless of any custom_tags that may have been added. These include: Vendor: Databricks, Creator: <username_of_creator>, ClusterName: <name_of_cluster>, ClusterId: <id_of_cluster>, Name:
- Id string
- The provider-assigned unique ID for this managed resource.
- State string
- (string) State of the cluster.
- Url string
- map[string]interface{}
- (map) Tags that are added by Databricks by default, regardless of any custom_tags that may have been added. These include: Vendor: Databricks, Creator: <username_of_creator>, ClusterName: <name_of_cluster>, ClusterId: <id_of_cluster>, Name:
- Id string
- The provider-assigned unique ID for this managed resource.
- State string
- (string) State of the cluster.
- Url string
- Map<String,Object>
- (map) Tags that are added by Databricks by default, regardless of any custom_tags that may have been added. These include: Vendor: Databricks, Creator: <username_of_creator>, ClusterName: <name_of_cluster>, ClusterId: <id_of_cluster>, Name:
- id String
- The provider-assigned unique ID for this managed resource.
- state String
- (string) State of the cluster.
- url String
- {[key: string]: any}
- (map) Tags that are added by Databricks by default, regardless of any custom_tags that may have been added. These include: Vendor: Databricks, Creator: <username_of_creator>, ClusterName: <name_of_cluster>, ClusterId: <id_of_cluster>, Name:
- id string
- The provider-assigned unique ID for this managed resource.
- state string
- (string) State of the cluster.
- url string
- Mapping[str, Any]
- (map) Tags that are added by Databricks by default, regardless of any custom_tags that may have been added. These include: Vendor: Databricks, Creator: <username_of_creator>, ClusterName: <name_of_cluster>, ClusterId: <id_of_cluster>, Name:
- id str
- The provider-assigned unique ID for this managed resource.
- state str
- (string) State of the cluster.
- url str
- Map<Any>
- (map) Tags that are added by Databricks by default, regardless of any custom_tags that may have been added. These include: Vendor: Databricks, Creator: <username_of_creator>, ClusterName: <name_of_cluster>, ClusterId: <id_of_cluster>, Name:
- id String
- The provider-assigned unique ID for this managed resource.
- state String
- (string) State of the cluster.
- url String
Look up Existing Cluster Resource
Get an existing Cluster resource’s state with the given name, ID, and optional extra properties used to qualify the lookup.
public static get(name: string, id: Input<ID>, state?: ClusterState, opts?: CustomResourceOptions): Cluster@staticmethod
def get(resource_name: str,
id: str,
opts: Optional[ResourceOptions] = None,
autoscale: Optional[ClusterAutoscaleArgs] = None,
autotermination_minutes: Optional[int] = None,
aws_attributes: Optional[ClusterAwsAttributesArgs] = None,
azure_attributes: Optional[ClusterAzureAttributesArgs] = None,
cluster_id: Optional[str] = None,
cluster_log_conf: Optional[ClusterClusterLogConfArgs] = None,
cluster_name: Optional[str] = None,
custom_tags: Optional[Mapping[str, Any]] = None,
data_security_mode: Optional[str] = None,
default_tags: Optional[Mapping[str, Any]] = None,
docker_image: Optional[ClusterDockerImageArgs] = None,
driver_instance_pool_id: Optional[str] = None,
driver_node_type_id: Optional[str] = None,
enable_elastic_disk: Optional[bool] = None,
enable_local_disk_encryption: Optional[bool] = None,
gcp_attributes: Optional[ClusterGcpAttributesArgs] = None,
idempotency_token: Optional[str] = None,
init_scripts: Optional[Sequence[ClusterInitScriptArgs]] = None,
instance_pool_id: Optional[str] = None,
is_pinned: Optional[bool] = None,
libraries: Optional[Sequence[ClusterLibraryArgs]] = None,
node_type_id: Optional[str] = None,
num_workers: Optional[int] = None,
policy_id: Optional[str] = None,
single_user_name: Optional[str] = None,
spark_conf: Optional[Mapping[str, Any]] = None,
spark_env_vars: Optional[Mapping[str, Any]] = None,
spark_version: Optional[str] = None,
ssh_public_keys: Optional[Sequence[str]] = None,
state: Optional[str] = None,
url: Optional[str] = None) -> Clusterfunc GetCluster(ctx *Context, name string, id IDInput, state *ClusterState, opts ...ResourceOption) (*Cluster, error)public static Cluster Get(string name, Input<string> id, ClusterState? state, CustomResourceOptions? opts = null)public static Cluster get(String name, Output<String> id, ClusterState state, CustomResourceOptions options)resources: _: type: databricks:Cluster get: id: ${id}- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- resource_name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- Autoscale
Cluster
Autoscale - Autotermination
Minutes int - Automatically terminate the cluster after being inactive for this time in minutes. If not set, Databricks won't automatically terminate an inactive cluster. If specified, the threshold must be between 10 and 10000 minutes. You can also set this value to 0 to explicitly disable automatic termination. We highly recommend having this setting present for Interactive/BI clusters.
- Aws
Attributes ClusterAws Attributes - Azure
Attributes ClusterAzure Attributes - Cluster
Id string - Cluster
Log ClusterConf Cluster Log Conf - Cluster
Name string - Cluster name, which doesn’t have to be unique. If not specified at creation, the cluster name will be an empty string.
- Dictionary<string, object>
- Additional tags for cluster resources. Databricks will tag all cluster resources (e.g., AWS EC2 instances and EBS volumes) with these tags in addition to
default_tags. - Data
Security stringMode - Select the security features of the cluster. Unity Catalog requires
SINGLE_USERorUSER_ISOLATIONmode.LEGACY_PASSTHROUGHfor passthrough cluster andLEGACY_TABLE_ACLfor Table ACL cluster. Default toNONE, i.e. no security feature enabled. - Dictionary<string, object>
- (map) Tags that are added by Databricks by default, regardless of any custom_tags that may have been added. These include: Vendor: Databricks, Creator: <username_of_creator>, ClusterName: <name_of_cluster>, ClusterId: <id_of_cluster>, Name:
- Docker
Image ClusterDocker Image - Driver
Instance stringPool Id - similar to
instance_pool_id, but for driver node. If omitted, andinstance_pool_idis specified, then driver will be allocated from that pool. - Driver
Node stringType Id - The node type of the Spark driver. This field is optional; if unset, API will set the driver node type to the same value as
node_type_iddefined above. - Enable
Elastic boolDisk - If you don’t want to allocate a fixed number of EBS volumes at cluster creation time, use autoscaling local storage. With autoscaling local storage, Databricks monitors the amount of free disk space available on your cluster’s Spark workers. If a worker begins to run too low on disk, Databricks automatically attaches a new EBS volume to the worker before it runs out of disk space. EBS volumes are attached up to a limit of 5 TB of total disk space per instance (including the instance’s local storage). To scale down EBS usage, make sure you have
autotermination_minutesandautoscaleattributes set. More documentation available at cluster configuration page. - Enable
Local boolDisk Encryption - Some instance types you use to run clusters may have locally attached disks. Databricks may store shuffle data or temporary data on these locally attached disks. To ensure that all data at rest is encrypted for all storage types, including shuffle data stored temporarily on your cluster’s local disks, you can enable local disk encryption. When local disk encryption is enabled, Databricks generates an encryption key locally unique to each cluster node and encrypting all data stored on local disks. The scope of the key is local to each cluster node and is destroyed along with the cluster node itself. During its lifetime, the key resides in memory for encryption and decryption and is stored encrypted on the disk. Your workloads may run more slowly because of the performance impact of reading and writing encrypted data to and from local volumes. This feature is not available for all Azure Databricks subscriptions. Contact your Microsoft or Databricks account representative to request access.
- Gcp
Attributes ClusterGcp Attributes - Idempotency
Token string - An optional token to guarantee the idempotency of cluster creation requests. If an active cluster with the provided token already exists, the request will not create a new cluster, but it will return the existing running cluster's ID instead. If you specify the idempotency token, upon failure, you can retry until the request succeeds. Databricks platform guarantees to launch exactly one cluster with that idempotency token. This token should have at most 64 characters.
- Init
Scripts List<ClusterInit Script> - Instance
Pool stringId - To reduce cluster start time, you can attach a cluster to a predefined pool of idle instances. When attached to a pool, a cluster allocates its driver and worker nodes from the pool. If the pool does not have sufficient idle resources to accommodate the cluster’s request, it expands by allocating new instances from the instance provider. When an attached cluster changes its state to
TERMINATED, the instances it used are returned to the pool and reused by a different cluster.
- To reduce cluster start time, you can attach a cluster to a predefined pool of idle instances. When attached to a pool, a cluster allocates its driver and worker nodes from the pool. If the pool does not have sufficient idle resources to accommodate the cluster’s request, it expands by allocating new instances from the instance provider. When an attached cluster changes its state to
- Is
Pinned bool - boolean value specifying if cluster is pinned (not pinned by default). You must be a Databricks administrator to use this. The pinned clusters' maximum number is limited to 70, so
applymay fail if you have more than that. - Libraries
List<Cluster
Library> - Node
Type stringId - Any supported databricks.getNodeType id. If
instance_pool_idis specified, this field is not needed. - Num
Workers int - Policy
Id string - Identifier of Cluster Policy to validate cluster and preset certain defaults. The primary use for cluster policies is to allow users to create policy-scoped clusters via UI rather than sharing configuration for API-created clusters. For example, when you specify
policy_idof external metastore policy, you still have to fill in relevant keys forspark_conf. - Single
User stringName - The optional user name of the user to assign to an interactive cluster. This field is required when using standard AAD Passthrough for Azure Data Lake Storage (ADLS) with a single-user cluster (i.e., not high-concurrency clusters).
- Spark
Conf Dictionary<string, object> - Map with key-value pairs to fine-tune Spark clusters, where you can provide custom Spark configuration properties in a cluster configuration.
- Spark
Env Dictionary<string, object>Vars - Map with environment variable key-value pairs to fine-tune Spark clusters. Key-value pairs of the form (X,Y) are exported (i.e., X='Y') while launching the driver and workers.
- Spark
Version string - Runtime version of the cluster. Any supported databricks.getSparkVersion id. We advise using Cluster Policies to restrict the list of versions for simplicity while maintaining enough control.
- Ssh
Public List<string>Keys - SSH public key contents that will be added to each Spark node in this cluster. The corresponding private keys can be used to login with the user name ubuntu on port 2200. You can specify up to 10 keys.
- State string
- (string) State of the cluster.
- Url string
- Autoscale
Cluster
Autoscale Args - Autotermination
Minutes int - Automatically terminate the cluster after being inactive for this time in minutes. If not set, Databricks won't automatically terminate an inactive cluster. If specified, the threshold must be between 10 and 10000 minutes. You can also set this value to 0 to explicitly disable automatic termination. We highly recommend having this setting present for Interactive/BI clusters.
- Aws
Attributes ClusterAws Attributes Args - Azure
Attributes ClusterAzure Attributes Args - Cluster
Id string - Cluster
Log ClusterConf Cluster Log Conf Args - Cluster
Name string - Cluster name, which doesn’t have to be unique. If not specified at creation, the cluster name will be an empty string.
- map[string]interface{}
- Additional tags for cluster resources. Databricks will tag all cluster resources (e.g., AWS EC2 instances and EBS volumes) with these tags in addition to
default_tags. - Data
Security stringMode - Select the security features of the cluster. Unity Catalog requires
SINGLE_USERorUSER_ISOLATIONmode.LEGACY_PASSTHROUGHfor passthrough cluster andLEGACY_TABLE_ACLfor Table ACL cluster. Default toNONE, i.e. no security feature enabled. - map[string]interface{}
- (map) Tags that are added by Databricks by default, regardless of any custom_tags that may have been added. These include: Vendor: Databricks, Creator: <username_of_creator>, ClusterName: <name_of_cluster>, ClusterId: <id_of_cluster>, Name:
- Docker
Image ClusterDocker Image Args - Driver
Instance stringPool Id - similar to
instance_pool_id, but for driver node. If omitted, andinstance_pool_idis specified, then driver will be allocated from that pool. - Driver
Node stringType Id - The node type of the Spark driver. This field is optional; if unset, API will set the driver node type to the same value as
node_type_iddefined above. - Enable
Elastic boolDisk - If you don’t want to allocate a fixed number of EBS volumes at cluster creation time, use autoscaling local storage. With autoscaling local storage, Databricks monitors the amount of free disk space available on your cluster’s Spark workers. If a worker begins to run too low on disk, Databricks automatically attaches a new EBS volume to the worker before it runs out of disk space. EBS volumes are attached up to a limit of 5 TB of total disk space per instance (including the instance’s local storage). To scale down EBS usage, make sure you have
autotermination_minutesandautoscaleattributes set. More documentation available at cluster configuration page. - Enable
Local boolDisk Encryption - Some instance types you use to run clusters may have locally attached disks. Databricks may store shuffle data or temporary data on these locally attached disks. To ensure that all data at rest is encrypted for all storage types, including shuffle data stored temporarily on your cluster’s local disks, you can enable local disk encryption. When local disk encryption is enabled, Databricks generates an encryption key locally unique to each cluster node and encrypting all data stored on local disks. The scope of the key is local to each cluster node and is destroyed along with the cluster node itself. During its lifetime, the key resides in memory for encryption and decryption and is stored encrypted on the disk. Your workloads may run more slowly because of the performance impact of reading and writing encrypted data to and from local volumes. This feature is not available for all Azure Databricks subscriptions. Contact your Microsoft or Databricks account representative to request access.
- Gcp
Attributes ClusterGcp Attributes Args - Idempotency
Token string - An optional token to guarantee the idempotency of cluster creation requests. If an active cluster with the provided token already exists, the request will not create a new cluster, but it will return the existing running cluster's ID instead. If you specify the idempotency token, upon failure, you can retry until the request succeeds. Databricks platform guarantees to launch exactly one cluster with that idempotency token. This token should have at most 64 characters.
- Init
Scripts []ClusterInit Script Args - Instance
Pool stringId - To reduce cluster start time, you can attach a cluster to a predefined pool of idle instances. When attached to a pool, a cluster allocates its driver and worker nodes from the pool. If the pool does not have sufficient idle resources to accommodate the cluster’s request, it expands by allocating new instances from the instance provider. When an attached cluster changes its state to
TERMINATED, the instances it used are returned to the pool and reused by a different cluster.
- To reduce cluster start time, you can attach a cluster to a predefined pool of idle instances. When attached to a pool, a cluster allocates its driver and worker nodes from the pool. If the pool does not have sufficient idle resources to accommodate the cluster’s request, it expands by allocating new instances from the instance provider. When an attached cluster changes its state to
- Is
Pinned bool - boolean value specifying if cluster is pinned (not pinned by default). You must be a Databricks administrator to use this. The pinned clusters' maximum number is limited to 70, so
applymay fail if you have more than that. - Libraries
[]Cluster
Library Args - Node
Type stringId - Any supported databricks.getNodeType id. If
instance_pool_idis specified, this field is not needed. - Num
Workers int - Policy
Id string - Identifier of Cluster Policy to validate cluster and preset certain defaults. The primary use for cluster policies is to allow users to create policy-scoped clusters via UI rather than sharing configuration for API-created clusters. For example, when you specify
policy_idof external metastore policy, you still have to fill in relevant keys forspark_conf. - Single
User stringName - The optional user name of the user to assign to an interactive cluster. This field is required when using standard AAD Passthrough for Azure Data Lake Storage (ADLS) with a single-user cluster (i.e., not high-concurrency clusters).
- Spark
Conf map[string]interface{} - Map with key-value pairs to fine-tune Spark clusters, where you can provide custom Spark configuration properties in a cluster configuration.
- Spark
Env map[string]interface{}Vars - Map with environment variable key-value pairs to fine-tune Spark clusters. Key-value pairs of the form (X,Y) are exported (i.e., X='Y') while launching the driver and workers.
- Spark
Version string - Runtime version of the cluster. Any supported databricks.getSparkVersion id. We advise using Cluster Policies to restrict the list of versions for simplicity while maintaining enough control.
- Ssh
Public []stringKeys - SSH public key contents that will be added to each Spark node in this cluster. The corresponding private keys can be used to login with the user name ubuntu on port 2200. You can specify up to 10 keys.
- State string
- (string) State of the cluster.
- Url string
- autoscale
Cluster
Autoscale - autotermination
Minutes Integer - Automatically terminate the cluster after being inactive for this time in minutes. If not set, Databricks won't automatically terminate an inactive cluster. If specified, the threshold must be between 10 and 10000 minutes. You can also set this value to 0 to explicitly disable automatic termination. We highly recommend having this setting present for Interactive/BI clusters.
- aws
Attributes ClusterAws Attributes - azure
Attributes ClusterAzure Attributes - cluster
Id String - cluster
Log ClusterConf Cluster Log Conf - cluster
Name String - Cluster name, which doesn’t have to be unique. If not specified at creation, the cluster name will be an empty string.
- Map<String,Object>
- Additional tags for cluster resources. Databricks will tag all cluster resources (e.g., AWS EC2 instances and EBS volumes) with these tags in addition to
default_tags. - data
Security StringMode - Select the security features of the cluster. Unity Catalog requires
SINGLE_USERorUSER_ISOLATIONmode.LEGACY_PASSTHROUGHfor passthrough cluster andLEGACY_TABLE_ACLfor Table ACL cluster. Default toNONE, i.e. no security feature enabled. - Map<String,Object>
- (map) Tags that are added by Databricks by default, regardless of any custom_tags that may have been added. These include: Vendor: Databricks, Creator: <username_of_creator>, ClusterName: <name_of_cluster>, ClusterId: <id_of_cluster>, Name:
- docker
Image ClusterDocker Image - driver
Instance StringPool Id - similar to
instance_pool_id, but for driver node. If omitted, andinstance_pool_idis specified, then driver will be allocated from that pool. - driver
Node StringType Id - The node type of the Spark driver. This field is optional; if unset, API will set the driver node type to the same value as
node_type_iddefined above. - enable
Elastic BooleanDisk - If you don’t want to allocate a fixed number of EBS volumes at cluster creation time, use autoscaling local storage. With autoscaling local storage, Databricks monitors the amount of free disk space available on your cluster’s Spark workers. If a worker begins to run too low on disk, Databricks automatically attaches a new EBS volume to the worker before it runs out of disk space. EBS volumes are attached up to a limit of 5 TB of total disk space per instance (including the instance’s local storage). To scale down EBS usage, make sure you have
autotermination_minutesandautoscaleattributes set. More documentation available at cluster configuration page. - enable
Local BooleanDisk Encryption - Some instance types you use to run clusters may have locally attached disks. Databricks may store shuffle data or temporary data on these locally attached disks. To ensure that all data at rest is encrypted for all storage types, including shuffle data stored temporarily on your cluster’s local disks, you can enable local disk encryption. When local disk encryption is enabled, Databricks generates an encryption key locally unique to each cluster node and encrypting all data stored on local disks. The scope of the key is local to each cluster node and is destroyed along with the cluster node itself. During its lifetime, the key resides in memory for encryption and decryption and is stored encrypted on the disk. Your workloads may run more slowly because of the performance impact of reading and writing encrypted data to and from local volumes. This feature is not available for all Azure Databricks subscriptions. Contact your Microsoft or Databricks account representative to request access.
- gcp
Attributes ClusterGcp Attributes - idempotency
Token String - An optional token to guarantee the idempotency of cluster creation requests. If an active cluster with the provided token already exists, the request will not create a new cluster, but it will return the existing running cluster's ID instead. If you specify the idempotency token, upon failure, you can retry until the request succeeds. Databricks platform guarantees to launch exactly one cluster with that idempotency token. This token should have at most 64 characters.
- init
Scripts List<ClusterInit Script> - instance
Pool StringId - To reduce cluster start time, you can attach a cluster to a predefined pool of idle instances. When attached to a pool, a cluster allocates its driver and worker nodes from the pool. If the pool does not have sufficient idle resources to accommodate the cluster’s request, it expands by allocating new instances from the instance provider. When an attached cluster changes its state to
TERMINATED, the instances it used are returned to the pool and reused by a different cluster.
- To reduce cluster start time, you can attach a cluster to a predefined pool of idle instances. When attached to a pool, a cluster allocates its driver and worker nodes from the pool. If the pool does not have sufficient idle resources to accommodate the cluster’s request, it expands by allocating new instances from the instance provider. When an attached cluster changes its state to
- is
Pinned Boolean - boolean value specifying if cluster is pinned (not pinned by default). You must be a Databricks administrator to use this. The pinned clusters' maximum number is limited to 70, so
applymay fail if you have more than that. - libraries
List<Cluster
Library> - node
Type StringId - Any supported databricks.getNodeType id. If
instance_pool_idis specified, this field is not needed. - num
Workers Integer - policy
Id String - Identifier of Cluster Policy to validate cluster and preset certain defaults. The primary use for cluster policies is to allow users to create policy-scoped clusters via UI rather than sharing configuration for API-created clusters. For example, when you specify
policy_idof external metastore policy, you still have to fill in relevant keys forspark_conf. - single
User StringName - The optional user name of the user to assign to an interactive cluster. This field is required when using standard AAD Passthrough for Azure Data Lake Storage (ADLS) with a single-user cluster (i.e., not high-concurrency clusters).
- spark
Conf Map<String,Object> - Map with key-value pairs to fine-tune Spark clusters, where you can provide custom Spark configuration properties in a cluster configuration.
- spark
Env Map<String,Object>Vars - Map with environment variable key-value pairs to fine-tune Spark clusters. Key-value pairs of the form (X,Y) are exported (i.e., X='Y') while launching the driver and workers.
- spark
Version String - Runtime version of the cluster. Any supported databricks.getSparkVersion id. We advise using Cluster Policies to restrict the list of versions for simplicity while maintaining enough control.
- ssh
Public List<String>Keys - SSH public key contents that will be added to each Spark node in this cluster. The corresponding private keys can be used to login with the user name ubuntu on port 2200. You can specify up to 10 keys.
- state String
- (string) State of the cluster.
- url String
- autoscale
Cluster
Autoscale - autotermination
Minutes number - Automatically terminate the cluster after being inactive for this time in minutes. If not set, Databricks won't automatically terminate an inactive cluster. If specified, the threshold must be between 10 and 10000 minutes. You can also set this value to 0 to explicitly disable automatic termination. We highly recommend having this setting present for Interactive/BI clusters.
- aws
Attributes ClusterAws Attributes - azure
Attributes ClusterAzure Attributes - cluster
Id string - cluster
Log ClusterConf Cluster Log Conf - cluster
Name string - Cluster name, which doesn’t have to be unique. If not specified at creation, the cluster name will be an empty string.
- {[key: string]: any}
- Additional tags for cluster resources. Databricks will tag all cluster resources (e.g., AWS EC2 instances and EBS volumes) with these tags in addition to
default_tags. - data
Security stringMode - Select the security features of the cluster. Unity Catalog requires
SINGLE_USERorUSER_ISOLATIONmode.LEGACY_PASSTHROUGHfor passthrough cluster andLEGACY_TABLE_ACLfor Table ACL cluster. Default toNONE, i.e. no security feature enabled. - {[key: string]: any}
- (map) Tags that are added by Databricks by default, regardless of any custom_tags that may have been added. These include: Vendor: Databricks, Creator: <username_of_creator>, ClusterName: <name_of_cluster>, ClusterId: <id_of_cluster>, Name:
- docker
Image ClusterDocker Image - driver
Instance stringPool Id - similar to
instance_pool_id, but for driver node. If omitted, andinstance_pool_idis specified, then driver will be allocated from that pool. - driver
Node stringType Id - The node type of the Spark driver. This field is optional; if unset, API will set the driver node type to the same value as
node_type_iddefined above. - enable
Elastic booleanDisk - If you don’t want to allocate a fixed number of EBS volumes at cluster creation time, use autoscaling local storage. With autoscaling local storage, Databricks monitors the amount of free disk space available on your cluster’s Spark workers. If a worker begins to run too low on disk, Databricks automatically attaches a new EBS volume to the worker before it runs out of disk space. EBS volumes are attached up to a limit of 5 TB of total disk space per instance (including the instance’s local storage). To scale down EBS usage, make sure you have
autotermination_minutesandautoscaleattributes set. More documentation available at cluster configuration page. - enable
Local booleanDisk Encryption - Some instance types you use to run clusters may have locally attached disks. Databricks may store shuffle data or temporary data on these locally attached disks. To ensure that all data at rest is encrypted for all storage types, including shuffle data stored temporarily on your cluster’s local disks, you can enable local disk encryption. When local disk encryption is enabled, Databricks generates an encryption key locally unique to each cluster node and encrypting all data stored on local disks. The scope of the key is local to each cluster node and is destroyed along with the cluster node itself. During its lifetime, the key resides in memory for encryption and decryption and is stored encrypted on the disk. Your workloads may run more slowly because of the performance impact of reading and writing encrypted data to and from local volumes. This feature is not available for all Azure Databricks subscriptions. Contact your Microsoft or Databricks account representative to request access.
- gcp
Attributes ClusterGcp Attributes - idempotency
Token string - An optional token to guarantee the idempotency of cluster creation requests. If an active cluster with the provided token already exists, the request will not create a new cluster, but it will return the existing running cluster's ID instead. If you specify the idempotency token, upon failure, you can retry until the request succeeds. Databricks platform guarantees to launch exactly one cluster with that idempotency token. This token should have at most 64 characters.
- init
Scripts ClusterInit Script[] - instance
Pool stringId - To reduce cluster start time, you can attach a cluster to a predefined pool of idle instances. When attached to a pool, a cluster allocates its driver and worker nodes from the pool. If the pool does not have sufficient idle resources to accommodate the cluster’s request, it expands by allocating new instances from the instance provider. When an attached cluster changes its state to
TERMINATED, the instances it used are returned to the pool and reused by a different cluster.
- To reduce cluster start time, you can attach a cluster to a predefined pool of idle instances. When attached to a pool, a cluster allocates its driver and worker nodes from the pool. If the pool does not have sufficient idle resources to accommodate the cluster’s request, it expands by allocating new instances from the instance provider. When an attached cluster changes its state to
- is
Pinned boolean - boolean value specifying if cluster is pinned (not pinned by default). You must be a Databricks administrator to use this. The pinned clusters' maximum number is limited to 70, so
applymay fail if you have more than that. - libraries
Cluster
Library[] - node
Type stringId - Any supported databricks.getNodeType id. If
instance_pool_idis specified, this field is not needed. - num
Workers number - policy
Id string - Identifier of Cluster Policy to validate cluster and preset certain defaults. The primary use for cluster policies is to allow users to create policy-scoped clusters via UI rather than sharing configuration for API-created clusters. For example, when you specify
policy_idof external metastore policy, you still have to fill in relevant keys forspark_conf. - single
User stringName - The optional user name of the user to assign to an interactive cluster. This field is required when using standard AAD Passthrough for Azure Data Lake Storage (ADLS) with a single-user cluster (i.e., not high-concurrency clusters).
- spark
Conf {[key: string]: any} - Map with key-value pairs to fine-tune Spark clusters, where you can provide custom Spark configuration properties in a cluster configuration.
- spark
Env {[key: string]: any}Vars - Map with environment variable key-value pairs to fine-tune Spark clusters. Key-value pairs of the form (X,Y) are exported (i.e., X='Y') while launching the driver and workers.
- spark
Version string - Runtime version of the cluster. Any supported databricks.getSparkVersion id. We advise using Cluster Policies to restrict the list of versions for simplicity while maintaining enough control.
- ssh
Public string[]Keys - SSH public key contents that will be added to each Spark node in this cluster. The corresponding private keys can be used to login with the user name ubuntu on port 2200. You can specify up to 10 keys.
- state string
- (string) State of the cluster.
- url string
- autoscale
Cluster
Autoscale Args - autotermination_
minutes int - Automatically terminate the cluster after being inactive for this time in minutes. If not set, Databricks won't automatically terminate an inactive cluster. If specified, the threshold must be between 10 and 10000 minutes. You can also set this value to 0 to explicitly disable automatic termination. We highly recommend having this setting present for Interactive/BI clusters.
- aws_
attributes ClusterAws Attributes Args - azure_
attributes ClusterAzure Attributes Args - cluster_
id str - cluster_
log_ Clusterconf Cluster Log Conf Args - cluster_
name str - Cluster name, which doesn’t have to be unique. If not specified at creation, the cluster name will be an empty string.
- Mapping[str, Any]
- Additional tags for cluster resources. Databricks will tag all cluster resources (e.g., AWS EC2 instances and EBS volumes) with these tags in addition to
default_tags. - data_
security_ strmode - Select the security features of the cluster. Unity Catalog requires
SINGLE_USERorUSER_ISOLATIONmode.LEGACY_PASSTHROUGHfor passthrough cluster andLEGACY_TABLE_ACLfor Table ACL cluster. Default toNONE, i.e. no security feature enabled. - Mapping[str, Any]
- (map) Tags that are added by Databricks by default, regardless of any custom_tags that may have been added. These include: Vendor: Databricks, Creator: <username_of_creator>, ClusterName: <name_of_cluster>, ClusterId: <id_of_cluster>, Name:
- docker_
image ClusterDocker Image Args - driver_
instance_ strpool_ id - similar to
instance_pool_id, but for driver node. If omitted, andinstance_pool_idis specified, then driver will be allocated from that pool. - driver_
node_ strtype_ id - The node type of the Spark driver. This field is optional; if unset, API will set the driver node type to the same value as
node_type_iddefined above. - enable_
elastic_ booldisk - If you don’t want to allocate a fixed number of EBS volumes at cluster creation time, use autoscaling local storage. With autoscaling local storage, Databricks monitors the amount of free disk space available on your cluster’s Spark workers. If a worker begins to run too low on disk, Databricks automatically attaches a new EBS volume to the worker before it runs out of disk space. EBS volumes are attached up to a limit of 5 TB of total disk space per instance (including the instance’s local storage). To scale down EBS usage, make sure you have
autotermination_minutesandautoscaleattributes set. More documentation available at cluster configuration page. - enable_
local_ booldisk_ encryption - Some instance types you use to run clusters may have locally attached disks. Databricks may store shuffle data or temporary data on these locally attached disks. To ensure that all data at rest is encrypted for all storage types, including shuffle data stored temporarily on your cluster’s local disks, you can enable local disk encryption. When local disk encryption is enabled, Databricks generates an encryption key locally unique to each cluster node and encrypting all data stored on local disks. The scope of the key is local to each cluster node and is destroyed along with the cluster node itself. During its lifetime, the key resides in memory for encryption and decryption and is stored encrypted on the disk. Your workloads may run more slowly because of the performance impact of reading and writing encrypted data to and from local volumes. This feature is not available for all Azure Databricks subscriptions. Contact your Microsoft or Databricks account representative to request access.
- gcp_
attributes ClusterGcp Attributes Args - idempotency_
token str - An optional token to guarantee the idempotency of cluster creation requests. If an active cluster with the provided token already exists, the request will not create a new cluster, but it will return the existing running cluster's ID instead. If you specify the idempotency token, upon failure, you can retry until the request succeeds. Databricks platform guarantees to launch exactly one cluster with that idempotency token. This token should have at most 64 characters.
- init_
scripts Sequence[ClusterInit Script Args] - instance_
pool_ strid - To reduce cluster start time, you can attach a cluster to a predefined pool of idle instances. When attached to a pool, a cluster allocates its driver and worker nodes from the pool. If the pool does not have sufficient idle resources to accommodate the cluster’s request, it expands by allocating new instances from the instance provider. When an attached cluster changes its state to
TERMINATED, the instances it used are returned to the pool and reused by a different cluster.
- To reduce cluster start time, you can attach a cluster to a predefined pool of idle instances. When attached to a pool, a cluster allocates its driver and worker nodes from the pool. If the pool does not have sufficient idle resources to accommodate the cluster’s request, it expands by allocating new instances from the instance provider. When an attached cluster changes its state to
- is_
pinned bool - boolean value specifying if cluster is pinned (not pinned by default). You must be a Databricks administrator to use this. The pinned clusters' maximum number is limited to 70, so
applymay fail if you have more than that. - libraries
Sequence[Cluster
Library Args] - node_
type_ strid - Any supported databricks.getNodeType id. If
instance_pool_idis specified, this field is not needed. - num_
workers int - policy_
id str - Identifier of Cluster Policy to validate cluster and preset certain defaults. The primary use for cluster policies is to allow users to create policy-scoped clusters via UI rather than sharing configuration for API-created clusters. For example, when you specify
policy_idof external metastore policy, you still have to fill in relevant keys forspark_conf. - single_
user_ strname - The optional user name of the user to assign to an interactive cluster. This field is required when using standard AAD Passthrough for Azure Data Lake Storage (ADLS) with a single-user cluster (i.e., not high-concurrency clusters).
- spark_
conf Mapping[str, Any] - Map with key-value pairs to fine-tune Spark clusters, where you can provide custom Spark configuration properties in a cluster configuration.
- spark_
env_ Mapping[str, Any]vars - Map with environment variable key-value pairs to fine-tune Spark clusters. Key-value pairs of the form (X,Y) are exported (i.e., X='Y') while launching the driver and workers.
- spark_
version str - Runtime version of the cluster. Any supported databricks.getSparkVersion id. We advise using Cluster Policies to restrict the list of versions for simplicity while maintaining enough control.
- ssh_
public_ Sequence[str]keys - SSH public key contents that will be added to each Spark node in this cluster. The corresponding private keys can be used to login with the user name ubuntu on port 2200. You can specify up to 10 keys.
- state str
- (string) State of the cluster.
- url str
- autoscale Property Map
- autotermination
Minutes Number - Automatically terminate the cluster after being inactive for this time in minutes. If not set, Databricks won't automatically terminate an inactive cluster. If specified, the threshold must be between 10 and 10000 minutes. You can also set this value to 0 to explicitly disable automatic termination. We highly recommend having this setting present for Interactive/BI clusters.
- aws
Attributes Property Map - azure
Attributes Property Map - cluster
Id String - cluster
Log Property MapConf - cluster
Name String - Cluster name, which doesn’t have to be unique. If not specified at creation, the cluster name will be an empty string.
- Map<Any>
- Additional tags for cluster resources. Databricks will tag all cluster resources (e.g., AWS EC2 instances and EBS volumes) with these tags in addition to
default_tags. - data
Security StringMode - Select the security features of the cluster. Unity Catalog requires
SINGLE_USERorUSER_ISOLATIONmode.LEGACY_PASSTHROUGHfor passthrough cluster andLEGACY_TABLE_ACLfor Table ACL cluster. Default toNONE, i.e. no security feature enabled. - Map<Any>
- (map) Tags that are added by Databricks by default, regardless of any custom_tags that may have been added. These include: Vendor: Databricks, Creator: <username_of_creator>, ClusterName: <name_of_cluster>, ClusterId: <id_of_cluster>, Name:
- docker
Image Property Map - driver
Instance StringPool Id - similar to
instance_pool_id, but for driver node. If omitted, andinstance_pool_idis specified, then driver will be allocated from that pool. - driver
Node StringType Id - The node type of the Spark driver. This field is optional; if unset, API will set the driver node type to the same value as
node_type_iddefined above. - enable
Elastic BooleanDisk - If you don’t want to allocate a fixed number of EBS volumes at cluster creation time, use autoscaling local storage. With autoscaling local storage, Databricks monitors the amount of free disk space available on your cluster’s Spark workers. If a worker begins to run too low on disk, Databricks automatically attaches a new EBS volume to the worker before it runs out of disk space. EBS volumes are attached up to a limit of 5 TB of total disk space per instance (including the instance’s local storage). To scale down EBS usage, make sure you have
autotermination_minutesandautoscaleattributes set. More documentation available at cluster configuration page. - enable
Local BooleanDisk Encryption - Some instance types you use to run clusters may have locally attached disks. Databricks may store shuffle data or temporary data on these locally attached disks. To ensure that all data at rest is encrypted for all storage types, including shuffle data stored temporarily on your cluster’s local disks, you can enable local disk encryption. When local disk encryption is enabled, Databricks generates an encryption key locally unique to each cluster node and encrypting all data stored on local disks. The scope of the key is local to each cluster node and is destroyed along with the cluster node itself. During its lifetime, the key resides in memory for encryption and decryption and is stored encrypted on the disk. Your workloads may run more slowly because of the performance impact of reading and writing encrypted data to and from local volumes. This feature is not available for all Azure Databricks subscriptions. Contact your Microsoft or Databricks account representative to request access.
- gcp
Attributes Property Map - idempotency
Token String - An optional token to guarantee the idempotency of cluster creation requests. If an active cluster with the provided token already exists, the request will not create a new cluster, but it will return the existing running cluster's ID instead. If you specify the idempotency token, upon failure, you can retry until the request succeeds. Databricks platform guarantees to launch exactly one cluster with that idempotency token. This token should have at most 64 characters.
- init
Scripts List<Property Map> - instance
Pool StringId - To reduce cluster start time, you can attach a cluster to a predefined pool of idle instances. When attached to a pool, a cluster allocates its driver and worker nodes from the pool. If the pool does not have sufficient idle resources to accommodate the cluster’s request, it expands by allocating new instances from the instance provider. When an attached cluster changes its state to
TERMINATED, the instances it used are returned to the pool and reused by a different cluster.
- To reduce cluster start time, you can attach a cluster to a predefined pool of idle instances. When attached to a pool, a cluster allocates its driver and worker nodes from the pool. If the pool does not have sufficient idle resources to accommodate the cluster’s request, it expands by allocating new instances from the instance provider. When an attached cluster changes its state to
- is
Pinned Boolean - boolean value specifying if cluster is pinned (not pinned by default). You must be a Databricks administrator to use this. The pinned clusters' maximum number is limited to 70, so
applymay fail if you have more than that. - libraries List<Property Map>
- node
Type StringId - Any supported databricks.getNodeType id. If
instance_pool_idis specified, this field is not needed. - num
Workers Number - policy
Id String - Identifier of Cluster Policy to validate cluster and preset certain defaults. The primary use for cluster policies is to allow users to create policy-scoped clusters via UI rather than sharing configuration for API-created clusters. For example, when you specify
policy_idof external metastore policy, you still have to fill in relevant keys forspark_conf. - single
User StringName - The optional user name of the user to assign to an interactive cluster. This field is required when using standard AAD Passthrough for Azure Data Lake Storage (ADLS) with a single-user cluster (i.e., not high-concurrency clusters).
- spark
Conf Map<Any> - Map with key-value pairs to fine-tune Spark clusters, where you can provide custom Spark configuration properties in a cluster configuration.
- spark
Env Map<Any>Vars - Map with environment variable key-value pairs to fine-tune Spark clusters. Key-value pairs of the form (X,Y) are exported (i.e., X='Y') while launching the driver and workers.
- spark
Version String - Runtime version of the cluster. Any supported databricks.getSparkVersion id. We advise using Cluster Policies to restrict the list of versions for simplicity while maintaining enough control.
- ssh
Public List<String>Keys - SSH public key contents that will be added to each Spark node in this cluster. The corresponding private keys can be used to login with the user name ubuntu on port 2200. You can specify up to 10 keys.
- state String
- (string) State of the cluster.
- url String
Supporting Types
ClusterAutoscale, ClusterAutoscaleArgs
- Max
Workers int - Min
Workers int
- Max
Workers int - Min
Workers int
- max
Workers Integer - min
Workers Integer
- max
Workers number - min
Workers number
- max_
workers int - min_
workers int
- max
Workers Number - min
Workers Number
ClusterAwsAttributes, ClusterAwsAttributesArgs
- Availability string
- Ebs
Volume intCount - Ebs
Volume intSize - Ebs
Volume stringType - First
On intDemand - Instance
Profile stringArn - Spot
Bid intPrice Percent - Zone
Id string
- Availability string
- Ebs
Volume intCount - Ebs
Volume intSize - Ebs
Volume stringType - First
On intDemand - Instance
Profile stringArn - Spot
Bid intPrice Percent - Zone
Id string
- availability String
- ebs
Volume IntegerCount - ebs
Volume IntegerSize - ebs
Volume StringType - first
On IntegerDemand - instance
Profile StringArn - spot
Bid IntegerPrice Percent - zone
Id String
- availability string
- ebs
Volume numberCount - ebs
Volume numberSize - ebs
Volume stringType - first
On numberDemand - instance
Profile stringArn - spot
Bid numberPrice Percent - zone
Id string
- availability str
- ebs_
volume_ intcount - ebs_
volume_ intsize - ebs_
volume_ strtype - first_
on_ intdemand - instance_
profile_ strarn - spot_
bid_ intprice_ percent - zone_
id str
- availability String
- ebs
Volume NumberCount - ebs
Volume NumberSize - ebs
Volume StringType - first
On NumberDemand - instance
Profile StringArn - spot
Bid NumberPrice Percent - zone
Id String
ClusterAzureAttributes, ClusterAzureAttributesArgs
- Availability string
- First
On intDemand - Spot
Bid doubleMax Price
- Availability string
- First
On intDemand - Spot
Bid float64Max Price
- availability String
- first
On IntegerDemand - spot
Bid DoubleMax Price
- availability string
- first
On numberDemand - spot
Bid numberMax Price
- availability str
- first_
on_ intdemand - spot_
bid_ floatmax_ price
- availability String
- first
On NumberDemand - spot
Bid NumberMax Price
ClusterClusterLogConf, ClusterClusterLogConfArgs
ClusterClusterLogConfDbfs, ClusterClusterLogConfDbfsArgs
- Destination string
- Destination string
- destination String
- destination string
- destination str
- destination String
ClusterClusterLogConfS3, ClusterClusterLogConfS3Args
- Destination string
- Canned
Acl string - Enable
Encryption bool - Encryption
Type string - Endpoint string
- Kms
Key string - Region string
- Destination string
- Canned
Acl string - Enable
Encryption bool - Encryption
Type string - Endpoint string
- Kms
Key string - Region string
- destination String
- canned
Acl String - enable
Encryption Boolean - encryption
Type String - endpoint String
- kms
Key String - region String
- destination string
- canned
Acl string - enable
Encryption boolean - encryption
Type string - endpoint string
- kms
Key string - region string
- destination str
- canned_
acl str - enable_
encryption bool - encryption_
type str - endpoint str
- kms_
key str - region str
- destination String
- canned
Acl String - enable
Encryption Boolean - encryption
Type String - endpoint String
- kms
Key String - region String
ClusterDockerImage, ClusterDockerImageArgs
- url String
- basic
Auth Property Map
ClusterDockerImageBasicAuth, ClusterDockerImageBasicAuthArgs
ClusterGcpAttributes, ClusterGcpAttributesArgs
- Availability string
- Boot
Disk intSize - Google
Service stringAccount - Use
Preemptible boolExecutors - Zone
Id string
- Availability string
- Boot
Disk intSize - Google
Service stringAccount - Use
Preemptible boolExecutors - Zone
Id string
- availability String
- boot
Disk IntegerSize - google
Service StringAccount - use
Preemptible BooleanExecutors - zone
Id String
- availability string
- boot
Disk numberSize - google
Service stringAccount - use
Preemptible booleanExecutors - zone
Id string
- availability str
- boot_
disk_ intsize - google_
service_ straccount - use_
preemptible_ boolexecutors - zone_
id str
- availability String
- boot
Disk NumberSize - google
Service StringAccount - use
Preemptible BooleanExecutors - zone
Id String
ClusterInitScript, ClusterInitScriptArgs
ClusterInitScriptDbfs, ClusterInitScriptDbfsArgs
- Destination string
- Destination string
- destination String
- destination string
- destination str
- destination String
ClusterInitScriptFile, ClusterInitScriptFileArgs
- Destination string
- Destination string
- destination String
- destination string
- destination str
- destination String
ClusterInitScriptS3, ClusterInitScriptS3Args
- Destination string
- Canned
Acl string - Enable
Encryption bool - Encryption
Type string - Endpoint string
- Kms
Key string - Region string
- Destination string
- Canned
Acl string - Enable
Encryption bool - Encryption
Type string - Endpoint string
- Kms
Key string - Region string
- destination String
- canned
Acl String - enable
Encryption Boolean - encryption
Type String - endpoint String
- kms
Key String - region String
- destination string
- canned
Acl string - enable
Encryption boolean - encryption
Type string - endpoint string
- kms
Key string - region string
- destination str
- canned_
acl str - enable_
encryption bool - encryption_
type str - endpoint str
- kms_
key str - region str
- destination String
- canned
Acl String - enable
Encryption Boolean - encryption
Type String - endpoint String
- kms
Key String - region String
ClusterLibrary, ClusterLibraryArgs
- Cran
Cluster
Library Cran - Egg string
- Jar string
- Maven
Cluster
Library Maven - Pypi
Cluster
Library Pypi - Whl string
- Cran
Cluster
Library Cran - Egg string
- Jar string
- Maven
Cluster
Library Maven - Pypi
Cluster
Library Pypi - Whl string
- cran
Cluster
Library Cran - egg String
- jar String
- maven
Cluster
Library Maven - pypi
Cluster
Library Pypi - whl String
- cran
Cluster
Library Cran - egg string
- jar string
- maven
Cluster
Library Maven - pypi
Cluster
Library Pypi - whl string
- cran Property Map
- egg String
- jar String
- maven Property Map
- pypi Property Map
- whl String
ClusterLibraryCran, ClusterLibraryCranArgs
ClusterLibraryMaven, ClusterLibraryMavenArgs
- Coordinates string
- Exclusions List<string>
- Repo string
- Coordinates string
- Exclusions []string
- Repo string
- coordinates String
- exclusions List<String>
- repo String
- coordinates string
- exclusions string[]
- repo string
- coordinates str
- exclusions Sequence[str]
- repo str
- coordinates String
- exclusions List<String>
- repo String
ClusterLibraryPypi, ClusterLibraryPypiArgs
Package Details
- Repository
- databricks pulumi/pulumi-databricks
- License
- Apache-2.0
- Notes
- This Pulumi package is based on the
databricksTerraform Provider.
published on Monday, Mar 9, 2026 by Pulumi
