We recommend using Azure Native.
published on Monday, Mar 9, 2026 by Pulumi
We recommend using Azure Native.
published on Monday, Mar 9, 2026 by Pulumi
Manages a Synapse Spark Pool.
Create SparkPool Resource
Resources are created with functions called constructors. To learn more about declaring and configuring resources, see Resources.
Constructor syntax
new SparkPool(name: string, args: SparkPoolArgs, opts?: CustomResourceOptions);@overload
def SparkPool(resource_name: str,
args: SparkPoolArgs,
opts: Optional[ResourceOptions] = None)
@overload
def SparkPool(resource_name: str,
opts: Optional[ResourceOptions] = None,
node_size: Optional[str] = None,
synapse_workspace_id: Optional[str] = None,
node_size_family: Optional[str] = None,
dynamic_executor_allocation_enabled: Optional[bool] = None,
session_level_packages_enabled: Optional[bool] = None,
library_requirement: Optional[SparkPoolLibraryRequirementArgs] = None,
name: Optional[str] = None,
node_count: Optional[int] = None,
compute_isolation_enabled: Optional[bool] = None,
cache_size: Optional[int] = None,
auto_pause: Optional[SparkPoolAutoPauseArgs] = None,
spark_config: Optional[SparkPoolSparkConfigArgs] = None,
spark_events_folder: Optional[str] = None,
spark_log_folder: Optional[str] = None,
spark_version: Optional[str] = None,
auto_scale: Optional[SparkPoolAutoScaleArgs] = None,
tags: Optional[Mapping[str, str]] = None)func NewSparkPool(ctx *Context, name string, args SparkPoolArgs, opts ...ResourceOption) (*SparkPool, error)public SparkPool(string name, SparkPoolArgs args, CustomResourceOptions? opts = null)
public SparkPool(String name, SparkPoolArgs args)
public SparkPool(String name, SparkPoolArgs args, CustomResourceOptions options)
type: azure:synapse:SparkPool
properties: # The arguments to resource properties.
options: # Bag of options to control resource's behavior.
Parameters
- name string
- The unique name of the resource.
- args SparkPoolArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- resource_name str
- The unique name of the resource.
- args SparkPoolArgs
- The arguments to resource properties.
- opts ResourceOptions
- Bag of options to control resource's behavior.
- ctx Context
- Context object for the current deployment.
- name string
- The unique name of the resource.
- args SparkPoolArgs
- The arguments to resource properties.
- opts ResourceOption
- Bag of options to control resource's behavior.
- name string
- The unique name of the resource.
- args SparkPoolArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- name String
- The unique name of the resource.
- args SparkPoolArgs
- The arguments to resource properties.
- options CustomResourceOptions
- Bag of options to control resource's behavior.
Constructor example
The following reference example uses placeholder values for all input properties.
var sparkPoolResource = new Azure.Synapse.SparkPool("sparkPoolResource", new()
{
NodeSize = "string",
SynapseWorkspaceId = "string",
NodeSizeFamily = "string",
DynamicExecutorAllocationEnabled = false,
SessionLevelPackagesEnabled = false,
LibraryRequirement = new Azure.Synapse.Inputs.SparkPoolLibraryRequirementArgs
{
Content = "string",
Filename = "string",
},
Name = "string",
NodeCount = 0,
ComputeIsolationEnabled = false,
CacheSize = 0,
AutoPause = new Azure.Synapse.Inputs.SparkPoolAutoPauseArgs
{
DelayInMinutes = 0,
},
SparkConfig = new Azure.Synapse.Inputs.SparkPoolSparkConfigArgs
{
Content = "string",
Filename = "string",
},
SparkEventsFolder = "string",
SparkLogFolder = "string",
SparkVersion = "string",
AutoScale = new Azure.Synapse.Inputs.SparkPoolAutoScaleArgs
{
MaxNodeCount = 0,
MinNodeCount = 0,
},
Tags =
{
{ "string", "string" },
},
});
example, err := synapse.NewSparkPool(ctx, "sparkPoolResource", &synapse.SparkPoolArgs{
NodeSize: pulumi.String("string"),
SynapseWorkspaceId: pulumi.String("string"),
NodeSizeFamily: pulumi.String("string"),
DynamicExecutorAllocationEnabled: pulumi.Bool(false),
SessionLevelPackagesEnabled: pulumi.Bool(false),
LibraryRequirement: &synapse.SparkPoolLibraryRequirementArgs{
Content: pulumi.String("string"),
Filename: pulumi.String("string"),
},
Name: pulumi.String("string"),
NodeCount: pulumi.Int(0),
ComputeIsolationEnabled: pulumi.Bool(false),
CacheSize: pulumi.Int(0),
AutoPause: &synapse.SparkPoolAutoPauseArgs{
DelayInMinutes: pulumi.Int(0),
},
SparkConfig: &synapse.SparkPoolSparkConfigArgs{
Content: pulumi.String("string"),
Filename: pulumi.String("string"),
},
SparkEventsFolder: pulumi.String("string"),
SparkLogFolder: pulumi.String("string"),
SparkVersion: pulumi.String("string"),
AutoScale: &synapse.SparkPoolAutoScaleArgs{
MaxNodeCount: pulumi.Int(0),
MinNodeCount: pulumi.Int(0),
},
Tags: pulumi.StringMap{
"string": pulumi.String("string"),
},
})
var sparkPoolResource = new SparkPool("sparkPoolResource", SparkPoolArgs.builder()
.nodeSize("string")
.synapseWorkspaceId("string")
.nodeSizeFamily("string")
.dynamicExecutorAllocationEnabled(false)
.sessionLevelPackagesEnabled(false)
.libraryRequirement(SparkPoolLibraryRequirementArgs.builder()
.content("string")
.filename("string")
.build())
.name("string")
.nodeCount(0)
.computeIsolationEnabled(false)
.cacheSize(0)
.autoPause(SparkPoolAutoPauseArgs.builder()
.delayInMinutes(0)
.build())
.sparkConfig(SparkPoolSparkConfigArgs.builder()
.content("string")
.filename("string")
.build())
.sparkEventsFolder("string")
.sparkLogFolder("string")
.sparkVersion("string")
.autoScale(SparkPoolAutoScaleArgs.builder()
.maxNodeCount(0)
.minNodeCount(0)
.build())
.tags(Map.of("string", "string"))
.build());
spark_pool_resource = azure.synapse.SparkPool("sparkPoolResource",
node_size="string",
synapse_workspace_id="string",
node_size_family="string",
dynamic_executor_allocation_enabled=False,
session_level_packages_enabled=False,
library_requirement={
"content": "string",
"filename": "string",
},
name="string",
node_count=0,
compute_isolation_enabled=False,
cache_size=0,
auto_pause={
"delay_in_minutes": 0,
},
spark_config={
"content": "string",
"filename": "string",
},
spark_events_folder="string",
spark_log_folder="string",
spark_version="string",
auto_scale={
"max_node_count": 0,
"min_node_count": 0,
},
tags={
"string": "string",
})
const sparkPoolResource = new azure.synapse.SparkPool("sparkPoolResource", {
nodeSize: "string",
synapseWorkspaceId: "string",
nodeSizeFamily: "string",
dynamicExecutorAllocationEnabled: false,
sessionLevelPackagesEnabled: false,
libraryRequirement: {
content: "string",
filename: "string",
},
name: "string",
nodeCount: 0,
computeIsolationEnabled: false,
cacheSize: 0,
autoPause: {
delayInMinutes: 0,
},
sparkConfig: {
content: "string",
filename: "string",
},
sparkEventsFolder: "string",
sparkLogFolder: "string",
sparkVersion: "string",
autoScale: {
maxNodeCount: 0,
minNodeCount: 0,
},
tags: {
string: "string",
},
});
type: azure:synapse:SparkPool
properties:
autoPause:
delayInMinutes: 0
autoScale:
maxNodeCount: 0
minNodeCount: 0
cacheSize: 0
computeIsolationEnabled: false
dynamicExecutorAllocationEnabled: false
libraryRequirement:
content: string
filename: string
name: string
nodeCount: 0
nodeSize: string
nodeSizeFamily: string
sessionLevelPackagesEnabled: false
sparkConfig:
content: string
filename: string
sparkEventsFolder: string
sparkLogFolder: string
sparkVersion: string
synapseWorkspaceId: string
tags:
string: string
SparkPool Resource Properties
To learn more about resource properties and how to use them, see Inputs and Outputs in the Architecture and Concepts docs.
Inputs
In Python, inputs that are objects can be passed either as argument classes or as dictionary literals.
The SparkPool resource accepts the following input properties:
- Node
Size string - The level of node in the Spark Pool. Possible value is
Small,MediumandLarge. - Node
Size stringFamily - The kind of nodes that the Spark Pool provides. Possible value is
MemoryOptimized. - Synapse
Workspace stringId - The ID of the Synapse Workspace where the Synapse Spark Pool should exist. Changing this forces a new Synapse Spark Pool to be created.
- Auto
Pause SparkPool Auto Pause - An
auto_pauseblock as defined below. - Auto
Scale SparkPool Auto Scale - An
auto_scaleblock as defined below. Exactly one ofnode_countorauto_scalemust be specified. - Cache
Size int - The cache size in the Spark Pool.
- Compute
Isolation boolEnabled - Indicates whether compute isolation is enabled or not. Defaults to
false. - Dynamic
Executor boolAllocation Enabled - Indicates whether Dynamic Executor Allocation is enabled or not. Defaults to
false. - Library
Requirement SparkPool Library Requirement - A
library_requirementblock as defined below. - Name string
- The name which should be used for this Synapse Spark Pool. Changing this forces a new Synapse Spark Pool to be created.
- Node
Count int - The number of nodes in the Spark Pool. Exactly one of
node_countorauto_scalemust be specified. - Session
Level boolPackages Enabled - Indicates whether session level packages are enabled or not. Defaults to
false. - Spark
Config SparkPool Spark Config - A
spark_configblock as defined below. - Spark
Events stringFolder - The Spark events folder. Defaults to
/events. - Spark
Log stringFolder - The default folder where Spark logs will be written. Defaults to
/logs. - Spark
Version string - The Apache Spark version. Possible values are
2.4and3.1. Defaults to2.4. - Dictionary<string, string>
- A mapping of tags which should be assigned to the Synapse Spark Pool.
- Node
Size string - The level of node in the Spark Pool. Possible value is
Small,MediumandLarge. - Node
Size stringFamily - The kind of nodes that the Spark Pool provides. Possible value is
MemoryOptimized. - Synapse
Workspace stringId - The ID of the Synapse Workspace where the Synapse Spark Pool should exist. Changing this forces a new Synapse Spark Pool to be created.
- Auto
Pause SparkPool Auto Pause Args - An
auto_pauseblock as defined below. - Auto
Scale SparkPool Auto Scale Args - An
auto_scaleblock as defined below. Exactly one ofnode_countorauto_scalemust be specified. - Cache
Size int - The cache size in the Spark Pool.
- Compute
Isolation boolEnabled - Indicates whether compute isolation is enabled or not. Defaults to
false. - Dynamic
Executor boolAllocation Enabled - Indicates whether Dynamic Executor Allocation is enabled or not. Defaults to
false. - Library
Requirement SparkPool Library Requirement Args - A
library_requirementblock as defined below. - Name string
- The name which should be used for this Synapse Spark Pool. Changing this forces a new Synapse Spark Pool to be created.
- Node
Count int - The number of nodes in the Spark Pool. Exactly one of
node_countorauto_scalemust be specified. - Session
Level boolPackages Enabled - Indicates whether session level packages are enabled or not. Defaults to
false. - Spark
Config SparkPool Spark Config Args - A
spark_configblock as defined below. - Spark
Events stringFolder - The Spark events folder. Defaults to
/events. - Spark
Log stringFolder - The default folder where Spark logs will be written. Defaults to
/logs. - Spark
Version string - The Apache Spark version. Possible values are
2.4and3.1. Defaults to2.4. - map[string]string
- A mapping of tags which should be assigned to the Synapse Spark Pool.
- node
Size String - The level of node in the Spark Pool. Possible value is
Small,MediumandLarge. - node
Size StringFamily - The kind of nodes that the Spark Pool provides. Possible value is
MemoryOptimized. - synapse
Workspace StringId - The ID of the Synapse Workspace where the Synapse Spark Pool should exist. Changing this forces a new Synapse Spark Pool to be created.
- auto
Pause SparkPool Auto Pause - An
auto_pauseblock as defined below. - auto
Scale SparkPool Auto Scale - An
auto_scaleblock as defined below. Exactly one ofnode_countorauto_scalemust be specified. - cache
Size Integer - The cache size in the Spark Pool.
- compute
Isolation BooleanEnabled - Indicates whether compute isolation is enabled or not. Defaults to
false. - dynamic
Executor BooleanAllocation Enabled - Indicates whether Dynamic Executor Allocation is enabled or not. Defaults to
false. - library
Requirement SparkPool Library Requirement - A
library_requirementblock as defined below. - name String
- The name which should be used for this Synapse Spark Pool. Changing this forces a new Synapse Spark Pool to be created.
- node
Count Integer - The number of nodes in the Spark Pool. Exactly one of
node_countorauto_scalemust be specified. - session
Level BooleanPackages Enabled - Indicates whether session level packages are enabled or not. Defaults to
false. - spark
Config SparkPool Spark Config - A
spark_configblock as defined below. - spark
Events StringFolder - The Spark events folder. Defaults to
/events. - spark
Log StringFolder - The default folder where Spark logs will be written. Defaults to
/logs. - spark
Version String - The Apache Spark version. Possible values are
2.4and3.1. Defaults to2.4. - Map<String,String>
- A mapping of tags which should be assigned to the Synapse Spark Pool.
- node
Size string - The level of node in the Spark Pool. Possible value is
Small,MediumandLarge. - node
Size stringFamily - The kind of nodes that the Spark Pool provides. Possible value is
MemoryOptimized. - synapse
Workspace stringId - The ID of the Synapse Workspace where the Synapse Spark Pool should exist. Changing this forces a new Synapse Spark Pool to be created.
- auto
Pause SparkPool Auto Pause - An
auto_pauseblock as defined below. - auto
Scale SparkPool Auto Scale - An
auto_scaleblock as defined below. Exactly one ofnode_countorauto_scalemust be specified. - cache
Size number - The cache size in the Spark Pool.
- compute
Isolation booleanEnabled - Indicates whether compute isolation is enabled or not. Defaults to
false. - dynamic
Executor booleanAllocation Enabled - Indicates whether Dynamic Executor Allocation is enabled or not. Defaults to
false. - library
Requirement SparkPool Library Requirement - A
library_requirementblock as defined below. - name string
- The name which should be used for this Synapse Spark Pool. Changing this forces a new Synapse Spark Pool to be created.
- node
Count number - The number of nodes in the Spark Pool. Exactly one of
node_countorauto_scalemust be specified. - session
Level booleanPackages Enabled - Indicates whether session level packages are enabled or not. Defaults to
false. - spark
Config SparkPool Spark Config - A
spark_configblock as defined below. - spark
Events stringFolder - The Spark events folder. Defaults to
/events. - spark
Log stringFolder - The default folder where Spark logs will be written. Defaults to
/logs. - spark
Version string - The Apache Spark version. Possible values are
2.4and3.1. Defaults to2.4. - {[key: string]: string}
- A mapping of tags which should be assigned to the Synapse Spark Pool.
- node_
size str - The level of node in the Spark Pool. Possible value is
Small,MediumandLarge. - node_
size_ strfamily - The kind of nodes that the Spark Pool provides. Possible value is
MemoryOptimized. - synapse_
workspace_ strid - The ID of the Synapse Workspace where the Synapse Spark Pool should exist. Changing this forces a new Synapse Spark Pool to be created.
- auto_
pause SparkPool Auto Pause Args - An
auto_pauseblock as defined below. - auto_
scale SparkPool Auto Scale Args - An
auto_scaleblock as defined below. Exactly one ofnode_countorauto_scalemust be specified. - cache_
size int - The cache size in the Spark Pool.
- compute_
isolation_ boolenabled - Indicates whether compute isolation is enabled or not. Defaults to
false. - dynamic_
executor_ boolallocation_ enabled - Indicates whether Dynamic Executor Allocation is enabled or not. Defaults to
false. - library_
requirement SparkPool Library Requirement Args - A
library_requirementblock as defined below. - name str
- The name which should be used for this Synapse Spark Pool. Changing this forces a new Synapse Spark Pool to be created.
- node_
count int - The number of nodes in the Spark Pool. Exactly one of
node_countorauto_scalemust be specified. - session_
level_ boolpackages_ enabled - Indicates whether session level packages are enabled or not. Defaults to
false. - spark_
config SparkPool Spark Config Args - A
spark_configblock as defined below. - spark_
events_ strfolder - The Spark events folder. Defaults to
/events. - spark_
log_ strfolder - The default folder where Spark logs will be written. Defaults to
/logs. - spark_
version str - The Apache Spark version. Possible values are
2.4and3.1. Defaults to2.4. - Mapping[str, str]
- A mapping of tags which should be assigned to the Synapse Spark Pool.
- node
Size String - The level of node in the Spark Pool. Possible value is
Small,MediumandLarge. - node
Size StringFamily - The kind of nodes that the Spark Pool provides. Possible value is
MemoryOptimized. - synapse
Workspace StringId - The ID of the Synapse Workspace where the Synapse Spark Pool should exist. Changing this forces a new Synapse Spark Pool to be created.
- auto
Pause Property Map - An
auto_pauseblock as defined below. - auto
Scale Property Map - An
auto_scaleblock as defined below. Exactly one ofnode_countorauto_scalemust be specified. - cache
Size Number - The cache size in the Spark Pool.
- compute
Isolation BooleanEnabled - Indicates whether compute isolation is enabled or not. Defaults to
false. - dynamic
Executor BooleanAllocation Enabled - Indicates whether Dynamic Executor Allocation is enabled or not. Defaults to
false. - library
Requirement Property Map - A
library_requirementblock as defined below. - name String
- The name which should be used for this Synapse Spark Pool. Changing this forces a new Synapse Spark Pool to be created.
- node
Count Number - The number of nodes in the Spark Pool. Exactly one of
node_countorauto_scalemust be specified. - session
Level BooleanPackages Enabled - Indicates whether session level packages are enabled or not. Defaults to
false. - spark
Config Property Map - A
spark_configblock as defined below. - spark
Events StringFolder - The Spark events folder. Defaults to
/events. - spark
Log StringFolder - The default folder where Spark logs will be written. Defaults to
/logs. - spark
Version String - The Apache Spark version. Possible values are
2.4and3.1. Defaults to2.4. - Map<String>
- A mapping of tags which should be assigned to the Synapse Spark Pool.
Outputs
All input properties are implicitly available as output properties. Additionally, the SparkPool resource produces the following output properties:
- Id string
- The provider-assigned unique ID for this managed resource.
- Id string
- The provider-assigned unique ID for this managed resource.
- id String
- The provider-assigned unique ID for this managed resource.
- id string
- The provider-assigned unique ID for this managed resource.
- id str
- The provider-assigned unique ID for this managed resource.
- id String
- The provider-assigned unique ID for this managed resource.
Look up Existing SparkPool Resource
Get an existing SparkPool resource’s state with the given name, ID, and optional extra properties used to qualify the lookup.
public static get(name: string, id: Input<ID>, state?: SparkPoolState, opts?: CustomResourceOptions): SparkPool@staticmethod
def get(resource_name: str,
id: str,
opts: Optional[ResourceOptions] = None,
auto_pause: Optional[SparkPoolAutoPauseArgs] = None,
auto_scale: Optional[SparkPoolAutoScaleArgs] = None,
cache_size: Optional[int] = None,
compute_isolation_enabled: Optional[bool] = None,
dynamic_executor_allocation_enabled: Optional[bool] = None,
library_requirement: Optional[SparkPoolLibraryRequirementArgs] = None,
name: Optional[str] = None,
node_count: Optional[int] = None,
node_size: Optional[str] = None,
node_size_family: Optional[str] = None,
session_level_packages_enabled: Optional[bool] = None,
spark_config: Optional[SparkPoolSparkConfigArgs] = None,
spark_events_folder: Optional[str] = None,
spark_log_folder: Optional[str] = None,
spark_version: Optional[str] = None,
synapse_workspace_id: Optional[str] = None,
tags: Optional[Mapping[str, str]] = None) -> SparkPoolfunc GetSparkPool(ctx *Context, name string, id IDInput, state *SparkPoolState, opts ...ResourceOption) (*SparkPool, error)public static SparkPool Get(string name, Input<string> id, SparkPoolState? state, CustomResourceOptions? opts = null)public static SparkPool get(String name, Output<String> id, SparkPoolState state, CustomResourceOptions options)resources: _: type: azure:synapse:SparkPool get: id: ${id}- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- resource_name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- Auto
Pause SparkPool Auto Pause - An
auto_pauseblock as defined below. - Auto
Scale SparkPool Auto Scale - An
auto_scaleblock as defined below. Exactly one ofnode_countorauto_scalemust be specified. - Cache
Size int - The cache size in the Spark Pool.
- Compute
Isolation boolEnabled - Indicates whether compute isolation is enabled or not. Defaults to
false. - Dynamic
Executor boolAllocation Enabled - Indicates whether Dynamic Executor Allocation is enabled or not. Defaults to
false. - Library
Requirement SparkPool Library Requirement - A
library_requirementblock as defined below. - Name string
- The name which should be used for this Synapse Spark Pool. Changing this forces a new Synapse Spark Pool to be created.
- Node
Count int - The number of nodes in the Spark Pool. Exactly one of
node_countorauto_scalemust be specified. - Node
Size string - The level of node in the Spark Pool. Possible value is
Small,MediumandLarge. - Node
Size stringFamily - The kind of nodes that the Spark Pool provides. Possible value is
MemoryOptimized. - Session
Level boolPackages Enabled - Indicates whether session level packages are enabled or not. Defaults to
false. - Spark
Config SparkPool Spark Config - A
spark_configblock as defined below. - Spark
Events stringFolder - The Spark events folder. Defaults to
/events. - Spark
Log stringFolder - The default folder where Spark logs will be written. Defaults to
/logs. - Spark
Version string - The Apache Spark version. Possible values are
2.4and3.1. Defaults to2.4. - Synapse
Workspace stringId - The ID of the Synapse Workspace where the Synapse Spark Pool should exist. Changing this forces a new Synapse Spark Pool to be created.
- Dictionary<string, string>
- A mapping of tags which should be assigned to the Synapse Spark Pool.
- Auto
Pause SparkPool Auto Pause Args - An
auto_pauseblock as defined below. - Auto
Scale SparkPool Auto Scale Args - An
auto_scaleblock as defined below. Exactly one ofnode_countorauto_scalemust be specified. - Cache
Size int - The cache size in the Spark Pool.
- Compute
Isolation boolEnabled - Indicates whether compute isolation is enabled or not. Defaults to
false. - Dynamic
Executor boolAllocation Enabled - Indicates whether Dynamic Executor Allocation is enabled or not. Defaults to
false. - Library
Requirement SparkPool Library Requirement Args - A
library_requirementblock as defined below. - Name string
- The name which should be used for this Synapse Spark Pool. Changing this forces a new Synapse Spark Pool to be created.
- Node
Count int - The number of nodes in the Spark Pool. Exactly one of
node_countorauto_scalemust be specified. - Node
Size string - The level of node in the Spark Pool. Possible value is
Small,MediumandLarge. - Node
Size stringFamily - The kind of nodes that the Spark Pool provides. Possible value is
MemoryOptimized. - Session
Level boolPackages Enabled - Indicates whether session level packages are enabled or not. Defaults to
false. - Spark
Config SparkPool Spark Config Args - A
spark_configblock as defined below. - Spark
Events stringFolder - The Spark events folder. Defaults to
/events. - Spark
Log stringFolder - The default folder where Spark logs will be written. Defaults to
/logs. - Spark
Version string - The Apache Spark version. Possible values are
2.4and3.1. Defaults to2.4. - Synapse
Workspace stringId - The ID of the Synapse Workspace where the Synapse Spark Pool should exist. Changing this forces a new Synapse Spark Pool to be created.
- map[string]string
- A mapping of tags which should be assigned to the Synapse Spark Pool.
- auto
Pause SparkPool Auto Pause - An
auto_pauseblock as defined below. - auto
Scale SparkPool Auto Scale - An
auto_scaleblock as defined below. Exactly one ofnode_countorauto_scalemust be specified. - cache
Size Integer - The cache size in the Spark Pool.
- compute
Isolation BooleanEnabled - Indicates whether compute isolation is enabled or not. Defaults to
false. - dynamic
Executor BooleanAllocation Enabled - Indicates whether Dynamic Executor Allocation is enabled or not. Defaults to
false. - library
Requirement SparkPool Library Requirement - A
library_requirementblock as defined below. - name String
- The name which should be used for this Synapse Spark Pool. Changing this forces a new Synapse Spark Pool to be created.
- node
Count Integer - The number of nodes in the Spark Pool. Exactly one of
node_countorauto_scalemust be specified. - node
Size String - The level of node in the Spark Pool. Possible value is
Small,MediumandLarge. - node
Size StringFamily - The kind of nodes that the Spark Pool provides. Possible value is
MemoryOptimized. - session
Level BooleanPackages Enabled - Indicates whether session level packages are enabled or not. Defaults to
false. - spark
Config SparkPool Spark Config - A
spark_configblock as defined below. - spark
Events StringFolder - The Spark events folder. Defaults to
/events. - spark
Log StringFolder - The default folder where Spark logs will be written. Defaults to
/logs. - spark
Version String - The Apache Spark version. Possible values are
2.4and3.1. Defaults to2.4. - synapse
Workspace StringId - The ID of the Synapse Workspace where the Synapse Spark Pool should exist. Changing this forces a new Synapse Spark Pool to be created.
- Map<String,String>
- A mapping of tags which should be assigned to the Synapse Spark Pool.
- auto
Pause SparkPool Auto Pause - An
auto_pauseblock as defined below. - auto
Scale SparkPool Auto Scale - An
auto_scaleblock as defined below. Exactly one ofnode_countorauto_scalemust be specified. - cache
Size number - The cache size in the Spark Pool.
- compute
Isolation booleanEnabled - Indicates whether compute isolation is enabled or not. Defaults to
false. - dynamic
Executor booleanAllocation Enabled - Indicates whether Dynamic Executor Allocation is enabled or not. Defaults to
false. - library
Requirement SparkPool Library Requirement - A
library_requirementblock as defined below. - name string
- The name which should be used for this Synapse Spark Pool. Changing this forces a new Synapse Spark Pool to be created.
- node
Count number - The number of nodes in the Spark Pool. Exactly one of
node_countorauto_scalemust be specified. - node
Size string - The level of node in the Spark Pool. Possible value is
Small,MediumandLarge. - node
Size stringFamily - The kind of nodes that the Spark Pool provides. Possible value is
MemoryOptimized. - session
Level booleanPackages Enabled - Indicates whether session level packages are enabled or not. Defaults to
false. - spark
Config SparkPool Spark Config - A
spark_configblock as defined below. - spark
Events stringFolder - The Spark events folder. Defaults to
/events. - spark
Log stringFolder - The default folder where Spark logs will be written. Defaults to
/logs. - spark
Version string - The Apache Spark version. Possible values are
2.4and3.1. Defaults to2.4. - synapse
Workspace stringId - The ID of the Synapse Workspace where the Synapse Spark Pool should exist. Changing this forces a new Synapse Spark Pool to be created.
- {[key: string]: string}
- A mapping of tags which should be assigned to the Synapse Spark Pool.
- auto_
pause SparkPool Auto Pause Args - An
auto_pauseblock as defined below. - auto_
scale SparkPool Auto Scale Args - An
auto_scaleblock as defined below. Exactly one ofnode_countorauto_scalemust be specified. - cache_
size int - The cache size in the Spark Pool.
- compute_
isolation_ boolenabled - Indicates whether compute isolation is enabled or not. Defaults to
false. - dynamic_
executor_ boolallocation_ enabled - Indicates whether Dynamic Executor Allocation is enabled or not. Defaults to
false. - library_
requirement SparkPool Library Requirement Args - A
library_requirementblock as defined below. - name str
- The name which should be used for this Synapse Spark Pool. Changing this forces a new Synapse Spark Pool to be created.
- node_
count int - The number of nodes in the Spark Pool. Exactly one of
node_countorauto_scalemust be specified. - node_
size str - The level of node in the Spark Pool. Possible value is
Small,MediumandLarge. - node_
size_ strfamily - The kind of nodes that the Spark Pool provides. Possible value is
MemoryOptimized. - session_
level_ boolpackages_ enabled - Indicates whether session level packages are enabled or not. Defaults to
false. - spark_
config SparkPool Spark Config Args - A
spark_configblock as defined below. - spark_
events_ strfolder - The Spark events folder. Defaults to
/events. - spark_
log_ strfolder - The default folder where Spark logs will be written. Defaults to
/logs. - spark_
version str - The Apache Spark version. Possible values are
2.4and3.1. Defaults to2.4. - synapse_
workspace_ strid - The ID of the Synapse Workspace where the Synapse Spark Pool should exist. Changing this forces a new Synapse Spark Pool to be created.
- Mapping[str, str]
- A mapping of tags which should be assigned to the Synapse Spark Pool.
- auto
Pause Property Map - An
auto_pauseblock as defined below. - auto
Scale Property Map - An
auto_scaleblock as defined below. Exactly one ofnode_countorauto_scalemust be specified. - cache
Size Number - The cache size in the Spark Pool.
- compute
Isolation BooleanEnabled - Indicates whether compute isolation is enabled or not. Defaults to
false. - dynamic
Executor BooleanAllocation Enabled - Indicates whether Dynamic Executor Allocation is enabled or not. Defaults to
false. - library
Requirement Property Map - A
library_requirementblock as defined below. - name String
- The name which should be used for this Synapse Spark Pool. Changing this forces a new Synapse Spark Pool to be created.
- node
Count Number - The number of nodes in the Spark Pool. Exactly one of
node_countorauto_scalemust be specified. - node
Size String - The level of node in the Spark Pool. Possible value is
Small,MediumandLarge. - node
Size StringFamily - The kind of nodes that the Spark Pool provides. Possible value is
MemoryOptimized. - session
Level BooleanPackages Enabled - Indicates whether session level packages are enabled or not. Defaults to
false. - spark
Config Property Map - A
spark_configblock as defined below. - spark
Events StringFolder - The Spark events folder. Defaults to
/events. - spark
Log StringFolder - The default folder where Spark logs will be written. Defaults to
/logs. - spark
Version String - The Apache Spark version. Possible values are
2.4and3.1. Defaults to2.4. - synapse
Workspace StringId - The ID of the Synapse Workspace where the Synapse Spark Pool should exist. Changing this forces a new Synapse Spark Pool to be created.
- Map<String>
- A mapping of tags which should be assigned to the Synapse Spark Pool.
Supporting Types
SparkPoolAutoPause, SparkPoolAutoPauseArgs
- Delay
In intMinutes - Number of minutes of idle time before the Spark Pool is automatically paused. Must be between
5and10080.
- Delay
In intMinutes - Number of minutes of idle time before the Spark Pool is automatically paused. Must be between
5and10080.
- delay
In IntegerMinutes - Number of minutes of idle time before the Spark Pool is automatically paused. Must be between
5and10080.
- delay
In numberMinutes - Number of minutes of idle time before the Spark Pool is automatically paused. Must be between
5and10080.
- delay_
in_ intminutes - Number of minutes of idle time before the Spark Pool is automatically paused. Must be between
5and10080.
- delay
In NumberMinutes - Number of minutes of idle time before the Spark Pool is automatically paused. Must be between
5and10080.
SparkPoolAutoScale, SparkPoolAutoScaleArgs
- Max
Node intCount - The maximum number of nodes the Spark Pool can support. Must be between
3and200. - Min
Node intCount - The minimum number of nodes the Spark Pool can support. Must be between
3and200.
- Max
Node intCount - The maximum number of nodes the Spark Pool can support. Must be between
3and200. - Min
Node intCount - The minimum number of nodes the Spark Pool can support. Must be between
3and200.
- max
Node IntegerCount - The maximum number of nodes the Spark Pool can support. Must be between
3and200. - min
Node IntegerCount - The minimum number of nodes the Spark Pool can support. Must be between
3and200.
- max
Node numberCount - The maximum number of nodes the Spark Pool can support. Must be between
3and200. - min
Node numberCount - The minimum number of nodes the Spark Pool can support. Must be between
3and200.
- max_
node_ intcount - The maximum number of nodes the Spark Pool can support. Must be between
3and200. - min_
node_ intcount - The minimum number of nodes the Spark Pool can support. Must be between
3and200.
- max
Node NumberCount - The maximum number of nodes the Spark Pool can support. Must be between
3and200. - min
Node NumberCount - The minimum number of nodes the Spark Pool can support. Must be between
3and200.
SparkPoolLibraryRequirement, SparkPoolLibraryRequirementArgs
SparkPoolSparkConfig, SparkPoolSparkConfigArgs
Import
Synapse Spark Pool can be imported using the resource id, e.g.
$ pulumi import azure:synapse/sparkPool:SparkPool example /subscriptions/00000000-0000-0000-0000-000000000000/resourceGroups/group1/providers/Microsoft.Synapse/workspaces/workspace1/bigDataPools/sparkPool1
To learn more about importing existing cloud resources, see Importing resources.
Package Details
- Repository
- Azure Classic pulumi/pulumi-azure
- License
- Apache-2.0
- Notes
- This Pulumi package is based on the
azurermTerraform Provider.
We recommend using Azure Native.
published on Monday, Mar 9, 2026 by Pulumi
