1. Packages
  2. Azure Classic
  3. API Docs
  4. streamanalytics
  5. OutputBlob

We recommend using Azure Native.

Viewing docs for Azure v4.42.0 (Older version)
published on Monday, Mar 9, 2026 by Pulumi
azure logo

We recommend using Azure Native.

Viewing docs for Azure v4.42.0 (Older version)
published on Monday, Mar 9, 2026 by Pulumi

    Manages a Stream Analytics Output to Blob Storage.

    Example Usage

    using Pulumi;
    using Azure = Pulumi.Azure;
    
    class MyStack : Stack
    {
        public MyStack()
        {
            var exampleResourceGroup = Output.Create(Azure.Core.GetResourceGroup.InvokeAsync(new Azure.Core.GetResourceGroupArgs
            {
                Name = "example-resources",
            }));
            var exampleJob = Output.Create(Azure.StreamAnalytics.GetJob.InvokeAsync(new Azure.StreamAnalytics.GetJobArgs
            {
                Name = "example-job",
                ResourceGroupName = azurerm_resource_group.Example.Name,
            }));
            var exampleAccount = new Azure.Storage.Account("exampleAccount", new Azure.Storage.AccountArgs
            {
                ResourceGroupName = exampleResourceGroup.Apply(exampleResourceGroup => exampleResourceGroup.Name),
                Location = exampleResourceGroup.Apply(exampleResourceGroup => exampleResourceGroup.Location),
                AccountTier = "Standard",
                AccountReplicationType = "LRS",
            });
            var exampleContainer = new Azure.Storage.Container("exampleContainer", new Azure.Storage.ContainerArgs
            {
                StorageAccountName = exampleAccount.Name,
                ContainerAccessType = "private",
            });
            var exampleOutputBlob = new Azure.StreamAnalytics.OutputBlob("exampleOutputBlob", new Azure.StreamAnalytics.OutputBlobArgs
            {
                StreamAnalyticsJobName = exampleJob.Apply(exampleJob => exampleJob.Name),
                ResourceGroupName = exampleJob.Apply(exampleJob => exampleJob.ResourceGroupName),
                StorageAccountName = exampleAccount.Name,
                StorageAccountKey = exampleAccount.PrimaryAccessKey,
                StorageContainerName = exampleContainer.Name,
                PathPattern = "some-pattern",
                DateFormat = "yyyy-MM-dd",
                TimeFormat = "HH",
                Serialization = new Azure.StreamAnalytics.Inputs.OutputBlobSerializationArgs
                {
                    Type = "Csv",
                    Encoding = "UTF8",
                    FieldDelimiter = ",",
                },
            });
        }
    
    }
    
    package main
    
    import (
    	"github.com/pulumi/pulumi-azure/sdk/v4/go/azure/core"
    	"github.com/pulumi/pulumi-azure/sdk/v4/go/azure/storage"
    	"github.com/pulumi/pulumi-azure/sdk/v4/go/azure/streamanalytics"
    	"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
    )
    
    func main() {
    	pulumi.Run(func(ctx *pulumi.Context) error {
    		exampleResourceGroup, err := core.LookupResourceGroup(ctx, &core.LookupResourceGroupArgs{
    			Name: "example-resources",
    		}, nil)
    		if err != nil {
    			return err
    		}
    		exampleJob, err := streamanalytics.LookupJob(ctx, &streamanalytics.LookupJobArgs{
    			Name:              "example-job",
    			ResourceGroupName: azurerm_resource_group.Example.Name,
    		}, nil)
    		if err != nil {
    			return err
    		}
    		exampleAccount, err := storage.NewAccount(ctx, "exampleAccount", &storage.AccountArgs{
    			ResourceGroupName:      pulumi.String(exampleResourceGroup.Name),
    			Location:               pulumi.String(exampleResourceGroup.Location),
    			AccountTier:            pulumi.String("Standard"),
    			AccountReplicationType: pulumi.String("LRS"),
    		})
    		if err != nil {
    			return err
    		}
    		exampleContainer, err := storage.NewContainer(ctx, "exampleContainer", &storage.ContainerArgs{
    			StorageAccountName:  exampleAccount.Name,
    			ContainerAccessType: pulumi.String("private"),
    		})
    		if err != nil {
    			return err
    		}
    		_, err = streamanalytics.NewOutputBlob(ctx, "exampleOutputBlob", &streamanalytics.OutputBlobArgs{
    			StreamAnalyticsJobName: pulumi.String(exampleJob.Name),
    			ResourceGroupName:      pulumi.String(exampleJob.ResourceGroupName),
    			StorageAccountName:     exampleAccount.Name,
    			StorageAccountKey:      exampleAccount.PrimaryAccessKey,
    			StorageContainerName:   exampleContainer.Name,
    			PathPattern:            pulumi.String("some-pattern"),
    			DateFormat:             pulumi.String("yyyy-MM-dd"),
    			TimeFormat:             pulumi.String("HH"),
    			Serialization: &streamanalytics.OutputBlobSerializationArgs{
    				Type:           pulumi.String("Csv"),
    				Encoding:       pulumi.String("UTF8"),
    				FieldDelimiter: pulumi.String(","),
    			},
    		})
    		if err != nil {
    			return err
    		}
    		return nil
    	})
    }
    

    Example coming soon!

    import * as pulumi from "@pulumi/pulumi";
    import * as azure from "@pulumi/azure";
    
    const exampleResourceGroup = azure.core.getResourceGroup({
        name: "example-resources",
    });
    const exampleJob = azure.streamanalytics.getJob({
        name: "example-job",
        resourceGroupName: azurerm_resource_group.example.name,
    });
    const exampleAccount = new azure.storage.Account("exampleAccount", {
        resourceGroupName: exampleResourceGroup.then(exampleResourceGroup => exampleResourceGroup.name),
        location: exampleResourceGroup.then(exampleResourceGroup => exampleResourceGroup.location),
        accountTier: "Standard",
        accountReplicationType: "LRS",
    });
    const exampleContainer = new azure.storage.Container("exampleContainer", {
        storageAccountName: exampleAccount.name,
        containerAccessType: "private",
    });
    const exampleOutputBlob = new azure.streamanalytics.OutputBlob("exampleOutputBlob", {
        streamAnalyticsJobName: exampleJob.then(exampleJob => exampleJob.name),
        resourceGroupName: exampleJob.then(exampleJob => exampleJob.resourceGroupName),
        storageAccountName: exampleAccount.name,
        storageAccountKey: exampleAccount.primaryAccessKey,
        storageContainerName: exampleContainer.name,
        pathPattern: "some-pattern",
        dateFormat: "yyyy-MM-dd",
        timeFormat: "HH",
        serialization: {
            type: "Csv",
            encoding: "UTF8",
            fieldDelimiter: ",",
        },
    });
    
    import pulumi
    import pulumi_azure as azure
    
    example_resource_group = azure.core.get_resource_group(name="example-resources")
    example_job = azure.streamanalytics.get_job(name="example-job",
        resource_group_name=azurerm_resource_group["example"]["name"])
    example_account = azure.storage.Account("exampleAccount",
        resource_group_name=example_resource_group.name,
        location=example_resource_group.location,
        account_tier="Standard",
        account_replication_type="LRS")
    example_container = azure.storage.Container("exampleContainer",
        storage_account_name=example_account.name,
        container_access_type="private")
    example_output_blob = azure.streamanalytics.OutputBlob("exampleOutputBlob",
        stream_analytics_job_name=example_job.name,
        resource_group_name=example_job.resource_group_name,
        storage_account_name=example_account.name,
        storage_account_key=example_account.primary_access_key,
        storage_container_name=example_container.name,
        path_pattern="some-pattern",
        date_format="yyyy-MM-dd",
        time_format="HH",
        serialization=azure.streamanalytics.OutputBlobSerializationArgs(
            type="Csv",
            encoding="UTF8",
            field_delimiter=",",
        ))
    

    Example coming soon!

    Create OutputBlob Resource

    Resources are created with functions called constructors. To learn more about declaring and configuring resources, see Resources.

    Constructor syntax

    new OutputBlob(name: string, args: OutputBlobArgs, opts?: CustomResourceOptions);
    @overload
    def OutputBlob(resource_name: str,
                   args: OutputBlobArgs,
                   opts: Optional[ResourceOptions] = None)
    
    @overload
    def OutputBlob(resource_name: str,
                   opts: Optional[ResourceOptions] = None,
                   date_format: Optional[str] = None,
                   path_pattern: Optional[str] = None,
                   resource_group_name: Optional[str] = None,
                   serialization: Optional[OutputBlobSerializationArgs] = None,
                   storage_account_key: Optional[str] = None,
                   storage_account_name: Optional[str] = None,
                   storage_container_name: Optional[str] = None,
                   stream_analytics_job_name: Optional[str] = None,
                   time_format: Optional[str] = None,
                   batch_max_wait_time: Optional[str] = None,
                   batch_min_rows: Optional[float] = None,
                   name: Optional[str] = None)
    func NewOutputBlob(ctx *Context, name string, args OutputBlobArgs, opts ...ResourceOption) (*OutputBlob, error)
    public OutputBlob(string name, OutputBlobArgs args, CustomResourceOptions? opts = null)
    public OutputBlob(String name, OutputBlobArgs args)
    public OutputBlob(String name, OutputBlobArgs args, CustomResourceOptions options)
    
    type: azure:streamanalytics:OutputBlob
    properties: # The arguments to resource properties.
    options: # Bag of options to control resource's behavior.
    
    

    Parameters

    name string
    The unique name of the resource.
    args OutputBlobArgs
    The arguments to resource properties.
    opts CustomResourceOptions
    Bag of options to control resource's behavior.
    resource_name str
    The unique name of the resource.
    args OutputBlobArgs
    The arguments to resource properties.
    opts ResourceOptions
    Bag of options to control resource's behavior.
    ctx Context
    Context object for the current deployment.
    name string
    The unique name of the resource.
    args OutputBlobArgs
    The arguments to resource properties.
    opts ResourceOption
    Bag of options to control resource's behavior.
    name string
    The unique name of the resource.
    args OutputBlobArgs
    The arguments to resource properties.
    opts CustomResourceOptions
    Bag of options to control resource's behavior.
    name String
    The unique name of the resource.
    args OutputBlobArgs
    The arguments to resource properties.
    options CustomResourceOptions
    Bag of options to control resource's behavior.

    Constructor example

    The following reference example uses placeholder values for all input properties.

    var outputBlobResource = new Azure.StreamAnalytics.OutputBlob("outputBlobResource", new()
    {
        DateFormat = "string",
        PathPattern = "string",
        ResourceGroupName = "string",
        Serialization = new Azure.StreamAnalytics.Inputs.OutputBlobSerializationArgs
        {
            Type = "string",
            Encoding = "string",
            FieldDelimiter = "string",
            Format = "string",
        },
        StorageAccountKey = "string",
        StorageAccountName = "string",
        StorageContainerName = "string",
        StreamAnalyticsJobName = "string",
        TimeFormat = "string",
        BatchMaxWaitTime = "string",
        BatchMinRows = 0,
        Name = "string",
    });
    
    example, err := streamanalytics.NewOutputBlob(ctx, "outputBlobResource", &streamanalytics.OutputBlobArgs{
    	DateFormat:        pulumi.String("string"),
    	PathPattern:       pulumi.String("string"),
    	ResourceGroupName: pulumi.String("string"),
    	Serialization: &streamanalytics.OutputBlobSerializationArgs{
    		Type:           pulumi.String("string"),
    		Encoding:       pulumi.String("string"),
    		FieldDelimiter: pulumi.String("string"),
    		Format:         pulumi.String("string"),
    	},
    	StorageAccountKey:      pulumi.String("string"),
    	StorageAccountName:     pulumi.String("string"),
    	StorageContainerName:   pulumi.String("string"),
    	StreamAnalyticsJobName: pulumi.String("string"),
    	TimeFormat:             pulumi.String("string"),
    	BatchMaxWaitTime:       pulumi.String("string"),
    	BatchMinRows:           pulumi.Float64(0),
    	Name:                   pulumi.String("string"),
    })
    
    var outputBlobResource = new OutputBlob("outputBlobResource", OutputBlobArgs.builder()
        .dateFormat("string")
        .pathPattern("string")
        .resourceGroupName("string")
        .serialization(OutputBlobSerializationArgs.builder()
            .type("string")
            .encoding("string")
            .fieldDelimiter("string")
            .format("string")
            .build())
        .storageAccountKey("string")
        .storageAccountName("string")
        .storageContainerName("string")
        .streamAnalyticsJobName("string")
        .timeFormat("string")
        .batchMaxWaitTime("string")
        .batchMinRows(0.0)
        .name("string")
        .build());
    
    output_blob_resource = azure.streamanalytics.OutputBlob("outputBlobResource",
        date_format="string",
        path_pattern="string",
        resource_group_name="string",
        serialization={
            "type": "string",
            "encoding": "string",
            "field_delimiter": "string",
            "format": "string",
        },
        storage_account_key="string",
        storage_account_name="string",
        storage_container_name="string",
        stream_analytics_job_name="string",
        time_format="string",
        batch_max_wait_time="string",
        batch_min_rows=0,
        name="string")
    
    const outputBlobResource = new azure.streamanalytics.OutputBlob("outputBlobResource", {
        dateFormat: "string",
        pathPattern: "string",
        resourceGroupName: "string",
        serialization: {
            type: "string",
            encoding: "string",
            fieldDelimiter: "string",
            format: "string",
        },
        storageAccountKey: "string",
        storageAccountName: "string",
        storageContainerName: "string",
        streamAnalyticsJobName: "string",
        timeFormat: "string",
        batchMaxWaitTime: "string",
        batchMinRows: 0,
        name: "string",
    });
    
    type: azure:streamanalytics:OutputBlob
    properties:
        batchMaxWaitTime: string
        batchMinRows: 0
        dateFormat: string
        name: string
        pathPattern: string
        resourceGroupName: string
        serialization:
            encoding: string
            fieldDelimiter: string
            format: string
            type: string
        storageAccountKey: string
        storageAccountName: string
        storageContainerName: string
        streamAnalyticsJobName: string
        timeFormat: string
    

    OutputBlob Resource Properties

    To learn more about resource properties and how to use them, see Inputs and Outputs in the Architecture and Concepts docs.

    Inputs

    In Python, inputs that are objects can be passed either as argument classes or as dictionary literals.

    The OutputBlob resource accepts the following input properties:

    DateFormat string
    The date format. Wherever {date} appears in path_pattern, the value of this property is used as the date format instead.
    PathPattern string
    The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
    ResourceGroupName string
    The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
    Serialization OutputBlobSerialization
    A serialization block as defined below.
    StorageAccountKey string
    The Access Key which should be used to connect to this Storage Account.
    StorageAccountName string
    The name of the Storage Account.
    StorageContainerName string
    The name of the Container within the Storage Account.
    StreamAnalyticsJobName string
    The name of the Stream Analytics Job. Changing this forces a new resource to be created.
    TimeFormat string
    The time format. Wherever {time} appears in path_pattern, the value of this property is used as the time format instead.
    BatchMaxWaitTime string
    The maximum wait time per batch in hh:mm:ss e.g. 00:02:00 for two minutes.
    BatchMinRows double
    The minimum number of rows per batch (must be between 0 and 10000).
    Name string
    The name of the Stream Output. Changing this forces a new resource to be created.
    DateFormat string
    The date format. Wherever {date} appears in path_pattern, the value of this property is used as the date format instead.
    PathPattern string
    The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
    ResourceGroupName string
    The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
    Serialization OutputBlobSerializationArgs
    A serialization block as defined below.
    StorageAccountKey string
    The Access Key which should be used to connect to this Storage Account.
    StorageAccountName string
    The name of the Storage Account.
    StorageContainerName string
    The name of the Container within the Storage Account.
    StreamAnalyticsJobName string
    The name of the Stream Analytics Job. Changing this forces a new resource to be created.
    TimeFormat string
    The time format. Wherever {time} appears in path_pattern, the value of this property is used as the time format instead.
    BatchMaxWaitTime string
    The maximum wait time per batch in hh:mm:ss e.g. 00:02:00 for two minutes.
    BatchMinRows float64
    The minimum number of rows per batch (must be between 0 and 10000).
    Name string
    The name of the Stream Output. Changing this forces a new resource to be created.
    dateFormat String
    The date format. Wherever {date} appears in path_pattern, the value of this property is used as the date format instead.
    pathPattern String
    The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
    resourceGroupName String
    The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
    serialization OutputBlobSerialization
    A serialization block as defined below.
    storageAccountKey String
    The Access Key which should be used to connect to this Storage Account.
    storageAccountName String
    The name of the Storage Account.
    storageContainerName String
    The name of the Container within the Storage Account.
    streamAnalyticsJobName String
    The name of the Stream Analytics Job. Changing this forces a new resource to be created.
    timeFormat String
    The time format. Wherever {time} appears in path_pattern, the value of this property is used as the time format instead.
    batchMaxWaitTime String
    The maximum wait time per batch in hh:mm:ss e.g. 00:02:00 for two minutes.
    batchMinRows Double
    The minimum number of rows per batch (must be between 0 and 10000).
    name String
    The name of the Stream Output. Changing this forces a new resource to be created.
    dateFormat string
    The date format. Wherever {date} appears in path_pattern, the value of this property is used as the date format instead.
    pathPattern string
    The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
    resourceGroupName string
    The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
    serialization OutputBlobSerialization
    A serialization block as defined below.
    storageAccountKey string
    The Access Key which should be used to connect to this Storage Account.
    storageAccountName string
    The name of the Storage Account.
    storageContainerName string
    The name of the Container within the Storage Account.
    streamAnalyticsJobName string
    The name of the Stream Analytics Job. Changing this forces a new resource to be created.
    timeFormat string
    The time format. Wherever {time} appears in path_pattern, the value of this property is used as the time format instead.
    batchMaxWaitTime string
    The maximum wait time per batch in hh:mm:ss e.g. 00:02:00 for two minutes.
    batchMinRows number
    The minimum number of rows per batch (must be between 0 and 10000).
    name string
    The name of the Stream Output. Changing this forces a new resource to be created.
    date_format str
    The date format. Wherever {date} appears in path_pattern, the value of this property is used as the date format instead.
    path_pattern str
    The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
    resource_group_name str
    The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
    serialization OutputBlobSerializationArgs
    A serialization block as defined below.
    storage_account_key str
    The Access Key which should be used to connect to this Storage Account.
    storage_account_name str
    The name of the Storage Account.
    storage_container_name str
    The name of the Container within the Storage Account.
    stream_analytics_job_name str
    The name of the Stream Analytics Job. Changing this forces a new resource to be created.
    time_format str
    The time format. Wherever {time} appears in path_pattern, the value of this property is used as the time format instead.
    batch_max_wait_time str
    The maximum wait time per batch in hh:mm:ss e.g. 00:02:00 for two minutes.
    batch_min_rows float
    The minimum number of rows per batch (must be between 0 and 10000).
    name str
    The name of the Stream Output. Changing this forces a new resource to be created.
    dateFormat String
    The date format. Wherever {date} appears in path_pattern, the value of this property is used as the date format instead.
    pathPattern String
    The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
    resourceGroupName String
    The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
    serialization Property Map
    A serialization block as defined below.
    storageAccountKey String
    The Access Key which should be used to connect to this Storage Account.
    storageAccountName String
    The name of the Storage Account.
    storageContainerName String
    The name of the Container within the Storage Account.
    streamAnalyticsJobName String
    The name of the Stream Analytics Job. Changing this forces a new resource to be created.
    timeFormat String
    The time format. Wherever {time} appears in path_pattern, the value of this property is used as the time format instead.
    batchMaxWaitTime String
    The maximum wait time per batch in hh:mm:ss e.g. 00:02:00 for two minutes.
    batchMinRows Number
    The minimum number of rows per batch (must be between 0 and 10000).
    name String
    The name of the Stream Output. Changing this forces a new resource to be created.

    Outputs

    All input properties are implicitly available as output properties. Additionally, the OutputBlob resource produces the following output properties:

    Id string
    The provider-assigned unique ID for this managed resource.
    Id string
    The provider-assigned unique ID for this managed resource.
    id String
    The provider-assigned unique ID for this managed resource.
    id string
    The provider-assigned unique ID for this managed resource.
    id str
    The provider-assigned unique ID for this managed resource.
    id String
    The provider-assigned unique ID for this managed resource.

    Look up Existing OutputBlob Resource

    Get an existing OutputBlob resource’s state with the given name, ID, and optional extra properties used to qualify the lookup.

    public static get(name: string, id: Input<ID>, state?: OutputBlobState, opts?: CustomResourceOptions): OutputBlob
    @staticmethod
    def get(resource_name: str,
            id: str,
            opts: Optional[ResourceOptions] = None,
            batch_max_wait_time: Optional[str] = None,
            batch_min_rows: Optional[float] = None,
            date_format: Optional[str] = None,
            name: Optional[str] = None,
            path_pattern: Optional[str] = None,
            resource_group_name: Optional[str] = None,
            serialization: Optional[OutputBlobSerializationArgs] = None,
            storage_account_key: Optional[str] = None,
            storage_account_name: Optional[str] = None,
            storage_container_name: Optional[str] = None,
            stream_analytics_job_name: Optional[str] = None,
            time_format: Optional[str] = None) -> OutputBlob
    func GetOutputBlob(ctx *Context, name string, id IDInput, state *OutputBlobState, opts ...ResourceOption) (*OutputBlob, error)
    public static OutputBlob Get(string name, Input<string> id, OutputBlobState? state, CustomResourceOptions? opts = null)
    public static OutputBlob get(String name, Output<String> id, OutputBlobState state, CustomResourceOptions options)
    resources:  _:    type: azure:streamanalytics:OutputBlob    get:      id: ${id}
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    resource_name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    The following state arguments are supported:
    BatchMaxWaitTime string
    The maximum wait time per batch in hh:mm:ss e.g. 00:02:00 for two minutes.
    BatchMinRows double
    The minimum number of rows per batch (must be between 0 and 10000).
    DateFormat string
    The date format. Wherever {date} appears in path_pattern, the value of this property is used as the date format instead.
    Name string
    The name of the Stream Output. Changing this forces a new resource to be created.
    PathPattern string
    The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
    ResourceGroupName string
    The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
    Serialization OutputBlobSerialization
    A serialization block as defined below.
    StorageAccountKey string
    The Access Key which should be used to connect to this Storage Account.
    StorageAccountName string
    The name of the Storage Account.
    StorageContainerName string
    The name of the Container within the Storage Account.
    StreamAnalyticsJobName string
    The name of the Stream Analytics Job. Changing this forces a new resource to be created.
    TimeFormat string
    The time format. Wherever {time} appears in path_pattern, the value of this property is used as the time format instead.
    BatchMaxWaitTime string
    The maximum wait time per batch in hh:mm:ss e.g. 00:02:00 for two minutes.
    BatchMinRows float64
    The minimum number of rows per batch (must be between 0 and 10000).
    DateFormat string
    The date format. Wherever {date} appears in path_pattern, the value of this property is used as the date format instead.
    Name string
    The name of the Stream Output. Changing this forces a new resource to be created.
    PathPattern string
    The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
    ResourceGroupName string
    The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
    Serialization OutputBlobSerializationArgs
    A serialization block as defined below.
    StorageAccountKey string
    The Access Key which should be used to connect to this Storage Account.
    StorageAccountName string
    The name of the Storage Account.
    StorageContainerName string
    The name of the Container within the Storage Account.
    StreamAnalyticsJobName string
    The name of the Stream Analytics Job. Changing this forces a new resource to be created.
    TimeFormat string
    The time format. Wherever {time} appears in path_pattern, the value of this property is used as the time format instead.
    batchMaxWaitTime String
    The maximum wait time per batch in hh:mm:ss e.g. 00:02:00 for two minutes.
    batchMinRows Double
    The minimum number of rows per batch (must be between 0 and 10000).
    dateFormat String
    The date format. Wherever {date} appears in path_pattern, the value of this property is used as the date format instead.
    name String
    The name of the Stream Output. Changing this forces a new resource to be created.
    pathPattern String
    The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
    resourceGroupName String
    The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
    serialization OutputBlobSerialization
    A serialization block as defined below.
    storageAccountKey String
    The Access Key which should be used to connect to this Storage Account.
    storageAccountName String
    The name of the Storage Account.
    storageContainerName String
    The name of the Container within the Storage Account.
    streamAnalyticsJobName String
    The name of the Stream Analytics Job. Changing this forces a new resource to be created.
    timeFormat String
    The time format. Wherever {time} appears in path_pattern, the value of this property is used as the time format instead.
    batchMaxWaitTime string
    The maximum wait time per batch in hh:mm:ss e.g. 00:02:00 for two minutes.
    batchMinRows number
    The minimum number of rows per batch (must be between 0 and 10000).
    dateFormat string
    The date format. Wherever {date} appears in path_pattern, the value of this property is used as the date format instead.
    name string
    The name of the Stream Output. Changing this forces a new resource to be created.
    pathPattern string
    The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
    resourceGroupName string
    The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
    serialization OutputBlobSerialization
    A serialization block as defined below.
    storageAccountKey string
    The Access Key which should be used to connect to this Storage Account.
    storageAccountName string
    The name of the Storage Account.
    storageContainerName string
    The name of the Container within the Storage Account.
    streamAnalyticsJobName string
    The name of the Stream Analytics Job. Changing this forces a new resource to be created.
    timeFormat string
    The time format. Wherever {time} appears in path_pattern, the value of this property is used as the time format instead.
    batch_max_wait_time str
    The maximum wait time per batch in hh:mm:ss e.g. 00:02:00 for two minutes.
    batch_min_rows float
    The minimum number of rows per batch (must be between 0 and 10000).
    date_format str
    The date format. Wherever {date} appears in path_pattern, the value of this property is used as the date format instead.
    name str
    The name of the Stream Output. Changing this forces a new resource to be created.
    path_pattern str
    The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
    resource_group_name str
    The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
    serialization OutputBlobSerializationArgs
    A serialization block as defined below.
    storage_account_key str
    The Access Key which should be used to connect to this Storage Account.
    storage_account_name str
    The name of the Storage Account.
    storage_container_name str
    The name of the Container within the Storage Account.
    stream_analytics_job_name str
    The name of the Stream Analytics Job. Changing this forces a new resource to be created.
    time_format str
    The time format. Wherever {time} appears in path_pattern, the value of this property is used as the time format instead.
    batchMaxWaitTime String
    The maximum wait time per batch in hh:mm:ss e.g. 00:02:00 for two minutes.
    batchMinRows Number
    The minimum number of rows per batch (must be between 0 and 10000).
    dateFormat String
    The date format. Wherever {date} appears in path_pattern, the value of this property is used as the date format instead.
    name String
    The name of the Stream Output. Changing this forces a new resource to be created.
    pathPattern String
    The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
    resourceGroupName String
    The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
    serialization Property Map
    A serialization block as defined below.
    storageAccountKey String
    The Access Key which should be used to connect to this Storage Account.
    storageAccountName String
    The name of the Storage Account.
    storageContainerName String
    The name of the Container within the Storage Account.
    streamAnalyticsJobName String
    The name of the Stream Analytics Job. Changing this forces a new resource to be created.
    timeFormat String
    The time format. Wherever {time} appears in path_pattern, the value of this property is used as the time format instead.

    Supporting Types

    OutputBlobSerialization, OutputBlobSerializationArgs

    Type string
    The serialization format used for outgoing data streams. Possible values are Avro, Csv, Json and Parquet.
    Encoding string
    The encoding of the incoming data in the case of input and the encoding of outgoing data in the case of output. It currently can only be set to UTF8.
    FieldDelimiter string
    The delimiter that will be used to separate comma-separated value (CSV) records. Possible values are (space), , (comma), (tab), | (pipe) and ;.
    Format string
    Specifies the format of the JSON the output will be written in. Possible values are Array and LineSeparated.
    Type string
    The serialization format used for outgoing data streams. Possible values are Avro, Csv, Json and Parquet.
    Encoding string
    The encoding of the incoming data in the case of input and the encoding of outgoing data in the case of output. It currently can only be set to UTF8.
    FieldDelimiter string
    The delimiter that will be used to separate comma-separated value (CSV) records. Possible values are (space), , (comma), (tab), | (pipe) and ;.
    Format string
    Specifies the format of the JSON the output will be written in. Possible values are Array and LineSeparated.
    type String
    The serialization format used for outgoing data streams. Possible values are Avro, Csv, Json and Parquet.
    encoding String
    The encoding of the incoming data in the case of input and the encoding of outgoing data in the case of output. It currently can only be set to UTF8.
    fieldDelimiter String
    The delimiter that will be used to separate comma-separated value (CSV) records. Possible values are (space), , (comma), (tab), | (pipe) and ;.
    format String
    Specifies the format of the JSON the output will be written in. Possible values are Array and LineSeparated.
    type string
    The serialization format used for outgoing data streams. Possible values are Avro, Csv, Json and Parquet.
    encoding string
    The encoding of the incoming data in the case of input and the encoding of outgoing data in the case of output. It currently can only be set to UTF8.
    fieldDelimiter string
    The delimiter that will be used to separate comma-separated value (CSV) records. Possible values are (space), , (comma), (tab), | (pipe) and ;.
    format string
    Specifies the format of the JSON the output will be written in. Possible values are Array and LineSeparated.
    type str
    The serialization format used for outgoing data streams. Possible values are Avro, Csv, Json and Parquet.
    encoding str
    The encoding of the incoming data in the case of input and the encoding of outgoing data in the case of output. It currently can only be set to UTF8.
    field_delimiter str
    The delimiter that will be used to separate comma-separated value (CSV) records. Possible values are (space), , (comma), (tab), | (pipe) and ;.
    format str
    Specifies the format of the JSON the output will be written in. Possible values are Array and LineSeparated.
    type String
    The serialization format used for outgoing data streams. Possible values are Avro, Csv, Json and Parquet.
    encoding String
    The encoding of the incoming data in the case of input and the encoding of outgoing data in the case of output. It currently can only be set to UTF8.
    fieldDelimiter String
    The delimiter that will be used to separate comma-separated value (CSV) records. Possible values are (space), , (comma), (tab), | (pipe) and ;.
    format String
    Specifies the format of the JSON the output will be written in. Possible values are Array and LineSeparated.

    Import

    Stream Analytics Outputs to Blob Storage can be imported using the resource id, e.g.

     $ pulumi import azure:streamanalytics/outputBlob:OutputBlob example /subscriptions/00000000-0000-0000-0000-000000000000/resourcegroups/group1/providers/Microsoft.StreamAnalytics/streamingjobs/job1/outputs/output1
    

    To learn more about importing existing cloud resources, see Importing resources.

    Package Details

    Repository
    Azure Classic pulumi/pulumi-azure
    License
    Apache-2.0
    Notes
    This Pulumi package is based on the azurerm Terraform Provider.
    azure logo

    We recommend using Azure Native.

    Viewing docs for Azure v4.42.0 (Older version)
    published on Monday, Mar 9, 2026 by Pulumi
      Try Pulumi Cloud free. Your team will thank you.