1. Packages
  2. Azure Classic
  3. API Docs
  4. streamanalytics
  5. StreamInputBlob

We recommend using Azure Native.

Azure Classic v5.70.0 published on Wednesday, Mar 27, 2024 by Pulumi

azure.streamanalytics.StreamInputBlob

Explore with Pulumi AI

azure logo

We recommend using Azure Native.

Azure Classic v5.70.0 published on Wednesday, Mar 27, 2024 by Pulumi

    Manages a Stream Analytics Stream Input Blob.

    Example Usage

    import * as pulumi from "@pulumi/pulumi";
    import * as azure from "@pulumi/azure";
    
    const exampleResourceGroup = new azure.core.ResourceGroup("example", {
        name: "example-resources",
        location: "West Europe",
    });
    const example = azure.streamanalytics.getJobOutput({
        name: "example-job",
        resourceGroupName: exampleResourceGroup.name,
    });
    const exampleAccount = new azure.storage.Account("example", {
        name: "examplestoracc",
        resourceGroupName: exampleResourceGroup.name,
        location: exampleResourceGroup.location,
        accountTier: "Standard",
        accountReplicationType: "LRS",
    });
    const exampleContainer = new azure.storage.Container("example", {
        name: "example",
        storageAccountName: exampleAccount.name,
        containerAccessType: "private",
    });
    const exampleStreamInputBlob = new azure.streamanalytics.StreamInputBlob("example", {
        name: "blob-stream-input",
        streamAnalyticsJobName: example.apply(example => example.name),
        resourceGroupName: example.apply(example => example.resourceGroupName),
        storageAccountName: exampleAccount.name,
        storageAccountKey: exampleAccount.primaryAccessKey,
        storageContainerName: exampleContainer.name,
        pathPattern: "some-random-pattern",
        dateFormat: "yyyy/MM/dd",
        timeFormat: "HH",
        serialization: {
            type: "Json",
            encoding: "UTF8",
        },
    });
    
    import pulumi
    import pulumi_azure as azure
    
    example_resource_group = azure.core.ResourceGroup("example",
        name="example-resources",
        location="West Europe")
    example = azure.streamanalytics.get_job_output(name="example-job",
        resource_group_name=example_resource_group.name)
    example_account = azure.storage.Account("example",
        name="examplestoracc",
        resource_group_name=example_resource_group.name,
        location=example_resource_group.location,
        account_tier="Standard",
        account_replication_type="LRS")
    example_container = azure.storage.Container("example",
        name="example",
        storage_account_name=example_account.name,
        container_access_type="private")
    example_stream_input_blob = azure.streamanalytics.StreamInputBlob("example",
        name="blob-stream-input",
        stream_analytics_job_name=example.name,
        resource_group_name=example.resource_group_name,
        storage_account_name=example_account.name,
        storage_account_key=example_account.primary_access_key,
        storage_container_name=example_container.name,
        path_pattern="some-random-pattern",
        date_format="yyyy/MM/dd",
        time_format="HH",
        serialization=azure.streamanalytics.StreamInputBlobSerializationArgs(
            type="Json",
            encoding="UTF8",
        ))
    
    package main
    
    import (
    	"github.com/pulumi/pulumi-azure/sdk/v5/go/azure/core"
    	"github.com/pulumi/pulumi-azure/sdk/v5/go/azure/storage"
    	"github.com/pulumi/pulumi-azure/sdk/v5/go/azure/streamanalytics"
    	"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
    )
    
    func main() {
    	pulumi.Run(func(ctx *pulumi.Context) error {
    		exampleResourceGroup, err := core.NewResourceGroup(ctx, "example", &core.ResourceGroupArgs{
    			Name:     pulumi.String("example-resources"),
    			Location: pulumi.String("West Europe"),
    		})
    		if err != nil {
    			return err
    		}
    		example := streamanalytics.LookupJobOutput(ctx, streamanalytics.GetJobOutputArgs{
    			Name:              pulumi.String("example-job"),
    			ResourceGroupName: exampleResourceGroup.Name,
    		}, nil)
    		exampleAccount, err := storage.NewAccount(ctx, "example", &storage.AccountArgs{
    			Name:                   pulumi.String("examplestoracc"),
    			ResourceGroupName:      exampleResourceGroup.Name,
    			Location:               exampleResourceGroup.Location,
    			AccountTier:            pulumi.String("Standard"),
    			AccountReplicationType: pulumi.String("LRS"),
    		})
    		if err != nil {
    			return err
    		}
    		exampleContainer, err := storage.NewContainer(ctx, "example", &storage.ContainerArgs{
    			Name:                pulumi.String("example"),
    			StorageAccountName:  exampleAccount.Name,
    			ContainerAccessType: pulumi.String("private"),
    		})
    		if err != nil {
    			return err
    		}
    		_, err = streamanalytics.NewStreamInputBlob(ctx, "example", &streamanalytics.StreamInputBlobArgs{
    			Name: pulumi.String("blob-stream-input"),
    			StreamAnalyticsJobName: example.ApplyT(func(example streamanalytics.GetJobResult) (*string, error) {
    				return &example.Name, nil
    			}).(pulumi.StringPtrOutput),
    			ResourceGroupName: example.ApplyT(func(example streamanalytics.GetJobResult) (*string, error) {
    				return &example.ResourceGroupName, nil
    			}).(pulumi.StringPtrOutput),
    			StorageAccountName:   exampleAccount.Name,
    			StorageAccountKey:    exampleAccount.PrimaryAccessKey,
    			StorageContainerName: exampleContainer.Name,
    			PathPattern:          pulumi.String("some-random-pattern"),
    			DateFormat:           pulumi.String("yyyy/MM/dd"),
    			TimeFormat:           pulumi.String("HH"),
    			Serialization: &streamanalytics.StreamInputBlobSerializationArgs{
    				Type:     pulumi.String("Json"),
    				Encoding: pulumi.String("UTF8"),
    			},
    		})
    		if err != nil {
    			return err
    		}
    		return nil
    	})
    }
    
    using System.Collections.Generic;
    using System.Linq;
    using Pulumi;
    using Azure = Pulumi.Azure;
    
    return await Deployment.RunAsync(() => 
    {
        var exampleResourceGroup = new Azure.Core.ResourceGroup("example", new()
        {
            Name = "example-resources",
            Location = "West Europe",
        });
    
        var example = Azure.StreamAnalytics.GetJob.Invoke(new()
        {
            Name = "example-job",
            ResourceGroupName = exampleResourceGroup.Name,
        });
    
        var exampleAccount = new Azure.Storage.Account("example", new()
        {
            Name = "examplestoracc",
            ResourceGroupName = exampleResourceGroup.Name,
            Location = exampleResourceGroup.Location,
            AccountTier = "Standard",
            AccountReplicationType = "LRS",
        });
    
        var exampleContainer = new Azure.Storage.Container("example", new()
        {
            Name = "example",
            StorageAccountName = exampleAccount.Name,
            ContainerAccessType = "private",
        });
    
        var exampleStreamInputBlob = new Azure.StreamAnalytics.StreamInputBlob("example", new()
        {
            Name = "blob-stream-input",
            StreamAnalyticsJobName = example.Apply(getJobResult => getJobResult.Name),
            ResourceGroupName = example.Apply(getJobResult => getJobResult.ResourceGroupName),
            StorageAccountName = exampleAccount.Name,
            StorageAccountKey = exampleAccount.PrimaryAccessKey,
            StorageContainerName = exampleContainer.Name,
            PathPattern = "some-random-pattern",
            DateFormat = "yyyy/MM/dd",
            TimeFormat = "HH",
            Serialization = new Azure.StreamAnalytics.Inputs.StreamInputBlobSerializationArgs
            {
                Type = "Json",
                Encoding = "UTF8",
            },
        });
    
    });
    
    package generated_program;
    
    import com.pulumi.Context;
    import com.pulumi.Pulumi;
    import com.pulumi.core.Output;
    import com.pulumi.azure.core.ResourceGroup;
    import com.pulumi.azure.core.ResourceGroupArgs;
    import com.pulumi.azure.streamanalytics.StreamanalyticsFunctions;
    import com.pulumi.azure.streamanalytics.inputs.GetJobArgs;
    import com.pulumi.azure.storage.Account;
    import com.pulumi.azure.storage.AccountArgs;
    import com.pulumi.azure.storage.Container;
    import com.pulumi.azure.storage.ContainerArgs;
    import com.pulumi.azure.streamanalytics.StreamInputBlob;
    import com.pulumi.azure.streamanalytics.StreamInputBlobArgs;
    import com.pulumi.azure.streamanalytics.inputs.StreamInputBlobSerializationArgs;
    import java.util.List;
    import java.util.ArrayList;
    import java.util.Map;
    import java.io.File;
    import java.nio.file.Files;
    import java.nio.file.Paths;
    
    public class App {
        public static void main(String[] args) {
            Pulumi.run(App::stack);
        }
    
        public static void stack(Context ctx) {
            var exampleResourceGroup = new ResourceGroup("exampleResourceGroup", ResourceGroupArgs.builder()        
                .name("example-resources")
                .location("West Europe")
                .build());
    
            final var example = StreamanalyticsFunctions.getJob(GetJobArgs.builder()
                .name("example-job")
                .resourceGroupName(exampleResourceGroup.name())
                .build());
    
            var exampleAccount = new Account("exampleAccount", AccountArgs.builder()        
                .name("examplestoracc")
                .resourceGroupName(exampleResourceGroup.name())
                .location(exampleResourceGroup.location())
                .accountTier("Standard")
                .accountReplicationType("LRS")
                .build());
    
            var exampleContainer = new Container("exampleContainer", ContainerArgs.builder()        
                .name("example")
                .storageAccountName(exampleAccount.name())
                .containerAccessType("private")
                .build());
    
            var exampleStreamInputBlob = new StreamInputBlob("exampleStreamInputBlob", StreamInputBlobArgs.builder()        
                .name("blob-stream-input")
                .streamAnalyticsJobName(example.applyValue(getJobResult -> getJobResult).applyValue(example -> example.applyValue(getJobResult -> getJobResult.name())))
                .resourceGroupName(example.applyValue(getJobResult -> getJobResult).applyValue(example -> example.applyValue(getJobResult -> getJobResult.resourceGroupName())))
                .storageAccountName(exampleAccount.name())
                .storageAccountKey(exampleAccount.primaryAccessKey())
                .storageContainerName(exampleContainer.name())
                .pathPattern("some-random-pattern")
                .dateFormat("yyyy/MM/dd")
                .timeFormat("HH")
                .serialization(StreamInputBlobSerializationArgs.builder()
                    .type("Json")
                    .encoding("UTF8")
                    .build())
                .build());
    
        }
    }
    
    resources:
      exampleResourceGroup:
        type: azure:core:ResourceGroup
        name: example
        properties:
          name: example-resources
          location: West Europe
      exampleAccount:
        type: azure:storage:Account
        name: example
        properties:
          name: examplestoracc
          resourceGroupName: ${exampleResourceGroup.name}
          location: ${exampleResourceGroup.location}
          accountTier: Standard
          accountReplicationType: LRS
      exampleContainer:
        type: azure:storage:Container
        name: example
        properties:
          name: example
          storageAccountName: ${exampleAccount.name}
          containerAccessType: private
      exampleStreamInputBlob:
        type: azure:streamanalytics:StreamInputBlob
        name: example
        properties:
          name: blob-stream-input
          streamAnalyticsJobName: ${example.name}
          resourceGroupName: ${example.resourceGroupName}
          storageAccountName: ${exampleAccount.name}
          storageAccountKey: ${exampleAccount.primaryAccessKey}
          storageContainerName: ${exampleContainer.name}
          pathPattern: some-random-pattern
          dateFormat: yyyy/MM/dd
          timeFormat: HH
          serialization:
            type: Json
            encoding: UTF8
    variables:
      example:
        fn::invoke:
          Function: azure:streamanalytics:getJob
          Arguments:
            name: example-job
            resourceGroupName: ${exampleResourceGroup.name}
    

    Create StreamInputBlob Resource

    new StreamInputBlob(name: string, args: StreamInputBlobArgs, opts?: CustomResourceOptions);
    @overload
    def StreamInputBlob(resource_name: str,
                        opts: Optional[ResourceOptions] = None,
                        date_format: Optional[str] = None,
                        name: Optional[str] = None,
                        path_pattern: Optional[str] = None,
                        resource_group_name: Optional[str] = None,
                        serialization: Optional[StreamInputBlobSerializationArgs] = None,
                        storage_account_key: Optional[str] = None,
                        storage_account_name: Optional[str] = None,
                        storage_container_name: Optional[str] = None,
                        stream_analytics_job_name: Optional[str] = None,
                        time_format: Optional[str] = None)
    @overload
    def StreamInputBlob(resource_name: str,
                        args: StreamInputBlobArgs,
                        opts: Optional[ResourceOptions] = None)
    func NewStreamInputBlob(ctx *Context, name string, args StreamInputBlobArgs, opts ...ResourceOption) (*StreamInputBlob, error)
    public StreamInputBlob(string name, StreamInputBlobArgs args, CustomResourceOptions? opts = null)
    public StreamInputBlob(String name, StreamInputBlobArgs args)
    public StreamInputBlob(String name, StreamInputBlobArgs args, CustomResourceOptions options)
    
    type: azure:streamanalytics:StreamInputBlob
    properties: # The arguments to resource properties.
    options: # Bag of options to control resource's behavior.
    
    
    name string
    The unique name of the resource.
    args StreamInputBlobArgs
    The arguments to resource properties.
    opts CustomResourceOptions
    Bag of options to control resource's behavior.
    resource_name str
    The unique name of the resource.
    args StreamInputBlobArgs
    The arguments to resource properties.
    opts ResourceOptions
    Bag of options to control resource's behavior.
    ctx Context
    Context object for the current deployment.
    name string
    The unique name of the resource.
    args StreamInputBlobArgs
    The arguments to resource properties.
    opts ResourceOption
    Bag of options to control resource's behavior.
    name string
    The unique name of the resource.
    args StreamInputBlobArgs
    The arguments to resource properties.
    opts CustomResourceOptions
    Bag of options to control resource's behavior.
    name String
    The unique name of the resource.
    args StreamInputBlobArgs
    The arguments to resource properties.
    options CustomResourceOptions
    Bag of options to control resource's behavior.

    StreamInputBlob Resource Properties

    To learn more about resource properties and how to use them, see Inputs and Outputs in the Architecture and Concepts docs.

    Inputs

    The StreamInputBlob resource accepts the following input properties:

    DateFormat string
    The date format. Wherever {date} appears in path_pattern, the value of this property is used as the date format instead.
    PathPattern string
    The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
    ResourceGroupName string
    The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
    Serialization StreamInputBlobSerialization
    A serialization block as defined below.
    StorageAccountKey string
    The Access Key which should be used to connect to this Storage Account.
    StorageAccountName string
    The name of the Storage Account.
    StorageContainerName string
    The name of the Container within the Storage Account.
    StreamAnalyticsJobName string
    The name of the Stream Analytics Job. Changing this forces a new resource to be created.
    TimeFormat string
    The time format. Wherever {time} appears in path_pattern, the value of this property is used as the time format instead.
    Name string
    The name of the Stream Input Blob. Changing this forces a new resource to be created.
    DateFormat string
    The date format. Wherever {date} appears in path_pattern, the value of this property is used as the date format instead.
    PathPattern string
    The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
    ResourceGroupName string
    The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
    Serialization StreamInputBlobSerializationArgs
    A serialization block as defined below.
    StorageAccountKey string
    The Access Key which should be used to connect to this Storage Account.
    StorageAccountName string
    The name of the Storage Account.
    StorageContainerName string
    The name of the Container within the Storage Account.
    StreamAnalyticsJobName string
    The name of the Stream Analytics Job. Changing this forces a new resource to be created.
    TimeFormat string
    The time format. Wherever {time} appears in path_pattern, the value of this property is used as the time format instead.
    Name string
    The name of the Stream Input Blob. Changing this forces a new resource to be created.
    dateFormat String
    The date format. Wherever {date} appears in path_pattern, the value of this property is used as the date format instead.
    pathPattern String
    The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
    resourceGroupName String
    The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
    serialization StreamInputBlobSerialization
    A serialization block as defined below.
    storageAccountKey String
    The Access Key which should be used to connect to this Storage Account.
    storageAccountName String
    The name of the Storage Account.
    storageContainerName String
    The name of the Container within the Storage Account.
    streamAnalyticsJobName String
    The name of the Stream Analytics Job. Changing this forces a new resource to be created.
    timeFormat String
    The time format. Wherever {time} appears in path_pattern, the value of this property is used as the time format instead.
    name String
    The name of the Stream Input Blob. Changing this forces a new resource to be created.
    dateFormat string
    The date format. Wherever {date} appears in path_pattern, the value of this property is used as the date format instead.
    pathPattern string
    The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
    resourceGroupName string
    The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
    serialization StreamInputBlobSerialization
    A serialization block as defined below.
    storageAccountKey string
    The Access Key which should be used to connect to this Storage Account.
    storageAccountName string
    The name of the Storage Account.
    storageContainerName string
    The name of the Container within the Storage Account.
    streamAnalyticsJobName string
    The name of the Stream Analytics Job. Changing this forces a new resource to be created.
    timeFormat string
    The time format. Wherever {time} appears in path_pattern, the value of this property is used as the time format instead.
    name string
    The name of the Stream Input Blob. Changing this forces a new resource to be created.
    date_format str
    The date format. Wherever {date} appears in path_pattern, the value of this property is used as the date format instead.
    path_pattern str
    The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
    resource_group_name str
    The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
    serialization StreamInputBlobSerializationArgs
    A serialization block as defined below.
    storage_account_key str
    The Access Key which should be used to connect to this Storage Account.
    storage_account_name str
    The name of the Storage Account.
    storage_container_name str
    The name of the Container within the Storage Account.
    stream_analytics_job_name str
    The name of the Stream Analytics Job. Changing this forces a new resource to be created.
    time_format str
    The time format. Wherever {time} appears in path_pattern, the value of this property is used as the time format instead.
    name str
    The name of the Stream Input Blob. Changing this forces a new resource to be created.
    dateFormat String
    The date format. Wherever {date} appears in path_pattern, the value of this property is used as the date format instead.
    pathPattern String
    The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
    resourceGroupName String
    The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
    serialization Property Map
    A serialization block as defined below.
    storageAccountKey String
    The Access Key which should be used to connect to this Storage Account.
    storageAccountName String
    The name of the Storage Account.
    storageContainerName String
    The name of the Container within the Storage Account.
    streamAnalyticsJobName String
    The name of the Stream Analytics Job. Changing this forces a new resource to be created.
    timeFormat String
    The time format. Wherever {time} appears in path_pattern, the value of this property is used as the time format instead.
    name String
    The name of the Stream Input Blob. Changing this forces a new resource to be created.

    Outputs

    All input properties are implicitly available as output properties. Additionally, the StreamInputBlob resource produces the following output properties:

    Id string
    The provider-assigned unique ID for this managed resource.
    Id string
    The provider-assigned unique ID for this managed resource.
    id String
    The provider-assigned unique ID for this managed resource.
    id string
    The provider-assigned unique ID for this managed resource.
    id str
    The provider-assigned unique ID for this managed resource.
    id String
    The provider-assigned unique ID for this managed resource.

    Look up Existing StreamInputBlob Resource

    Get an existing StreamInputBlob resource’s state with the given name, ID, and optional extra properties used to qualify the lookup.

    public static get(name: string, id: Input<ID>, state?: StreamInputBlobState, opts?: CustomResourceOptions): StreamInputBlob
    @staticmethod
    def get(resource_name: str,
            id: str,
            opts: Optional[ResourceOptions] = None,
            date_format: Optional[str] = None,
            name: Optional[str] = None,
            path_pattern: Optional[str] = None,
            resource_group_name: Optional[str] = None,
            serialization: Optional[StreamInputBlobSerializationArgs] = None,
            storage_account_key: Optional[str] = None,
            storage_account_name: Optional[str] = None,
            storage_container_name: Optional[str] = None,
            stream_analytics_job_name: Optional[str] = None,
            time_format: Optional[str] = None) -> StreamInputBlob
    func GetStreamInputBlob(ctx *Context, name string, id IDInput, state *StreamInputBlobState, opts ...ResourceOption) (*StreamInputBlob, error)
    public static StreamInputBlob Get(string name, Input<string> id, StreamInputBlobState? state, CustomResourceOptions? opts = null)
    public static StreamInputBlob get(String name, Output<String> id, StreamInputBlobState state, CustomResourceOptions options)
    Resource lookup is not supported in YAML
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    resource_name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    The following state arguments are supported:
    DateFormat string
    The date format. Wherever {date} appears in path_pattern, the value of this property is used as the date format instead.
    Name string
    The name of the Stream Input Blob. Changing this forces a new resource to be created.
    PathPattern string
    The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
    ResourceGroupName string
    The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
    Serialization StreamInputBlobSerialization
    A serialization block as defined below.
    StorageAccountKey string
    The Access Key which should be used to connect to this Storage Account.
    StorageAccountName string
    The name of the Storage Account.
    StorageContainerName string
    The name of the Container within the Storage Account.
    StreamAnalyticsJobName string
    The name of the Stream Analytics Job. Changing this forces a new resource to be created.
    TimeFormat string
    The time format. Wherever {time} appears in path_pattern, the value of this property is used as the time format instead.
    DateFormat string
    The date format. Wherever {date} appears in path_pattern, the value of this property is used as the date format instead.
    Name string
    The name of the Stream Input Blob. Changing this forces a new resource to be created.
    PathPattern string
    The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
    ResourceGroupName string
    The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
    Serialization StreamInputBlobSerializationArgs
    A serialization block as defined below.
    StorageAccountKey string
    The Access Key which should be used to connect to this Storage Account.
    StorageAccountName string
    The name of the Storage Account.
    StorageContainerName string
    The name of the Container within the Storage Account.
    StreamAnalyticsJobName string
    The name of the Stream Analytics Job. Changing this forces a new resource to be created.
    TimeFormat string
    The time format. Wherever {time} appears in path_pattern, the value of this property is used as the time format instead.
    dateFormat String
    The date format. Wherever {date} appears in path_pattern, the value of this property is used as the date format instead.
    name String
    The name of the Stream Input Blob. Changing this forces a new resource to be created.
    pathPattern String
    The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
    resourceGroupName String
    The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
    serialization StreamInputBlobSerialization
    A serialization block as defined below.
    storageAccountKey String
    The Access Key which should be used to connect to this Storage Account.
    storageAccountName String
    The name of the Storage Account.
    storageContainerName String
    The name of the Container within the Storage Account.
    streamAnalyticsJobName String
    The name of the Stream Analytics Job. Changing this forces a new resource to be created.
    timeFormat String
    The time format. Wherever {time} appears in path_pattern, the value of this property is used as the time format instead.
    dateFormat string
    The date format. Wherever {date} appears in path_pattern, the value of this property is used as the date format instead.
    name string
    The name of the Stream Input Blob. Changing this forces a new resource to be created.
    pathPattern string
    The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
    resourceGroupName string
    The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
    serialization StreamInputBlobSerialization
    A serialization block as defined below.
    storageAccountKey string
    The Access Key which should be used to connect to this Storage Account.
    storageAccountName string
    The name of the Storage Account.
    storageContainerName string
    The name of the Container within the Storage Account.
    streamAnalyticsJobName string
    The name of the Stream Analytics Job. Changing this forces a new resource to be created.
    timeFormat string
    The time format. Wherever {time} appears in path_pattern, the value of this property is used as the time format instead.
    date_format str
    The date format. Wherever {date} appears in path_pattern, the value of this property is used as the date format instead.
    name str
    The name of the Stream Input Blob. Changing this forces a new resource to be created.
    path_pattern str
    The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
    resource_group_name str
    The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
    serialization StreamInputBlobSerializationArgs
    A serialization block as defined below.
    storage_account_key str
    The Access Key which should be used to connect to this Storage Account.
    storage_account_name str
    The name of the Storage Account.
    storage_container_name str
    The name of the Container within the Storage Account.
    stream_analytics_job_name str
    The name of the Stream Analytics Job. Changing this forces a new resource to be created.
    time_format str
    The time format. Wherever {time} appears in path_pattern, the value of this property is used as the time format instead.
    dateFormat String
    The date format. Wherever {date} appears in path_pattern, the value of this property is used as the date format instead.
    name String
    The name of the Stream Input Blob. Changing this forces a new resource to be created.
    pathPattern String
    The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
    resourceGroupName String
    The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
    serialization Property Map
    A serialization block as defined below.
    storageAccountKey String
    The Access Key which should be used to connect to this Storage Account.
    storageAccountName String
    The name of the Storage Account.
    storageContainerName String
    The name of the Container within the Storage Account.
    streamAnalyticsJobName String
    The name of the Stream Analytics Job. Changing this forces a new resource to be created.
    timeFormat String
    The time format. Wherever {time} appears in path_pattern, the value of this property is used as the time format instead.

    Supporting Types

    StreamInputBlobSerialization, StreamInputBlobSerializationArgs

    Type string
    The serialization format used for incoming data streams. Possible values are Avro, Csv and Json.
    Encoding string

    The encoding of the incoming data in the case of input and the encoding of outgoing data in the case of output. It currently can only be set to UTF8.

    NOTE: This is required when type is set to Csv or Json.

    FieldDelimiter string

    The delimiter that will be used to separate comma-separated value (CSV) records. Possible values are (space), , (comma), (tab), | (pipe) and ;.

    NOTE: This is required when type is set to Csv.

    Type string
    The serialization format used for incoming data streams. Possible values are Avro, Csv and Json.
    Encoding string

    The encoding of the incoming data in the case of input and the encoding of outgoing data in the case of output. It currently can only be set to UTF8.

    NOTE: This is required when type is set to Csv or Json.

    FieldDelimiter string

    The delimiter that will be used to separate comma-separated value (CSV) records. Possible values are (space), , (comma), (tab), | (pipe) and ;.

    NOTE: This is required when type is set to Csv.

    type String
    The serialization format used for incoming data streams. Possible values are Avro, Csv and Json.
    encoding String

    The encoding of the incoming data in the case of input and the encoding of outgoing data in the case of output. It currently can only be set to UTF8.

    NOTE: This is required when type is set to Csv or Json.

    fieldDelimiter String

    The delimiter that will be used to separate comma-separated value (CSV) records. Possible values are (space), , (comma), (tab), | (pipe) and ;.

    NOTE: This is required when type is set to Csv.

    type string
    The serialization format used for incoming data streams. Possible values are Avro, Csv and Json.
    encoding string

    The encoding of the incoming data in the case of input and the encoding of outgoing data in the case of output. It currently can only be set to UTF8.

    NOTE: This is required when type is set to Csv or Json.

    fieldDelimiter string

    The delimiter that will be used to separate comma-separated value (CSV) records. Possible values are (space), , (comma), (tab), | (pipe) and ;.

    NOTE: This is required when type is set to Csv.

    type str
    The serialization format used for incoming data streams. Possible values are Avro, Csv and Json.
    encoding str

    The encoding of the incoming data in the case of input and the encoding of outgoing data in the case of output. It currently can only be set to UTF8.

    NOTE: This is required when type is set to Csv or Json.

    field_delimiter str

    The delimiter that will be used to separate comma-separated value (CSV) records. Possible values are (space), , (comma), (tab), | (pipe) and ;.

    NOTE: This is required when type is set to Csv.

    type String
    The serialization format used for incoming data streams. Possible values are Avro, Csv and Json.
    encoding String

    The encoding of the incoming data in the case of input and the encoding of outgoing data in the case of output. It currently can only be set to UTF8.

    NOTE: This is required when type is set to Csv or Json.

    fieldDelimiter String

    The delimiter that will be used to separate comma-separated value (CSV) records. Possible values are (space), , (comma), (tab), | (pipe) and ;.

    NOTE: This is required when type is set to Csv.

    Import

    Stream Analytics Stream Input Blob’s can be imported using the resource id, e.g.

    $ pulumi import azure:streamanalytics/streamInputBlob:StreamInputBlob example /subscriptions/00000000-0000-0000-0000-000000000000/resourceGroups/group1/providers/Microsoft.StreamAnalytics/streamingJobs/job1/inputs/input1
    

    Package Details

    Repository
    Azure Classic pulumi/pulumi-azure
    License
    Apache-2.0
    Notes
    This Pulumi package is based on the azurerm Terraform Provider.
    azure logo

    We recommend using Azure Native.

    Azure Classic v5.70.0 published on Wednesday, Mar 27, 2024 by Pulumi