azure-native.media.Transform

A Transform encapsulates the rules or instructions for generating desired outputs from input media, such as by transcoding or by extracting insights. After the Transform is created, it can be applied to input media by creating Jobs. API Version: 2020-05-01.

Example Usage

Create or update a Transform

using System.Collections.Generic;
using Pulumi;
using AzureNative = Pulumi.AzureNative;

return await Deployment.RunAsync(() => 
{
    var transform = new AzureNative.Media.Transform("transform", new()
    {
        AccountName = "contosomedia",
        Description = "Example Transform to illustrate create and update.",
        Outputs = new[]
        {
            new AzureNative.Media.Inputs.TransformOutputArgs
            {
                Preset = new AzureNative.Media.Inputs.BuiltInStandardEncoderPresetArgs
                {
                    OdataType = "#Microsoft.Media.BuiltInStandardEncoderPreset",
                    PresetName = "AdaptiveStreaming",
                },
            },
        },
        ResourceGroupName = "contosoresources",
        TransformName = "createdTransform",
    });

});
package main

import (
	media "github.com/pulumi/pulumi-azure-native/sdk/go/azure/media"
	"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
)

func main() {
	pulumi.Run(func(ctx *pulumi.Context) error {
		_, err := media.NewTransform(ctx, "transform", &media.TransformArgs{
			AccountName: pulumi.String("contosomedia"),
			Description: pulumi.String("Example Transform to illustrate create and update."),
			Outputs: []media.TransformOutputTypeArgs{
				{
					Preset: {
						OdataType:  "#Microsoft.Media.BuiltInStandardEncoderPreset",
						PresetName: "AdaptiveStreaming",
					},
				},
			},
			ResourceGroupName: pulumi.String("contosoresources"),
			TransformName:     pulumi.String("createdTransform"),
		})
		if err != nil {
			return err
		}
		return nil
	})
}
package generated_program;

import com.pulumi.Context;
import com.pulumi.Pulumi;
import com.pulumi.core.Output;
import com.pulumi.azurenative.media.Transform;
import com.pulumi.azurenative.media.TransformArgs;
import java.util.List;
import java.util.ArrayList;
import java.util.Map;
import java.io.File;
import java.nio.file.Files;
import java.nio.file.Paths;

public class App {
    public static void main(String[] args) {
        Pulumi.run(App::stack);
    }

    public static void stack(Context ctx) {
        var transform = new Transform("transform", TransformArgs.builder()        
            .accountName("contosomedia")
            .description("Example Transform to illustrate create and update.")
            .outputs(Map.of("preset", Map.ofEntries(
                Map.entry("odataType", "#Microsoft.Media.BuiltInStandardEncoderPreset"),
                Map.entry("presetName", "AdaptiveStreaming")
            )))
            .resourceGroupName("contosoresources")
            .transformName("createdTransform")
            .build());

    }
}
import pulumi
import pulumi_azure_native as azure_native

transform = azure_native.media.Transform("transform",
    account_name="contosomedia",
    description="Example Transform to illustrate create and update.",
    outputs=[azure_native.media.TransformOutputArgs(
        preset=azure_native.media.BuiltInStandardEncoderPresetArgs(
            odata_type="#Microsoft.Media.BuiltInStandardEncoderPreset",
            preset_name="AdaptiveStreaming",
        ),
    )],
    resource_group_name="contosoresources",
    transform_name="createdTransform")
import * as pulumi from "@pulumi/pulumi";
import * as azure_native from "@pulumi/azure-native";

const transform = new azure_native.media.Transform("transform", {
    accountName: "contosomedia",
    description: "Example Transform to illustrate create and update.",
    outputs: [{
        preset: {
            odataType: "#Microsoft.Media.BuiltInStandardEncoderPreset",
            presetName: "AdaptiveStreaming",
        },
    }],
    resourceGroupName: "contosoresources",
    transformName: "createdTransform",
});
resources:
  transform:
    type: azure-native:media:Transform
    properties:
      accountName: contosomedia
      description: Example Transform to illustrate create and update.
      outputs:
        - preset:
            odataType: '#Microsoft.Media.BuiltInStandardEncoderPreset'
            presetName: AdaptiveStreaming
      resourceGroupName: contosoresources
      transformName: createdTransform

Create Transform Resource

new Transform(name: string, args: TransformArgs, opts?: CustomResourceOptions);
@overload
def Transform(resource_name: str,
              opts: Optional[ResourceOptions] = None,
              account_name: Optional[str] = None,
              description: Optional[str] = None,
              outputs: Optional[Sequence[TransformOutputArgs]] = None,
              resource_group_name: Optional[str] = None,
              transform_name: Optional[str] = None)
@overload
def Transform(resource_name: str,
              args: TransformArgs,
              opts: Optional[ResourceOptions] = None)
func NewTransform(ctx *Context, name string, args TransformArgs, opts ...ResourceOption) (*Transform, error)
public Transform(string name, TransformArgs args, CustomResourceOptions? opts = null)
public Transform(String name, TransformArgs args)
public Transform(String name, TransformArgs args, CustomResourceOptions options)
type: azure-native:media:Transform
properties: # The arguments to resource properties.
options: # Bag of options to control resource's behavior.

name string
The unique name of the resource.
args TransformArgs
The arguments to resource properties.
opts CustomResourceOptions
Bag of options to control resource's behavior.
resource_name str
The unique name of the resource.
args TransformArgs
The arguments to resource properties.
opts ResourceOptions
Bag of options to control resource's behavior.
ctx Context
Context object for the current deployment.
name string
The unique name of the resource.
args TransformArgs
The arguments to resource properties.
opts ResourceOption
Bag of options to control resource's behavior.
name string
The unique name of the resource.
args TransformArgs
The arguments to resource properties.
opts CustomResourceOptions
Bag of options to control resource's behavior.
name String
The unique name of the resource.
args TransformArgs
The arguments to resource properties.
options CustomResourceOptions
Bag of options to control resource's behavior.

Transform Resource Properties

To learn more about resource properties and how to use them, see Inputs and Outputs in the Architecture and Concepts docs.

Inputs

The Transform resource accepts the following input properties:

AccountName string

The Media Services account name.

Outputs List<Pulumi.AzureNative.Media.Inputs.TransformOutputArgs>

An array of one or more TransformOutputs that the Transform should generate.

ResourceGroupName string

The name of the resource group within the Azure subscription.

Description string

An optional verbose description of the Transform.

TransformName string

The Transform name.

AccountName string

The Media Services account name.

Outputs []TransformOutputTypeArgs

An array of one or more TransformOutputs that the Transform should generate.

ResourceGroupName string

The name of the resource group within the Azure subscription.

Description string

An optional verbose description of the Transform.

TransformName string

The Transform name.

accountName String

The Media Services account name.

outputs List<TransformOutputArgs>

An array of one or more TransformOutputs that the Transform should generate.

resourceGroupName String

The name of the resource group within the Azure subscription.

description String

An optional verbose description of the Transform.

transformName String

The Transform name.

accountName string

The Media Services account name.

outputs TransformOutputArgs[]

An array of one or more TransformOutputs that the Transform should generate.

resourceGroupName string

The name of the resource group within the Azure subscription.

description string

An optional verbose description of the Transform.

transformName string

The Transform name.

account_name str

The Media Services account name.

outputs Sequence[TransformOutputArgs]

An array of one or more TransformOutputs that the Transform should generate.

resource_group_name str

The name of the resource group within the Azure subscription.

description str

An optional verbose description of the Transform.

transform_name str

The Transform name.

accountName String

The Media Services account name.

outputs List<Property Map>

An array of one or more TransformOutputs that the Transform should generate.

resourceGroupName String

The name of the resource group within the Azure subscription.

description String

An optional verbose description of the Transform.

transformName String

The Transform name.

Outputs

All input properties are implicitly available as output properties. Additionally, the Transform resource produces the following output properties:

Created string

The UTC date and time when the Transform was created, in 'YYYY-MM-DDThh:mm:ssZ' format.

Id string

The provider-assigned unique ID for this managed resource.

LastModified string

The UTC date and time when the Transform was last updated, in 'YYYY-MM-DDThh:mm:ssZ' format.

Name string

The name of the resource

SystemData Pulumi.AzureNative.Media.Outputs.SystemDataResponse

The system metadata relating to this resource.

Type string

The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or "Microsoft.Storage/storageAccounts"

Created string

The UTC date and time when the Transform was created, in 'YYYY-MM-DDThh:mm:ssZ' format.

Id string

The provider-assigned unique ID for this managed resource.

LastModified string

The UTC date and time when the Transform was last updated, in 'YYYY-MM-DDThh:mm:ssZ' format.

Name string

The name of the resource

SystemData SystemDataResponse

The system metadata relating to this resource.

Type string

The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or "Microsoft.Storage/storageAccounts"

created String

The UTC date and time when the Transform was created, in 'YYYY-MM-DDThh:mm:ssZ' format.

id String

The provider-assigned unique ID for this managed resource.

lastModified String

The UTC date and time when the Transform was last updated, in 'YYYY-MM-DDThh:mm:ssZ' format.

name String

The name of the resource

systemData SystemDataResponse

The system metadata relating to this resource.

type String

The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or "Microsoft.Storage/storageAccounts"

created string

The UTC date and time when the Transform was created, in 'YYYY-MM-DDThh:mm:ssZ' format.

id string

The provider-assigned unique ID for this managed resource.

lastModified string

The UTC date and time when the Transform was last updated, in 'YYYY-MM-DDThh:mm:ssZ' format.

name string

The name of the resource

systemData SystemDataResponse

The system metadata relating to this resource.

type string

The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or "Microsoft.Storage/storageAccounts"

created str

The UTC date and time when the Transform was created, in 'YYYY-MM-DDThh:mm:ssZ' format.

id str

The provider-assigned unique ID for this managed resource.

last_modified str

The UTC date and time when the Transform was last updated, in 'YYYY-MM-DDThh:mm:ssZ' format.

name str

The name of the resource

system_data SystemDataResponse

The system metadata relating to this resource.

type str

The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or "Microsoft.Storage/storageAccounts"

created String

The UTC date and time when the Transform was created, in 'YYYY-MM-DDThh:mm:ssZ' format.

id String

The provider-assigned unique ID for this managed resource.

lastModified String

The UTC date and time when the Transform was last updated, in 'YYYY-MM-DDThh:mm:ssZ' format.

name String

The name of the resource

systemData Property Map

The system metadata relating to this resource.

type String

The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or "Microsoft.Storage/storageAccounts"

Supporting Types

AacAudio

Bitrate int

The bitrate, in bits per second, of the output encoded audio.

Channels int

The number of channels in the audio.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

Profile string | Pulumi.AzureNative.Media.AacAudioProfile

The encoding profile to be used when encoding audio with AAC.

SamplingRate int

The sampling rate to use for encoding in hertz.

Bitrate int

The bitrate, in bits per second, of the output encoded audio.

Channels int

The number of channels in the audio.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

Profile string | AacAudioProfile

The encoding profile to be used when encoding audio with AAC.

SamplingRate int

The sampling rate to use for encoding in hertz.

bitrate Integer

The bitrate, in bits per second, of the output encoded audio.

channels Integer

The number of channels in the audio.

label String

An optional label for the codec. The label can be used to control muxing behavior.

profile String | AacAudioProfile

The encoding profile to be used when encoding audio with AAC.

samplingRate Integer

The sampling rate to use for encoding in hertz.

bitrate number

The bitrate, in bits per second, of the output encoded audio.

channels number

The number of channels in the audio.

label string

An optional label for the codec. The label can be used to control muxing behavior.

profile string | AacAudioProfile

The encoding profile to be used when encoding audio with AAC.

samplingRate number

The sampling rate to use for encoding in hertz.

bitrate int

The bitrate, in bits per second, of the output encoded audio.

channels int

The number of channels in the audio.

label str

An optional label for the codec. The label can be used to control muxing behavior.

profile str | AacAudioProfile

The encoding profile to be used when encoding audio with AAC.

sampling_rate int

The sampling rate to use for encoding in hertz.

bitrate Number

The bitrate, in bits per second, of the output encoded audio.

channels Number

The number of channels in the audio.

label String

An optional label for the codec. The label can be used to control muxing behavior.

profile String | "AacLc" | "HeAacV1" | "HeAacV2"

The encoding profile to be used when encoding audio with AAC.

samplingRate Number

The sampling rate to use for encoding in hertz.

AacAudioProfile

AacLc
AacLc

Specifies that the output audio is to be encoded into AAC Low Complexity profile (AAC-LC).

HeAacV1
HeAacV1

Specifies that the output audio is to be encoded into HE-AAC v1 profile.

HeAacV2
HeAacV2

Specifies that the output audio is to be encoded into HE-AAC v2 profile.

AacAudioProfileAacLc
AacLc

Specifies that the output audio is to be encoded into AAC Low Complexity profile (AAC-LC).

AacAudioProfileHeAacV1
HeAacV1

Specifies that the output audio is to be encoded into HE-AAC v1 profile.

AacAudioProfileHeAacV2
HeAacV2

Specifies that the output audio is to be encoded into HE-AAC v2 profile.

AacLc
AacLc

Specifies that the output audio is to be encoded into AAC Low Complexity profile (AAC-LC).

HeAacV1
HeAacV1

Specifies that the output audio is to be encoded into HE-AAC v1 profile.

HeAacV2
HeAacV2

Specifies that the output audio is to be encoded into HE-AAC v2 profile.

AacLc
AacLc

Specifies that the output audio is to be encoded into AAC Low Complexity profile (AAC-LC).

HeAacV1
HeAacV1

Specifies that the output audio is to be encoded into HE-AAC v1 profile.

HeAacV2
HeAacV2

Specifies that the output audio is to be encoded into HE-AAC v2 profile.

AAC_LC
AacLc

Specifies that the output audio is to be encoded into AAC Low Complexity profile (AAC-LC).

HE_AAC_V1
HeAacV1

Specifies that the output audio is to be encoded into HE-AAC v1 profile.

HE_AAC_V2
HeAacV2

Specifies that the output audio is to be encoded into HE-AAC v2 profile.

"AacLc"
AacLc

Specifies that the output audio is to be encoded into AAC Low Complexity profile (AAC-LC).

"HeAacV1"
HeAacV1

Specifies that the output audio is to be encoded into HE-AAC v1 profile.

"HeAacV2"
HeAacV2

Specifies that the output audio is to be encoded into HE-AAC v2 profile.

AacAudioResponse

Bitrate int

The bitrate, in bits per second, of the output encoded audio.

Channels int

The number of channels in the audio.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

Profile string

The encoding profile to be used when encoding audio with AAC.

SamplingRate int

The sampling rate to use for encoding in hertz.

Bitrate int

The bitrate, in bits per second, of the output encoded audio.

Channels int

The number of channels in the audio.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

Profile string

The encoding profile to be used when encoding audio with AAC.

SamplingRate int

The sampling rate to use for encoding in hertz.

bitrate Integer

The bitrate, in bits per second, of the output encoded audio.

channels Integer

The number of channels in the audio.

label String

An optional label for the codec. The label can be used to control muxing behavior.

profile String

The encoding profile to be used when encoding audio with AAC.

samplingRate Integer

The sampling rate to use for encoding in hertz.

bitrate number

The bitrate, in bits per second, of the output encoded audio.

channels number

The number of channels in the audio.

label string

An optional label for the codec. The label can be used to control muxing behavior.

profile string

The encoding profile to be used when encoding audio with AAC.

samplingRate number

The sampling rate to use for encoding in hertz.

bitrate int

The bitrate, in bits per second, of the output encoded audio.

channels int

The number of channels in the audio.

label str

An optional label for the codec. The label can be used to control muxing behavior.

profile str

The encoding profile to be used when encoding audio with AAC.

sampling_rate int

The sampling rate to use for encoding in hertz.

bitrate Number

The bitrate, in bits per second, of the output encoded audio.

channels Number

The number of channels in the audio.

label String

An optional label for the codec. The label can be used to control muxing behavior.

profile String

The encoding profile to be used when encoding audio with AAC.

samplingRate Number

The sampling rate to use for encoding in hertz.

AnalysisResolution

SourceResolution
SourceResolution
StandardDefinition
StandardDefinition
AnalysisResolutionSourceResolution
SourceResolution
AnalysisResolutionStandardDefinition
StandardDefinition
SourceResolution
SourceResolution
StandardDefinition
StandardDefinition
SourceResolution
SourceResolution
StandardDefinition
StandardDefinition
SOURCE_RESOLUTION
SourceResolution
STANDARD_DEFINITION
StandardDefinition
"SourceResolution"
SourceResolution
"StandardDefinition"
StandardDefinition

Audio

Bitrate int

The bitrate, in bits per second, of the output encoded audio.

Channels int

The number of channels in the audio.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

SamplingRate int

The sampling rate to use for encoding in hertz.

Bitrate int

The bitrate, in bits per second, of the output encoded audio.

Channels int

The number of channels in the audio.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

SamplingRate int

The sampling rate to use for encoding in hertz.

bitrate Integer

The bitrate, in bits per second, of the output encoded audio.

channels Integer

The number of channels in the audio.

label String

An optional label for the codec. The label can be used to control muxing behavior.

samplingRate Integer

The sampling rate to use for encoding in hertz.

bitrate number

The bitrate, in bits per second, of the output encoded audio.

channels number

The number of channels in the audio.

label string

An optional label for the codec. The label can be used to control muxing behavior.

samplingRate number

The sampling rate to use for encoding in hertz.

bitrate int

The bitrate, in bits per second, of the output encoded audio.

channels int

The number of channels in the audio.

label str

An optional label for the codec. The label can be used to control muxing behavior.

sampling_rate int

The sampling rate to use for encoding in hertz.

bitrate Number

The bitrate, in bits per second, of the output encoded audio.

channels Number

The number of channels in the audio.

label String

An optional label for the codec. The label can be used to control muxing behavior.

samplingRate Number

The sampling rate to use for encoding in hertz.

AudioAnalysisMode

Standard
Standard

Performs all operations included in the Basic mode, additionally performing language detection and speaker diarization.

Basic
Basic

This mode performs speech-to-text transcription and generation of a VTT subtitle/caption file. The output of this mode includes an Insights JSON file including only the keywords, transcription,and timing information. Automatic language detection and speaker diarization are not included in this mode.

AudioAnalysisModeStandard
Standard

Performs all operations included in the Basic mode, additionally performing language detection and speaker diarization.

AudioAnalysisModeBasic
Basic

This mode performs speech-to-text transcription and generation of a VTT subtitle/caption file. The output of this mode includes an Insights JSON file including only the keywords, transcription,and timing information. Automatic language detection and speaker diarization are not included in this mode.

Standard
Standard

Performs all operations included in the Basic mode, additionally performing language detection and speaker diarization.

Basic
Basic

This mode performs speech-to-text transcription and generation of a VTT subtitle/caption file. The output of this mode includes an Insights JSON file including only the keywords, transcription,and timing information. Automatic language detection and speaker diarization are not included in this mode.

Standard
Standard

Performs all operations included in the Basic mode, additionally performing language detection and speaker diarization.

Basic
Basic

This mode performs speech-to-text transcription and generation of a VTT subtitle/caption file. The output of this mode includes an Insights JSON file including only the keywords, transcription,and timing information. Automatic language detection and speaker diarization are not included in this mode.

STANDARD
Standard

Performs all operations included in the Basic mode, additionally performing language detection and speaker diarization.

BASIC
Basic

This mode performs speech-to-text transcription and generation of a VTT subtitle/caption file. The output of this mode includes an Insights JSON file including only the keywords, transcription,and timing information. Automatic language detection and speaker diarization are not included in this mode.

"Standard"
Standard

Performs all operations included in the Basic mode, additionally performing language detection and speaker diarization.

"Basic"
Basic

This mode performs speech-to-text transcription and generation of a VTT subtitle/caption file. The output of this mode includes an Insights JSON file including only the keywords, transcription,and timing information. Automatic language detection and speaker diarization are not included in this mode.

AudioAnalyzerPreset

AudioLanguage string

The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463

ExperimentalOptions Dictionary<string, string>

Dictionary containing key value pairs for parameters not exposed in the preset itself

Mode string | Pulumi.AzureNative.Media.AudioAnalysisMode

Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.

AudioLanguage string

The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463

ExperimentalOptions map[string]string

Dictionary containing key value pairs for parameters not exposed in the preset itself

Mode string | AudioAnalysisMode

Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.

audioLanguage String

The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463

experimentalOptions Map<String,String>

Dictionary containing key value pairs for parameters not exposed in the preset itself

mode String | AudioAnalysisMode

Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.

audioLanguage string

The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463

experimentalOptions {[key: string]: string}

Dictionary containing key value pairs for parameters not exposed in the preset itself

mode string | AudioAnalysisMode

Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.

audio_language str

The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463

experimental_options Mapping[str, str]

Dictionary containing key value pairs for parameters not exposed in the preset itself

mode str | AudioAnalysisMode

Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.

audioLanguage String

The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463

experimentalOptions Map<String>

Dictionary containing key value pairs for parameters not exposed in the preset itself

mode String | "Standard" | "Basic"

Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.

AudioAnalyzerPresetResponse

AudioLanguage string

The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463

ExperimentalOptions Dictionary<string, string>

Dictionary containing key value pairs for parameters not exposed in the preset itself

Mode string

Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.

AudioLanguage string

The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463

ExperimentalOptions map[string]string

Dictionary containing key value pairs for parameters not exposed in the preset itself

Mode string

Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.

audioLanguage String

The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463

experimentalOptions Map<String,String>

Dictionary containing key value pairs for parameters not exposed in the preset itself

mode String

Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.

audioLanguage string

The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463

experimentalOptions {[key: string]: string}

Dictionary containing key value pairs for parameters not exposed in the preset itself

mode string

Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.

audio_language str

The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463

experimental_options Mapping[str, str]

Dictionary containing key value pairs for parameters not exposed in the preset itself

mode str

Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.

audioLanguage String

The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463

experimentalOptions Map<String>

Dictionary containing key value pairs for parameters not exposed in the preset itself

mode String

Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.

AudioOverlay

InputLabel string

The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.

AudioGainLevel double

The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.

End string

The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.

FadeInDuration string

The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).

FadeOutDuration string

The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).

Start string

The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.

InputLabel string

The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.

AudioGainLevel float64

The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.

End string

The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.

FadeInDuration string

The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).

FadeOutDuration string

The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).

Start string

The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.

inputLabel String

The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.

audioGainLevel Double

The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.

end String

The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.

fadeInDuration String

The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).

fadeOutDuration String

The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).

start String

The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.

inputLabel string

The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.

audioGainLevel number

The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.

end string

The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.

fadeInDuration string

The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).

fadeOutDuration string

The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).

start string

The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.

input_label str

The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.

audio_gain_level float

The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.

end str

The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.

fade_in_duration str

The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).

fade_out_duration str

The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).

start str

The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.

inputLabel String

The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.

audioGainLevel Number

The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.

end String

The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.

fadeInDuration String

The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).

fadeOutDuration String

The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).

start String

The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.

AudioOverlayResponse

InputLabel string

The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.

AudioGainLevel double

The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.

End string

The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.

FadeInDuration string

The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).

FadeOutDuration string

The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).

Start string

The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.

InputLabel string

The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.

AudioGainLevel float64

The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.

End string

The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.

FadeInDuration string

The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).

FadeOutDuration string

The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).

Start string

The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.

inputLabel String

The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.

audioGainLevel Double

The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.

end String

The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.

fadeInDuration String

The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).

fadeOutDuration String

The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).

start String

The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.

inputLabel string

The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.

audioGainLevel number

The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.

end string

The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.

fadeInDuration string

The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).

fadeOutDuration string

The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).

start string

The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.

input_label str

The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.

audio_gain_level float

The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.

end str

The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.

fade_in_duration str

The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).

fade_out_duration str

The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).

start str

The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.

inputLabel String

The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.

audioGainLevel Number

The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.

end String

The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.

fadeInDuration String

The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).

fadeOutDuration String

The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).

start String

The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.

AudioResponse

Bitrate int

The bitrate, in bits per second, of the output encoded audio.

Channels int

The number of channels in the audio.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

SamplingRate int

The sampling rate to use for encoding in hertz.

Bitrate int

The bitrate, in bits per second, of the output encoded audio.

Channels int

The number of channels in the audio.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

SamplingRate int

The sampling rate to use for encoding in hertz.

bitrate Integer

The bitrate, in bits per second, of the output encoded audio.

channels Integer

The number of channels in the audio.

label String

An optional label for the codec. The label can be used to control muxing behavior.

samplingRate Integer

The sampling rate to use for encoding in hertz.

bitrate number

The bitrate, in bits per second, of the output encoded audio.

channels number

The number of channels in the audio.

label string

An optional label for the codec. The label can be used to control muxing behavior.

samplingRate number

The sampling rate to use for encoding in hertz.

bitrate int

The bitrate, in bits per second, of the output encoded audio.

channels int

The number of channels in the audio.

label str

An optional label for the codec. The label can be used to control muxing behavior.

sampling_rate int

The sampling rate to use for encoding in hertz.

bitrate Number

The bitrate, in bits per second, of the output encoded audio.

channels Number

The number of channels in the audio.

label String

An optional label for the codec. The label can be used to control muxing behavior.

samplingRate Number

The sampling rate to use for encoding in hertz.

BlurType

Box
Box

Box: debug filter, bounding box only

Low
Low

Low: box-car blur filter

Med
Med

Med: Gaussian blur filter

High
High

High: Confuse blur filter

Black
Black

Black: Black out filter

BlurTypeBox
Box

Box: debug filter, bounding box only

BlurTypeLow
Low

Low: box-car blur filter

BlurTypeMed
Med

Med: Gaussian blur filter

BlurTypeHigh
High

High: Confuse blur filter

BlurTypeBlack
Black

Black: Black out filter

Box
Box

Box: debug filter, bounding box only

Low
Low

Low: box-car blur filter

Med
Med

Med: Gaussian blur filter

High
High

High: Confuse blur filter

Black
Black

Black: Black out filter

Box
Box

Box: debug filter, bounding box only

Low
Low

Low: box-car blur filter

Med
Med

Med: Gaussian blur filter

High
High

High: Confuse blur filter

Black
Black

Black: Black out filter

BOX
Box

Box: debug filter, bounding box only

LOW
Low

Low: box-car blur filter

MED
Med

Med: Gaussian blur filter

HIGH
High

High: Confuse blur filter

BLACK
Black

Black: Black out filter

"Box"
Box

Box: debug filter, bounding box only

"Low"
Low

Low: box-car blur filter

"Med"
Med

Med: Gaussian blur filter

"High"
High

High: Confuse blur filter

"Black"
Black

Black: Black out filter

BuiltInStandardEncoderPreset

PresetName string | Pulumi.AzureNative.Media.EncoderNamedPreset

The built-in preset to be used for encoding videos.

PresetName string | EncoderNamedPreset

The built-in preset to be used for encoding videos.

presetName String | EncoderNamedPreset

The built-in preset to be used for encoding videos.

presetName string | EncoderNamedPreset

The built-in preset to be used for encoding videos.

preset_name str | EncoderNamedPreset

The built-in preset to be used for encoding videos.

BuiltInStandardEncoderPresetResponse

PresetName string

The built-in preset to be used for encoding videos.

PresetName string

The built-in preset to be used for encoding videos.

presetName String

The built-in preset to be used for encoding videos.

presetName string

The built-in preset to be used for encoding videos.

preset_name str

The built-in preset to be used for encoding videos.

presetName String

The built-in preset to be used for encoding videos.

CopyAudio

Label string

An optional label for the codec. The label can be used to control muxing behavior.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

label String

An optional label for the codec. The label can be used to control muxing behavior.

label string

An optional label for the codec. The label can be used to control muxing behavior.

label str

An optional label for the codec. The label can be used to control muxing behavior.

label String

An optional label for the codec. The label can be used to control muxing behavior.

CopyAudioResponse

Label string

An optional label for the codec. The label can be used to control muxing behavior.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

label String

An optional label for the codec. The label can be used to control muxing behavior.

label string

An optional label for the codec. The label can be used to control muxing behavior.

label str

An optional label for the codec. The label can be used to control muxing behavior.

label String

An optional label for the codec. The label can be used to control muxing behavior.

CopyVideo

Label string

An optional label for the codec. The label can be used to control muxing behavior.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

label String

An optional label for the codec. The label can be used to control muxing behavior.

label string

An optional label for the codec. The label can be used to control muxing behavior.

label str

An optional label for the codec. The label can be used to control muxing behavior.

label String

An optional label for the codec. The label can be used to control muxing behavior.

CopyVideoResponse

Label string

An optional label for the codec. The label can be used to control muxing behavior.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

label String

An optional label for the codec. The label can be used to control muxing behavior.

label string

An optional label for the codec. The label can be used to control muxing behavior.

label str

An optional label for the codec. The label can be used to control muxing behavior.

label String

An optional label for the codec. The label can be used to control muxing behavior.

Deinterlace

Mode string | Pulumi.AzureNative.Media.DeinterlaceMode

The deinterlacing mode. Defaults to AutoPixelAdaptive.

Parity string | Pulumi.AzureNative.Media.DeinterlaceParity

The field parity for de-interlacing, defaults to Auto.

Mode string | DeinterlaceMode

The deinterlacing mode. Defaults to AutoPixelAdaptive.

Parity string | DeinterlaceParity

The field parity for de-interlacing, defaults to Auto.

mode String | DeinterlaceMode

The deinterlacing mode. Defaults to AutoPixelAdaptive.

parity String | DeinterlaceParity

The field parity for de-interlacing, defaults to Auto.

mode string | DeinterlaceMode

The deinterlacing mode. Defaults to AutoPixelAdaptive.

parity string | DeinterlaceParity

The field parity for de-interlacing, defaults to Auto.

mode str | DeinterlaceMode

The deinterlacing mode. Defaults to AutoPixelAdaptive.

parity str | DeinterlaceParity

The field parity for de-interlacing, defaults to Auto.

mode String | "Off" | "AutoPixelAdaptive"

The deinterlacing mode. Defaults to AutoPixelAdaptive.

parity String | "Auto" | "TopFieldFirst" | "BottomFieldFirst"

The field parity for de-interlacing, defaults to Auto.

DeinterlaceMode

Off
Off

Disables de-interlacing of the source video.

AutoPixelAdaptive
AutoPixelAdaptive

Apply automatic pixel adaptive de-interlacing on each frame in the input video.

DeinterlaceModeOff
Off

Disables de-interlacing of the source video.

DeinterlaceModeAutoPixelAdaptive
AutoPixelAdaptive

Apply automatic pixel adaptive de-interlacing on each frame in the input video.

Off
Off

Disables de-interlacing of the source video.

AutoPixelAdaptive
AutoPixelAdaptive

Apply automatic pixel adaptive de-interlacing on each frame in the input video.

Off
Off

Disables de-interlacing of the source video.

AutoPixelAdaptive
AutoPixelAdaptive

Apply automatic pixel adaptive de-interlacing on each frame in the input video.

OFF
Off

Disables de-interlacing of the source video.

AUTO_PIXEL_ADAPTIVE
AutoPixelAdaptive

Apply automatic pixel adaptive de-interlacing on each frame in the input video.

"Off"
Off

Disables de-interlacing of the source video.

"AutoPixelAdaptive"
AutoPixelAdaptive

Apply automatic pixel adaptive de-interlacing on each frame in the input video.

DeinterlaceParity

Auto
Auto

Automatically detect the order of fields

TopFieldFirst
TopFieldFirst

Apply top field first processing of input video.

BottomFieldFirst
BottomFieldFirst

Apply bottom field first processing of input video.

DeinterlaceParityAuto
Auto

Automatically detect the order of fields

DeinterlaceParityTopFieldFirst
TopFieldFirst

Apply top field first processing of input video.

DeinterlaceParityBottomFieldFirst
BottomFieldFirst

Apply bottom field first processing of input video.

Auto
Auto

Automatically detect the order of fields

TopFieldFirst
TopFieldFirst

Apply top field first processing of input video.

BottomFieldFirst
BottomFieldFirst

Apply bottom field first processing of input video.

Auto
Auto

Automatically detect the order of fields

TopFieldFirst
TopFieldFirst

Apply top field first processing of input video.

BottomFieldFirst
BottomFieldFirst

Apply bottom field first processing of input video.

AUTO
Auto

Automatically detect the order of fields

TOP_FIELD_FIRST
TopFieldFirst

Apply top field first processing of input video.

BOTTOM_FIELD_FIRST
BottomFieldFirst

Apply bottom field first processing of input video.

"Auto"
Auto

Automatically detect the order of fields

"TopFieldFirst"
TopFieldFirst

Apply top field first processing of input video.

"BottomFieldFirst"
BottomFieldFirst

Apply bottom field first processing of input video.

DeinterlaceResponse

Mode string

The deinterlacing mode. Defaults to AutoPixelAdaptive.

Parity string

The field parity for de-interlacing, defaults to Auto.

Mode string

The deinterlacing mode. Defaults to AutoPixelAdaptive.

Parity string

The field parity for de-interlacing, defaults to Auto.

mode String

The deinterlacing mode. Defaults to AutoPixelAdaptive.

parity String

The field parity for de-interlacing, defaults to Auto.

mode string

The deinterlacing mode. Defaults to AutoPixelAdaptive.

parity string

The field parity for de-interlacing, defaults to Auto.

mode str

The deinterlacing mode. Defaults to AutoPixelAdaptive.

parity str

The field parity for de-interlacing, defaults to Auto.

mode String

The deinterlacing mode. Defaults to AutoPixelAdaptive.

parity String

The field parity for de-interlacing, defaults to Auto.

EncoderNamedPreset

H264SingleBitrateSD
H264SingleBitrateSD

Produces an MP4 file where the video is encoded with H.264 codec at 2200 kbps and a picture height of 480 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

H264SingleBitrate720p
H264SingleBitrate720p

Produces an MP4 file where the video is encoded with H.264 codec at 4500 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

H264SingleBitrate1080p
H264SingleBitrate1080p

Produces an MP4 file where the video is encoded with H.264 codec at 6750 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

AdaptiveStreaming
AdaptiveStreaming

Produces a set of GOP aligned MP4 files with H.264 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.

AACGoodQualityAudio
AACGoodQualityAudio

Produces a single MP4 file containing only stereo audio encoded at 192 kbps.

ContentAwareEncodingExperimental
ContentAwareEncodingExperimental

Exposes an experimental preset for content-aware encoding. Given any input content, the service attempts to automatically determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. The underlying algorithms will continue to evolve over time. The output will contain MP4 files with video and audio interleaved.

ContentAwareEncoding
ContentAwareEncoding

Produces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.

CopyAllBitrateNonInterleaved
CopyAllBitrateNonInterleaved

Copy all video and audio streams from the input asset as non-interleaved video and audio output files. This preset can be used to clip an existing asset or convert a group of key frame (GOP) aligned MP4 files as an asset that can be streamed.

H264MultipleBitrate1080p
H264MultipleBitrate1080p

Produces a set of 8 GOP-aligned MP4 files, ranging from 6000 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 1080p and goes down to 180p.

H264MultipleBitrate720p
H264MultipleBitrate720p

Produces a set of 6 GOP-aligned MP4 files, ranging from 3400 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 720p and goes down to 180p.

H264MultipleBitrateSD
H264MultipleBitrateSD

Produces a set of 5 GOP-aligned MP4 files, ranging from 1900kbps to 400 kbps, and stereo AAC audio. Resolution starts at 480p and goes down to 240p.

H265ContentAwareEncoding
H265ContentAwareEncoding

Produces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.

H265AdaptiveStreaming
H265AdaptiveStreaming

Produces a set of GOP aligned MP4 files with H.265 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.

H265SingleBitrate720p
H265SingleBitrate720p

Produces an MP4 file where the video is encoded with H.265 codec at 1800 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

H265SingleBitrate1080p
H265SingleBitrate1080p

Produces an MP4 file where the video is encoded with H.265 codec at 3500 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

H265SingleBitrate4K
H265SingleBitrate4K

Produces an MP4 file where the video is encoded with H.265 codec at 9500 kbps and a picture height of 2160 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

EncoderNamedPresetH264SingleBitrateSD
H264SingleBitrateSD

Produces an MP4 file where the video is encoded with H.264 codec at 2200 kbps and a picture height of 480 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

EncoderNamedPresetH264SingleBitrate720p
H264SingleBitrate720p

Produces an MP4 file where the video is encoded with H.264 codec at 4500 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

EncoderNamedPresetH264SingleBitrate1080p
H264SingleBitrate1080p

Produces an MP4 file where the video is encoded with H.264 codec at 6750 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

EncoderNamedPresetAdaptiveStreaming
AdaptiveStreaming

Produces a set of GOP aligned MP4 files with H.264 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.

EncoderNamedPresetAACGoodQualityAudio
AACGoodQualityAudio

Produces a single MP4 file containing only stereo audio encoded at 192 kbps.

EncoderNamedPresetContentAwareEncodingExperimental
ContentAwareEncodingExperimental

Exposes an experimental preset for content-aware encoding. Given any input content, the service attempts to automatically determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. The underlying algorithms will continue to evolve over time. The output will contain MP4 files with video and audio interleaved.

EncoderNamedPresetContentAwareEncoding
ContentAwareEncoding

Produces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.

EncoderNamedPresetCopyAllBitrateNonInterleaved
CopyAllBitrateNonInterleaved

Copy all video and audio streams from the input asset as non-interleaved video and audio output files. This preset can be used to clip an existing asset or convert a group of key frame (GOP) aligned MP4 files as an asset that can be streamed.

EncoderNamedPresetH264MultipleBitrate1080p
H264MultipleBitrate1080p

Produces a set of 8 GOP-aligned MP4 files, ranging from 6000 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 1080p and goes down to 180p.

EncoderNamedPresetH264MultipleBitrate720p
H264MultipleBitrate720p

Produces a set of 6 GOP-aligned MP4 files, ranging from 3400 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 720p and goes down to 180p.

EncoderNamedPresetH264MultipleBitrateSD
H264MultipleBitrateSD

Produces a set of 5 GOP-aligned MP4 files, ranging from 1900kbps to 400 kbps, and stereo AAC audio. Resolution starts at 480p and goes down to 240p.

EncoderNamedPresetH265ContentAwareEncoding
H265ContentAwareEncoding

Produces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.

EncoderNamedPresetH265AdaptiveStreaming
H265AdaptiveStreaming

Produces a set of GOP aligned MP4 files with H.265 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.

EncoderNamedPresetH265SingleBitrate720p
H265SingleBitrate720p

Produces an MP4 file where the video is encoded with H.265 codec at 1800 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

EncoderNamedPresetH265SingleBitrate1080p
H265SingleBitrate1080p

Produces an MP4 file where the video is encoded with H.265 codec at 3500 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

EncoderNamedPresetH265SingleBitrate4K
H265SingleBitrate4K

Produces an MP4 file where the video is encoded with H.265 codec at 9500 kbps and a picture height of 2160 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

H264SingleBitrateSD
H264SingleBitrateSD

Produces an MP4 file where the video is encoded with H.264 codec at 2200 kbps and a picture height of 480 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

H264SingleBitrate720p
H264SingleBitrate720p

Produces an MP4 file where the video is encoded with H.264 codec at 4500 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

H264SingleBitrate1080p
H264SingleBitrate1080p

Produces an MP4 file where the video is encoded with H.264 codec at 6750 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

AdaptiveStreaming
AdaptiveStreaming

Produces a set of GOP aligned MP4 files with H.264 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.

AACGoodQualityAudio
AACGoodQualityAudio

Produces a single MP4 file containing only stereo audio encoded at 192 kbps.

ContentAwareEncodingExperimental
ContentAwareEncodingExperimental

Exposes an experimental preset for content-aware encoding. Given any input content, the service attempts to automatically determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. The underlying algorithms will continue to evolve over time. The output will contain MP4 files with video and audio interleaved.

ContentAwareEncoding
ContentAwareEncoding

Produces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.

CopyAllBitrateNonInterleaved
CopyAllBitrateNonInterleaved

Copy all video and audio streams from the input asset as non-interleaved video and audio output files. This preset can be used to clip an existing asset or convert a group of key frame (GOP) aligned MP4 files as an asset that can be streamed.

H264MultipleBitrate1080p
H264MultipleBitrate1080p

Produces a set of 8 GOP-aligned MP4 files, ranging from 6000 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 1080p and goes down to 180p.

H264MultipleBitrate720p
H264MultipleBitrate720p

Produces a set of 6 GOP-aligned MP4 files, ranging from 3400 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 720p and goes down to 180p.

H264MultipleBitrateSD
H264MultipleBitrateSD

Produces a set of 5 GOP-aligned MP4 files, ranging from 1900kbps to 400 kbps, and stereo AAC audio. Resolution starts at 480p and goes down to 240p.

H265ContentAwareEncoding
H265ContentAwareEncoding

Produces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.

H265AdaptiveStreaming
H265AdaptiveStreaming

Produces a set of GOP aligned MP4 files with H.265 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.

H265SingleBitrate720p
H265SingleBitrate720p

Produces an MP4 file where the video is encoded with H.265 codec at 1800 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

H265SingleBitrate1080p
H265SingleBitrate1080p

Produces an MP4 file where the video is encoded with H.265 codec at 3500 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

H265SingleBitrate4K
H265SingleBitrate4K

Produces an MP4 file where the video is encoded with H.265 codec at 9500 kbps and a picture height of 2160 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

H264SingleBitrateSD
H264SingleBitrateSD

Produces an MP4 file where the video is encoded with H.264 codec at 2200 kbps and a picture height of 480 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

H264SingleBitrate720p
H264SingleBitrate720p

Produces an MP4 file where the video is encoded with H.264 codec at 4500 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

H264SingleBitrate1080p
H264SingleBitrate1080p

Produces an MP4 file where the video is encoded with H.264 codec at 6750 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

AdaptiveStreaming
AdaptiveStreaming

Produces a set of GOP aligned MP4 files with H.264 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.

AACGoodQualityAudio
AACGoodQualityAudio

Produces a single MP4 file containing only stereo audio encoded at 192 kbps.

ContentAwareEncodingExperimental
ContentAwareEncodingExperimental

Exposes an experimental preset for content-aware encoding. Given any input content, the service attempts to automatically determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. The underlying algorithms will continue to evolve over time. The output will contain MP4 files with video and audio interleaved.

ContentAwareEncoding
ContentAwareEncoding

Produces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.

CopyAllBitrateNonInterleaved
CopyAllBitrateNonInterleaved

Copy all video and audio streams from the input asset as non-interleaved video and audio output files. This preset can be used to clip an existing asset or convert a group of key frame (GOP) aligned MP4 files as an asset that can be streamed.

H264MultipleBitrate1080p
H264MultipleBitrate1080p

Produces a set of 8 GOP-aligned MP4 files, ranging from 6000 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 1080p and goes down to 180p.

H264MultipleBitrate720p
H264MultipleBitrate720p

Produces a set of 6 GOP-aligned MP4 files, ranging from 3400 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 720p and goes down to 180p.

H264MultipleBitrateSD
H264MultipleBitrateSD

Produces a set of 5 GOP-aligned MP4 files, ranging from 1900kbps to 400 kbps, and stereo AAC audio. Resolution starts at 480p and goes down to 240p.

H265ContentAwareEncoding
H265ContentAwareEncoding

Produces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.

H265AdaptiveStreaming
H265AdaptiveStreaming

Produces a set of GOP aligned MP4 files with H.265 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.

H265SingleBitrate720p
H265SingleBitrate720p

Produces an MP4 file where the video is encoded with H.265 codec at 1800 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

H265SingleBitrate1080p
H265SingleBitrate1080p

Produces an MP4 file where the video is encoded with H.265 codec at 3500 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

H265SingleBitrate4K
H265SingleBitrate4K

Produces an MP4 file where the video is encoded with H.265 codec at 9500 kbps and a picture height of 2160 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

H264_SINGLE_BITRATE_SD
H264SingleBitrateSD

Produces an MP4 file where the video is encoded with H.264 codec at 2200 kbps and a picture height of 480 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

H264_SINGLE_BITRATE720P
H264SingleBitrate720p

Produces an MP4 file where the video is encoded with H.264 codec at 4500 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

H264_SINGLE_BITRATE1080P
H264SingleBitrate1080p

Produces an MP4 file where the video is encoded with H.264 codec at 6750 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

ADAPTIVE_STREAMING
AdaptiveStreaming

Produces a set of GOP aligned MP4 files with H.264 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.

AAC_GOOD_QUALITY_AUDIO
AACGoodQualityAudio

Produces a single MP4 file containing only stereo audio encoded at 192 kbps.

CONTENT_AWARE_ENCODING_EXPERIMENTAL
ContentAwareEncodingExperimental

Exposes an experimental preset for content-aware encoding. Given any input content, the service attempts to automatically determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. The underlying algorithms will continue to evolve over time. The output will contain MP4 files with video and audio interleaved.

CONTENT_AWARE_ENCODING
ContentAwareEncoding

Produces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.

COPY_ALL_BITRATE_NON_INTERLEAVED
CopyAllBitrateNonInterleaved

Copy all video and audio streams from the input asset as non-interleaved video and audio output files. This preset can be used to clip an existing asset or convert a group of key frame (GOP) aligned MP4 files as an asset that can be streamed.

H264_MULTIPLE_BITRATE1080P
H264MultipleBitrate1080p

Produces a set of 8 GOP-aligned MP4 files, ranging from 6000 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 1080p and goes down to 180p.

H264_MULTIPLE_BITRATE720P
H264MultipleBitrate720p

Produces a set of 6 GOP-aligned MP4 files, ranging from 3400 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 720p and goes down to 180p.

H264_MULTIPLE_BITRATE_SD
H264MultipleBitrateSD

Produces a set of 5 GOP-aligned MP4 files, ranging from 1900kbps to 400 kbps, and stereo AAC audio. Resolution starts at 480p and goes down to 240p.

H265_CONTENT_AWARE_ENCODING
H265ContentAwareEncoding

Produces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.

H265_ADAPTIVE_STREAMING
H265AdaptiveStreaming

Produces a set of GOP aligned MP4 files with H.265 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.

H265_SINGLE_BITRATE720P
H265SingleBitrate720p

Produces an MP4 file where the video is encoded with H.265 codec at 1800 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

H265_SINGLE_BITRATE1080P
H265SingleBitrate1080p

Produces an MP4 file where the video is encoded with H.265 codec at 3500 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

H265_SINGLE_BITRATE4_K
H265SingleBitrate4K

Produces an MP4 file where the video is encoded with H.265 codec at 9500 kbps and a picture height of 2160 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

"H264SingleBitrateSD"
H264SingleBitrateSD

Produces an MP4 file where the video is encoded with H.264 codec at 2200 kbps and a picture height of 480 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

"H264SingleBitrate720p"
H264SingleBitrate720p

Produces an MP4 file where the video is encoded with H.264 codec at 4500 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

"H264SingleBitrate1080p"
H264SingleBitrate1080p

Produces an MP4 file where the video is encoded with H.264 codec at 6750 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

"AdaptiveStreaming"
AdaptiveStreaming

Produces a set of GOP aligned MP4 files with H.264 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.

"AACGoodQualityAudio"
AACGoodQualityAudio

Produces a single MP4 file containing only stereo audio encoded at 192 kbps.

"ContentAwareEncodingExperimental"
ContentAwareEncodingExperimental

Exposes an experimental preset for content-aware encoding. Given any input content, the service attempts to automatically determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. The underlying algorithms will continue to evolve over time. The output will contain MP4 files with video and audio interleaved.

"ContentAwareEncoding"
ContentAwareEncoding

Produces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.

"CopyAllBitrateNonInterleaved"
CopyAllBitrateNonInterleaved

Copy all video and audio streams from the input asset as non-interleaved video and audio output files. This preset can be used to clip an existing asset or convert a group of key frame (GOP) aligned MP4 files as an asset that can be streamed.

"H264MultipleBitrate1080p"
H264MultipleBitrate1080p

Produces a set of 8 GOP-aligned MP4 files, ranging from 6000 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 1080p and goes down to 180p.

"H264MultipleBitrate720p"
H264MultipleBitrate720p

Produces a set of 6 GOP-aligned MP4 files, ranging from 3400 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 720p and goes down to 180p.

"H264MultipleBitrateSD"
H264MultipleBitrateSD

Produces a set of 5 GOP-aligned MP4 files, ranging from 1900kbps to 400 kbps, and stereo AAC audio. Resolution starts at 480p and goes down to 240p.

"H265ContentAwareEncoding"
H265ContentAwareEncoding

Produces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.

"H265AdaptiveStreaming"
H265AdaptiveStreaming

Produces a set of GOP aligned MP4 files with H.265 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.

"H265SingleBitrate720p"
H265SingleBitrate720p

Produces an MP4 file where the video is encoded with H.265 codec at 1800 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

"H265SingleBitrate1080p"
H265SingleBitrate1080p

Produces an MP4 file where the video is encoded with H.265 codec at 3500 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

"H265SingleBitrate4K"
H265SingleBitrate4K

Produces an MP4 file where the video is encoded with H.265 codec at 9500 kbps and a picture height of 2160 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

EntropyMode

Cabac
Cabac

Context Adaptive Binary Arithmetic Coder (CABAC) entropy encoding.

Cavlc
Cavlc

Context Adaptive Variable Length Coder (CAVLC) entropy encoding.

EntropyModeCabac
Cabac

Context Adaptive Binary Arithmetic Coder (CABAC) entropy encoding.

EntropyModeCavlc
Cavlc

Context Adaptive Variable Length Coder (CAVLC) entropy encoding.

Cabac
Cabac

Context Adaptive Binary Arithmetic Coder (CABAC) entropy encoding.

Cavlc
Cavlc

Context Adaptive Variable Length Coder (CAVLC) entropy encoding.

Cabac
Cabac

Context Adaptive Binary Arithmetic Coder (CABAC) entropy encoding.

Cavlc
Cavlc

Context Adaptive Variable Length Coder (CAVLC) entropy encoding.

CABAC
Cabac

Context Adaptive Binary Arithmetic Coder (CABAC) entropy encoding.

CAVLC
Cavlc

Context Adaptive Variable Length Coder (CAVLC) entropy encoding.

"Cabac"
Cabac

Context Adaptive Binary Arithmetic Coder (CABAC) entropy encoding.

"Cavlc"
Cavlc

Context Adaptive Variable Length Coder (CAVLC) entropy encoding.

FaceDetectorPreset

BlurType string | Pulumi.AzureNative.Media.BlurType

Blur type

ExperimentalOptions Dictionary<string, string>

Dictionary containing key value pairs for parameters not exposed in the preset itself

Mode string | Pulumi.AzureNative.Media.FaceRedactorMode

This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.

Resolution string | Pulumi.AzureNative.Media.AnalysisResolution

Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.

BlurType string | BlurType

Blur type

ExperimentalOptions map[string]string

Dictionary containing key value pairs for parameters not exposed in the preset itself

Mode string | FaceRedactorMode

This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.

Resolution string | AnalysisResolution

Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.

blurType String | BlurType

Blur type

experimentalOptions Map<String,String>

Dictionary containing key value pairs for parameters not exposed in the preset itself

mode String | FaceRedactorMode

This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.

resolution String | AnalysisResolution

Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.

blurType string | BlurType

Blur type

experimentalOptions {[key: string]: string}

Dictionary containing key value pairs for parameters not exposed in the preset itself

mode string | FaceRedactorMode

This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.

resolution string | AnalysisResolution

Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.

blur_type str | BlurType

Blur type

experimental_options Mapping[str, str]

Dictionary containing key value pairs for parameters not exposed in the preset itself

mode str | FaceRedactorMode

This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.

resolution str | AnalysisResolution

Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.

blurType String | "Box" | "Low" | "Med" | "High" | "Black"

Blur type

experimentalOptions Map<String>

Dictionary containing key value pairs for parameters not exposed in the preset itself

mode String | "Analyze" | "Redact" | "Combined"

This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.

resolution String | "SourceResolution" | "StandardDefinition"

Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.

FaceDetectorPresetResponse

BlurType string

Blur type

ExperimentalOptions Dictionary<string, string>

Dictionary containing key value pairs for parameters not exposed in the preset itself

Mode string

This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.

Resolution string

Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.

BlurType string

Blur type

ExperimentalOptions map[string]string

Dictionary containing key value pairs for parameters not exposed in the preset itself

Mode string

This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.

Resolution string

Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.

blurType String

Blur type

experimentalOptions Map<String,String>

Dictionary containing key value pairs for parameters not exposed in the preset itself

mode String

This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.

resolution String

Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.

blurType string

Blur type

experimentalOptions {[key: string]: string}

Dictionary containing key value pairs for parameters not exposed in the preset itself

mode string

This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.

resolution string

Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.

blur_type str

Blur type

experimental_options Mapping[str, str]

Dictionary containing key value pairs for parameters not exposed in the preset itself

mode str

This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.

resolution str

Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.

blurType String

Blur type

experimentalOptions Map<String>

Dictionary containing key value pairs for parameters not exposed in the preset itself

mode String

This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.

resolution String

Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.

FaceRedactorMode

Analyze
Analyze

Analyze mode detects faces and outputs a metadata file with the results. Allows editing of the metadata file before faces are blurred with Redact mode.

Redact
Redact

Redact mode consumes the metadata file from Analyze mode and redacts the faces found.

Combined
Combined

Combined mode does the Analyze and Redact steps in one pass when editing the analyzed faces is not desired.

FaceRedactorModeAnalyze
Analyze

Analyze mode detects faces and outputs a metadata file with the results. Allows editing of the metadata file before faces are blurred with Redact mode.

FaceRedactorModeRedact
Redact

Redact mode consumes the metadata file from Analyze mode and redacts the faces found.

FaceRedactorModeCombined
Combined

Combined mode does the Analyze and Redact steps in one pass when editing the analyzed faces is not desired.

Analyze
Analyze

Analyze mode detects faces and outputs a metadata file with the results. Allows editing of the metadata file before faces are blurred with Redact mode.

Redact
Redact

Redact mode consumes the metadata file from Analyze mode and redacts the faces found.

Combined
Combined

Combined mode does the Analyze and Redact steps in one pass when editing the analyzed faces is not desired.

Analyze
Analyze

Analyze mode detects faces and outputs a metadata file with the results. Allows editing of the metadata file before faces are blurred with Redact mode.

Redact
Redact

Redact mode consumes the metadata file from Analyze mode and redacts the faces found.

Combined
Combined

Combined mode does the Analyze and Redact steps in one pass when editing the analyzed faces is not desired.

ANALYZE
Analyze

Analyze mode detects faces and outputs a metadata file with the results. Allows editing of the metadata file before faces are blurred with Redact mode.

REDACT
Redact

Redact mode consumes the metadata file from Analyze mode and redacts the faces found.

COMBINED
Combined

Combined mode does the Analyze and Redact steps in one pass when editing the analyzed faces is not desired.

"Analyze"
Analyze

Analyze mode detects faces and outputs a metadata file with the results. Allows editing of the metadata file before faces are blurred with Redact mode.

"Redact"
Redact

Redact mode consumes the metadata file from Analyze mode and redacts the faces found.

"Combined"
Combined

Combined mode does the Analyze and Redact steps in one pass when editing the analyzed faces is not desired.

Filters

Crop Pulumi.AzureNative.Media.Inputs.Rectangle

The parameters for the rectangular window with which to crop the input video.

Deinterlace Pulumi.AzureNative.Media.Inputs.Deinterlace

The de-interlacing settings.

Overlays List<Union<Pulumi.AzureNative.Media.Inputs.AudioOverlay, Pulumi.AzureNative.Media.Inputs.VideoOverlay>>

The properties of overlays to be applied to the input video. These could be audio, image or video overlays.

Rotation string | Pulumi.AzureNative.Media.Rotation

The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto

Crop Rectangle

The parameters for the rectangular window with which to crop the input video.

Deinterlace Deinterlace

The de-interlacing settings.

Overlays []interface{}

The properties of overlays to be applied to the input video. These could be audio, image or video overlays.

Rotation string | Rotation

The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto

crop Rectangle

The parameters for the rectangular window with which to crop the input video.

deinterlace Deinterlace

The de-interlacing settings.

overlays List<Either<AudioOverlay,VideoOverlay>>

The properties of overlays to be applied to the input video. These could be audio, image or video overlays.

rotation String | Rotation

The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto

crop Rectangle

The parameters for the rectangular window with which to crop the input video.

deinterlace Deinterlace

The de-interlacing settings.

overlays (AudioOverlay | VideoOverlay)[]

The properties of overlays to be applied to the input video. These could be audio, image or video overlays.

rotation string | Rotation

The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto

crop Rectangle

The parameters for the rectangular window with which to crop the input video.

deinterlace Deinterlace

The de-interlacing settings.

overlays Sequence[Union[AudioOverlay, VideoOverlay]]

The properties of overlays to be applied to the input video. These could be audio, image or video overlays.

rotation str | Rotation

The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto

crop Property Map

The parameters for the rectangular window with which to crop the input video.

deinterlace Property Map

The de-interlacing settings.

overlays List<Property Map | Property Map>

The properties of overlays to be applied to the input video. These could be audio, image or video overlays.

rotation String | "Auto" | "None" | "Rotate0" | "Rotate90" | "Rotate180" | "Rotate270"

The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto

FiltersResponse

Crop Pulumi.AzureNative.Media.Inputs.RectangleResponse

The parameters for the rectangular window with which to crop the input video.

Deinterlace Pulumi.AzureNative.Media.Inputs.DeinterlaceResponse

The de-interlacing settings.

Overlays List<Union<Pulumi.AzureNative.Media.Inputs.AudioOverlayResponse, Pulumi.AzureNative.Media.Inputs.VideoOverlayResponse>>

The properties of overlays to be applied to the input video. These could be audio, image or video overlays.

Rotation string

The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto

Crop RectangleResponse

The parameters for the rectangular window with which to crop the input video.

Deinterlace DeinterlaceResponse

The de-interlacing settings.

Overlays []interface{}

The properties of overlays to be applied to the input video. These could be audio, image or video overlays.

Rotation string

The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto

crop RectangleResponse

The parameters for the rectangular window with which to crop the input video.

deinterlace DeinterlaceResponse

The de-interlacing settings.

overlays List<Either<AudioOverlayResponse,VideoOverlayResponse>>

The properties of overlays to be applied to the input video. These could be audio, image or video overlays.

rotation String

The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto

crop RectangleResponse

The parameters for the rectangular window with which to crop the input video.

deinterlace DeinterlaceResponse

The de-interlacing settings.

overlays (AudioOverlayResponse | VideoOverlayResponse)[]

The properties of overlays to be applied to the input video. These could be audio, image or video overlays.

rotation string

The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto

crop RectangleResponse

The parameters for the rectangular window with which to crop the input video.

deinterlace DeinterlaceResponse

The de-interlacing settings.

overlays Sequence[Union[AudioOverlayResponse, VideoOverlayResponse]]

The properties of overlays to be applied to the input video. These could be audio, image or video overlays.

rotation str

The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto

crop Property Map

The parameters for the rectangular window with which to crop the input video.

deinterlace Property Map

The de-interlacing settings.

overlays List<Property Map | Property Map>

The properties of overlays to be applied to the input video. These could be audio, image or video overlays.

rotation String

The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto

H264Complexity

Speed
Speed

Tells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.

Balanced
Balanced

Tells the encoder to use settings that achieve a balance between speed and quality.

Quality
Quality

Tells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.

H264ComplexitySpeed
Speed

Tells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.

H264ComplexityBalanced
Balanced

Tells the encoder to use settings that achieve a balance between speed and quality.

H264ComplexityQuality
Quality

Tells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.

Speed
Speed

Tells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.

Balanced
Balanced

Tells the encoder to use settings that achieve a balance between speed and quality.

Quality
Quality

Tells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.

Speed
Speed

Tells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.

Balanced
Balanced

Tells the encoder to use settings that achieve a balance between speed and quality.

Quality
Quality

Tells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.

SPEED
Speed

Tells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.

BALANCED
Balanced

Tells the encoder to use settings that achieve a balance between speed and quality.

QUALITY
Quality

Tells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.

"Speed"
Speed

Tells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.

"Balanced"
Balanced

Tells the encoder to use settings that achieve a balance between speed and quality.

"Quality"
Quality

Tells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.

H264Layer

Bitrate int

The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.

AdaptiveBFrame bool

Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.

BFrames int

The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.

BufferWindow string

The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).

EntropyMode string | Pulumi.AzureNative.Media.EntropyMode

The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.

FrameRate string

The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.

Height string

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

Label string

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

Level string

We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.

MaxBitrate int

The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.

Profile string | Pulumi.AzureNative.Media.H264VideoProfile

We currently support Baseline, Main, High, High422, High444. Default is Auto.

ReferenceFrames int

The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.

Slices int

The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.

Width string

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

Bitrate int

The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.

AdaptiveBFrame bool

Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.

BFrames int

The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.

BufferWindow string

The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).

EntropyMode string | EntropyMode

The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.

FrameRate string

The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.

Height string

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

Label string

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

Level string

We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.

MaxBitrate int

The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.

Profile string | H264VideoProfile

We currently support Baseline, Main, High, High422, High444. Default is Auto.

ReferenceFrames int

The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.

Slices int

The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.

Width string

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

bitrate Integer

The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.

adaptiveBFrame Boolean

Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.

bFrames Integer

The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.

bufferWindow String

The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).

entropyMode String | EntropyMode

The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.

frameRate String

The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.

height String

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label String

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

level String

We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.

maxBitrate Integer

The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.

profile String | H264VideoProfile

We currently support Baseline, Main, High, High422, High444. Default is Auto.

referenceFrames Integer

The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.

slices Integer

The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.

width String

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

bitrate number

The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.

adaptiveBFrame boolean

Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.

bFrames number

The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.

bufferWindow string

The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).

entropyMode string | EntropyMode

The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.

frameRate string

The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.

height string

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label string

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

level string

We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.

maxBitrate number

The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.

profile string | H264VideoProfile

We currently support Baseline, Main, High, High422, High444. Default is Auto.

referenceFrames number

The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.

slices number

The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.

width string

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

bitrate int

The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.

adaptive_b_frame bool

Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.

b_frames int

The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.

buffer_window str

The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).

entropy_mode str | EntropyMode

The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.

frame_rate str

The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.

height str

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label str

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

level str

We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.

max_bitrate int

The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.

profile str | H264VideoProfile

We currently support Baseline, Main, High, High422, High444. Default is Auto.

reference_frames int

The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.

slices int

The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.

width str

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

bitrate Number

The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.

adaptiveBFrame Boolean

Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.

bFrames Number

The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.

bufferWindow String

The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).

entropyMode String | "Cabac" | "Cavlc"

The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.

frameRate String

The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.

height String

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label String

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

level String

We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.

maxBitrate Number

The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.

profile String | "Auto" | "Baseline" | "Main" | "High" | "High422" | "High444"

We currently support Baseline, Main, High, High422, High444. Default is Auto.

referenceFrames Number

The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.

slices Number

The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.

width String

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

H264LayerResponse

Bitrate int

The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.

AdaptiveBFrame bool

Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.

BFrames int

The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.

BufferWindow string

The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).

EntropyMode string

The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.

FrameRate string

The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.

Height string

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

Label string

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

Level string

We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.

MaxBitrate int

The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.

Profile string

We currently support Baseline, Main, High, High422, High444. Default is Auto.

ReferenceFrames int

The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.

Slices int

The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.

Width string

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

Bitrate int

The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.

AdaptiveBFrame bool

Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.

BFrames int

The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.

BufferWindow string

The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).

EntropyMode string

The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.

FrameRate string

The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.

Height string

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

Label string

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

Level string

We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.

MaxBitrate int

The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.

Profile string

We currently support Baseline, Main, High, High422, High444. Default is Auto.

ReferenceFrames int

The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.

Slices int

The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.

Width string

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

bitrate Integer

The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.

adaptiveBFrame Boolean

Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.

bFrames Integer

The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.

bufferWindow String

The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).

entropyMode String

The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.

frameRate String

The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.

height String

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label String

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

level String

We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.

maxBitrate Integer

The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.

profile String

We currently support Baseline, Main, High, High422, High444. Default is Auto.

referenceFrames Integer

The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.

slices Integer

The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.

width String

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

bitrate number

The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.

adaptiveBFrame boolean

Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.

bFrames number

The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.

bufferWindow string

The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).

entropyMode string

The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.

frameRate string

The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.

height string

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label string

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

level string

We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.

maxBitrate number

The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.

profile string

We currently support Baseline, Main, High, High422, High444. Default is Auto.

referenceFrames number

The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.

slices number

The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.

width string

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

bitrate int

The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.

adaptive_b_frame bool

Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.

b_frames int

The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.

buffer_window str

The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).

entropy_mode str

The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.

frame_rate str

The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.

height str

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label str

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

level str

We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.

max_bitrate int

The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.

profile str

We currently support Baseline, Main, High, High422, High444. Default is Auto.

reference_frames int

The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.

slices int

The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.

width str

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

bitrate Number

The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.

adaptiveBFrame Boolean

Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.

bFrames Number

The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.

bufferWindow String

The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).

entropyMode String

The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.

frameRate String

The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.

height String

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label String

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

level String

We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.

maxBitrate Number

The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.

profile String

We currently support Baseline, Main, High, High422, High444. Default is Auto.

referenceFrames Number

The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.

slices Number

The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.

width String

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

H264Video

Complexity string | Pulumi.AzureNative.Media.H264Complexity

Tells the encoder how to choose its encoding settings. The default value is Balanced.

KeyFrameInterval string

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

Layers List<Pulumi.AzureNative.Media.Inputs.H264Layer>

The collection of output H.264 layers to be produced by the encoder.

SceneChangeDetection bool

Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.

StretchMode string | Pulumi.AzureNative.Media.StretchMode

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

SyncMode string | Pulumi.AzureNative.Media.VideoSyncMode

The Video Sync Mode

Complexity string | H264Complexity

Tells the encoder how to choose its encoding settings. The default value is Balanced.

KeyFrameInterval string

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

Layers []H264Layer

The collection of output H.264 layers to be produced by the encoder.

SceneChangeDetection bool

Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.

StretchMode string | StretchMode

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

SyncMode string | VideoSyncMode

The Video Sync Mode

complexity String | H264Complexity

Tells the encoder how to choose its encoding settings. The default value is Balanced.

keyFrameInterval String

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label String

An optional label for the codec. The label can be used to control muxing behavior.

layers List<H264Layer>

The collection of output H.264 layers to be produced by the encoder.

sceneChangeDetection Boolean

Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.

stretchMode String | StretchMode

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

syncMode String | VideoSyncMode

The Video Sync Mode

complexity string | H264Complexity

Tells the encoder how to choose its encoding settings. The default value is Balanced.

keyFrameInterval string

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label string

An optional label for the codec. The label can be used to control muxing behavior.

layers H264Layer[]

The collection of output H.264 layers to be produced by the encoder.

sceneChangeDetection boolean

Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.

stretchMode string | StretchMode

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

syncMode string | VideoSyncMode

The Video Sync Mode

complexity str | H264Complexity

Tells the encoder how to choose its encoding settings. The default value is Balanced.

key_frame_interval str

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label str

An optional label for the codec. The label can be used to control muxing behavior.

layers Sequence[H264Layer]

The collection of output H.264 layers to be produced by the encoder.

scene_change_detection bool

Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.

stretch_mode str | StretchMode

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

sync_mode str | VideoSyncMode

The Video Sync Mode

complexity String | "Speed" | "Balanced" | "Quality"

Tells the encoder how to choose its encoding settings. The default value is Balanced.

keyFrameInterval String

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label String

An optional label for the codec. The label can be used to control muxing behavior.

layers List<Property Map>

The collection of output H.264 layers to be produced by the encoder.

sceneChangeDetection Boolean

Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.

stretchMode String | "None" | "AutoSize" | "AutoFit"

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

syncMode String | "Auto" | "Passthrough" | "Cfr" | "Vfr"

The Video Sync Mode

H264VideoProfile

Auto
Auto

Tells the encoder to automatically determine the appropriate H.264 profile.

Baseline
Baseline

Baseline profile

Main
Main

Main profile

High
High

High profile.

High422
High422

High 4:2:2 profile.

High444
High444

High 4:4:4 predictive profile.

H264VideoProfileAuto
Auto

Tells the encoder to automatically determine the appropriate H.264 profile.

H264VideoProfileBaseline
Baseline

Baseline profile

H264VideoProfileMain
Main

Main profile

H264VideoProfileHigh
High

High profile.

H264VideoProfileHigh422
High422

High 4:2:2 profile.

H264VideoProfileHigh444
High444

High 4:4:4 predictive profile.

Auto
Auto

Tells the encoder to automatically determine the appropriate H.264 profile.

Baseline
Baseline

Baseline profile

Main
Main

Main profile

High
High

High profile.

High422
High422

High 4:2:2 profile.

High444
High444

High 4:4:4 predictive profile.

Auto
Auto

Tells the encoder to automatically determine the appropriate H.264 profile.

Baseline
Baseline

Baseline profile

Main
Main

Main profile

High
High

High profile.

High422
High422

High 4:2:2 profile.

High444
High444

High 4:4:4 predictive profile.

AUTO
Auto

Tells the encoder to automatically determine the appropriate H.264 profile.

BASELINE
Baseline

Baseline profile

MAIN
Main

Main profile

HIGH
High

High profile.

HIGH422
High422

High 4:2:2 profile.

HIGH444
High444

High 4:4:4 predictive profile.

"Auto"
Auto

Tells the encoder to automatically determine the appropriate H.264 profile.

"Baseline"
Baseline

Baseline profile

"Main"
Main

Main profile

"High"
High

High profile.

"High422"
High422

High 4:2:2 profile.

"High444"
High444

High 4:4:4 predictive profile.

H264VideoResponse

Complexity string

Tells the encoder how to choose its encoding settings. The default value is Balanced.

KeyFrameInterval string

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

Layers List<Pulumi.AzureNative.Media.Inputs.H264LayerResponse>

The collection of output H.264 layers to be produced by the encoder.

SceneChangeDetection bool

Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.

StretchMode string

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

SyncMode string

The Video Sync Mode

Complexity string

Tells the encoder how to choose its encoding settings. The default value is Balanced.

KeyFrameInterval string

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

Layers []H264LayerResponse

The collection of output H.264 layers to be produced by the encoder.

SceneChangeDetection bool

Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.

StretchMode string

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

SyncMode string

The Video Sync Mode

complexity String

Tells the encoder how to choose its encoding settings. The default value is Balanced.

keyFrameInterval String

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label String

An optional label for the codec. The label can be used to control muxing behavior.

layers List<H264LayerResponse>

The collection of output H.264 layers to be produced by the encoder.

sceneChangeDetection Boolean

Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.

stretchMode String

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

syncMode String

The Video Sync Mode

complexity string

Tells the encoder how to choose its encoding settings. The default value is Balanced.

keyFrameInterval string

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label string

An optional label for the codec. The label can be used to control muxing behavior.

layers H264LayerResponse[]

The collection of output H.264 layers to be produced by the encoder.

sceneChangeDetection boolean

Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.

stretchMode string

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

syncMode string

The Video Sync Mode

complexity str

Tells the encoder how to choose its encoding settings. The default value is Balanced.

key_frame_interval str

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label str

An optional label for the codec. The label can be used to control muxing behavior.

layers Sequence[H264LayerResponse]

The collection of output H.264 layers to be produced by the encoder.

scene_change_detection bool

Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.

stretch_mode str

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

sync_mode str

The Video Sync Mode

complexity String

Tells the encoder how to choose its encoding settings. The default value is Balanced.

keyFrameInterval String

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label String

An optional label for the codec. The label can be used to control muxing behavior.

layers List<Property Map>

The collection of output H.264 layers to be produced by the encoder.

sceneChangeDetection Boolean

Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.

stretchMode String

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

syncMode String

The Video Sync Mode

H265Complexity

Speed
Speed

Tells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.

Balanced
Balanced

Tells the encoder to use settings that achieve a balance between speed and quality.

Quality
Quality

Tells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.

H265ComplexitySpeed
Speed

Tells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.

H265ComplexityBalanced
Balanced

Tells the encoder to use settings that achieve a balance between speed and quality.

H265ComplexityQuality
Quality

Tells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.

Speed
Speed

Tells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.

Balanced
Balanced

Tells the encoder to use settings that achieve a balance between speed and quality.

Quality
Quality

Tells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.

Speed
Speed

Tells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.

Balanced
Balanced

Tells the encoder to use settings that achieve a balance between speed and quality.

Quality
Quality

Tells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.

SPEED
Speed

Tells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.

BALANCED
Balanced

Tells the encoder to use settings that achieve a balance between speed and quality.

QUALITY
Quality

Tells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.

"Speed"
Speed

Tells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.

"Balanced"
Balanced

Tells the encoder to use settings that achieve a balance between speed and quality.

"Quality"
Quality

Tells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.

H265Layer

Bitrate int

The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.

AdaptiveBFrame bool

Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.

BFrames int

The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.

BufferWindow string

The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).

FrameRate string

The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.

Height string

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

Label string

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

Level string

We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.

MaxBitrate int

The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.

Profile string | Pulumi.AzureNative.Media.H265VideoProfile

We currently support Main. Default is Auto.

ReferenceFrames int

The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.

Slices int

The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.

Width string

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

Bitrate int

The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.

AdaptiveBFrame bool

Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.

BFrames int

The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.

BufferWindow string

The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).

FrameRate string

The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.

Height string

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

Label string

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

Level string

We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.

MaxBitrate int

The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.

Profile string | H265VideoProfile

We currently support Main. Default is Auto.

ReferenceFrames int

The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.

Slices int

The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.

Width string

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

bitrate Integer

The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.

adaptiveBFrame Boolean

Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.

bFrames Integer

The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.

bufferWindow String

The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).

frameRate String

The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.

height String

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label String

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

level String

We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.

maxBitrate Integer

The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.

profile String | H265VideoProfile

We currently support Main. Default is Auto.

referenceFrames Integer

The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.

slices Integer

The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.

width String

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

bitrate number

The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.

adaptiveBFrame boolean

Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.

bFrames number

The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.

bufferWindow string

The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).

frameRate string

The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.

height string

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label string

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

level string

We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.

maxBitrate number

The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.

profile string | H265VideoProfile

We currently support Main. Default is Auto.

referenceFrames number

The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.

slices number

The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.

width string

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

bitrate int

The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.

adaptive_b_frame bool

Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.

b_frames int

The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.

buffer_window str

The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).

frame_rate str

The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.

height str

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label str

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

level str

We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.

max_bitrate int

The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.

profile str | H265VideoProfile

We currently support Main. Default is Auto.

reference_frames int

The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.

slices int

The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.

width str

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

bitrate Number

The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.

adaptiveBFrame Boolean

Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.

bFrames Number

The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.

bufferWindow String

The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).

frameRate String

The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.

height String

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label String

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

level String

We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.

maxBitrate Number

The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.

profile String | "Auto" | "Main"

We currently support Main. Default is Auto.

referenceFrames Number

The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.

slices Number

The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.

width String

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

H265LayerResponse

Bitrate int

The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.

AdaptiveBFrame bool

Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.

BFrames int

The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.

BufferWindow string

The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).

FrameRate string

The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.

Height string

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

Label string

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

Level string

We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.

MaxBitrate int

The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.

Profile string

We currently support Main. Default is Auto.

ReferenceFrames int

The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.

Slices int

The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.

Width string

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

Bitrate int

The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.

AdaptiveBFrame bool

Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.

BFrames int

The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.

BufferWindow string

The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).

FrameRate string

The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.

Height string

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

Label string

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

Level string

We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.

MaxBitrate int

The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.

Profile string

We currently support Main. Default is Auto.

ReferenceFrames int

The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.

Slices int

The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.

Width string

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

bitrate Integer

The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.

adaptiveBFrame Boolean

Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.

bFrames Integer

The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.

bufferWindow String

The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).

frameRate String

The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.

height String

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label String

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

level String

We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.

maxBitrate Integer

The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.

profile String

We currently support Main. Default is Auto.

referenceFrames Integer

The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.

slices Integer

The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.

width String

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

bitrate number

The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.

adaptiveBFrame boolean

Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.

bFrames number

The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.

bufferWindow string

The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).

frameRate string

The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.

height string

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label string

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

level string

We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.

maxBitrate number

The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.

profile string

We currently support Main. Default is Auto.

referenceFrames number

The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.

slices number

The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.

width string

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

bitrate int

The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.

adaptive_b_frame bool

Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.

b_frames int

The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.

buffer_window str

The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).

frame_rate str

The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.

height str

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label str

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

level str

We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.

max_bitrate int

The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.

profile str

We currently support Main. Default is Auto.

reference_frames int

The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.

slices int

The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.

width str

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

bitrate Number

The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.

adaptiveBFrame Boolean

Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.

bFrames Number

The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.

bufferWindow String

The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).

frameRate String

The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.

height String

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label String

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

level String

We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.

maxBitrate Number

The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.

profile String

We currently support Main. Default is Auto.

referenceFrames Number

The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.

slices Number

The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.

width String

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

H265Video

Complexity string | Pulumi.AzureNative.Media.H265Complexity

Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.

KeyFrameInterval string

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

Layers List<Pulumi.AzureNative.Media.Inputs.H265Layer>

The collection of output H.265 layers to be produced by the encoder.

SceneChangeDetection bool

Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.

StretchMode string | Pulumi.AzureNative.Media.StretchMode

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

SyncMode string | Pulumi.AzureNative.Media.VideoSyncMode

The Video Sync Mode

Complexity string | H265Complexity

Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.

KeyFrameInterval string

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

Layers []H265Layer

The collection of output H.265 layers to be produced by the encoder.

SceneChangeDetection bool

Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.

StretchMode string | StretchMode

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

SyncMode string | VideoSyncMode

The Video Sync Mode

complexity String | H265Complexity

Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.

keyFrameInterval String

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label String

An optional label for the codec. The label can be used to control muxing behavior.

layers List<H265Layer>

The collection of output H.265 layers to be produced by the encoder.

sceneChangeDetection Boolean

Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.

stretchMode String | StretchMode

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

syncMode String | VideoSyncMode

The Video Sync Mode

complexity string | H265Complexity

Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.

keyFrameInterval string

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label string

An optional label for the codec. The label can be used to control muxing behavior.

layers H265Layer[]

The collection of output H.265 layers to be produced by the encoder.

sceneChangeDetection boolean

Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.

stretchMode string | StretchMode

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

syncMode string | VideoSyncMode

The Video Sync Mode

complexity str | H265Complexity

Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.

key_frame_interval str

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label str

An optional label for the codec. The label can be used to control muxing behavior.

layers Sequence[H265Layer]

The collection of output H.265 layers to be produced by the encoder.

scene_change_detection bool

Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.

stretch_mode str | StretchMode

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

sync_mode str | VideoSyncMode

The Video Sync Mode

complexity String | "Speed" | "Balanced" | "Quality"

Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.

keyFrameInterval String

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label String

An optional label for the codec. The label can be used to control muxing behavior.

layers List<Property Map>

The collection of output H.265 layers to be produced by the encoder.

sceneChangeDetection Boolean

Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.

stretchMode String | "None" | "AutoSize" | "AutoFit"

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

syncMode String | "Auto" | "Passthrough" | "Cfr" | "Vfr"

The Video Sync Mode

H265VideoProfile

Auto
Auto

Tells the encoder to automatically determine the appropriate H.265 profile.

Main
Main

Main profile (https://x265.readthedocs.io/en/default/cli.html?highlight=profile#profile-level-tier)

H265VideoProfileAuto
Auto

Tells the encoder to automatically determine the appropriate H.265 profile.

H265VideoProfileMain
Main

Main profile (https://x265.readthedocs.io/en/default/cli.html?highlight=profile#profile-level-tier)

Auto
Auto

Tells the encoder to automatically determine the appropriate H.265 profile.

Main
Main

Main profile (https://x265.readthedocs.io/en/default/cli.html?highlight=profile#profile-level-tier)

Auto
Auto

Tells the encoder to automatically determine the appropriate H.265 profile.

Main
Main

Main profile (https://x265.readthedocs.io/en/default/cli.html?highlight=profile#profile-level-tier)

AUTO
Auto

Tells the encoder to automatically determine the appropriate H.265 profile.

MAIN
Main

Main profile (https://x265.readthedocs.io/en/default/cli.html?highlight=profile#profile-level-tier)

"Auto"
Auto

Tells the encoder to automatically determine the appropriate H.265 profile.

"Main"
Main

Main profile (https://x265.readthedocs.io/en/default/cli.html?highlight=profile#profile-level-tier)

H265VideoResponse

Complexity string

Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.

KeyFrameInterval string

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

Layers List<Pulumi.AzureNative.Media.Inputs.H265LayerResponse>

The collection of output H.265 layers to be produced by the encoder.

SceneChangeDetection bool

Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.

StretchMode string

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

SyncMode string

The Video Sync Mode

Complexity string

Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.

KeyFrameInterval string

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

Layers []H265LayerResponse

The collection of output H.265 layers to be produced by the encoder.

SceneChangeDetection bool

Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.

StretchMode string

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

SyncMode string

The Video Sync Mode

complexity String

Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.

keyFrameInterval String

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label String

An optional label for the codec. The label can be used to control muxing behavior.

layers List<H265LayerResponse>

The collection of output H.265 layers to be produced by the encoder.

sceneChangeDetection Boolean

Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.

stretchMode String

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

syncMode String

The Video Sync Mode

complexity string

Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.

keyFrameInterval string

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label string

An optional label for the codec. The label can be used to control muxing behavior.

layers H265LayerResponse[]

The collection of output H.265 layers to be produced by the encoder.

sceneChangeDetection boolean

Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.

stretchMode string

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

syncMode string

The Video Sync Mode

complexity str

Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.

key_frame_interval str

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label str

An optional label for the codec. The label can be used to control muxing behavior.

layers Sequence[H265LayerResponse]

The collection of output H.265 layers to be produced by the encoder.

scene_change_detection bool

Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.

stretch_mode str

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

sync_mode str

The Video Sync Mode

complexity String

Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.

keyFrameInterval String

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label String

An optional label for the codec. The label can be used to control muxing behavior.

layers List<Property Map>

The collection of output H.265 layers to be produced by the encoder.

sceneChangeDetection Boolean

Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.

stretchMode String

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

syncMode String

The Video Sync Mode

Image

Start string

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

KeyFrameInterval string

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

Range string

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

Step string

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

StretchMode string | Pulumi.AzureNative.Media.StretchMode

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

SyncMode string | Pulumi.AzureNative.Media.VideoSyncMode

The Video Sync Mode

Start string

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

KeyFrameInterval string

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

Range string

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

Step string

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

StretchMode string | StretchMode

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

SyncMode string | VideoSyncMode

The Video Sync Mode

start String

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

keyFrameInterval String

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label String

An optional label for the codec. The label can be used to control muxing behavior.

range String

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

step String

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

stretchMode String | StretchMode

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

syncMode String | VideoSyncMode

The Video Sync Mode

start string

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

keyFrameInterval string

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label string

An optional label for the codec. The label can be used to control muxing behavior.

range string

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

step string

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

stretchMode string | StretchMode

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

syncMode string | VideoSyncMode

The Video Sync Mode

start str

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

key_frame_interval str

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label str

An optional label for the codec. The label can be used to control muxing behavior.

range str

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

step str

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

stretch_mode str | StretchMode

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

sync_mode str | VideoSyncMode

The Video Sync Mode

start String

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

keyFrameInterval String

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label String

An optional label for the codec. The label can be used to control muxing behavior.

range String

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

step String

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

stretchMode String | "None" | "AutoSize" | "AutoFit"

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

syncMode String | "Auto" | "Passthrough" | "Cfr" | "Vfr"

The Video Sync Mode

ImageFormat

FilenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

FilenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

filenamePattern String

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

filenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

filename_pattern str

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

filenamePattern String

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

ImageFormatResponse

FilenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

FilenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

filenamePattern String

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

filenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

filename_pattern str

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

filenamePattern String

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

ImageResponse

Start string

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

KeyFrameInterval string

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

Range string

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

Step string

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

StretchMode string

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

SyncMode string

The Video Sync Mode

Start string

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

KeyFrameInterval string

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

Range string

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

Step string

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

StretchMode string

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

SyncMode string

The Video Sync Mode

start String

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

keyFrameInterval String

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label String

An optional label for the codec. The label can be used to control muxing behavior.

range String

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

step String

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

stretchMode String

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

syncMode String

The Video Sync Mode

start string

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

keyFrameInterval string

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label string

An optional label for the codec. The label can be used to control muxing behavior.

range string

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

step string

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

stretchMode string

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

syncMode string

The Video Sync Mode

start str

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

key_frame_interval str

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label str

An optional label for the codec. The label can be used to control muxing behavior.

range str

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

step str

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

stretch_mode str

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

sync_mode str

The Video Sync Mode

start String

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

keyFrameInterval String

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label String

An optional label for the codec. The label can be used to control muxing behavior.

range String

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

step String

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

stretchMode String

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

syncMode String

The Video Sync Mode

InsightsType

AudioInsightsOnly
AudioInsightsOnly

Generate audio only insights. Ignore video even if present. Fails if no audio is present.

VideoInsightsOnly
VideoInsightsOnly

Generate video only insights. Ignore audio if present. Fails if no video is present.

AllInsights
AllInsights

Generate both audio and video insights. Fails if either audio or video Insights fail.

InsightsTypeAudioInsightsOnly
AudioInsightsOnly

Generate audio only insights. Ignore video even if present. Fails if no audio is present.

InsightsTypeVideoInsightsOnly
VideoInsightsOnly

Generate video only insights. Ignore audio if present. Fails if no video is present.

InsightsTypeAllInsights
AllInsights

Generate both audio and video insights. Fails if either audio or video Insights fail.

AudioInsightsOnly
AudioInsightsOnly

Generate audio only insights. Ignore video even if present. Fails if no audio is present.

VideoInsightsOnly
VideoInsightsOnly

Generate video only insights. Ignore audio if present. Fails if no video is present.

AllInsights
AllInsights

Generate both audio and video insights. Fails if either audio or video Insights fail.

AudioInsightsOnly
AudioInsightsOnly

Generate audio only insights. Ignore video even if present. Fails if no audio is present.

VideoInsightsOnly
VideoInsightsOnly

Generate video only insights. Ignore audio if present. Fails if no video is present.

AllInsights
AllInsights

Generate both audio and video insights. Fails if either audio or video Insights fail.

AUDIO_INSIGHTS_ONLY
AudioInsightsOnly

Generate audio only insights. Ignore video even if present. Fails if no audio is present.

VIDEO_INSIGHTS_ONLY
VideoInsightsOnly

Generate video only insights. Ignore audio if present. Fails if no video is present.

ALL_INSIGHTS
AllInsights

Generate both audio and video insights. Fails if either audio or video Insights fail.

"AudioInsightsOnly"
AudioInsightsOnly

Generate audio only insights. Ignore video even if present. Fails if no audio is present.

"VideoInsightsOnly"
VideoInsightsOnly

Generate video only insights. Ignore audio if present. Fails if no video is present.

"AllInsights"
AllInsights

Generate both audio and video insights. Fails if either audio or video Insights fail.

JpgFormat

FilenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

FilenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

filenamePattern String

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

filenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

filename_pattern str

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

filenamePattern String

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

JpgFormatResponse

FilenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

FilenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

filenamePattern String

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

filenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

filename_pattern str

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

filenamePattern String

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

JpgImage

Start string

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

KeyFrameInterval string

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

Layers List<Pulumi.AzureNative.Media.Inputs.JpgLayer>

A collection of output JPEG image layers to be produced by the encoder.

Range string

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

SpriteColumn int

Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.

Step string

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

StretchMode string | Pulumi.AzureNative.Media.StretchMode

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

SyncMode string | Pulumi.AzureNative.Media.VideoSyncMode

The Video Sync Mode

Start string

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

KeyFrameInterval string

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

Layers []JpgLayer

A collection of output JPEG image layers to be produced by the encoder.

Range string

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

SpriteColumn int

Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.

Step string

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

StretchMode string | StretchMode

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

SyncMode string | VideoSyncMode

The Video Sync Mode

start String

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

keyFrameInterval String

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label String

An optional label for the codec. The label can be used to control muxing behavior.

layers List<JpgLayer>

A collection of output JPEG image layers to be produced by the encoder.

range String

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

spriteColumn Integer

Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.

step String

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

stretchMode String | StretchMode

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

syncMode String | VideoSyncMode

The Video Sync Mode

start string

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

keyFrameInterval string

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label string

An optional label for the codec. The label can be used to control muxing behavior.

layers JpgLayer[]

A collection of output JPEG image layers to be produced by the encoder.

range string

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

spriteColumn number

Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.

step string

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

stretchMode string | StretchMode

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

syncMode string | VideoSyncMode

The Video Sync Mode

start str

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

key_frame_interval str

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label str

An optional label for the codec. The label can be used to control muxing behavior.

layers Sequence[JpgLayer]

A collection of output JPEG image layers to be produced by the encoder.

range str

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

sprite_column int

Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.

step str

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

stretch_mode str | StretchMode

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

sync_mode str | VideoSyncMode

The Video Sync Mode

start String

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

keyFrameInterval String

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label String

An optional label for the codec. The label can be used to control muxing behavior.

layers List<Property Map>

A collection of output JPEG image layers to be produced by the encoder.

range String

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

spriteColumn Number

Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.

step String

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

stretchMode String | "None" | "AutoSize" | "AutoFit"

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

syncMode String | "Auto" | "Passthrough" | "Cfr" | "Vfr"

The Video Sync Mode

JpgImageResponse

Start string

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

KeyFrameInterval string

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

Layers List<Pulumi.AzureNative.Media.Inputs.JpgLayerResponse>

A collection of output JPEG image layers to be produced by the encoder.

Range string

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

SpriteColumn int

Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.

Step string

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

StretchMode string

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

SyncMode string

The Video Sync Mode

Start string

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

KeyFrameInterval string

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

Layers []JpgLayerResponse

A collection of output JPEG image layers to be produced by the encoder.

Range string

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

SpriteColumn int

Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.

Step string

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

StretchMode string

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

SyncMode string

The Video Sync Mode

start String

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

keyFrameInterval String

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label String

An optional label for the codec. The label can be used to control muxing behavior.

layers List<JpgLayerResponse>

A collection of output JPEG image layers to be produced by the encoder.

range String

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

spriteColumn Integer

Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.

step String

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

stretchMode String

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

syncMode String

The Video Sync Mode

start string

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

keyFrameInterval string

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label string

An optional label for the codec. The label can be used to control muxing behavior.

layers JpgLayerResponse[]

A collection of output JPEG image layers to be produced by the encoder.

range string

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

spriteColumn number

Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.

step string

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

stretchMode string

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

syncMode string

The Video Sync Mode

start str

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

key_frame_interval str

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label str

An optional label for the codec. The label can be used to control muxing behavior.

layers Sequence[JpgLayerResponse]

A collection of output JPEG image layers to be produced by the encoder.

range str

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

sprite_column int

Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.

step str

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

stretch_mode str

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

sync_mode str

The Video Sync Mode

start String

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

keyFrameInterval String

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label String

An optional label for the codec. The label can be used to control muxing behavior.

layers List<Property Map>

A collection of output JPEG image layers to be produced by the encoder.

range String

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

spriteColumn Number

Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.

step String

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

stretchMode String

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

syncMode String

The Video Sync Mode

JpgLayer

Height string

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

Label string

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

Quality int

The compression quality of the JPEG output. Range is from 0-100 and the default is 70.

Width string

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

Height string

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

Label string

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

Quality int

The compression quality of the JPEG output. Range is from 0-100 and the default is 70.

Width string

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

height String

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label String

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

quality Integer

The compression quality of the JPEG output. Range is from 0-100 and the default is 70.

width String

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

height string

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label string

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

quality number

The compression quality of the JPEG output. Range is from 0-100 and the default is 70.

width string

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

height str

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label str

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

quality int

The compression quality of the JPEG output. Range is from 0-100 and the default is 70.

width str

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

height String

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label String

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

quality Number

The compression quality of the JPEG output. Range is from 0-100 and the default is 70.

width String

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

JpgLayerResponse

Height string

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

Label string

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

Quality int

The compression quality of the JPEG output. Range is from 0-100 and the default is 70.

Width string

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

Height string

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

Label string

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

Quality int

The compression quality of the JPEG output. Range is from 0-100 and the default is 70.

Width string

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

height String

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label String

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

quality Integer

The compression quality of the JPEG output. Range is from 0-100 and the default is 70.

width String

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

height string

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label string

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

quality number

The compression quality of the JPEG output. Range is from 0-100 and the default is 70.

width string

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

height str

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label str

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

quality int

The compression quality of the JPEG output. Range is from 0-100 and the default is 70.

width str

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

height String

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label String

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

quality Number

The compression quality of the JPEG output. Range is from 0-100 and the default is 70.

width String

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

Mp4Format

FilenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

OutputFiles List<Pulumi.AzureNative.Media.Inputs.OutputFile>

The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

FilenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

OutputFiles []OutputFile

The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

filenamePattern String

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

outputFiles List<OutputFile>

The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

filenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

outputFiles OutputFile[]

The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

filename_pattern str

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

output_files Sequence[OutputFile]

The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

filenamePattern String

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

outputFiles List<Property Map>

The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

Mp4FormatResponse

FilenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

OutputFiles List<Pulumi.AzureNative.Media.Inputs.OutputFileResponse>

The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

FilenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

OutputFiles []OutputFileResponse

The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

filenamePattern String

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

outputFiles List<OutputFileResponse>

The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

filenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

outputFiles OutputFileResponse[]

The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

filename_pattern str

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

output_files Sequence[OutputFileResponse]

The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

filenamePattern String

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

outputFiles List<Property Map>

The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

MultiBitrateFormat

FilenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

OutputFiles List<Pulumi.AzureNative.Media.Inputs.OutputFile>

The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

FilenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

OutputFiles []OutputFile

The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

filenamePattern String

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

outputFiles List<OutputFile>

The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

filenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

outputFiles OutputFile[]

The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

filename_pattern str

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

output_files Sequence[OutputFile]

The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

filenamePattern String

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

outputFiles List<Property Map>

The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

MultiBitrateFormatResponse

FilenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

OutputFiles List<Pulumi.AzureNative.Media.Inputs.OutputFileResponse>

The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

FilenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

OutputFiles []OutputFileResponse

The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

filenamePattern String

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

outputFiles List<OutputFileResponse>

The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

filenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

outputFiles OutputFileResponse[]

The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

filename_pattern str

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

output_files Sequence[OutputFileResponse]

The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

filenamePattern String

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

outputFiles List<Property Map>

The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

OnErrorType

StopProcessingJob
StopProcessingJob

Tells the service that if this TransformOutput fails, then any other incomplete TransformOutputs can be stopped.

ContinueJob
ContinueJob

Tells the service that if this TransformOutput fails, then allow any other TransformOutput to continue.

OnErrorTypeStopProcessingJob
StopProcessingJob

Tells the service that if this TransformOutput fails, then any other incomplete TransformOutputs can be stopped.

OnErrorTypeContinueJob
ContinueJob

Tells the service that if this TransformOutput fails, then allow any other TransformOutput to continue.

StopProcessingJob
StopProcessingJob

Tells the service that if this TransformOutput fails, then any other incomplete TransformOutputs can be stopped.

ContinueJob
ContinueJob

Tells the service that if this TransformOutput fails, then allow any other TransformOutput to continue.

StopProcessingJob
StopProcessingJob

Tells the service that if this TransformOutput fails, then any other incomplete TransformOutputs can be stopped.

ContinueJob
ContinueJob

Tells the service that if this TransformOutput fails, then allow any other TransformOutput to continue.

STOP_PROCESSING_JOB
StopProcessingJob

Tells the service that if this TransformOutput fails, then any other incomplete TransformOutputs can be stopped.

CONTINUE_JOB
ContinueJob

Tells the service that if this TransformOutput fails, then allow any other TransformOutput to continue.

"StopProcessingJob"
StopProcessingJob

Tells the service that if this TransformOutput fails, then any other incomplete TransformOutputs can be stopped.

"ContinueJob"
ContinueJob

Tells the service that if this TransformOutput fails, then allow any other TransformOutput to continue.

OutputFile

Labels List<string>

The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.

Labels []string

The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.

labels List<String>

The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.

labels string[]

The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.

labels Sequence[str]

The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.

labels List<String>

The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.

OutputFileResponse

Labels List<string>

The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.

Labels []string

The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.

labels List<String>

The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.

labels string[]

The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.

labels Sequence[str]

The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.

labels List<String>

The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.

PngFormat

FilenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

FilenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

filenamePattern String

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

filenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

filename_pattern str

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

filenamePattern String

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

PngFormatResponse

FilenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

FilenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

filenamePattern String

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

filenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

filename_pattern str

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

filenamePattern String

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

PngImage

Start string

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

KeyFrameInterval string

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

Layers List<Pulumi.AzureNative.Media.Inputs.PngLayer>

A collection of output PNG image layers to be produced by the encoder.

Range string

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

Step string

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

StretchMode string | Pulumi.AzureNative.Media.StretchMode

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

SyncMode string | Pulumi.AzureNative.Media.VideoSyncMode

The Video Sync Mode

Start string

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

KeyFrameInterval string

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

Layers []PngLayer

A collection of output PNG image layers to be produced by the encoder.

Range string

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

Step string

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

StretchMode string | StretchMode

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

SyncMode string | VideoSyncMode

The Video Sync Mode

start String

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

keyFrameInterval String

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label String

An optional label for the codec. The label can be used to control muxing behavior.

layers List<PngLayer>

A collection of output PNG image layers to be produced by the encoder.

range String

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

step String

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

stretchMode String | StretchMode

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

syncMode String | VideoSyncMode

The Video Sync Mode

start string

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

keyFrameInterval string

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label string

An optional label for the codec. The label can be used to control muxing behavior.

layers PngLayer[]

A collection of output PNG image layers to be produced by the encoder.

range string

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

step string

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

stretchMode string | StretchMode

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

syncMode string | VideoSyncMode

The Video Sync Mode

start str

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

key_frame_interval str

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label str

An optional label for the codec. The label can be used to control muxing behavior.

layers Sequence[PngLayer]

A collection of output PNG image layers to be produced by the encoder.

range str

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

step str

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

stretch_mode str | StretchMode

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

sync_mode str | VideoSyncMode

The Video Sync Mode

start String

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

keyFrameInterval String

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label String

An optional label for the codec. The label can be used to control muxing behavior.

layers List<Property Map>

A collection of output PNG image layers to be produced by the encoder.

range String

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

step String

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

stretchMode String | "None" | "AutoSize" | "AutoFit"

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

syncMode String | "Auto" | "Passthrough" | "Cfr" | "Vfr"

The Video Sync Mode

PngImageResponse

Start string

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

KeyFrameInterval string

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

Layers List<Pulumi.AzureNative.Media.Inputs.PngLayerResponse>

A collection of output PNG image layers to be produced by the encoder.

Range string

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

Step string

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

StretchMode string

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

SyncMode string

The Video Sync Mode

Start string

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

KeyFrameInterval string

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

Label string

An optional label for the codec. The label can be used to control muxing behavior.

Layers []PngLayerResponse

A collection of output PNG image layers to be produced by the encoder.

Range string

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

Step string

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

StretchMode string

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

SyncMode string

The Video Sync Mode

start String

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

keyFrameInterval String

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label String

An optional label for the codec. The label can be used to control muxing behavior.

layers List<PngLayerResponse>

A collection of output PNG image layers to be produced by the encoder.

range String

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

step String

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

stretchMode String

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

syncMode String

The Video Sync Mode

start string

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

keyFrameInterval string

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label string

An optional label for the codec. The label can be used to control muxing behavior.

layers PngLayerResponse[]

A collection of output PNG image layers to be produced by the encoder.

range string

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

step string

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

stretchMode string

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

syncMode string

The Video Sync Mode

start str

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

key_frame_interval str

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label str

An optional label for the codec. The label can be used to control muxing behavior.

layers Sequence[PngLayerResponse]

A collection of output PNG image layers to be produced by the encoder.

range str

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

step str

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

stretch_mode str

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

sync_mode str

The Video Sync Mode

start String

The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.

keyFrameInterval String

The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.

label String

An optional label for the codec. The label can be used to control muxing behavior.

layers List<Property Map>

A collection of output PNG image layers to be produced by the encoder.

range String

The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.

step String

The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.

stretchMode String

The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize

syncMode String

The Video Sync Mode

PngLayer

Height string

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

Label string

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

Width string

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

Height string

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

Label string

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

Width string

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

height String

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label String

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

width String

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

height string

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label string

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

width string

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

height str

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label str

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

width str

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

height String

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label String

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

width String

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

PngLayerResponse

Height string

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

Label string

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

Width string

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

Height string

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

Label string

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

Width string

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

height String

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label String

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

width String

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

height string

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label string

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

width string

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

height str

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label str

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

width str

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

height String

The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.

label String

The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.

width String

The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

Priority

Low
Low

Used for TransformOutputs that can be generated after Normal and High priority TransformOutputs.

Normal
Normal

Used for TransformOutputs that can be generated at Normal priority.

High
High

Used for TransformOutputs that should take precedence over others.

PriorityLow
Low

Used for TransformOutputs that can be generated after Normal and High priority TransformOutputs.

PriorityNormal
Normal

Used for TransformOutputs that can be generated at Normal priority.

PriorityHigh
High

Used for TransformOutputs that should take precedence over others.

Low
Low

Used for TransformOutputs that can be generated after Normal and High priority TransformOutputs.

Normal
Normal

Used for TransformOutputs that can be generated at Normal priority.

High
High

Used for TransformOutputs that should take precedence over others.

Low
Low

Used for TransformOutputs that can be generated after Normal and High priority TransformOutputs.

Normal
Normal

Used for TransformOutputs that can be generated at Normal priority.

High
High

Used for TransformOutputs that should take precedence over others.

LOW
Low

Used for TransformOutputs that can be generated after Normal and High priority TransformOutputs.

NORMAL
Normal

Used for TransformOutputs that can be generated at Normal priority.

HIGH
High

Used for TransformOutputs that should take precedence over others.

"Low"
Low

Used for TransformOutputs that can be generated after Normal and High priority TransformOutputs.

"Normal"
Normal

Used for TransformOutputs that can be generated at Normal priority.

"High"
High

Used for TransformOutputs that should take precedence over others.

Rectangle

Height string

The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

Left string

The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

Top string

The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

Width string

The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

Height string

The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

Left string

The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

Top string

The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

Width string

The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

height String

The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

left String

The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

top String

The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

width String

The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

height string

The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

left string

The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

top string

The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

width string

The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

height str

The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

left str

The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

top str

The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

width str

The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

height String

The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

left String

The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

top String

The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

width String

The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

RectangleResponse

Height string

The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

Left string

The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

Top string

The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

Width string

The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

Height string

The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

Left string

The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

Top string

The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

Width string

The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

height String

The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

left String

The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

top String

The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

width String

The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

height string

The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

left string

The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

top string

The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

width string

The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

height str

The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

left str

The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

top str

The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

width str

The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

height String

The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

left String

The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

top String

The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

width String

The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

Rotation

Auto
Auto

Automatically detect and rotate as needed.

None
None

Do not rotate the video. If the output format supports it, any metadata about rotation is kept intact.

Rotate0
Rotate0

Do not rotate the video but remove any metadata about the rotation.

Rotate90
Rotate90

Rotate 90 degrees clockwise.

Rotate180
Rotate180

Rotate 180 degrees clockwise.

Rotate270
Rotate270

Rotate 270 degrees clockwise.

RotationAuto
Auto

Automatically detect and rotate as needed.

RotationNone
None

Do not rotate the video. If the output format supports it, any metadata about rotation is kept intact.

RotationRotate0
Rotate0

Do not rotate the video but remove any metadata about the rotation.

RotationRotate90
Rotate90

Rotate 90 degrees clockwise.

RotationRotate180
Rotate180

Rotate 180 degrees clockwise.

RotationRotate270
Rotate270

Rotate 270 degrees clockwise.

Auto
Auto

Automatically detect and rotate as needed.

None
None

Do not rotate the video. If the output format supports it, any metadata about rotation is kept intact.

Rotate0
Rotate0

Do not rotate the video but remove any metadata about the rotation.

Rotate90
Rotate90

Rotate 90 degrees clockwise.

Rotate180
Rotate180

Rotate 180 degrees clockwise.

Rotate270
Rotate270

Rotate 270 degrees clockwise.

Auto
Auto

Automatically detect and rotate as needed.

None
None

Do not rotate the video. If the output format supports it, any metadata about rotation is kept intact.

Rotate0
Rotate0

Do not rotate the video but remove any metadata about the rotation.

Rotate90
Rotate90

Rotate 90 degrees clockwise.

Rotate180
Rotate180

Rotate 180 degrees clockwise.

Rotate270
Rotate270

Rotate 270 degrees clockwise.

AUTO
Auto

Automatically detect and rotate as needed.

NONE
None

Do not rotate the video. If the output format supports it, any metadata about rotation is kept intact.

ROTATE0
Rotate0

Do not rotate the video but remove any metadata about the rotation.

ROTATE90
Rotate90

Rotate 90 degrees clockwise.

ROTATE180
Rotate180

Rotate 180 degrees clockwise.

ROTATE270
Rotate270

Rotate 270 degrees clockwise.

"Auto"
Auto

Automatically detect and rotate as needed.

"None"
None

Do not rotate the video. If the output format supports it, any metadata about rotation is kept intact.

"Rotate0"
Rotate0

Do not rotate the video but remove any metadata about the rotation.

"Rotate90"
Rotate90

Rotate 90 degrees clockwise.

"Rotate180"
Rotate180

Rotate 180 degrees clockwise.

"Rotate270"
Rotate270

Rotate 270 degrees clockwise.

StandardEncoderPreset

Codecs List<object>

The list of codecs to be used when encoding the input video.

Formats List<object>

The list of outputs to be produced by the encoder.

Filters Pulumi.AzureNative.Media.Inputs.Filters

One or more filtering operations that are applied to the input media before encoding.

Codecs []interface{}

The list of codecs to be used when encoding the input video.

Formats []interface{}

The list of outputs to be produced by the encoder.

Filters Filters

One or more filtering operations that are applied to the input media before encoding.

codecs List<Object>

The list of codecs to be used when encoding the input video.

formats List<Object>

The list of outputs to be produced by the encoder.

filters Filters

One or more filtering operations that are applied to the input media before encoding.

codecs (AacAudio | Audio | CopyAudio | CopyVideo | H264Video | H265Video | Image | JpgImage | PngImage | Video)[]

The list of codecs to be used when encoding the input video.

formats (ImageFormat | JpgFormat | Mp4Format | MultiBitrateFormat | PngFormat | TransportStreamFormat)[]

The list of outputs to be produced by the encoder.

filters Filters

One or more filtering operations that are applied to the input media before encoding.

codecs Sequence[Union[AacAudio, Audio, CopyAudio, CopyVideo, H264Video, H265Video, Image, JpgImage, PngImage, Video]]

The list of codecs to be used when encoding the input video.

formats Sequence[Union[ImageFormat, JpgFormat, Mp4Format, MultiBitrateFormat, PngFormat, TransportStreamFormat]]

The list of outputs to be produced by the encoder.

filters Filters

One or more filtering operations that are applied to the input media before encoding.

codecs List<Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map>

The list of codecs to be used when encoding the input video.

formats List<Property Map | Property Map | Property Map | Property Map | Property Map | Property Map>

The list of outputs to be produced by the encoder.

filters Property Map

One or more filtering operations that are applied to the input media before encoding.

StandardEncoderPresetResponse

Codecs List<object>

The list of codecs to be used when encoding the input video.

Formats List<object>

The list of outputs to be produced by the encoder.

Filters Pulumi.AzureNative.Media.Inputs.FiltersResponse

One or more filtering operations that are applied to the input media before encoding.

Codecs []interface{}

The list of codecs to be used when encoding the input video.

Formats []interface{}

The list of outputs to be produced by the encoder.

Filters FiltersResponse

One or more filtering operations that are applied to the input media before encoding.

codecs List<Object>

The list of codecs to be used when encoding the input video.

formats List<Object>

The list of outputs to be produced by the encoder.

filters FiltersResponse

One or more filtering operations that are applied to the input media before encoding.

codecs (AacAudioResponse | AudioResponse | CopyAudioResponse | CopyVideoResponse | H264VideoResponse | H265VideoResponse | ImageResponse | JpgImageResponse | PngImageResponse | VideoResponse)[]

The list of codecs to be used when encoding the input video.

formats (ImageFormatResponse | JpgFormatResponse | Mp4FormatResponse | MultiBitrateFormatResponse | PngFormatResponse | TransportStreamFormatResponse)[]

The list of outputs to be produced by the encoder.

filters FiltersResponse

One or more filtering operations that are applied to the input media before encoding.

codecs Sequence[Union[AacAudioResponse, AudioResponse, CopyAudioResponse, CopyVideoResponse, H264VideoResponse, H265VideoResponse, ImageResponse, JpgImageResponse, PngImageResponse, VideoResponse]]

The list of codecs to be used when encoding the input video.

formats Sequence[Union[ImageFormatResponse, JpgFormatResponse, Mp4FormatResponse, MultiBitrateFormatResponse, PngFormatResponse, TransportStreamFormatResponse]]

The list of outputs to be produced by the encoder.

filters FiltersResponse

One or more filtering operations that are applied to the input media before encoding.

codecs List<Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map>

The list of codecs to be used when encoding the input video.

formats List<Property Map | Property Map | Property Map | Property Map | Property Map | Property Map>

The list of outputs to be produced by the encoder.

filters Property Map

One or more filtering operations that are applied to the input media before encoding.

StretchMode

None
None

Strictly respect the output resolution without considering the pixel aspect ratio or display aspect ratio of the input video.

AutoSize
AutoSize

Override the output resolution, and change it to match the display aspect ratio of the input, without padding. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the value in the preset is overridden, and the output will be at 1280x720, which maintains the input aspect ratio of 16:9.

AutoFit
AutoFit

Pad the output (with either letterbox or pillar box) to honor the output resolution, while ensuring that the active video region in the output has the same aspect ratio as the input. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the output will be at 1280x1280, which contains an inner rectangle of 1280x720 at aspect ratio of 16:9, and pillar box regions 280 pixels wide at the left and right.

StretchModeNone
None

Strictly respect the output resolution without considering the pixel aspect ratio or display aspect ratio of the input video.

StretchModeAutoSize
AutoSize

Override the output resolution, and change it to match the display aspect ratio of the input, without padding. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the value in the preset is overridden, and the output will be at 1280x720, which maintains the input aspect ratio of 16:9.

StretchModeAutoFit
AutoFit

Pad the output (with either letterbox or pillar box) to honor the output resolution, while ensuring that the active video region in the output has the same aspect ratio as the input. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the output will be at 1280x1280, which contains an inner rectangle of 1280x720 at aspect ratio of 16:9, and pillar box regions 280 pixels wide at the left and right.

None
None

Strictly respect the output resolution without considering the pixel aspect ratio or display aspect ratio of the input video.

AutoSize
AutoSize

Override the output resolution, and change it to match the display aspect ratio of the input, without padding. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the value in the preset is overridden, and the output will be at 1280x720, which maintains the input aspect ratio of 16:9.

AutoFit
AutoFit

Pad the output (with either letterbox or pillar box) to honor the output resolution, while ensuring that the active video region in the output has the same aspect ratio as the input. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the output will be at 1280x1280, which contains an inner rectangle of 1280x720 at aspect ratio of 16:9, and pillar box regions 280 pixels wide at the left and right.

None
None

Strictly respect the output resolution without considering the pixel aspect ratio or display aspect ratio of the input video.

AutoSize
AutoSize

Override the output resolution, and change it to match the display aspect ratio of the input, without padding. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the value in the preset is overridden, and the output will be at 1280x720, which maintains the input aspect ratio of 16:9.

AutoFit
AutoFit

Pad the output (with either letterbox or pillar box) to honor the output resolution, while ensuring that the active video region in the output has the same aspect ratio as the input. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the output will be at 1280x1280, which contains an inner rectangle of 1280x720 at aspect ratio of 16:9, and pillar box regions 280 pixels wide at the left and right.

NONE
None

Strictly respect the output resolution without considering the pixel aspect ratio or display aspect ratio of the input video.

AUTO_SIZE
AutoSize

Override the output resolution, and change it to match the display aspect ratio of the input, without padding. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the value in the preset is overridden, and the output will be at 1280x720, which maintains the input aspect ratio of 16:9.

AUTO_FIT
AutoFit

Pad the output (with either letterbox or pillar box) to honor the output resolution, while ensuring that the active video region in the output has the same aspect ratio as the input. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the output will be at 1280x1280, which contains an inner rectangle of 1280x720 at aspect ratio of 16:9, and pillar box regions 280 pixels wide at the left and right.

"None"
None

Strictly respect the output resolution without considering the pixel aspect ratio or display aspect ratio of the input video.

"AutoSize"
AutoSize

Override the output resolution, and change it to match the display aspect ratio of the input, without padding. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the value in the preset is overridden, and the output will be at 1280x720, which maintains the input aspect ratio of 16:9.

"AutoFit"
AutoFit

Pad the output (with either letterbox or pillar box) to honor the output resolution, while ensuring that the active video region in the output has the same aspect ratio as the input. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the output will be at 1280x1280, which contains an inner rectangle of 1280x720 at aspect ratio of 16:9, and pillar box regions 280 pixels wide at the left and right.

SystemDataResponse

CreatedAt string

The timestamp of resource creation (UTC).

CreatedBy string

The identity that created the resource.

CreatedByType string

The type of identity that created the resource.

LastModifiedAt string

The timestamp of resource last modification (UTC)

LastModifiedBy string

The identity that last modified the resource.

LastModifiedByType string

The type of identity that last modified the resource.

CreatedAt string

The timestamp of resource creation (UTC).

CreatedBy string

The identity that created the resource.

CreatedByType string

The type of identity that created the resource.

LastModifiedAt string

The timestamp of resource last modification (UTC)

LastModifiedBy string

The identity that last modified the resource.

LastModifiedByType string

The type of identity that last modified the resource.

createdAt String

The timestamp of resource creation (UTC).

createdBy String

The identity that created the resource.

createdByType String

The type of identity that created the resource.

lastModifiedAt String

The timestamp of resource last modification (UTC)

lastModifiedBy String

The identity that last modified the resource.

lastModifiedByType String

The type of identity that last modified the resource.

createdAt string

The timestamp of resource creation (UTC).

createdBy string

The identity that created the resource.

createdByType string

The type of identity that created the resource.

lastModifiedAt string

The timestamp of resource last modification (UTC)

lastModifiedBy string

The identity that last modified the resource.

lastModifiedByType string

The type of identity that last modified the resource.

created_at str

The timestamp of resource creation (UTC).

created_by str

The identity that created the resource.

created_by_type str

The type of identity that created the resource.

last_modified_at str

The timestamp of resource last modification (UTC)

last_modified_by str

The identity that last modified the resource.

last_modified_by_type str

The type of identity that last modified the resource.

createdAt String

The timestamp of resource creation (UTC).

createdBy String

The identity that created the resource.

createdByType String

The type of identity that created the resource.

lastModifiedAt String

The timestamp of resource last modification (UTC)

lastModifiedBy String

The identity that last modified the resource.

lastModifiedByType String

The type of identity that last modified the resource.

TransformOutput

Preset Pulumi.AzureNative.Media.Inputs.AudioAnalyzerPreset | Pulumi.AzureNative.Media.Inputs.BuiltInStandardEncoderPreset | Pulumi.AzureNative.Media.Inputs.FaceDetectorPreset | Pulumi.AzureNative.Media.Inputs.StandardEncoderPreset | Pulumi.AzureNative.Media.Inputs.VideoAnalyzerPreset

Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.

OnError string | Pulumi.AzureNative.Media.OnErrorType

A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.

RelativePriority string | Pulumi.AzureNative.Media.Priority

Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.

Preset AudioAnalyzerPreset | BuiltInStandardEncoderPreset | FaceDetectorPreset | StandardEncoderPreset | VideoAnalyzerPreset

Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.

OnError string | OnErrorType

A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.

RelativePriority string | Priority

Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.

preset AudioAnalyzerPreset | BuiltInStandardEncoderPreset | FaceDetectorPreset | StandardEncoderPreset | VideoAnalyzerPreset

Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.

onError String | OnErrorType

A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.

relativePriority String | Priority

Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.

preset AudioAnalyzerPreset | BuiltInStandardEncoderPreset | FaceDetectorPreset | StandardEncoderPreset | VideoAnalyzerPreset

Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.

onError string | OnErrorType

A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.

relativePriority string | Priority

Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.

preset AudioAnalyzerPreset | BuiltInStandardEncoderPreset | FaceDetectorPreset | StandardEncoderPreset | VideoAnalyzerPreset

Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.

on_error str | OnErrorType

A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.

relative_priority str | Priority

Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.

preset Property Map | Property Map | Property Map | Property Map | Property Map

Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.

onError String | "StopProcessingJob" | "ContinueJob"

A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.

relativePriority String | "Low" | "Normal" | "High"

Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.

TransformOutputResponse

Preset Pulumi.AzureNative.Media.Inputs.AudioAnalyzerPresetResponse | Pulumi.AzureNative.Media.Inputs.BuiltInStandardEncoderPresetResponse | Pulumi.AzureNative.Media.Inputs.FaceDetectorPresetResponse | Pulumi.AzureNative.Media.Inputs.StandardEncoderPresetResponse | Pulumi.AzureNative.Media.Inputs.VideoAnalyzerPresetResponse

Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.

OnError string

A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.

RelativePriority string

Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.

Preset AudioAnalyzerPresetResponse | BuiltInStandardEncoderPresetResponse | FaceDetectorPresetResponse | StandardEncoderPresetResponse | VideoAnalyzerPresetResponse

Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.

OnError string

A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.

RelativePriority string

Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.

preset AudioAnalyzerPresetResponse | BuiltInStandardEncoderPresetResponse | FaceDetectorPresetResponse | StandardEncoderPresetResponse | VideoAnalyzerPresetResponse

Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.

onError String

A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.

relativePriority String

Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.

preset AudioAnalyzerPresetResponse | BuiltInStandardEncoderPresetResponse | FaceDetectorPresetResponse | StandardEncoderPresetResponse | VideoAnalyzerPresetResponse

Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.

onError string

A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.

relativePriority string

Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.

preset AudioAnalyzerPresetResponse | BuiltInStandardEncoderPresetResponse | FaceDetectorPresetResponse | StandardEncoderPresetResponse | VideoAnalyzerPresetResponse

Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.

on_error str

A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.

relative_priority str

Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.

preset Property Map | Property Map | Property Map | Property Map | Property Map

Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.

onError String

A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.

relativePriority String

Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.

TransportStreamFormat

FilenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

OutputFiles List<Pulumi.AzureNative.Media.Inputs.OutputFile>

The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

FilenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

OutputFiles []OutputFile

The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

filenamePattern String

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

outputFiles List<OutputFile>

The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

filenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

outputFiles OutputFile[]

The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

filename_pattern str

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

output_files Sequence[OutputFile]

The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

filenamePattern String

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

outputFiles List<Property Map>

The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

TransportStreamFormatResponse

FilenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

OutputFiles List<Pulumi.AzureNative.Media.Inputs.OutputFileResponse>

The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

FilenamePattern string

The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.