azure-native.media.Transform
A Transform encapsulates the rules or instructions for generating desired outputs from input media, such as by transcoding or by extracting insights. After the Transform is created, it can be applied to input media by creating Jobs. API Version: 2020-05-01.
Example Usage
Create or update a Transform
using System.Collections.Generic;
using Pulumi;
using AzureNative = Pulumi.AzureNative;
return await Deployment.RunAsync(() =>
{
var transform = new AzureNative.Media.Transform("transform", new()
{
AccountName = "contosomedia",
Description = "Example Transform to illustrate create and update.",
Outputs = new[]
{
new AzureNative.Media.Inputs.TransformOutputArgs
{
Preset = new AzureNative.Media.Inputs.BuiltInStandardEncoderPresetArgs
{
OdataType = "#Microsoft.Media.BuiltInStandardEncoderPreset",
PresetName = "AdaptiveStreaming",
},
},
},
ResourceGroupName = "contosoresources",
TransformName = "createdTransform",
});
});
package main
import (
media "github.com/pulumi/pulumi-azure-native/sdk/go/azure/media"
"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
)
func main() {
pulumi.Run(func(ctx *pulumi.Context) error {
_, err := media.NewTransform(ctx, "transform", &media.TransformArgs{
AccountName: pulumi.String("contosomedia"),
Description: pulumi.String("Example Transform to illustrate create and update."),
Outputs: []media.TransformOutputTypeArgs{
{
Preset: {
OdataType: "#Microsoft.Media.BuiltInStandardEncoderPreset",
PresetName: "AdaptiveStreaming",
},
},
},
ResourceGroupName: pulumi.String("contosoresources"),
TransformName: pulumi.String("createdTransform"),
})
if err != nil {
return err
}
return nil
})
}
package generated_program;
import com.pulumi.Context;
import com.pulumi.Pulumi;
import com.pulumi.core.Output;
import com.pulumi.azurenative.media.Transform;
import com.pulumi.azurenative.media.TransformArgs;
import java.util.List;
import java.util.ArrayList;
import java.util.Map;
import java.io.File;
import java.nio.file.Files;
import java.nio.file.Paths;
public class App {
public static void main(String[] args) {
Pulumi.run(App::stack);
}
public static void stack(Context ctx) {
var transform = new Transform("transform", TransformArgs.builder()
.accountName("contosomedia")
.description("Example Transform to illustrate create and update.")
.outputs(Map.of("preset", Map.ofEntries(
Map.entry("odataType", "#Microsoft.Media.BuiltInStandardEncoderPreset"),
Map.entry("presetName", "AdaptiveStreaming")
)))
.resourceGroupName("contosoresources")
.transformName("createdTransform")
.build());
}
}
import pulumi
import pulumi_azure_native as azure_native
transform = azure_native.media.Transform("transform",
account_name="contosomedia",
description="Example Transform to illustrate create and update.",
outputs=[azure_native.media.TransformOutputArgs(
preset=azure_native.media.BuiltInStandardEncoderPresetArgs(
odata_type="#Microsoft.Media.BuiltInStandardEncoderPreset",
preset_name="AdaptiveStreaming",
),
)],
resource_group_name="contosoresources",
transform_name="createdTransform")
import * as pulumi from "@pulumi/pulumi";
import * as azure_native from "@pulumi/azure-native";
const transform = new azure_native.media.Transform("transform", {
accountName: "contosomedia",
description: "Example Transform to illustrate create and update.",
outputs: [{
preset: {
odataType: "#Microsoft.Media.BuiltInStandardEncoderPreset",
presetName: "AdaptiveStreaming",
},
}],
resourceGroupName: "contosoresources",
transformName: "createdTransform",
});
resources:
transform:
type: azure-native:media:Transform
properties:
accountName: contosomedia
description: Example Transform to illustrate create and update.
outputs:
- preset:
odataType: '#Microsoft.Media.BuiltInStandardEncoderPreset'
presetName: AdaptiveStreaming
resourceGroupName: contosoresources
transformName: createdTransform
Create Transform Resource
new Transform(name: string, args: TransformArgs, opts?: CustomResourceOptions);
@overload
def Transform(resource_name: str,
opts: Optional[ResourceOptions] = None,
account_name: Optional[str] = None,
description: Optional[str] = None,
outputs: Optional[Sequence[TransformOutputArgs]] = None,
resource_group_name: Optional[str] = None,
transform_name: Optional[str] = None)
@overload
def Transform(resource_name: str,
args: TransformArgs,
opts: Optional[ResourceOptions] = None)
func NewTransform(ctx *Context, name string, args TransformArgs, opts ...ResourceOption) (*Transform, error)
public Transform(string name, TransformArgs args, CustomResourceOptions? opts = null)
public Transform(String name, TransformArgs args)
public Transform(String name, TransformArgs args, CustomResourceOptions options)
type: azure-native:media:Transform
properties: # The arguments to resource properties.
options: # Bag of options to control resource's behavior.
- name string
- The unique name of the resource.
- args TransformArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- resource_name str
- The unique name of the resource.
- args TransformArgs
- The arguments to resource properties.
- opts ResourceOptions
- Bag of options to control resource's behavior.
- ctx Context
- Context object for the current deployment.
- name string
- The unique name of the resource.
- args TransformArgs
- The arguments to resource properties.
- opts ResourceOption
- Bag of options to control resource's behavior.
- name string
- The unique name of the resource.
- args TransformArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- name String
- The unique name of the resource.
- args TransformArgs
- The arguments to resource properties.
- options CustomResourceOptions
- Bag of options to control resource's behavior.
Transform Resource Properties
To learn more about resource properties and how to use them, see Inputs and Outputs in the Architecture and Concepts docs.
Inputs
The Transform resource accepts the following input properties:
- Account
Name string The Media Services account name.
- Outputs
List<Pulumi.
Azure Native. Media. Inputs. Transform Output Args> An array of one or more TransformOutputs that the Transform should generate.
- Resource
Group stringName The name of the resource group within the Azure subscription.
- Description string
An optional verbose description of the Transform.
- Transform
Name string The Transform name.
- Account
Name string The Media Services account name.
- Outputs
[]Transform
Output Type Args An array of one or more TransformOutputs that the Transform should generate.
- Resource
Group stringName The name of the resource group within the Azure subscription.
- Description string
An optional verbose description of the Transform.
- Transform
Name string The Transform name.
- account
Name String The Media Services account name.
- outputs
List<Transform
Output Args> An array of one or more TransformOutputs that the Transform should generate.
- resource
Group StringName The name of the resource group within the Azure subscription.
- description String
An optional verbose description of the Transform.
- transform
Name String The Transform name.
- account
Name string The Media Services account name.
- outputs
Transform
Output Args[] An array of one or more TransformOutputs that the Transform should generate.
- resource
Group stringName The name of the resource group within the Azure subscription.
- description string
An optional verbose description of the Transform.
- transform
Name string The Transform name.
- account_
name str The Media Services account name.
- outputs
Sequence[Transform
Output Args] An array of one or more TransformOutputs that the Transform should generate.
- resource_
group_ strname The name of the resource group within the Azure subscription.
- description str
An optional verbose description of the Transform.
- transform_
name str The Transform name.
- account
Name String The Media Services account name.
- outputs List<Property Map>
An array of one or more TransformOutputs that the Transform should generate.
- resource
Group StringName The name of the resource group within the Azure subscription.
- description String
An optional verbose description of the Transform.
- transform
Name String The Transform name.
Outputs
All input properties are implicitly available as output properties. Additionally, the Transform resource produces the following output properties:
- Created string
The UTC date and time when the Transform was created, in 'YYYY-MM-DDThh:mm:ssZ' format.
- Id string
The provider-assigned unique ID for this managed resource.
- Last
Modified string The UTC date and time when the Transform was last updated, in 'YYYY-MM-DDThh:mm:ssZ' format.
- Name string
The name of the resource
- System
Data Pulumi.Azure Native. Media. Outputs. System Data Response The system metadata relating to this resource.
- Type string
The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or "Microsoft.Storage/storageAccounts"
- Created string
The UTC date and time when the Transform was created, in 'YYYY-MM-DDThh:mm:ssZ' format.
- Id string
The provider-assigned unique ID for this managed resource.
- Last
Modified string The UTC date and time when the Transform was last updated, in 'YYYY-MM-DDThh:mm:ssZ' format.
- Name string
The name of the resource
- System
Data SystemData Response The system metadata relating to this resource.
- Type string
The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or "Microsoft.Storage/storageAccounts"
- created String
The UTC date and time when the Transform was created, in 'YYYY-MM-DDThh:mm:ssZ' format.
- id String
The provider-assigned unique ID for this managed resource.
- last
Modified String The UTC date and time when the Transform was last updated, in 'YYYY-MM-DDThh:mm:ssZ' format.
- name String
The name of the resource
- system
Data SystemData Response The system metadata relating to this resource.
- type String
The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or "Microsoft.Storage/storageAccounts"
- created string
The UTC date and time when the Transform was created, in 'YYYY-MM-DDThh:mm:ssZ' format.
- id string
The provider-assigned unique ID for this managed resource.
- last
Modified string The UTC date and time when the Transform was last updated, in 'YYYY-MM-DDThh:mm:ssZ' format.
- name string
The name of the resource
- system
Data SystemData Response The system metadata relating to this resource.
- type string
The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or "Microsoft.Storage/storageAccounts"
- created str
The UTC date and time when the Transform was created, in 'YYYY-MM-DDThh:mm:ssZ' format.
- id str
The provider-assigned unique ID for this managed resource.
- last_
modified str The UTC date and time when the Transform was last updated, in 'YYYY-MM-DDThh:mm:ssZ' format.
- name str
The name of the resource
- system_
data SystemData Response The system metadata relating to this resource.
- type str
The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or "Microsoft.Storage/storageAccounts"
- created String
The UTC date and time when the Transform was created, in 'YYYY-MM-DDThh:mm:ssZ' format.
- id String
The provider-assigned unique ID for this managed resource.
- last
Modified String The UTC date and time when the Transform was last updated, in 'YYYY-MM-DDThh:mm:ssZ' format.
- name String
The name of the resource
- system
Data Property Map The system metadata relating to this resource.
- type String
The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or "Microsoft.Storage/storageAccounts"
Supporting Types
AacAudio
- Bitrate int
The bitrate, in bits per second, of the output encoded audio.
- Channels int
The number of channels in the audio.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Profile
string | Pulumi.
Azure Native. Media. Aac Audio Profile The encoding profile to be used when encoding audio with AAC.
- Sampling
Rate int The sampling rate to use for encoding in hertz.
- Bitrate int
The bitrate, in bits per second, of the output encoded audio.
- Channels int
The number of channels in the audio.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Profile
string | Aac
Audio Profile The encoding profile to be used when encoding audio with AAC.
- Sampling
Rate int The sampling rate to use for encoding in hertz.
- bitrate Integer
The bitrate, in bits per second, of the output encoded audio.
- channels Integer
The number of channels in the audio.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- profile
String | Aac
Audio Profile The encoding profile to be used when encoding audio with AAC.
- sampling
Rate Integer The sampling rate to use for encoding in hertz.
- bitrate number
The bitrate, in bits per second, of the output encoded audio.
- channels number
The number of channels in the audio.
- label string
An optional label for the codec. The label can be used to control muxing behavior.
- profile
string | Aac
Audio Profile The encoding profile to be used when encoding audio with AAC.
- sampling
Rate number The sampling rate to use for encoding in hertz.
- bitrate int
The bitrate, in bits per second, of the output encoded audio.
- channels int
The number of channels in the audio.
- label str
An optional label for the codec. The label can be used to control muxing behavior.
- profile
str | Aac
Audio Profile The encoding profile to be used when encoding audio with AAC.
- sampling_
rate int The sampling rate to use for encoding in hertz.
- bitrate Number
The bitrate, in bits per second, of the output encoded audio.
- channels Number
The number of channels in the audio.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- profile
String | "Aac
Lc" | "He Aac V1" | "He Aac V2" The encoding profile to be used when encoding audio with AAC.
- sampling
Rate Number The sampling rate to use for encoding in hertz.
AacAudioProfile
- Aac
Lc - AacLc
Specifies that the output audio is to be encoded into AAC Low Complexity profile (AAC-LC).
- He
Aac V1 - HeAacV1
Specifies that the output audio is to be encoded into HE-AAC v1 profile.
- He
Aac V2 - HeAacV2
Specifies that the output audio is to be encoded into HE-AAC v2 profile.
- Aac
Audio Profile Aac Lc - AacLc
Specifies that the output audio is to be encoded into AAC Low Complexity profile (AAC-LC).
- Aac
Audio Profile He Aac V1 - HeAacV1
Specifies that the output audio is to be encoded into HE-AAC v1 profile.
- Aac
Audio Profile He Aac V2 - HeAacV2
Specifies that the output audio is to be encoded into HE-AAC v2 profile.
- Aac
Lc - AacLc
Specifies that the output audio is to be encoded into AAC Low Complexity profile (AAC-LC).
- He
Aac V1 - HeAacV1
Specifies that the output audio is to be encoded into HE-AAC v1 profile.
- He
Aac V2 - HeAacV2
Specifies that the output audio is to be encoded into HE-AAC v2 profile.
- Aac
Lc - AacLc
Specifies that the output audio is to be encoded into AAC Low Complexity profile (AAC-LC).
- He
Aac V1 - HeAacV1
Specifies that the output audio is to be encoded into HE-AAC v1 profile.
- He
Aac V2 - HeAacV2
Specifies that the output audio is to be encoded into HE-AAC v2 profile.
- AAC_LC
- AacLc
Specifies that the output audio is to be encoded into AAC Low Complexity profile (AAC-LC).
- HE_AAC_V1
- HeAacV1
Specifies that the output audio is to be encoded into HE-AAC v1 profile.
- HE_AAC_V2
- HeAacV2
Specifies that the output audio is to be encoded into HE-AAC v2 profile.
- "Aac
Lc" - AacLc
Specifies that the output audio is to be encoded into AAC Low Complexity profile (AAC-LC).
- "He
Aac V1" - HeAacV1
Specifies that the output audio is to be encoded into HE-AAC v1 profile.
- "He
Aac V2" - HeAacV2
Specifies that the output audio is to be encoded into HE-AAC v2 profile.
AacAudioResponse
- Bitrate int
The bitrate, in bits per second, of the output encoded audio.
- Channels int
The number of channels in the audio.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Profile string
The encoding profile to be used when encoding audio with AAC.
- Sampling
Rate int The sampling rate to use for encoding in hertz.
- Bitrate int
The bitrate, in bits per second, of the output encoded audio.
- Channels int
The number of channels in the audio.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Profile string
The encoding profile to be used when encoding audio with AAC.
- Sampling
Rate int The sampling rate to use for encoding in hertz.
- bitrate Integer
The bitrate, in bits per second, of the output encoded audio.
- channels Integer
The number of channels in the audio.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- profile String
The encoding profile to be used when encoding audio with AAC.
- sampling
Rate Integer The sampling rate to use for encoding in hertz.
- bitrate number
The bitrate, in bits per second, of the output encoded audio.
- channels number
The number of channels in the audio.
- label string
An optional label for the codec. The label can be used to control muxing behavior.
- profile string
The encoding profile to be used when encoding audio with AAC.
- sampling
Rate number The sampling rate to use for encoding in hertz.
- bitrate int
The bitrate, in bits per second, of the output encoded audio.
- channels int
The number of channels in the audio.
- label str
An optional label for the codec. The label can be used to control muxing behavior.
- profile str
The encoding profile to be used when encoding audio with AAC.
- sampling_
rate int The sampling rate to use for encoding in hertz.
- bitrate Number
The bitrate, in bits per second, of the output encoded audio.
- channels Number
The number of channels in the audio.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- profile String
The encoding profile to be used when encoding audio with AAC.
- sampling
Rate Number The sampling rate to use for encoding in hertz.
AnalysisResolution
- Source
Resolution - SourceResolution
- Standard
Definition - StandardDefinition
- Analysis
Resolution Source Resolution - SourceResolution
- Analysis
Resolution Standard Definition - StandardDefinition
- Source
Resolution - SourceResolution
- Standard
Definition - StandardDefinition
- Source
Resolution - SourceResolution
- Standard
Definition - StandardDefinition
- SOURCE_RESOLUTION
- SourceResolution
- STANDARD_DEFINITION
- StandardDefinition
- "Source
Resolution" - SourceResolution
- "Standard
Definition" - StandardDefinition
Audio
- Bitrate int
The bitrate, in bits per second, of the output encoded audio.
- Channels int
The number of channels in the audio.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Sampling
Rate int The sampling rate to use for encoding in hertz.
- Bitrate int
The bitrate, in bits per second, of the output encoded audio.
- Channels int
The number of channels in the audio.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Sampling
Rate int The sampling rate to use for encoding in hertz.
- bitrate Integer
The bitrate, in bits per second, of the output encoded audio.
- channels Integer
The number of channels in the audio.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- sampling
Rate Integer The sampling rate to use for encoding in hertz.
- bitrate number
The bitrate, in bits per second, of the output encoded audio.
- channels number
The number of channels in the audio.
- label string
An optional label for the codec. The label can be used to control muxing behavior.
- sampling
Rate number The sampling rate to use for encoding in hertz.
- bitrate int
The bitrate, in bits per second, of the output encoded audio.
- channels int
The number of channels in the audio.
- label str
An optional label for the codec. The label can be used to control muxing behavior.
- sampling_
rate int The sampling rate to use for encoding in hertz.
- bitrate Number
The bitrate, in bits per second, of the output encoded audio.
- channels Number
The number of channels in the audio.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- sampling
Rate Number The sampling rate to use for encoding in hertz.
AudioAnalysisMode
- Standard
- Standard
Performs all operations included in the Basic mode, additionally performing language detection and speaker diarization.
- Basic
- Basic
This mode performs speech-to-text transcription and generation of a VTT subtitle/caption file. The output of this mode includes an Insights JSON file including only the keywords, transcription,and timing information. Automatic language detection and speaker diarization are not included in this mode.
- Audio
Analysis Mode Standard - Standard
Performs all operations included in the Basic mode, additionally performing language detection and speaker diarization.
- Audio
Analysis Mode Basic - Basic
This mode performs speech-to-text transcription and generation of a VTT subtitle/caption file. The output of this mode includes an Insights JSON file including only the keywords, transcription,and timing information. Automatic language detection and speaker diarization are not included in this mode.
- Standard
- Standard
Performs all operations included in the Basic mode, additionally performing language detection and speaker diarization.
- Basic
- Basic
This mode performs speech-to-text transcription and generation of a VTT subtitle/caption file. The output of this mode includes an Insights JSON file including only the keywords, transcription,and timing information. Automatic language detection and speaker diarization are not included in this mode.
- Standard
- Standard
Performs all operations included in the Basic mode, additionally performing language detection and speaker diarization.
- Basic
- Basic
This mode performs speech-to-text transcription and generation of a VTT subtitle/caption file. The output of this mode includes an Insights JSON file including only the keywords, transcription,and timing information. Automatic language detection and speaker diarization are not included in this mode.
- STANDARD
- Standard
Performs all operations included in the Basic mode, additionally performing language detection and speaker diarization.
- BASIC
- Basic
This mode performs speech-to-text transcription and generation of a VTT subtitle/caption file. The output of this mode includes an Insights JSON file including only the keywords, transcription,and timing information. Automatic language detection and speaker diarization are not included in this mode.
- "Standard"
- Standard
Performs all operations included in the Basic mode, additionally performing language detection and speaker diarization.
- "Basic"
- Basic
This mode performs speech-to-text transcription and generation of a VTT subtitle/caption file. The output of this mode includes an Insights JSON file including only the keywords, transcription,and timing information. Automatic language detection and speaker diarization are not included in this mode.
AudioAnalyzerPreset
- Audio
Language string The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- Experimental
Options Dictionary<string, string> Dictionary containing key value pairs for parameters not exposed in the preset itself
- Mode
string | Pulumi.
Azure Native. Media. Audio Analysis Mode Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
- Audio
Language string The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- Experimental
Options map[string]string Dictionary containing key value pairs for parameters not exposed in the preset itself
- Mode
string | Audio
Analysis Mode Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
- audio
Language String The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- experimental
Options Map<String,String> Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode
String | Audio
Analysis Mode Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
- audio
Language string The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- experimental
Options {[key: string]: string} Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode
string | Audio
Analysis Mode Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
- audio_
language str The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- experimental_
options Mapping[str, str] Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode
str | Audio
Analysis Mode Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
- audio
Language String The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- experimental
Options Map<String> Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode String | "Standard" | "Basic"
Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
AudioAnalyzerPresetResponse
- Audio
Language string The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- Experimental
Options Dictionary<string, string> Dictionary containing key value pairs for parameters not exposed in the preset itself
- Mode string
Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
- Audio
Language string The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- Experimental
Options map[string]string Dictionary containing key value pairs for parameters not exposed in the preset itself
- Mode string
Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
- audio
Language String The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- experimental
Options Map<String,String> Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode String
Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
- audio
Language string The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- experimental
Options {[key: string]: string} Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode string
Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
- audio_
language str The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- experimental_
options Mapping[str, str] Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode str
Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
- audio
Language String The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- experimental
Options Map<String> Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode String
Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
AudioOverlay
- Input
Label string The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- Audio
Gain doubleLevel The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- End string
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- Fade
In stringDuration The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- Fade
Out stringDuration The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- Start string
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
- Input
Label string The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- Audio
Gain float64Level The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- End string
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- Fade
In stringDuration The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- Fade
Out stringDuration The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- Start string
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
- input
Label String The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- audio
Gain DoubleLevel The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- end String
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- fade
In StringDuration The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- fade
Out StringDuration The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- start String
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
- input
Label string The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- audio
Gain numberLevel The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- end string
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- fade
In stringDuration The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- fade
Out stringDuration The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- start string
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
- input_
label str The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- audio_
gain_ floatlevel The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- end str
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- fade_
in_ strduration The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- fade_
out_ strduration The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- start str
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
- input
Label String The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- audio
Gain NumberLevel The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- end String
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- fade
In StringDuration The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- fade
Out StringDuration The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- start String
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
AudioOverlayResponse
- Input
Label string The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- Audio
Gain doubleLevel The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- End string
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- Fade
In stringDuration The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- Fade
Out stringDuration The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- Start string
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
- Input
Label string The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- Audio
Gain float64Level The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- End string
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- Fade
In stringDuration The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- Fade
Out stringDuration The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- Start string
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
- input
Label String The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- audio
Gain DoubleLevel The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- end String
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- fade
In StringDuration The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- fade
Out StringDuration The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- start String
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
- input
Label string The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- audio
Gain numberLevel The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- end string
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- fade
In stringDuration The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- fade
Out stringDuration The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- start string
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
- input_
label str The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- audio_
gain_ floatlevel The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- end str
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- fade_
in_ strduration The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- fade_
out_ strduration The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- start str
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
- input
Label String The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- audio
Gain NumberLevel The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- end String
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- fade
In StringDuration The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- fade
Out StringDuration The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- start String
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
AudioResponse
- Bitrate int
The bitrate, in bits per second, of the output encoded audio.
- Channels int
The number of channels in the audio.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Sampling
Rate int The sampling rate to use for encoding in hertz.
- Bitrate int
The bitrate, in bits per second, of the output encoded audio.
- Channels int
The number of channels in the audio.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Sampling
Rate int The sampling rate to use for encoding in hertz.
- bitrate Integer
The bitrate, in bits per second, of the output encoded audio.
- channels Integer
The number of channels in the audio.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- sampling
Rate Integer The sampling rate to use for encoding in hertz.
- bitrate number
The bitrate, in bits per second, of the output encoded audio.
- channels number
The number of channels in the audio.
- label string
An optional label for the codec. The label can be used to control muxing behavior.
- sampling
Rate number The sampling rate to use for encoding in hertz.
- bitrate int
The bitrate, in bits per second, of the output encoded audio.
- channels int
The number of channels in the audio.
- label str
An optional label for the codec. The label can be used to control muxing behavior.
- sampling_
rate int The sampling rate to use for encoding in hertz.
- bitrate Number
The bitrate, in bits per second, of the output encoded audio.
- channels Number
The number of channels in the audio.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- sampling
Rate Number The sampling rate to use for encoding in hertz.
BlurType
- Box
- Box
Box: debug filter, bounding box only
- Low
- Low
Low: box-car blur filter
- Med
- Med
Med: Gaussian blur filter
- High
- High
High: Confuse blur filter
- Black
- Black
Black: Black out filter
- Blur
Type Box - Box
Box: debug filter, bounding box only
- Blur
Type Low - Low
Low: box-car blur filter
- Blur
Type Med - Med
Med: Gaussian blur filter
- Blur
Type High - High
High: Confuse blur filter
- Blur
Type Black - Black
Black: Black out filter
- Box
- Box
Box: debug filter, bounding box only
- Low
- Low
Low: box-car blur filter
- Med
- Med
Med: Gaussian blur filter
- High
- High
High: Confuse blur filter
- Black
- Black
Black: Black out filter
- Box
- Box
Box: debug filter, bounding box only
- Low
- Low
Low: box-car blur filter
- Med
- Med
Med: Gaussian blur filter
- High
- High
High: Confuse blur filter
- Black
- Black
Black: Black out filter
- BOX
- Box
Box: debug filter, bounding box only
- LOW
- Low
Low: box-car blur filter
- MED
- Med
Med: Gaussian blur filter
- HIGH
- High
High: Confuse blur filter
- BLACK
- Black
Black: Black out filter
- "Box"
- Box
Box: debug filter, bounding box only
- "Low"
- Low
Low: box-car blur filter
- "Med"
- Med
Med: Gaussian blur filter
- "High"
- High
High: Confuse blur filter
- "Black"
- Black
Black: Black out filter
BuiltInStandardEncoderPreset
- Preset
Name string | Pulumi.Azure Native. Media. Encoder Named Preset The built-in preset to be used for encoding videos.
- Preset
Name string | EncoderNamed Preset The built-in preset to be used for encoding videos.
- preset
Name String | EncoderNamed Preset The built-in preset to be used for encoding videos.
- preset
Name string | EncoderNamed Preset The built-in preset to be used for encoding videos.
- preset_
name str | EncoderNamed Preset The built-in preset to be used for encoding videos.
- preset
Name String | "H264SingleBitrate SD" | "H264Single Bitrate720p" | "H264Single Bitrate1080p" | "Adaptive Streaming" | "AACGood Quality Audio" | "Content Aware Encoding Experimental" | "Content Aware Encoding" | "Copy All Bitrate Non Interleaved" | "H264Multiple Bitrate1080p" | "H264Multiple Bitrate720p" | "H264Multiple Bitrate SD" | "H265Content Aware Encoding" | "H265Adaptive Streaming" | "H265Single Bitrate720p" | "H265Single Bitrate1080p" | "H265Single Bitrate4K" The built-in preset to be used for encoding videos.
BuiltInStandardEncoderPresetResponse
- Preset
Name string The built-in preset to be used for encoding videos.
- Preset
Name string The built-in preset to be used for encoding videos.
- preset
Name String The built-in preset to be used for encoding videos.
- preset
Name string The built-in preset to be used for encoding videos.
- preset_
name str The built-in preset to be used for encoding videos.
- preset
Name String The built-in preset to be used for encoding videos.
CopyAudio
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- label string
An optional label for the codec. The label can be used to control muxing behavior.
- label str
An optional label for the codec. The label can be used to control muxing behavior.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
CopyAudioResponse
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- label string
An optional label for the codec. The label can be used to control muxing behavior.
- label str
An optional label for the codec. The label can be used to control muxing behavior.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
CopyVideo
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- label string
An optional label for the codec. The label can be used to control muxing behavior.
- label str
An optional label for the codec. The label can be used to control muxing behavior.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
CopyVideoResponse
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- label string
An optional label for the codec. The label can be used to control muxing behavior.
- label str
An optional label for the codec. The label can be used to control muxing behavior.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
Deinterlace
- Mode
string | Pulumi.
Azure Native. Media. Deinterlace Mode The deinterlacing mode. Defaults to AutoPixelAdaptive.
- Parity
string | Pulumi.
Azure Native. Media. Deinterlace Parity The field parity for de-interlacing, defaults to Auto.
- Mode
string | Deinterlace
Mode The deinterlacing mode. Defaults to AutoPixelAdaptive.
- Parity
string | Deinterlace
Parity The field parity for de-interlacing, defaults to Auto.
- mode
String | Deinterlace
Mode The deinterlacing mode. Defaults to AutoPixelAdaptive.
- parity
String | Deinterlace
Parity The field parity for de-interlacing, defaults to Auto.
- mode
string | Deinterlace
Mode The deinterlacing mode. Defaults to AutoPixelAdaptive.
- parity
string | Deinterlace
Parity The field parity for de-interlacing, defaults to Auto.
- mode
str | Deinterlace
Mode The deinterlacing mode. Defaults to AutoPixelAdaptive.
- parity
str | Deinterlace
Parity The field parity for de-interlacing, defaults to Auto.
- mode
String | "Off" | "Auto
Pixel Adaptive" The deinterlacing mode. Defaults to AutoPixelAdaptive.
- parity
String | "Auto" | "Top
Field First" | "Bottom Field First" The field parity for de-interlacing, defaults to Auto.
DeinterlaceMode
- Off
- Off
Disables de-interlacing of the source video.
- Auto
Pixel Adaptive - AutoPixelAdaptive
Apply automatic pixel adaptive de-interlacing on each frame in the input video.
- Deinterlace
Mode Off - Off
Disables de-interlacing of the source video.
- Deinterlace
Mode Auto Pixel Adaptive - AutoPixelAdaptive
Apply automatic pixel adaptive de-interlacing on each frame in the input video.
- Off
- Off
Disables de-interlacing of the source video.
- Auto
Pixel Adaptive - AutoPixelAdaptive
Apply automatic pixel adaptive de-interlacing on each frame in the input video.
- Off
- Off
Disables de-interlacing of the source video.
- Auto
Pixel Adaptive - AutoPixelAdaptive
Apply automatic pixel adaptive de-interlacing on each frame in the input video.
- OFF
- Off
Disables de-interlacing of the source video.
- AUTO_PIXEL_ADAPTIVE
- AutoPixelAdaptive
Apply automatic pixel adaptive de-interlacing on each frame in the input video.
- "Off"
- Off
Disables de-interlacing of the source video.
- "Auto
Pixel Adaptive" - AutoPixelAdaptive
Apply automatic pixel adaptive de-interlacing on each frame in the input video.
DeinterlaceParity
- Auto
- Auto
Automatically detect the order of fields
- Top
Field First - TopFieldFirst
Apply top field first processing of input video.
- Bottom
Field First - BottomFieldFirst
Apply bottom field first processing of input video.
- Deinterlace
Parity Auto - Auto
Automatically detect the order of fields
- Deinterlace
Parity Top Field First - TopFieldFirst
Apply top field first processing of input video.
- Deinterlace
Parity Bottom Field First - BottomFieldFirst
Apply bottom field first processing of input video.
- Auto
- Auto
Automatically detect the order of fields
- Top
Field First - TopFieldFirst
Apply top field first processing of input video.
- Bottom
Field First - BottomFieldFirst
Apply bottom field first processing of input video.
- Auto
- Auto
Automatically detect the order of fields
- Top
Field First - TopFieldFirst
Apply top field first processing of input video.
- Bottom
Field First - BottomFieldFirst
Apply bottom field first processing of input video.
- AUTO
- Auto
Automatically detect the order of fields
- TOP_FIELD_FIRST
- TopFieldFirst
Apply top field first processing of input video.
- BOTTOM_FIELD_FIRST
- BottomFieldFirst
Apply bottom field first processing of input video.
- "Auto"
- Auto
Automatically detect the order of fields
- "Top
Field First" - TopFieldFirst
Apply top field first processing of input video.
- "Bottom
Field First" - BottomFieldFirst
Apply bottom field first processing of input video.
DeinterlaceResponse
EncoderNamedPreset
- H264Single
Bitrate SD - H264SingleBitrateSD
Produces an MP4 file where the video is encoded with H.264 codec at 2200 kbps and a picture height of 480 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H264Single
Bitrate720p - H264SingleBitrate720p
Produces an MP4 file where the video is encoded with H.264 codec at 4500 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H264Single
Bitrate1080p - H264SingleBitrate1080p
Produces an MP4 file where the video is encoded with H.264 codec at 6750 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- Adaptive
Streaming - AdaptiveStreaming
Produces a set of GOP aligned MP4 files with H.264 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
- AACGood
Quality Audio - AACGoodQualityAudio
Produces a single MP4 file containing only stereo audio encoded at 192 kbps.
- Content
Aware Encoding Experimental - ContentAwareEncodingExperimental
Exposes an experimental preset for content-aware encoding. Given any input content, the service attempts to automatically determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. The underlying algorithms will continue to evolve over time. The output will contain MP4 files with video and audio interleaved.
- Content
Aware Encoding - ContentAwareEncoding
Produces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
- Copy
All Bitrate Non Interleaved - CopyAllBitrateNonInterleaved
Copy all video and audio streams from the input asset as non-interleaved video and audio output files. This preset can be used to clip an existing asset or convert a group of key frame (GOP) aligned MP4 files as an asset that can be streamed.
- H264Multiple
Bitrate1080p - H264MultipleBitrate1080p
Produces a set of 8 GOP-aligned MP4 files, ranging from 6000 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 1080p and goes down to 180p.
- H264Multiple
Bitrate720p - H264MultipleBitrate720p
Produces a set of 6 GOP-aligned MP4 files, ranging from 3400 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 720p and goes down to 180p.
- H264Multiple
Bitrate SD - H264MultipleBitrateSD
Produces a set of 5 GOP-aligned MP4 files, ranging from 1900kbps to 400 kbps, and stereo AAC audio. Resolution starts at 480p and goes down to 240p.
- H265Content
Aware Encoding - H265ContentAwareEncoding
Produces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
- H265Adaptive
Streaming - H265AdaptiveStreaming
Produces a set of GOP aligned MP4 files with H.265 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
- H265Single
Bitrate720p - H265SingleBitrate720p
Produces an MP4 file where the video is encoded with H.265 codec at 1800 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H265Single
Bitrate1080p - H265SingleBitrate1080p
Produces an MP4 file where the video is encoded with H.265 codec at 3500 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H265Single
Bitrate4K - H265SingleBitrate4K
Produces an MP4 file where the video is encoded with H.265 codec at 9500 kbps and a picture height of 2160 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- Encoder
Named Preset H264Single Bitrate SD - H264SingleBitrateSD
Produces an MP4 file where the video is encoded with H.264 codec at 2200 kbps and a picture height of 480 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- Encoder
Named Preset H264Single Bitrate720p - H264SingleBitrate720p
Produces an MP4 file where the video is encoded with H.264 codec at 4500 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- Encoder
Named Preset H264Single Bitrate1080p - H264SingleBitrate1080p
Produces an MP4 file where the video is encoded with H.264 codec at 6750 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- Encoder
Named Preset Adaptive Streaming - AdaptiveStreaming
Produces a set of GOP aligned MP4 files with H.264 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
- Encoder
Named Preset AACGood Quality Audio - AACGoodQualityAudio
Produces a single MP4 file containing only stereo audio encoded at 192 kbps.
- Encoder
Named Preset Content Aware Encoding Experimental - ContentAwareEncodingExperimental
Exposes an experimental preset for content-aware encoding. Given any input content, the service attempts to automatically determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. The underlying algorithms will continue to evolve over time. The output will contain MP4 files with video and audio interleaved.
- Encoder
Named Preset Content Aware Encoding - ContentAwareEncoding
Produces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
- Encoder
Named Preset Copy All Bitrate Non Interleaved - CopyAllBitrateNonInterleaved
Copy all video and audio streams from the input asset as non-interleaved video and audio output files. This preset can be used to clip an existing asset or convert a group of key frame (GOP) aligned MP4 files as an asset that can be streamed.
- Encoder
Named Preset H264Multiple Bitrate1080p - H264MultipleBitrate1080p
Produces a set of 8 GOP-aligned MP4 files, ranging from 6000 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 1080p and goes down to 180p.
- Encoder
Named Preset H264Multiple Bitrate720p - H264MultipleBitrate720p
Produces a set of 6 GOP-aligned MP4 files, ranging from 3400 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 720p and goes down to 180p.
- Encoder
Named Preset H264Multiple Bitrate SD - H264MultipleBitrateSD
Produces a set of 5 GOP-aligned MP4 files, ranging from 1900kbps to 400 kbps, and stereo AAC audio. Resolution starts at 480p and goes down to 240p.
- Encoder
Named Preset H265Content Aware Encoding - H265ContentAwareEncoding
Produces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
- Encoder
Named Preset H265Adaptive Streaming - H265AdaptiveStreaming
Produces a set of GOP aligned MP4 files with H.265 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
- Encoder
Named Preset H265Single Bitrate720p - H265SingleBitrate720p
Produces an MP4 file where the video is encoded with H.265 codec at 1800 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- Encoder
Named Preset H265Single Bitrate1080p - H265SingleBitrate1080p
Produces an MP4 file where the video is encoded with H.265 codec at 3500 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- Encoder
Named Preset H265Single Bitrate4K - H265SingleBitrate4K
Produces an MP4 file where the video is encoded with H.265 codec at 9500 kbps and a picture height of 2160 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H264Single
Bitrate SD - H264SingleBitrateSD
Produces an MP4 file where the video is encoded with H.264 codec at 2200 kbps and a picture height of 480 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H264Single
Bitrate720p - H264SingleBitrate720p
Produces an MP4 file where the video is encoded with H.264 codec at 4500 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H264Single
Bitrate1080p - H264SingleBitrate1080p
Produces an MP4 file where the video is encoded with H.264 codec at 6750 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- Adaptive
Streaming - AdaptiveStreaming
Produces a set of GOP aligned MP4 files with H.264 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
- AACGood
Quality Audio - AACGoodQualityAudio
Produces a single MP4 file containing only stereo audio encoded at 192 kbps.
- Content
Aware Encoding Experimental - ContentAwareEncodingExperimental
Exposes an experimental preset for content-aware encoding. Given any input content, the service attempts to automatically determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. The underlying algorithms will continue to evolve over time. The output will contain MP4 files with video and audio interleaved.
- Content
Aware Encoding - ContentAwareEncoding
Produces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
- Copy
All Bitrate Non Interleaved - CopyAllBitrateNonInterleaved
Copy all video and audio streams from the input asset as non-interleaved video and audio output files. This preset can be used to clip an existing asset or convert a group of key frame (GOP) aligned MP4 files as an asset that can be streamed.
- H264Multiple
Bitrate1080p - H264MultipleBitrate1080p
Produces a set of 8 GOP-aligned MP4 files, ranging from 6000 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 1080p and goes down to 180p.
- H264Multiple
Bitrate720p - H264MultipleBitrate720p
Produces a set of 6 GOP-aligned MP4 files, ranging from 3400 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 720p and goes down to 180p.
- H264Multiple
Bitrate SD - H264MultipleBitrateSD
Produces a set of 5 GOP-aligned MP4 files, ranging from 1900kbps to 400 kbps, and stereo AAC audio. Resolution starts at 480p and goes down to 240p.
- H265Content
Aware Encoding - H265ContentAwareEncoding
Produces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
- H265Adaptive
Streaming - H265AdaptiveStreaming
Produces a set of GOP aligned MP4 files with H.265 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
- H265Single
Bitrate720p - H265SingleBitrate720p
Produces an MP4 file where the video is encoded with H.265 codec at 1800 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H265Single
Bitrate1080p - H265SingleBitrate1080p
Produces an MP4 file where the video is encoded with H.265 codec at 3500 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H265Single
Bitrate4K - H265SingleBitrate4K
Produces an MP4 file where the video is encoded with H.265 codec at 9500 kbps and a picture height of 2160 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H264Single
Bitrate SD - H264SingleBitrateSD
Produces an MP4 file where the video is encoded with H.264 codec at 2200 kbps and a picture height of 480 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H264Single
Bitrate720p - H264SingleBitrate720p
Produces an MP4 file where the video is encoded with H.264 codec at 4500 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H264Single
Bitrate1080p - H264SingleBitrate1080p
Produces an MP4 file where the video is encoded with H.264 codec at 6750 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- Adaptive
Streaming - AdaptiveStreaming
Produces a set of GOP aligned MP4 files with H.264 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
- AACGood
Quality Audio - AACGoodQualityAudio
Produces a single MP4 file containing only stereo audio encoded at 192 kbps.
- Content
Aware Encoding Experimental - ContentAwareEncodingExperimental
Exposes an experimental preset for content-aware encoding. Given any input content, the service attempts to automatically determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. The underlying algorithms will continue to evolve over time. The output will contain MP4 files with video and audio interleaved.
- Content
Aware Encoding - ContentAwareEncoding
Produces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
- Copy
All Bitrate Non Interleaved - CopyAllBitrateNonInterleaved
Copy all video and audio streams from the input asset as non-interleaved video and audio output files. This preset can be used to clip an existing asset or convert a group of key frame (GOP) aligned MP4 files as an asset that can be streamed.
- H264Multiple
Bitrate1080p - H264MultipleBitrate1080p
Produces a set of 8 GOP-aligned MP4 files, ranging from 6000 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 1080p and goes down to 180p.
- H264Multiple
Bitrate720p - H264MultipleBitrate720p
Produces a set of 6 GOP-aligned MP4 files, ranging from 3400 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 720p and goes down to 180p.
- H264Multiple
Bitrate SD - H264MultipleBitrateSD
Produces a set of 5 GOP-aligned MP4 files, ranging from 1900kbps to 400 kbps, and stereo AAC audio. Resolution starts at 480p and goes down to 240p.
- H265Content
Aware Encoding - H265ContentAwareEncoding
Produces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
- H265Adaptive
Streaming - H265AdaptiveStreaming
Produces a set of GOP aligned MP4 files with H.265 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
- H265Single
Bitrate720p - H265SingleBitrate720p
Produces an MP4 file where the video is encoded with H.265 codec at 1800 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H265Single
Bitrate1080p - H265SingleBitrate1080p
Produces an MP4 file where the video is encoded with H.265 codec at 3500 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H265Single
Bitrate4K - H265SingleBitrate4K
Produces an MP4 file where the video is encoded with H.265 codec at 9500 kbps and a picture height of 2160 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H264_SINGLE_BITRATE_SD
- H264SingleBitrateSD
Produces an MP4 file where the video is encoded with H.264 codec at 2200 kbps and a picture height of 480 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H264_SINGLE_BITRATE720P
- H264SingleBitrate720p
Produces an MP4 file where the video is encoded with H.264 codec at 4500 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H264_SINGLE_BITRATE1080P
- H264SingleBitrate1080p
Produces an MP4 file where the video is encoded with H.264 codec at 6750 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- ADAPTIVE_STREAMING
- AdaptiveStreaming
Produces a set of GOP aligned MP4 files with H.264 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
- AAC_GOOD_QUALITY_AUDIO
- AACGoodQualityAudio
Produces a single MP4 file containing only stereo audio encoded at 192 kbps.
- CONTENT_AWARE_ENCODING_EXPERIMENTAL
- ContentAwareEncodingExperimental
Exposes an experimental preset for content-aware encoding. Given any input content, the service attempts to automatically determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. The underlying algorithms will continue to evolve over time. The output will contain MP4 files with video and audio interleaved.
- CONTENT_AWARE_ENCODING
- ContentAwareEncoding
Produces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
- COPY_ALL_BITRATE_NON_INTERLEAVED
- CopyAllBitrateNonInterleaved
Copy all video and audio streams from the input asset as non-interleaved video and audio output files. This preset can be used to clip an existing asset or convert a group of key frame (GOP) aligned MP4 files as an asset that can be streamed.
- H264_MULTIPLE_BITRATE1080P
- H264MultipleBitrate1080p
Produces a set of 8 GOP-aligned MP4 files, ranging from 6000 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 1080p and goes down to 180p.
- H264_MULTIPLE_BITRATE720P
- H264MultipleBitrate720p
Produces a set of 6 GOP-aligned MP4 files, ranging from 3400 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 720p and goes down to 180p.
- H264_MULTIPLE_BITRATE_SD
- H264MultipleBitrateSD
Produces a set of 5 GOP-aligned MP4 files, ranging from 1900kbps to 400 kbps, and stereo AAC audio. Resolution starts at 480p and goes down to 240p.
- H265_CONTENT_AWARE_ENCODING
- H265ContentAwareEncoding
Produces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
- H265_ADAPTIVE_STREAMING
- H265AdaptiveStreaming
Produces a set of GOP aligned MP4 files with H.265 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
- H265_SINGLE_BITRATE720P
- H265SingleBitrate720p
Produces an MP4 file where the video is encoded with H.265 codec at 1800 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H265_SINGLE_BITRATE1080P
- H265SingleBitrate1080p
Produces an MP4 file where the video is encoded with H.265 codec at 3500 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H265_SINGLE_BITRATE4_K
- H265SingleBitrate4K
Produces an MP4 file where the video is encoded with H.265 codec at 9500 kbps and a picture height of 2160 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- "H264Single
Bitrate SD" - H264SingleBitrateSD
Produces an MP4 file where the video is encoded with H.264 codec at 2200 kbps and a picture height of 480 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- "H264Single
Bitrate720p" - H264SingleBitrate720p
Produces an MP4 file where the video is encoded with H.264 codec at 4500 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- "H264Single
Bitrate1080p" - H264SingleBitrate1080p
Produces an MP4 file where the video is encoded with H.264 codec at 6750 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- "Adaptive
Streaming" - AdaptiveStreaming
Produces a set of GOP aligned MP4 files with H.264 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
- "AACGood
Quality Audio" - AACGoodQualityAudio
Produces a single MP4 file containing only stereo audio encoded at 192 kbps.
- "Content
Aware Encoding Experimental" - ContentAwareEncodingExperimental
Exposes an experimental preset for content-aware encoding. Given any input content, the service attempts to automatically determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. The underlying algorithms will continue to evolve over time. The output will contain MP4 files with video and audio interleaved.
- "Content
Aware Encoding" - ContentAwareEncoding
Produces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
- "Copy
All Bitrate Non Interleaved" - CopyAllBitrateNonInterleaved
Copy all video and audio streams from the input asset as non-interleaved video and audio output files. This preset can be used to clip an existing asset or convert a group of key frame (GOP) aligned MP4 files as an asset that can be streamed.
- "H264Multiple
Bitrate1080p" - H264MultipleBitrate1080p
Produces a set of 8 GOP-aligned MP4 files, ranging from 6000 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 1080p and goes down to 180p.
- "H264Multiple
Bitrate720p" - H264MultipleBitrate720p
Produces a set of 6 GOP-aligned MP4 files, ranging from 3400 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 720p and goes down to 180p.
- "H264Multiple
Bitrate SD" - H264MultipleBitrateSD
Produces a set of 5 GOP-aligned MP4 files, ranging from 1900kbps to 400 kbps, and stereo AAC audio. Resolution starts at 480p and goes down to 240p.
- "H265Content
Aware Encoding" - H265ContentAwareEncoding
Produces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
- "H265Adaptive
Streaming" - H265AdaptiveStreaming
Produces a set of GOP aligned MP4 files with H.265 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
- "H265Single
Bitrate720p" - H265SingleBitrate720p
Produces an MP4 file where the video is encoded with H.265 codec at 1800 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- "H265Single
Bitrate1080p" - H265SingleBitrate1080p
Produces an MP4 file where the video is encoded with H.265 codec at 3500 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- "H265Single
Bitrate4K" - H265SingleBitrate4K
Produces an MP4 file where the video is encoded with H.265 codec at 9500 kbps and a picture height of 2160 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
EntropyMode
- Cabac
- Cabac
Context Adaptive Binary Arithmetic Coder (CABAC) entropy encoding.
- Cavlc
- Cavlc
Context Adaptive Variable Length Coder (CAVLC) entropy encoding.
- Entropy
Mode Cabac - Cabac
Context Adaptive Binary Arithmetic Coder (CABAC) entropy encoding.
- Entropy
Mode Cavlc - Cavlc
Context Adaptive Variable Length Coder (CAVLC) entropy encoding.
- Cabac
- Cabac
Context Adaptive Binary Arithmetic Coder (CABAC) entropy encoding.
- Cavlc
- Cavlc
Context Adaptive Variable Length Coder (CAVLC) entropy encoding.
- Cabac
- Cabac
Context Adaptive Binary Arithmetic Coder (CABAC) entropy encoding.
- Cavlc
- Cavlc
Context Adaptive Variable Length Coder (CAVLC) entropy encoding.
- CABAC
- Cabac
Context Adaptive Binary Arithmetic Coder (CABAC) entropy encoding.
- CAVLC
- Cavlc
Context Adaptive Variable Length Coder (CAVLC) entropy encoding.
- "Cabac"
- Cabac
Context Adaptive Binary Arithmetic Coder (CABAC) entropy encoding.
- "Cavlc"
- Cavlc
Context Adaptive Variable Length Coder (CAVLC) entropy encoding.
FaceDetectorPreset
- Blur
Type string | Pulumi.Azure Native. Media. Blur Type Blur type
- Experimental
Options Dictionary<string, string> Dictionary containing key value pairs for parameters not exposed in the preset itself
- Mode
string | Pulumi.
Azure Native. Media. Face Redactor Mode This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
- Resolution
string | Pulumi.
Azure Native. Media. Analysis Resolution Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
- Blur
Type string | BlurType Blur type
- Experimental
Options map[string]string Dictionary containing key value pairs for parameters not exposed in the preset itself
- Mode
string | Face
Redactor Mode This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
- Resolution
string | Analysis
Resolution Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
- blur
Type String | BlurType Blur type
- experimental
Options Map<String,String> Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode
String | Face
Redactor Mode This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
- resolution
String | Analysis
Resolution Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
- blur
Type string | BlurType Blur type
- experimental
Options {[key: string]: string} Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode
string | Face
Redactor Mode This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
- resolution
string | Analysis
Resolution Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
- blur_
type str | BlurType Blur type
- experimental_
options Mapping[str, str] Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode
str | Face
Redactor Mode This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
- resolution
str | Analysis
Resolution Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
- blur
Type String | "Box" | "Low" | "Med" | "High" | "Black" Blur type
- experimental
Options Map<String> Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode String | "Analyze" | "Redact" | "Combined"
This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
- resolution
String | "Source
Resolution" | "Standard Definition" Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
FaceDetectorPresetResponse
- Blur
Type string Blur type
- Experimental
Options Dictionary<string, string> Dictionary containing key value pairs for parameters not exposed in the preset itself
- Mode string
This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
- Resolution string
Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
- Blur
Type string Blur type
- Experimental
Options map[string]string Dictionary containing key value pairs for parameters not exposed in the preset itself
- Mode string
This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
- Resolution string
Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
- blur
Type String Blur type
- experimental
Options Map<String,String> Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode String
This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
- resolution String
Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
- blur
Type string Blur type
- experimental
Options {[key: string]: string} Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode string
This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
- resolution string
Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
- blur_
type str Blur type
- experimental_
options Mapping[str, str] Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode str
This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
- resolution str
Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
- blur
Type String Blur type
- experimental
Options Map<String> Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode String
This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
- resolution String
Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
FaceRedactorMode
- Analyze
- Analyze
Analyze mode detects faces and outputs a metadata file with the results. Allows editing of the metadata file before faces are blurred with Redact mode.
- Redact
- Redact
Redact mode consumes the metadata file from Analyze mode and redacts the faces found.
- Combined
- Combined
Combined mode does the Analyze and Redact steps in one pass when editing the analyzed faces is not desired.
- Face
Redactor Mode Analyze - Analyze
Analyze mode detects faces and outputs a metadata file with the results. Allows editing of the metadata file before faces are blurred with Redact mode.
- Face
Redactor Mode Redact - Redact
Redact mode consumes the metadata file from Analyze mode and redacts the faces found.
- Face
Redactor Mode Combined - Combined
Combined mode does the Analyze and Redact steps in one pass when editing the analyzed faces is not desired.
- Analyze
- Analyze
Analyze mode detects faces and outputs a metadata file with the results. Allows editing of the metadata file before faces are blurred with Redact mode.
- Redact
- Redact
Redact mode consumes the metadata file from Analyze mode and redacts the faces found.
- Combined
- Combined
Combined mode does the Analyze and Redact steps in one pass when editing the analyzed faces is not desired.
- Analyze
- Analyze
Analyze mode detects faces and outputs a metadata file with the results. Allows editing of the metadata file before faces are blurred with Redact mode.
- Redact
- Redact
Redact mode consumes the metadata file from Analyze mode and redacts the faces found.
- Combined
- Combined
Combined mode does the Analyze and Redact steps in one pass when editing the analyzed faces is not desired.
- ANALYZE
- Analyze
Analyze mode detects faces and outputs a metadata file with the results. Allows editing of the metadata file before faces are blurred with Redact mode.
- REDACT
- Redact
Redact mode consumes the metadata file from Analyze mode and redacts the faces found.
- COMBINED
- Combined
Combined mode does the Analyze and Redact steps in one pass when editing the analyzed faces is not desired.
- "Analyze"
- Analyze
Analyze mode detects faces and outputs a metadata file with the results. Allows editing of the metadata file before faces are blurred with Redact mode.
- "Redact"
- Redact
Redact mode consumes the metadata file from Analyze mode and redacts the faces found.
- "Combined"
- Combined
Combined mode does the Analyze and Redact steps in one pass when editing the analyzed faces is not desired.
Filters
- Crop
Pulumi.
Azure Native. Media. Inputs. Rectangle The parameters for the rectangular window with which to crop the input video.
- Deinterlace
Pulumi.
Azure Native. Media. Inputs. Deinterlace The de-interlacing settings.
- Overlays
List<Union<Pulumi.
Azure Native. Media. Inputs. Audio Overlay, Pulumi. Azure Native. Media. Inputs. Video Overlay>> The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
- Rotation
string | Pulumi.
Azure Native. Media. Rotation The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
- Crop Rectangle
The parameters for the rectangular window with which to crop the input video.
- Deinterlace Deinterlace
The de-interlacing settings.
- Overlays []interface{}
The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
- Rotation string | Rotation
The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
- crop Rectangle
The parameters for the rectangular window with which to crop the input video.
- deinterlace Deinterlace
The de-interlacing settings.
- overlays
List<Either<Audio
Overlay,Video Overlay>> The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
- rotation String | Rotation
The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
- crop Rectangle
The parameters for the rectangular window with which to crop the input video.
- deinterlace Deinterlace
The de-interlacing settings.
- overlays
(Audio
Overlay | Video Overlay)[] The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
- rotation string | Rotation
The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
- crop Rectangle
The parameters for the rectangular window with which to crop the input video.
- deinterlace Deinterlace
The de-interlacing settings.
- overlays
Sequence[Union[Audio
Overlay, Video Overlay]] The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
- rotation str | Rotation
The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
- crop Property Map
The parameters for the rectangular window with which to crop the input video.
- deinterlace Property Map
The de-interlacing settings.
- overlays List<Property Map | Property Map>
The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
- rotation String | "Auto" | "None" | "Rotate0" | "Rotate90" | "Rotate180" | "Rotate270"
The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
FiltersResponse
- Crop
Pulumi.
Azure Native. Media. Inputs. Rectangle Response The parameters for the rectangular window with which to crop the input video.
- Deinterlace
Pulumi.
Azure Native. Media. Inputs. Deinterlace Response The de-interlacing settings.
- Overlays
List<Union<Pulumi.
Azure Native. Media. Inputs. Audio Overlay Response, Pulumi. Azure Native. Media. Inputs. Video Overlay Response>> The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
- Rotation string
The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
- Crop
Rectangle
Response The parameters for the rectangular window with which to crop the input video.
- Deinterlace
Deinterlace
Response The de-interlacing settings.
- Overlays []interface{}
The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
- Rotation string
The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
- crop
Rectangle
Response The parameters for the rectangular window with which to crop the input video.
- deinterlace
Deinterlace
Response The de-interlacing settings.
- overlays
List<Either<Audio
Overlay Response,Video Overlay Response>> The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
- rotation String
The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
- crop
Rectangle
Response The parameters for the rectangular window with which to crop the input video.
- deinterlace
Deinterlace
Response The de-interlacing settings.
- overlays
(Audio
Overlay Response | Video Overlay Response)[] The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
- rotation string
The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
- crop
Rectangle
Response The parameters for the rectangular window with which to crop the input video.
- deinterlace
Deinterlace
Response The de-interlacing settings.
- overlays
Sequence[Union[Audio
Overlay Response, Video Overlay Response]] The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
- rotation str
The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
- crop Property Map
The parameters for the rectangular window with which to crop the input video.
- deinterlace Property Map
The de-interlacing settings.
- overlays List<Property Map | Property Map>
The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
- rotation String
The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
H264Complexity
- Speed
- Speed
Tells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
- Balanced
- Balanced
Tells the encoder to use settings that achieve a balance between speed and quality.
- Quality
- Quality
Tells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
- H264Complexity
Speed - Speed
Tells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
- H264Complexity
Balanced - Balanced
Tells the encoder to use settings that achieve a balance between speed and quality.
- H264Complexity
Quality - Quality
Tells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
- Speed
- Speed
Tells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
- Balanced
- Balanced
Tells the encoder to use settings that achieve a balance between speed and quality.
- Quality
- Quality
Tells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
- Speed
- Speed
Tells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
- Balanced
- Balanced
Tells the encoder to use settings that achieve a balance between speed and quality.
- Quality
- Quality
Tells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
- SPEED
- Speed
Tells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
- BALANCED
- Balanced
Tells the encoder to use settings that achieve a balance between speed and quality.
- QUALITY
- Quality
Tells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
- "Speed"
- Speed
Tells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
- "Balanced"
- Balanced
Tells the encoder to use settings that achieve a balance between speed and quality.
- "Quality"
- Quality
Tells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
H264Layer
- Bitrate int
The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
- Adaptive
BFrame bool Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- BFrames int
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- Buffer
Window string The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- Entropy
Mode string | Pulumi.Azure Native. Media. Entropy Mode The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
- Frame
Rate string The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Level string
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- Max
Bitrate int The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- Profile
string | Pulumi.
Azure Native. Media. H264Video Profile We currently support Baseline, Main, High, High422, High444. Default is Auto.
- Reference
Frames int The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- Slices int
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- Bitrate int
The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
- Adaptive
BFrame bool Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- BFrames int
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- Buffer
Window string The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- Entropy
Mode string | EntropyMode The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
- Frame
Rate string The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Level string
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- Max
Bitrate int The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- Profile
string | H264Video
Profile We currently support Baseline, Main, High, High422, High444. Default is Auto.
- Reference
Frames int The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- Slices int
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate Integer
The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
- adaptive
BFrame Boolean Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b
Frames Integer The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer
Window String The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- entropy
Mode String | EntropyMode The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
- frame
Rate String The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level String
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max
Bitrate Integer The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile
String | H264Video
Profile We currently support Baseline, Main, High, High422, High444. Default is Auto.
- reference
Frames Integer The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices Integer
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate number
The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
- adaptive
BFrame boolean Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b
Frames number The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer
Window string The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- entropy
Mode string | EntropyMode The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
- frame
Rate string The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level string
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max
Bitrate number The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile
string | H264Video
Profile We currently support Baseline, Main, High, High422, High444. Default is Auto.
- reference
Frames number The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices number
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate int
The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
- adaptive_
b_ boolframe Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b_
frames int The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer_
window str The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- entropy_
mode str | EntropyMode The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
- frame_
rate str The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height str
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label str
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level str
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max_
bitrate int The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile
str | H264Video
Profile We currently support Baseline, Main, High, High422, High444. Default is Auto.
- reference_
frames int The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices int
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width str
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate Number
The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
- adaptive
BFrame Boolean Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b
Frames Number The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer
Window String The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- entropy
Mode String | "Cabac" | "Cavlc" The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
- frame
Rate String The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level String
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max
Bitrate Number The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile String | "Auto" | "Baseline" | "Main" | "High" | "High422" | "High444"
We currently support Baseline, Main, High, High422, High444. Default is Auto.
- reference
Frames Number The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices Number
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
H264LayerResponse
- Bitrate int
The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
- Adaptive
BFrame bool Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- BFrames int
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- Buffer
Window string The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- Entropy
Mode string The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
- Frame
Rate string The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Level string
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- Max
Bitrate int The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- Profile string
We currently support Baseline, Main, High, High422, High444. Default is Auto.
- Reference
Frames int The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- Slices int
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- Bitrate int
The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
- Adaptive
BFrame bool Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- BFrames int
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- Buffer
Window string The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- Entropy
Mode string The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
- Frame
Rate string The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Level string
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- Max
Bitrate int The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- Profile string
We currently support Baseline, Main, High, High422, High444. Default is Auto.
- Reference
Frames int The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- Slices int
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate Integer
The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
- adaptive
BFrame Boolean Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b
Frames Integer The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer
Window String The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- entropy
Mode String The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
- frame
Rate String The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level String
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max
Bitrate Integer The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile String
We currently support Baseline, Main, High, High422, High444. Default is Auto.
- reference
Frames Integer The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices Integer
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate number
The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
- adaptive
BFrame boolean Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b
Frames number The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer
Window string The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- entropy
Mode string The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
- frame
Rate string The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level string
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max
Bitrate number The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile string
We currently support Baseline, Main, High, High422, High444. Default is Auto.
- reference
Frames number The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices number
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate int
The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
- adaptive_
b_ boolframe Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b_
frames int The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer_
window str The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- entropy_
mode str The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
- frame_
rate str The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height str
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label str
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level str
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max_
bitrate int The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile str
We currently support Baseline, Main, High, High422, High444. Default is Auto.
- reference_
frames int The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices int
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width str
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate Number
The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
- adaptive
BFrame Boolean Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b
Frames Number The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer
Window String The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- entropy
Mode String The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
- frame
Rate String The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level String
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max
Bitrate Number The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile String
We currently support Baseline, Main, High, High422, High444. Default is Auto.
- reference
Frames Number The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices Number
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
H264Video
- Complexity
string | Pulumi.
Azure Native. Media. H264Complexity Tells the encoder how to choose its encoding settings. The default value is Balanced.
- Key
Frame stringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Layers
List<Pulumi.
Azure Native. Media. Inputs. H264Layer> The collection of output H.264 layers to be produced by the encoder.
- Scene
Change boolDetection Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- Stretch
Mode string | Pulumi.Azure Native. Media. Stretch Mode The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string | Pulumi.Azure Native. Media. Video Sync Mode The Video Sync Mode
- Complexity string | H264Complexity
Tells the encoder how to choose its encoding settings. The default value is Balanced.
- Key
Frame stringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Layers []H264Layer
The collection of output H.264 layers to be produced by the encoder.
- Scene
Change boolDetection Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- Stretch
Mode string | StretchMode The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string | VideoSync Mode The Video Sync Mode
- complexity String | H264Complexity
Tells the encoder how to choose its encoding settings. The default value is Balanced.
- key
Frame StringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- layers List<H264Layer>
The collection of output H.264 layers to be produced by the encoder.
- scene
Change BooleanDetection Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch
Mode String | StretchMode The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String | VideoSync Mode The Video Sync Mode
- complexity string | H264Complexity
Tells the encoder how to choose its encoding settings. The default value is Balanced.
- key
Frame stringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label string
An optional label for the codec. The label can be used to control muxing behavior.
- layers H264Layer[]
The collection of output H.264 layers to be produced by the encoder.
- scene
Change booleanDetection Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch
Mode string | StretchMode The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode string | VideoSync Mode The Video Sync Mode
- complexity str | H264Complexity
Tells the encoder how to choose its encoding settings. The default value is Balanced.
- key_
frame_ strinterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label str
An optional label for the codec. The label can be used to control muxing behavior.
- layers Sequence[H264Layer]
The collection of output H.264 layers to be produced by the encoder.
- scene_
change_ booldetection Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch_
mode str | StretchMode The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync_
mode str | VideoSync Mode The Video Sync Mode
- complexity String | "Speed" | "Balanced" | "Quality"
Tells the encoder how to choose its encoding settings. The default value is Balanced.
- key
Frame StringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- layers List<Property Map>
The collection of output H.264 layers to be produced by the encoder.
- scene
Change BooleanDetection Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch
Mode String | "None" | "AutoSize" | "Auto Fit" The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String | "Auto" | "Passthrough" | "Cfr" | "Vfr" The Video Sync Mode
H264VideoProfile
- Auto
- Auto
Tells the encoder to automatically determine the appropriate H.264 profile.
- Baseline
- Baseline
Baseline profile
- Main
- Main
Main profile
- High
- High
High profile.
- High422
- High422
High 4:2:2 profile.
- High444
- High444
High 4:4:4 predictive profile.
- H264Video
Profile Auto - Auto
Tells the encoder to automatically determine the appropriate H.264 profile.
- H264Video
Profile Baseline - Baseline
Baseline profile
- H264Video
Profile Main - Main
Main profile
- H264Video
Profile High - High
High profile.
- H264Video
Profile High422 - High422
High 4:2:2 profile.
- H264Video
Profile High444 - High444
High 4:4:4 predictive profile.
- Auto
- Auto
Tells the encoder to automatically determine the appropriate H.264 profile.
- Baseline
- Baseline
Baseline profile
- Main
- Main
Main profile
- High
- High
High profile.
- High422
- High422
High 4:2:2 profile.
- High444
- High444
High 4:4:4 predictive profile.
- Auto
- Auto
Tells the encoder to automatically determine the appropriate H.264 profile.
- Baseline
- Baseline
Baseline profile
- Main
- Main
Main profile
- High
- High
High profile.
- High422
- High422
High 4:2:2 profile.
- High444
- High444
High 4:4:4 predictive profile.
- AUTO
- Auto
Tells the encoder to automatically determine the appropriate H.264 profile.
- BASELINE
- Baseline
Baseline profile
- MAIN
- Main
Main profile
- HIGH
- High
High profile.
- HIGH422
- High422
High 4:2:2 profile.
- HIGH444
- High444
High 4:4:4 predictive profile.
- "Auto"
- Auto
Tells the encoder to automatically determine the appropriate H.264 profile.
- "Baseline"
- Baseline
Baseline profile
- "Main"
- Main
Main profile
- "High"
- High
High profile.
- "High422"
- High422
High 4:2:2 profile.
- "High444"
- High444
High 4:4:4 predictive profile.
H264VideoResponse
- Complexity string
Tells the encoder how to choose its encoding settings. The default value is Balanced.
- Key
Frame stringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Layers
List<Pulumi.
Azure Native. Media. Inputs. H264Layer Response> The collection of output H.264 layers to be produced by the encoder.
- Scene
Change boolDetection Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- Stretch
Mode string The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string The Video Sync Mode
- Complexity string
Tells the encoder how to choose its encoding settings. The default value is Balanced.
- Key
Frame stringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Layers
[]H264Layer
Response The collection of output H.264 layers to be produced by the encoder.
- Scene
Change boolDetection Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- Stretch
Mode string The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string The Video Sync Mode
- complexity String
Tells the encoder how to choose its encoding settings. The default value is Balanced.
- key
Frame StringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- layers
List<H264Layer
Response> The collection of output H.264 layers to be produced by the encoder.
- scene
Change BooleanDetection Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch
Mode String The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String The Video Sync Mode
- complexity string
Tells the encoder how to choose its encoding settings. The default value is Balanced.
- key
Frame stringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label string
An optional label for the codec. The label can be used to control muxing behavior.
- layers
H264Layer
Response[] The collection of output H.264 layers to be produced by the encoder.
- scene
Change booleanDetection Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch
Mode string The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode string The Video Sync Mode
- complexity str
Tells the encoder how to choose its encoding settings. The default value is Balanced.
- key_
frame_ strinterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label str
An optional label for the codec. The label can be used to control muxing behavior.
- layers
Sequence[H264Layer
Response] The collection of output H.264 layers to be produced by the encoder.
- scene_
change_ booldetection Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch_
mode str The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync_
mode str The Video Sync Mode
- complexity String
Tells the encoder how to choose its encoding settings. The default value is Balanced.
- key
Frame StringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- layers List<Property Map>
The collection of output H.264 layers to be produced by the encoder.
- scene
Change BooleanDetection Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch
Mode String The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String The Video Sync Mode
H265Complexity
- Speed
- Speed
Tells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
- Balanced
- Balanced
Tells the encoder to use settings that achieve a balance between speed and quality.
- Quality
- Quality
Tells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
- H265Complexity
Speed - Speed
Tells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
- H265Complexity
Balanced - Balanced
Tells the encoder to use settings that achieve a balance between speed and quality.
- H265Complexity
Quality - Quality
Tells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
- Speed
- Speed
Tells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
- Balanced
- Balanced
Tells the encoder to use settings that achieve a balance between speed and quality.
- Quality
- Quality
Tells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
- Speed
- Speed
Tells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
- Balanced
- Balanced
Tells the encoder to use settings that achieve a balance between speed and quality.
- Quality
- Quality
Tells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
- SPEED
- Speed
Tells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
- BALANCED
- Balanced
Tells the encoder to use settings that achieve a balance between speed and quality.
- QUALITY
- Quality
Tells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
- "Speed"
- Speed
Tells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
- "Balanced"
- Balanced
Tells the encoder to use settings that achieve a balance between speed and quality.
- "Quality"
- Quality
Tells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
H265Layer
- Bitrate int
The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
- Adaptive
BFrame bool Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- BFrames int
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- Buffer
Window string The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- Frame
Rate string The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Level string
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- Max
Bitrate int The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- Profile
string | Pulumi.
Azure Native. Media. H265Video Profile We currently support Main. Default is Auto.
- Reference
Frames int The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- Slices int
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- Bitrate int
The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
- Adaptive
BFrame bool Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- BFrames int
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- Buffer
Window string The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- Frame
Rate string The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Level string
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- Max
Bitrate int The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- Profile
string | H265Video
Profile We currently support Main. Default is Auto.
- Reference
Frames int The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- Slices int
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate Integer
The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
- adaptive
BFrame Boolean Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b
Frames Integer The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer
Window String The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- frame
Rate String The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level String
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max
Bitrate Integer The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile
String | H265Video
Profile We currently support Main. Default is Auto.
- reference
Frames Integer The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices Integer
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate number
The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
- adaptive
BFrame boolean Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b
Frames number The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer
Window string The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- frame
Rate string The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level string
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max
Bitrate number The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile
string | H265Video
Profile We currently support Main. Default is Auto.
- reference
Frames number The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices number
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate int
The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
- adaptive_
b_ boolframe Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b_
frames int The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer_
window str The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- frame_
rate str The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height str
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label str
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level str
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max_
bitrate int The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile
str | H265Video
Profile We currently support Main. Default is Auto.
- reference_
frames int The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices int
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width str
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate Number
The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
- adaptive
BFrame Boolean Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b
Frames Number The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer
Window String The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- frame
Rate String The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level String
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max
Bitrate Number The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile String | "Auto" | "Main"
We currently support Main. Default is Auto.
- reference
Frames Number The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices Number
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
H265LayerResponse
- Bitrate int
The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
- Adaptive
BFrame bool Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- BFrames int
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- Buffer
Window string The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- Frame
Rate string The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Level string
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- Max
Bitrate int The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- Profile string
We currently support Main. Default is Auto.
- Reference
Frames int The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- Slices int
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- Bitrate int
The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
- Adaptive
BFrame bool Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- BFrames int
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- Buffer
Window string The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- Frame
Rate string The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Level string
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- Max
Bitrate int The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- Profile string
We currently support Main. Default is Auto.
- Reference
Frames int The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- Slices int
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate Integer
The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
- adaptive
BFrame Boolean Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b
Frames Integer The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer
Window String The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- frame
Rate String The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level String
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max
Bitrate Integer The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile String
We currently support Main. Default is Auto.
- reference
Frames Integer The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices Integer
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate number
The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
- adaptive
BFrame boolean Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b
Frames number The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer
Window string The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- frame
Rate string The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level string
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max
Bitrate number The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile string
We currently support Main. Default is Auto.
- reference
Frames number The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices number
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate int
The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
- adaptive_
b_ boolframe Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b_
frames int The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer_
window str The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- frame_
rate str The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height str
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label str
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level str
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max_
bitrate int The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile str
We currently support Main. Default is Auto.
- reference_
frames int The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices int
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width str
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate Number
The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
- adaptive
BFrame Boolean Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b
Frames Number The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer
Window String The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- frame
Rate String The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level String
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max
Bitrate Number The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile String
We currently support Main. Default is Auto.
- reference
Frames Number The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices Number
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
H265Video
- Complexity
string | Pulumi.
Azure Native. Media. H265Complexity Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
- Key
Frame stringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Layers
List<Pulumi.
Azure Native. Media. Inputs. H265Layer> The collection of output H.265 layers to be produced by the encoder.
- Scene
Change boolDetection Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- Stretch
Mode string | Pulumi.Azure Native. Media. Stretch Mode The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string | Pulumi.Azure Native. Media. Video Sync Mode The Video Sync Mode
- Complexity string | H265Complexity
Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
- Key
Frame stringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Layers []H265Layer
The collection of output H.265 layers to be produced by the encoder.
- Scene
Change boolDetection Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- Stretch
Mode string | StretchMode The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string | VideoSync Mode The Video Sync Mode
- complexity String | H265Complexity
Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
- key
Frame StringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- layers List<H265Layer>
The collection of output H.265 layers to be produced by the encoder.
- scene
Change BooleanDetection Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch
Mode String | StretchMode The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String | VideoSync Mode The Video Sync Mode
- complexity string | H265Complexity
Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
- key
Frame stringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label string
An optional label for the codec. The label can be used to control muxing behavior.
- layers H265Layer[]
The collection of output H.265 layers to be produced by the encoder.
- scene
Change booleanDetection Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch
Mode string | StretchMode The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode string | VideoSync Mode The Video Sync Mode
- complexity str | H265Complexity
Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
- key_
frame_ strinterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label str
An optional label for the codec. The label can be used to control muxing behavior.
- layers Sequence[H265Layer]
The collection of output H.265 layers to be produced by the encoder.
- scene_
change_ booldetection Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch_
mode str | StretchMode The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync_
mode str | VideoSync Mode The Video Sync Mode
- complexity String | "Speed" | "Balanced" | "Quality"
Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
- key
Frame StringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- layers List<Property Map>
The collection of output H.265 layers to be produced by the encoder.
- scene
Change BooleanDetection Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch
Mode String | "None" | "AutoSize" | "Auto Fit" The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String | "Auto" | "Passthrough" | "Cfr" | "Vfr" The Video Sync Mode
H265VideoProfile
- Auto
- Auto
Tells the encoder to automatically determine the appropriate H.265 profile.
- Main
- Main
Main profile (https://x265.readthedocs.io/en/default/cli.html?highlight=profile#profile-level-tier)
- H265Video
Profile Auto - Auto
Tells the encoder to automatically determine the appropriate H.265 profile.
- H265Video
Profile Main - Main
Main profile (https://x265.readthedocs.io/en/default/cli.html?highlight=profile#profile-level-tier)
- Auto
- Auto
Tells the encoder to automatically determine the appropriate H.265 profile.
- Main
- Main
Main profile (https://x265.readthedocs.io/en/default/cli.html?highlight=profile#profile-level-tier)
- Auto
- Auto
Tells the encoder to automatically determine the appropriate H.265 profile.
- Main
- Main
Main profile (https://x265.readthedocs.io/en/default/cli.html?highlight=profile#profile-level-tier)
- AUTO
- Auto
Tells the encoder to automatically determine the appropriate H.265 profile.
- MAIN
- Main
Main profile (https://x265.readthedocs.io/en/default/cli.html?highlight=profile#profile-level-tier)
- "Auto"
- Auto
Tells the encoder to automatically determine the appropriate H.265 profile.
- "Main"
- Main
Main profile (https://x265.readthedocs.io/en/default/cli.html?highlight=profile#profile-level-tier)
H265VideoResponse
- Complexity string
Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
- Key
Frame stringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Layers
List<Pulumi.
Azure Native. Media. Inputs. H265Layer Response> The collection of output H.265 layers to be produced by the encoder.
- Scene
Change boolDetection Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- Stretch
Mode string The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string The Video Sync Mode
- Complexity string
Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
- Key
Frame stringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Layers
[]H265Layer
Response The collection of output H.265 layers to be produced by the encoder.
- Scene
Change boolDetection Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- Stretch
Mode string The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string The Video Sync Mode
- complexity String
Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
- key
Frame StringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- layers
List<H265Layer
Response> The collection of output H.265 layers to be produced by the encoder.
- scene
Change BooleanDetection Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch
Mode String The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String The Video Sync Mode
- complexity string
Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
- key
Frame stringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label string
An optional label for the codec. The label can be used to control muxing behavior.
- layers
H265Layer
Response[] The collection of output H.265 layers to be produced by the encoder.
- scene
Change booleanDetection Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch
Mode string The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode string The Video Sync Mode
- complexity str
Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
- key_
frame_ strinterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label str
An optional label for the codec. The label can be used to control muxing behavior.
- layers
Sequence[H265Layer
Response] The collection of output H.265 layers to be produced by the encoder.
- scene_
change_ booldetection Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch_
mode str The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync_
mode str The Video Sync Mode
- complexity String
Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
- key
Frame StringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- layers List<Property Map>
The collection of output H.265 layers to be produced by the encoder.
- scene
Change BooleanDetection Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch
Mode String The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String The Video Sync Mode
Image
- Start string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- Key
Frame stringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- Step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- Stretch
Mode string | Pulumi.Azure Native. Media. Stretch Mode The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string | Pulumi.Azure Native. Media. Video Sync Mode The Video Sync Mode
- Start string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- Key
Frame stringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- Step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- Stretch
Mode string | StretchMode The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string | VideoSync Mode The Video Sync Mode
- start String
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame StringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- range String
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step String
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode String | StretchMode The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String | VideoSync Mode The Video Sync Mode
- start string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame stringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label string
An optional label for the codec. The label can be used to control muxing behavior.
- range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode string | StretchMode The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode string | VideoSync Mode The Video Sync Mode
- start str
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key_
frame_ strinterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label str
An optional label for the codec. The label can be used to control muxing behavior.
- range str
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step str
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch_
mode str | StretchMode The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync_
mode str | VideoSync Mode The Video Sync Mode
- start String
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame StringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- range String
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step String
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode String | "None" | "AutoSize" | "Auto Fit" The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String | "Auto" | "Passthrough" | "Cfr" | "Vfr" The Video Sync Mode
ImageFormat
- Filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern String The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename_
pattern str The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern String The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
ImageFormatResponse
- Filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern String The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename_
pattern str The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern String The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
ImageResponse
- Start string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- Key
Frame stringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- Step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- Stretch
Mode string The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string The Video Sync Mode
- Start string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- Key
Frame stringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- Step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- Stretch
Mode string The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string The Video Sync Mode
- start String
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame StringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- range String
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step String
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode String The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String The Video Sync Mode
- start string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame stringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label string
An optional label for the codec. The label can be used to control muxing behavior.
- range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode string The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode string The Video Sync Mode
- start str
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key_
frame_ strinterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label str
An optional label for the codec. The label can be used to control muxing behavior.
- range str
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step str
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch_
mode str The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync_
mode str The Video Sync Mode
- start String
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame StringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- range String
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step String
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode String The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String The Video Sync Mode
InsightsType
- Audio
Insights Only - AudioInsightsOnly
Generate audio only insights. Ignore video even if present. Fails if no audio is present.
- Video
Insights Only - VideoInsightsOnly
Generate video only insights. Ignore audio if present. Fails if no video is present.
- All
Insights - AllInsights
Generate both audio and video insights. Fails if either audio or video Insights fail.
- Insights
Type Audio Insights Only - AudioInsightsOnly
Generate audio only insights. Ignore video even if present. Fails if no audio is present.
- Insights
Type Video Insights Only - VideoInsightsOnly
Generate video only insights. Ignore audio if present. Fails if no video is present.
- Insights
Type All Insights - AllInsights
Generate both audio and video insights. Fails if either audio or video Insights fail.
- Audio
Insights Only - AudioInsightsOnly
Generate audio only insights. Ignore video even if present. Fails if no audio is present.
- Video
Insights Only - VideoInsightsOnly
Generate video only insights. Ignore audio if present. Fails if no video is present.
- All
Insights - AllInsights
Generate both audio and video insights. Fails if either audio or video Insights fail.
- Audio
Insights Only - AudioInsightsOnly
Generate audio only insights. Ignore video even if present. Fails if no audio is present.
- Video
Insights Only - VideoInsightsOnly
Generate video only insights. Ignore audio if present. Fails if no video is present.
- All
Insights - AllInsights
Generate both audio and video insights. Fails if either audio or video Insights fail.
- AUDIO_INSIGHTS_ONLY
- AudioInsightsOnly
Generate audio only insights. Ignore video even if present. Fails if no audio is present.
- VIDEO_INSIGHTS_ONLY
- VideoInsightsOnly
Generate video only insights. Ignore audio if present. Fails if no video is present.
- ALL_INSIGHTS
- AllInsights
Generate both audio and video insights. Fails if either audio or video Insights fail.
- "Audio
Insights Only" - AudioInsightsOnly
Generate audio only insights. Ignore video even if present. Fails if no audio is present.
- "Video
Insights Only" - VideoInsightsOnly
Generate video only insights. Ignore audio if present. Fails if no video is present.
- "All
Insights" - AllInsights
Generate both audio and video insights. Fails if either audio or video Insights fail.
JpgFormat
- Filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern String The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename_
pattern str The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern String The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
JpgFormatResponse
- Filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern String The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename_
pattern str The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern String The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
JpgImage
- Start string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- Key
Frame stringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Layers
List<Pulumi.
Azure Native. Media. Inputs. Jpg Layer> A collection of output JPEG image layers to be produced by the encoder.
- Range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- Sprite
Column int Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
- Step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- Stretch
Mode string | Pulumi.Azure Native. Media. Stretch Mode The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string | Pulumi.Azure Native. Media. Video Sync Mode The Video Sync Mode
- Start string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- Key
Frame stringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Layers
[]Jpg
Layer A collection of output JPEG image layers to be produced by the encoder.
- Range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- Sprite
Column int Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
- Step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- Stretch
Mode string | StretchMode The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string | VideoSync Mode The Video Sync Mode
- start String
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame StringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- layers
List<Jpg
Layer> A collection of output JPEG image layers to be produced by the encoder.
- range String
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- sprite
Column Integer Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
- step String
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode String | StretchMode The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String | VideoSync Mode The Video Sync Mode
- start string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame stringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label string
An optional label for the codec. The label can be used to control muxing behavior.
- layers
Jpg
Layer[] A collection of output JPEG image layers to be produced by the encoder.
- range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- sprite
Column number Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
- step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode string | StretchMode The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode string | VideoSync Mode The Video Sync Mode
- start str
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key_
frame_ strinterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label str
An optional label for the codec. The label can be used to control muxing behavior.
- layers
Sequence[Jpg
Layer] A collection of output JPEG image layers to be produced by the encoder.
- range str
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- sprite_
column int Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
- step str
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch_
mode str | StretchMode The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync_
mode str | VideoSync Mode The Video Sync Mode
- start String
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame StringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- layers List<Property Map>
A collection of output JPEG image layers to be produced by the encoder.
- range String
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- sprite
Column Number Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
- step String
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode String | "None" | "AutoSize" | "Auto Fit" The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String | "Auto" | "Passthrough" | "Cfr" | "Vfr" The Video Sync Mode
JpgImageResponse
- Start string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- Key
Frame stringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Layers
List<Pulumi.
Azure Native. Media. Inputs. Jpg Layer Response> A collection of output JPEG image layers to be produced by the encoder.
- Range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- Sprite
Column int Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
- Step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- Stretch
Mode string The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string The Video Sync Mode
- Start string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- Key
Frame stringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Layers
[]Jpg
Layer Response A collection of output JPEG image layers to be produced by the encoder.
- Range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- Sprite
Column int Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
- Step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- Stretch
Mode string The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string The Video Sync Mode
- start String
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame StringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- layers
List<Jpg
Layer Response> A collection of output JPEG image layers to be produced by the encoder.
- range String
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- sprite
Column Integer Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
- step String
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode String The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String The Video Sync Mode
- start string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame stringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label string
An optional label for the codec. The label can be used to control muxing behavior.
- layers
Jpg
Layer Response[] A collection of output JPEG image layers to be produced by the encoder.
- range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- sprite
Column number Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
- step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode string The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode string The Video Sync Mode
- start str
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key_
frame_ strinterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label str
An optional label for the codec. The label can be used to control muxing behavior.
- layers
Sequence[Jpg
Layer Response] A collection of output JPEG image layers to be produced by the encoder.
- range str
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- sprite_
column int Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
- step str
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch_
mode str The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync_
mode str The Video Sync Mode
- start String
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame StringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- layers List<Property Map>
A collection of output JPEG image layers to be produced by the encoder.
- range String
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- sprite
Column Number Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
- step String
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode String The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String The Video Sync Mode
JpgLayer
- Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Quality int
The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
- Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Quality int
The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
- Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- quality Integer
The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
- width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- quality number
The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
- width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height str
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label str
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- quality int
The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
- width str
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- quality Number
The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
- width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
JpgLayerResponse
- Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Quality int
The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
- Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Quality int
The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
- Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- quality Integer
The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
- width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- quality number
The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
- width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height str
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label str
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- quality int
The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
- width str
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- quality Number
The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
- width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
Mp4Format
- Filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Output
Files List<Pulumi.Azure Native. Media. Inputs. Output File> The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- Filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Output
Files []OutputFile The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern String The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files List<OutputFile> The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files OutputFile[] The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename_
pattern str The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output_
files Sequence[OutputFile] The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern String The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files List<Property Map> The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
Mp4FormatResponse
- Filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Output
Files List<Pulumi.Azure Native. Media. Inputs. Output File Response> The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- Filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Output
Files []OutputFile Response The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern String The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files List<OutputFile Response> The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files OutputFile Response[] The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename_
pattern str The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output_
files Sequence[OutputFile Response] The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern String The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files List<Property Map> The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
MultiBitrateFormat
- Filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Output
Files List<Pulumi.Azure Native. Media. Inputs. Output File> The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- Filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Output
Files []OutputFile The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern String The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files List<OutputFile> The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files OutputFile[] The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename_
pattern str The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output_
files Sequence[OutputFile] The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern String The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files List<Property Map> The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
MultiBitrateFormatResponse
- Filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Output
Files List<Pulumi.Azure Native. Media. Inputs. Output File Response> The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- Filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Output
Files []OutputFile Response The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern String The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files List<OutputFile Response> The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files OutputFile Response[] The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename_
pattern str The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output_
files Sequence[OutputFile Response] The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern String The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files List<Property Map> The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
OnErrorType
- Stop
Processing Job - StopProcessingJob
Tells the service that if this TransformOutput fails, then any other incomplete TransformOutputs can be stopped.
- Continue
Job - ContinueJob
Tells the service that if this TransformOutput fails, then allow any other TransformOutput to continue.
- On
Error Type Stop Processing Job - StopProcessingJob
Tells the service that if this TransformOutput fails, then any other incomplete TransformOutputs can be stopped.
- On
Error Type Continue Job - ContinueJob
Tells the service that if this TransformOutput fails, then allow any other TransformOutput to continue.
- Stop
Processing Job - StopProcessingJob
Tells the service that if this TransformOutput fails, then any other incomplete TransformOutputs can be stopped.
- Continue
Job - ContinueJob
Tells the service that if this TransformOutput fails, then allow any other TransformOutput to continue.
- Stop
Processing Job - StopProcessingJob
Tells the service that if this TransformOutput fails, then any other incomplete TransformOutputs can be stopped.
- Continue
Job - ContinueJob
Tells the service that if this TransformOutput fails, then allow any other TransformOutput to continue.
- STOP_PROCESSING_JOB
- StopProcessingJob
Tells the service that if this TransformOutput fails, then any other incomplete TransformOutputs can be stopped.
- CONTINUE_JOB
- ContinueJob
Tells the service that if this TransformOutput fails, then allow any other TransformOutput to continue.
- "Stop
Processing Job" - StopProcessingJob
Tells the service that if this TransformOutput fails, then any other incomplete TransformOutputs can be stopped.
- "Continue
Job" - ContinueJob
Tells the service that if this TransformOutput fails, then allow any other TransformOutput to continue.
OutputFile
- Labels List<string>
The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
- Labels []string
The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
- labels List<String>
The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
- labels string[]
The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
- labels Sequence[str]
The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
- labels List<String>
The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
OutputFileResponse
- Labels List<string>
The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
- Labels []string
The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
- labels List<String>
The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
- labels string[]
The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
- labels Sequence[str]
The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
- labels List<String>
The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
PngFormat
- Filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern String The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename_
pattern str The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern String The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
PngFormatResponse
- Filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern String The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename_
pattern str The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern String The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
PngImage
- Start string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- Key
Frame stringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Layers
List<Pulumi.
Azure Native. Media. Inputs. Png Layer> A collection of output PNG image layers to be produced by the encoder.
- Range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- Step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- Stretch
Mode string | Pulumi.Azure Native. Media. Stretch Mode The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string | Pulumi.Azure Native. Media. Video Sync Mode The Video Sync Mode
- Start string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- Key
Frame stringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Layers
[]Png
Layer A collection of output PNG image layers to be produced by the encoder.
- Range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- Step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- Stretch
Mode string | StretchMode The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string | VideoSync Mode The Video Sync Mode
- start String
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame StringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- layers
List<Png
Layer> A collection of output PNG image layers to be produced by the encoder.
- range String
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step String
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode String | StretchMode The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String | VideoSync Mode The Video Sync Mode
- start string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame stringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label string
An optional label for the codec. The label can be used to control muxing behavior.
- layers
Png
Layer[] A collection of output PNG image layers to be produced by the encoder.
- range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode string | StretchMode The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode string | VideoSync Mode The Video Sync Mode
- start str
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key_
frame_ strinterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label str
An optional label for the codec. The label can be used to control muxing behavior.
- layers
Sequence[Png
Layer] A collection of output PNG image layers to be produced by the encoder.
- range str
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step str
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch_
mode str | StretchMode The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync_
mode str | VideoSync Mode The Video Sync Mode
- start String
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame StringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- layers List<Property Map>
A collection of output PNG image layers to be produced by the encoder.
- range String
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step String
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode String | "None" | "AutoSize" | "Auto Fit" The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String | "Auto" | "Passthrough" | "Cfr" | "Vfr" The Video Sync Mode
PngImageResponse
- Start string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- Key
Frame stringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Layers
List<Pulumi.
Azure Native. Media. Inputs. Png Layer Response> A collection of output PNG image layers to be produced by the encoder.
- Range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- Step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- Stretch
Mode string The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string The Video Sync Mode
- Start string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- Key
Frame stringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
An optional label for the codec. The label can be used to control muxing behavior.
- Layers
[]Png
Layer Response A collection of output PNG image layers to be produced by the encoder.
- Range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- Step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- Stretch
Mode string The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string The Video Sync Mode
- start String
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame StringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- layers
List<Png
Layer Response> A collection of output PNG image layers to be produced by the encoder.
- range String
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step String
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode String The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String The Video Sync Mode
- start string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame stringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label string
An optional label for the codec. The label can be used to control muxing behavior.
- layers
Png
Layer Response[] A collection of output PNG image layers to be produced by the encoder.
- range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode string The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode string The Video Sync Mode
- start str
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key_
frame_ strinterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label str
An optional label for the codec. The label can be used to control muxing behavior.
- layers
Sequence[Png
Layer Response] A collection of output PNG image layers to be produced by the encoder.
- range str
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step str
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch_
mode str The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync_
mode str The Video Sync Mode
- start String
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame StringInterval The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
An optional label for the codec. The label can be used to control muxing behavior.
- layers List<Property Map>
A collection of output PNG image layers to be produced by the encoder.
- range String
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step String
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode String The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String The Video Sync Mode
PngLayer
- Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height str
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label str
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- width str
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
PngLayerResponse
- Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height str
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label str
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- width str
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
Priority
- Low
- Low
Used for TransformOutputs that can be generated after Normal and High priority TransformOutputs.
- Normal
- Normal
Used for TransformOutputs that can be generated at Normal priority.
- High
- High
Used for TransformOutputs that should take precedence over others.
- Priority
Low - Low
Used for TransformOutputs that can be generated after Normal and High priority TransformOutputs.
- Priority
Normal - Normal
Used for TransformOutputs that can be generated at Normal priority.
- Priority
High - High
Used for TransformOutputs that should take precedence over others.
- Low
- Low
Used for TransformOutputs that can be generated after Normal and High priority TransformOutputs.
- Normal
- Normal
Used for TransformOutputs that can be generated at Normal priority.
- High
- High
Used for TransformOutputs that should take precedence over others.
- Low
- Low
Used for TransformOutputs that can be generated after Normal and High priority TransformOutputs.
- Normal
- Normal
Used for TransformOutputs that can be generated at Normal priority.
- High
- High
Used for TransformOutputs that should take precedence over others.
- LOW
- Low
Used for TransformOutputs that can be generated after Normal and High priority TransformOutputs.
- NORMAL
- Normal
Used for TransformOutputs that can be generated at Normal priority.
- HIGH
- High
Used for TransformOutputs that should take precedence over others.
- "Low"
- Low
Used for TransformOutputs that can be generated after Normal and High priority TransformOutputs.
- "Normal"
- Normal
Used for TransformOutputs that can be generated at Normal priority.
- "High"
- High
Used for TransformOutputs that should take precedence over others.
Rectangle
- Height string
The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- Left string
The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- Top string
The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- Width string
The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- Height string
The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- Left string
The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- Top string
The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- Width string
The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- height String
The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- left String
The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- top String
The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- width String
The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- height string
The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- left string
The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- top string
The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- width string
The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- height str
The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- left str
The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- top str
The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- width str
The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- height String
The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- left String
The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- top String
The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- width String
The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
RectangleResponse
- Height string
The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- Left string
The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- Top string
The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- Width string
The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- Height string
The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- Left string
The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- Top string
The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- Width string
The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- height String
The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- left String
The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- top String
The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- width String
The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- height string
The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- left string
The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- top string
The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- width string
The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- height str
The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- left str
The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- top str
The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- width str
The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- height String
The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- left String
The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- top String
The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- width String
The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
Rotation
- Auto
- Auto
Automatically detect and rotate as needed.
- None
- None
Do not rotate the video. If the output format supports it, any metadata about rotation is kept intact.
- Rotate0
- Rotate0
Do not rotate the video but remove any metadata about the rotation.
- Rotate90
- Rotate90
Rotate 90 degrees clockwise.
- Rotate180
- Rotate180
Rotate 180 degrees clockwise.
- Rotate270
- Rotate270
Rotate 270 degrees clockwise.
- Rotation
Auto - Auto
Automatically detect and rotate as needed.
- Rotation
None - None
Do not rotate the video. If the output format supports it, any metadata about rotation is kept intact.
- Rotation
Rotate0 - Rotate0
Do not rotate the video but remove any metadata about the rotation.
- Rotation
Rotate90 - Rotate90
Rotate 90 degrees clockwise.
- Rotation
Rotate180 - Rotate180
Rotate 180 degrees clockwise.
- Rotation
Rotate270 - Rotate270
Rotate 270 degrees clockwise.
- Auto
- Auto
Automatically detect and rotate as needed.
- None
- None
Do not rotate the video. If the output format supports it, any metadata about rotation is kept intact.
- Rotate0
- Rotate0
Do not rotate the video but remove any metadata about the rotation.
- Rotate90
- Rotate90
Rotate 90 degrees clockwise.
- Rotate180
- Rotate180
Rotate 180 degrees clockwise.
- Rotate270
- Rotate270
Rotate 270 degrees clockwise.
- Auto
- Auto
Automatically detect and rotate as needed.
- None
- None
Do not rotate the video. If the output format supports it, any metadata about rotation is kept intact.
- Rotate0
- Rotate0
Do not rotate the video but remove any metadata about the rotation.
- Rotate90
- Rotate90
Rotate 90 degrees clockwise.
- Rotate180
- Rotate180
Rotate 180 degrees clockwise.
- Rotate270
- Rotate270
Rotate 270 degrees clockwise.
- AUTO
- Auto
Automatically detect and rotate as needed.
- NONE
- None
Do not rotate the video. If the output format supports it, any metadata about rotation is kept intact.
- ROTATE0
- Rotate0
Do not rotate the video but remove any metadata about the rotation.
- ROTATE90
- Rotate90
Rotate 90 degrees clockwise.
- ROTATE180
- Rotate180
Rotate 180 degrees clockwise.
- ROTATE270
- Rotate270
Rotate 270 degrees clockwise.
- "Auto"
- Auto
Automatically detect and rotate as needed.
- "None"
- None
Do not rotate the video. If the output format supports it, any metadata about rotation is kept intact.
- "Rotate0"
- Rotate0
Do not rotate the video but remove any metadata about the rotation.
- "Rotate90"
- Rotate90
Rotate 90 degrees clockwise.
- "Rotate180"
- Rotate180
Rotate 180 degrees clockwise.
- "Rotate270"
- Rotate270
Rotate 270 degrees clockwise.
StandardEncoderPreset
- Codecs List<object>
The list of codecs to be used when encoding the input video.
- Formats List<object>
The list of outputs to be produced by the encoder.
- Filters
Pulumi.
Azure Native. Media. Inputs. Filters One or more filtering operations that are applied to the input media before encoding.
- codecs
(Aac
Audio | Audio | Copy Audio | Copy Video | H264Video | H265Video | Image | Jpg Image | Png Image | Video)[] The list of codecs to be used when encoding the input video.
- formats
(Image
Format | Jpg Format | Mp4Format | Multi Bitrate Format | Png Format | Transport Stream Format)[] The list of outputs to be produced by the encoder.
- filters Filters
One or more filtering operations that are applied to the input media before encoding.
- codecs
Sequence[Union[Aac
Audio, Audio, Copy Audio, Copy Video, H264Video, H265Video, Image, Jpg Image, Png Image, Video]] The list of codecs to be used when encoding the input video.
- formats
Sequence[Union[Image
Format, Jpg Format, Mp4Format, Multi Bitrate Format, Png Format, Transport Stream Format]] The list of outputs to be produced by the encoder.
- filters Filters
One or more filtering operations that are applied to the input media before encoding.
- codecs List<Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map>
The list of codecs to be used when encoding the input video.
- formats List<Property Map | Property Map | Property Map | Property Map | Property Map | Property Map>
The list of outputs to be produced by the encoder.
- filters Property Map
One or more filtering operations that are applied to the input media before encoding.
StandardEncoderPresetResponse
- Codecs List<object>
The list of codecs to be used when encoding the input video.
- Formats List<object>
The list of outputs to be produced by the encoder.
- Filters
Pulumi.
Azure Native. Media. Inputs. Filters Response One or more filtering operations that are applied to the input media before encoding.
- Codecs []interface{}
The list of codecs to be used when encoding the input video.
- Formats []interface{}
The list of outputs to be produced by the encoder.
- Filters
Filters
Response One or more filtering operations that are applied to the input media before encoding.
- codecs List<Object>
The list of codecs to be used when encoding the input video.
- formats List<Object>
The list of outputs to be produced by the encoder.
- filters
Filters
Response One or more filtering operations that are applied to the input media before encoding.
- codecs
(Aac
Audio Response | Audio Response | Copy Audio Response | Copy Video Response | H264Video Response | H265Video Response | Image Response | Jpg Image Response | Png Image Response | Video Response)[] The list of codecs to be used when encoding the input video.
- formats
(Image
Format Response | Jpg Format Response | Mp4Format Response | Multi Bitrate Format Response | Png Format Response | Transport Stream Format Response)[] The list of outputs to be produced by the encoder.
- filters
Filters
Response One or more filtering operations that are applied to the input media before encoding.
- codecs
Sequence[Union[Aac
Audio Response, Audio Response, Copy Audio Response, Copy Video Response, H264Video Response, H265Video Response, Image Response, Jpg Image Response, Png Image Response, Video Response]] The list of codecs to be used when encoding the input video.
- formats
Sequence[Union[Image
Format Response, Jpg Format Response, Mp4Format Response, Multi Bitrate Format Response, Png Format Response, Transport Stream Format Response]] The list of outputs to be produced by the encoder.
- filters
Filters
Response One or more filtering operations that are applied to the input media before encoding.
- codecs List<Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map>
The list of codecs to be used when encoding the input video.
- formats List<Property Map | Property Map | Property Map | Property Map | Property Map | Property Map>
The list of outputs to be produced by the encoder.
- filters Property Map
One or more filtering operations that are applied to the input media before encoding.
StretchMode
- None
- None
Strictly respect the output resolution without considering the pixel aspect ratio or display aspect ratio of the input video.
- Auto
Size - AutoSize
Override the output resolution, and change it to match the display aspect ratio of the input, without padding. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the value in the preset is overridden, and the output will be at 1280x720, which maintains the input aspect ratio of 16:9.
- Auto
Fit - AutoFit
Pad the output (with either letterbox or pillar box) to honor the output resolution, while ensuring that the active video region in the output has the same aspect ratio as the input. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the output will be at 1280x1280, which contains an inner rectangle of 1280x720 at aspect ratio of 16:9, and pillar box regions 280 pixels wide at the left and right.
- Stretch
Mode None - None
Strictly respect the output resolution without considering the pixel aspect ratio or display aspect ratio of the input video.
- Stretch
Mode Auto Size - AutoSize
Override the output resolution, and change it to match the display aspect ratio of the input, without padding. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the value in the preset is overridden, and the output will be at 1280x720, which maintains the input aspect ratio of 16:9.
- Stretch
Mode Auto Fit - AutoFit
Pad the output (with either letterbox or pillar box) to honor the output resolution, while ensuring that the active video region in the output has the same aspect ratio as the input. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the output will be at 1280x1280, which contains an inner rectangle of 1280x720 at aspect ratio of 16:9, and pillar box regions 280 pixels wide at the left and right.
- None
- None
Strictly respect the output resolution without considering the pixel aspect ratio or display aspect ratio of the input video.
- Auto
Size - AutoSize
Override the output resolution, and change it to match the display aspect ratio of the input, without padding. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the value in the preset is overridden, and the output will be at 1280x720, which maintains the input aspect ratio of 16:9.
- Auto
Fit - AutoFit
Pad the output (with either letterbox or pillar box) to honor the output resolution, while ensuring that the active video region in the output has the same aspect ratio as the input. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the output will be at 1280x1280, which contains an inner rectangle of 1280x720 at aspect ratio of 16:9, and pillar box regions 280 pixels wide at the left and right.
- None
- None
Strictly respect the output resolution without considering the pixel aspect ratio or display aspect ratio of the input video.
- Auto
Size - AutoSize
Override the output resolution, and change it to match the display aspect ratio of the input, without padding. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the value in the preset is overridden, and the output will be at 1280x720, which maintains the input aspect ratio of 16:9.
- Auto
Fit - AutoFit
Pad the output (with either letterbox or pillar box) to honor the output resolution, while ensuring that the active video region in the output has the same aspect ratio as the input. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the output will be at 1280x1280, which contains an inner rectangle of 1280x720 at aspect ratio of 16:9, and pillar box regions 280 pixels wide at the left and right.
- NONE
- None
Strictly respect the output resolution without considering the pixel aspect ratio or display aspect ratio of the input video.
- AUTO_SIZE
- AutoSize
Override the output resolution, and change it to match the display aspect ratio of the input, without padding. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the value in the preset is overridden, and the output will be at 1280x720, which maintains the input aspect ratio of 16:9.
- AUTO_FIT
- AutoFit
Pad the output (with either letterbox or pillar box) to honor the output resolution, while ensuring that the active video region in the output has the same aspect ratio as the input. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the output will be at 1280x1280, which contains an inner rectangle of 1280x720 at aspect ratio of 16:9, and pillar box regions 280 pixels wide at the left and right.
- "None"
- None
Strictly respect the output resolution without considering the pixel aspect ratio or display aspect ratio of the input video.
- "Auto
Size" - AutoSize
Override the output resolution, and change it to match the display aspect ratio of the input, without padding. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the value in the preset is overridden, and the output will be at 1280x720, which maintains the input aspect ratio of 16:9.
- "Auto
Fit" - AutoFit
Pad the output (with either letterbox or pillar box) to honor the output resolution, while ensuring that the active video region in the output has the same aspect ratio as the input. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the output will be at 1280x1280, which contains an inner rectangle of 1280x720 at aspect ratio of 16:9, and pillar box regions 280 pixels wide at the left and right.
SystemDataResponse
- Created
At string The timestamp of resource creation (UTC).
- Created
By string The identity that created the resource.
- Created
By stringType The type of identity that created the resource.
- Last
Modified stringAt The timestamp of resource last modification (UTC)
- Last
Modified stringBy The identity that last modified the resource.
- Last
Modified stringBy Type The type of identity that last modified the resource.
- Created
At string The timestamp of resource creation (UTC).
- Created
By string The identity that created the resource.
- Created
By stringType The type of identity that created the resource.
- Last
Modified stringAt The timestamp of resource last modification (UTC)
- Last
Modified stringBy The identity that last modified the resource.
- Last
Modified stringBy Type The type of identity that last modified the resource.
- created
At String The timestamp of resource creation (UTC).
- created
By String The identity that created the resource.
- created
By StringType The type of identity that created the resource.
- last
Modified StringAt The timestamp of resource last modification (UTC)
- last
Modified StringBy The identity that last modified the resource.
- last
Modified StringBy Type The type of identity that last modified the resource.
- created
At string The timestamp of resource creation (UTC).
- created
By string The identity that created the resource.
- created
By stringType The type of identity that created the resource.
- last
Modified stringAt The timestamp of resource last modification (UTC)
- last
Modified stringBy The identity that last modified the resource.
- last
Modified stringBy Type The type of identity that last modified the resource.
- created_
at str The timestamp of resource creation (UTC).
- created_
by str The identity that created the resource.
- created_
by_ strtype The type of identity that created the resource.
- last_
modified_ strat The timestamp of resource last modification (UTC)
- last_
modified_ strby The identity that last modified the resource.
- last_
modified_ strby_ type The type of identity that last modified the resource.
- created
At String The timestamp of resource creation (UTC).
- created
By String The identity that created the resource.
- created
By StringType The type of identity that created the resource.
- last
Modified StringAt The timestamp of resource last modification (UTC)
- last
Modified StringBy The identity that last modified the resource.
- last
Modified StringBy Type The type of identity that last modified the resource.
TransformOutput
- Preset
Pulumi.
Azure | Pulumi.Native. Media. Inputs. Audio Analyzer Preset Azure | Pulumi.Native. Media. Inputs. Built In Standard Encoder Preset Azure | Pulumi.Native. Media. Inputs. Face Detector Preset Azure | Pulumi.Native. Media. Inputs. Standard Encoder Preset Azure Native. Media. Inputs. Video Analyzer Preset Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
- On
Error string | Pulumi.Azure Native. Media. On Error Type A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
- Relative
Priority string | Pulumi.Azure Native. Media. Priority Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
- Preset
Audio
Analyzer | BuiltPreset In | FaceStandard Encoder Preset Detector | StandardPreset Encoder | VideoPreset Analyzer Preset Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
- On
Error string | OnError Type A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
- Relative
Priority string | Priority Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
- preset
Audio
Analyzer | BuiltPreset In | FaceStandard Encoder Preset Detector | StandardPreset Encoder | VideoPreset Analyzer Preset Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
- on
Error String | OnError Type A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
- relative
Priority String | Priority Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
- preset
Audio
Analyzer | BuiltPreset In | FaceStandard Encoder Preset Detector | StandardPreset Encoder | VideoPreset Analyzer Preset Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
- on
Error string | OnError Type A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
- relative
Priority string | Priority Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
- preset
Audio
Analyzer | BuiltPreset In | FaceStandard Encoder Preset Detector | StandardPreset Encoder | VideoPreset Analyzer Preset Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
- on_
error str | OnError Type A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
- relative_
priority str | Priority Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
- preset Property Map | Property Map | Property Map | Property Map | Property Map
Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
- on
Error String | "StopProcessing Job" | "Continue Job" A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
- relative
Priority String | "Low" | "Normal" | "High" Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
TransformOutputResponse
- Preset
Pulumi.
Azure | Pulumi.Native. Media. Inputs. Audio Analyzer Preset Response Azure | Pulumi.Native. Media. Inputs. Built In Standard Encoder Preset Response Azure | Pulumi.Native. Media. Inputs. Face Detector Preset Response Azure | Pulumi.Native. Media. Inputs. Standard Encoder Preset Response Azure Native. Media. Inputs. Video Analyzer Preset Response Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
- On
Error string A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
- Relative
Priority string Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
- Preset
Audio
Analyzer | BuiltPreset Response In | FaceStandard Encoder Preset Response Detector | StandardPreset Response Encoder | VideoPreset Response Analyzer Preset Response Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
- On
Error string A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
- Relative
Priority string Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
- preset
Audio
Analyzer | BuiltPreset Response In | FaceStandard Encoder Preset Response Detector | StandardPreset Response Encoder | VideoPreset Response Analyzer Preset Response Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
- on
Error String A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
- relative
Priority String Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
- preset
Audio
Analyzer | BuiltPreset Response In | FaceStandard Encoder Preset Response Detector | StandardPreset Response Encoder | VideoPreset Response Analyzer Preset Response Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
- on
Error string A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
- relative
Priority string Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
- preset
Audio
Analyzer | BuiltPreset Response In | FaceStandard Encoder Preset Response Detector | StandardPreset Response Encoder | VideoPreset Response Analyzer Preset Response Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
- on_
error str A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
- relative_
priority str Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
- preset Property Map | Property Map | Property Map | Property Map | Property Map
Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
- on
Error String A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
- relative
Priority String Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
TransportStreamFormat
- Filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Output
Files List<Pulumi.Azure Native. Media. Inputs. Output File> The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- Filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Output
Files []OutputFile The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern String The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files List<OutputFile> The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files OutputFile[] The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename_
pattern str The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output_
files Sequence[OutputFile] The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern String The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files List<Property Map> The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
TransportStreamFormatResponse
- Filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Output
Files List<Pulumi.Azure Native. Media. Inputs. Output File Response> The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- Filename
Pattern string The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.