1. Packages
  2. Azure Native
  3. API Docs
  4. media
  5. Transform
This is the latest version of Azure Native. Use the Azure Native v2 docs if using the v2 version of this package.
Azure Native v3.2.0 published on Monday, Apr 14, 2025 by Pulumi

azure-native.media.Transform

Explore with Pulumi AI

This is the latest version of Azure Native. Use the Azure Native v2 docs if using the v2 version of this package.
Azure Native v3.2.0 published on Monday, Apr 14, 2025 by Pulumi

A Transform encapsulates the rules or instructions for generating desired outputs from input media, such as by transcoding or by extracting insights. After the Transform is created, it can be applied to input media by creating Jobs.

Uses Azure REST API version 2022-07-01. In version 2.x of the Azure Native provider, it used API version 2022-07-01.

Other available API versions: 2018-03-30-preview, 2018-06-01-preview, 2018-07-01, 2020-05-01, 2021-06-01, 2021-11-01, 2022-05-01-preview. These can be accessed by generating a local SDK package using the CLI command pulumi package add azure-native media [ApiVersion]. See the version guide for details.

Example Usage

Create or update a Transform

using System.Collections.Generic;
using System.Linq;
using Pulumi;
using AzureNative = Pulumi.AzureNative;

return await Deployment.RunAsync(() => 
{
    var transform = new AzureNative.Media.Transform("transform", new()
    {
        AccountName = "contosomedia",
        Description = "Example Transform to illustrate create and update.",
        Outputs = new[]
        {
            new AzureNative.Media.Inputs.TransformOutputArgs
            {
                Preset = new AzureNative.Media.Inputs.BuiltInStandardEncoderPresetArgs
                {
                    OdataType = "#Microsoft.Media.BuiltInStandardEncoderPreset",
                    PresetName = AzureNative.Media.EncoderNamedPreset.AdaptiveStreaming,
                },
            },
        },
        ResourceGroupName = "contosoresources",
        TransformName = "createdTransform",
    });

});
Copy
package main

import (
	media "github.com/pulumi/pulumi-azure-native-sdk/media/v3"
	"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
)

func main() {
	pulumi.Run(func(ctx *pulumi.Context) error {
		_, err := media.NewTransform(ctx, "transform", &media.TransformArgs{
			AccountName: pulumi.String("contosomedia"),
			Description: pulumi.String("Example Transform to illustrate create and update."),
			Outputs: media.TransformOutputTypeArray{
				&media.TransformOutputTypeArgs{
					Preset: media.BuiltInStandardEncoderPreset{
						OdataType:  "#Microsoft.Media.BuiltInStandardEncoderPreset",
						PresetName: media.EncoderNamedPresetAdaptiveStreaming,
					},
				},
			},
			ResourceGroupName: pulumi.String("contosoresources"),
			TransformName:     pulumi.String("createdTransform"),
		})
		if err != nil {
			return err
		}
		return nil
	})
}
Copy
package generated_program;

import com.pulumi.Context;
import com.pulumi.Pulumi;
import com.pulumi.core.Output;
import com.pulumi.azurenative.media.Transform;
import com.pulumi.azurenative.media.TransformArgs;
import com.pulumi.azurenative.media.inputs.TransformOutputArgs;
import java.util.List;
import java.util.ArrayList;
import java.util.Map;
import java.io.File;
import java.nio.file.Files;
import java.nio.file.Paths;

public class App {
    public static void main(String[] args) {
        Pulumi.run(App::stack);
    }

    public static void stack(Context ctx) {
        var transform = new Transform("transform", TransformArgs.builder()
            .accountName("contosomedia")
            .description("Example Transform to illustrate create and update.")
            .outputs(TransformOutputArgs.builder()
                .preset(BuiltInStandardEncoderPresetArgs.builder()
                    .odataType("#Microsoft.Media.BuiltInStandardEncoderPreset")
                    .presetName("AdaptiveStreaming")
                    .build())
                .build())
            .resourceGroupName("contosoresources")
            .transformName("createdTransform")
            .build());

    }
}
Copy
import * as pulumi from "@pulumi/pulumi";
import * as azure_native from "@pulumi/azure-native";

const transform = new azure_native.media.Transform("transform", {
    accountName: "contosomedia",
    description: "Example Transform to illustrate create and update.",
    outputs: [{
        preset: {
            odataType: "#Microsoft.Media.BuiltInStandardEncoderPreset",
            presetName: azure_native.media.EncoderNamedPreset.AdaptiveStreaming,
        },
    }],
    resourceGroupName: "contosoresources",
    transformName: "createdTransform",
});
Copy
import pulumi
import pulumi_azure_native as azure_native

transform = azure_native.media.Transform("transform",
    account_name="contosomedia",
    description="Example Transform to illustrate create and update.",
    outputs=[{
        "preset": {
            "odata_type": "#Microsoft.Media.BuiltInStandardEncoderPreset",
            "preset_name": azure_native.media.EncoderNamedPreset.ADAPTIVE_STREAMING,
        },
    }],
    resource_group_name="contosoresources",
    transform_name="createdTransform")
Copy
resources:
  transform:
    type: azure-native:media:Transform
    properties:
      accountName: contosomedia
      description: Example Transform to illustrate create and update.
      outputs:
        - preset:
            odataType: '#Microsoft.Media.BuiltInStandardEncoderPreset'
            presetName: AdaptiveStreaming
      resourceGroupName: contosoresources
      transformName: createdTransform
Copy

Create Transform Resource

Resources are created with functions called constructors. To learn more about declaring and configuring resources, see Resources.

Constructor syntax

new Transform(name: string, args: TransformArgs, opts?: CustomResourceOptions);
@overload
def Transform(resource_name: str,
              args: TransformArgs,
              opts: Optional[ResourceOptions] = None)

@overload
def Transform(resource_name: str,
              opts: Optional[ResourceOptions] = None,
              account_name: Optional[str] = None,
              outputs: Optional[Sequence[TransformOutputArgs]] = None,
              resource_group_name: Optional[str] = None,
              description: Optional[str] = None,
              transform_name: Optional[str] = None)
func NewTransform(ctx *Context, name string, args TransformArgs, opts ...ResourceOption) (*Transform, error)
public Transform(string name, TransformArgs args, CustomResourceOptions? opts = null)
public Transform(String name, TransformArgs args)
public Transform(String name, TransformArgs args, CustomResourceOptions options)
type: azure-native:media:Transform
properties: # The arguments to resource properties.
options: # Bag of options to control resource's behavior.

Parameters

name This property is required. string
The unique name of the resource.
args This property is required. TransformArgs
The arguments to resource properties.
opts CustomResourceOptions
Bag of options to control resource's behavior.
resource_name This property is required. str
The unique name of the resource.
args This property is required. TransformArgs
The arguments to resource properties.
opts ResourceOptions
Bag of options to control resource's behavior.
ctx Context
Context object for the current deployment.
name This property is required. string
The unique name of the resource.
args This property is required. TransformArgs
The arguments to resource properties.
opts ResourceOption
Bag of options to control resource's behavior.
name This property is required. string
The unique name of the resource.
args This property is required. TransformArgs
The arguments to resource properties.
opts CustomResourceOptions
Bag of options to control resource's behavior.
name This property is required. String
The unique name of the resource.
args This property is required. TransformArgs
The arguments to resource properties.
options CustomResourceOptions
Bag of options to control resource's behavior.

Constructor example

The following reference example uses placeholder values for all input properties.

var transformResource = new AzureNative.Media.Transform("transformResource", new()
{
    AccountName = "string",
    Outputs = new[]
    {
        new AzureNative.Media.Inputs.TransformOutputArgs
        {
            Preset = new AzureNative.Media.Inputs.AudioAnalyzerPresetArgs
            {
                OdataType = "#Microsoft.Media.AudioAnalyzerPreset",
                AudioLanguage = "string",
                ExperimentalOptions = 
                {
                    { "string", "string" },
                },
                Mode = "string",
            },
            OnError = "string",
            RelativePriority = "string",
        },
    },
    ResourceGroupName = "string",
    Description = "string",
    TransformName = "string",
});
Copy
example, err := media.NewTransform(ctx, "transformResource", &media.TransformArgs{
	AccountName: pulumi.String("string"),
	Outputs: media.TransformOutputTypeArray{
		&media.TransformOutputTypeArgs{
			Preset: media.AudioAnalyzerPreset{
				OdataType:     "#Microsoft.Media.AudioAnalyzerPreset",
				AudioLanguage: "string",
				ExperimentalOptions: map[string]interface{}{
					"string": "string",
				},
				Mode: "string",
			},
			OnError:          pulumi.String("string"),
			RelativePriority: pulumi.String("string"),
		},
	},
	ResourceGroupName: pulumi.String("string"),
	Description:       pulumi.String("string"),
	TransformName:     pulumi.String("string"),
})
Copy
var transformResource = new Transform("transformResource", TransformArgs.builder()
    .accountName("string")
    .outputs(TransformOutputArgs.builder()
        .preset(AudioAnalyzerPresetArgs.builder()
            .odataType("#Microsoft.Media.AudioAnalyzerPreset")
            .audioLanguage("string")
            .experimentalOptions(Map.of("string", "string"))
            .mode("string")
            .build())
        .onError("string")
        .relativePriority("string")
        .build())
    .resourceGroupName("string")
    .description("string")
    .transformName("string")
    .build());
Copy
transform_resource = azure_native.media.Transform("transformResource",
    account_name="string",
    outputs=[{
        "preset": {
            "odata_type": "#Microsoft.Media.AudioAnalyzerPreset",
            "audio_language": "string",
            "experimental_options": {
                "string": "string",
            },
            "mode": "string",
        },
        "on_error": "string",
        "relative_priority": "string",
    }],
    resource_group_name="string",
    description="string",
    transform_name="string")
Copy
const transformResource = new azure_native.media.Transform("transformResource", {
    accountName: "string",
    outputs: [{
        preset: {
            odataType: "#Microsoft.Media.AudioAnalyzerPreset",
            audioLanguage: "string",
            experimentalOptions: {
                string: "string",
            },
            mode: "string",
        },
        onError: "string",
        relativePriority: "string",
    }],
    resourceGroupName: "string",
    description: "string",
    transformName: "string",
});
Copy
type: azure-native:media:Transform
properties:
    accountName: string
    description: string
    outputs:
        - onError: string
          preset:
            audioLanguage: string
            experimentalOptions:
                string: string
            mode: string
            odataType: '#Microsoft.Media.AudioAnalyzerPreset'
          relativePriority: string
    resourceGroupName: string
    transformName: string
Copy

Transform Resource Properties

To learn more about resource properties and how to use them, see Inputs and Outputs in the Architecture and Concepts docs.

Inputs

In Python, inputs that are objects can be passed either as argument classes or as dictionary literals.

The Transform resource accepts the following input properties:

AccountName
This property is required.
Changes to this property will trigger replacement.
string
The Media Services account name.
Outputs This property is required. List<Pulumi.AzureNative.Media.Inputs.TransformOutput>
An array of one or more TransformOutputs that the Transform should generate.
ResourceGroupName
This property is required.
Changes to this property will trigger replacement.
string
The name of the resource group within the Azure subscription.
Description string
An optional verbose description of the Transform.
TransformName Changes to this property will trigger replacement. string
The Transform name.
AccountName
This property is required.
Changes to this property will trigger replacement.
string
The Media Services account name.
Outputs This property is required. []TransformOutputTypeArgs
An array of one or more TransformOutputs that the Transform should generate.
ResourceGroupName
This property is required.
Changes to this property will trigger replacement.
string
The name of the resource group within the Azure subscription.
Description string
An optional verbose description of the Transform.
TransformName Changes to this property will trigger replacement. string
The Transform name.
accountName
This property is required.
Changes to this property will trigger replacement.
String
The Media Services account name.
outputs This property is required. List<TransformOutput>
An array of one or more TransformOutputs that the Transform should generate.
resourceGroupName
This property is required.
Changes to this property will trigger replacement.
String
The name of the resource group within the Azure subscription.
description String
An optional verbose description of the Transform.
transformName Changes to this property will trigger replacement. String
The Transform name.
accountName
This property is required.
Changes to this property will trigger replacement.
string
The Media Services account name.
outputs This property is required. TransformOutput[]
An array of one or more TransformOutputs that the Transform should generate.
resourceGroupName
This property is required.
Changes to this property will trigger replacement.
string
The name of the resource group within the Azure subscription.
description string
An optional verbose description of the Transform.
transformName Changes to this property will trigger replacement. string
The Transform name.
account_name
This property is required.
Changes to this property will trigger replacement.
str
The Media Services account name.
outputs This property is required. Sequence[TransformOutputArgs]
An array of one or more TransformOutputs that the Transform should generate.
resource_group_name
This property is required.
Changes to this property will trigger replacement.
str
The name of the resource group within the Azure subscription.
description str
An optional verbose description of the Transform.
transform_name Changes to this property will trigger replacement. str
The Transform name.
accountName
This property is required.
Changes to this property will trigger replacement.
String
The Media Services account name.
outputs This property is required. List<Property Map>
An array of one or more TransformOutputs that the Transform should generate.
resourceGroupName
This property is required.
Changes to this property will trigger replacement.
String
The name of the resource group within the Azure subscription.
description String
An optional verbose description of the Transform.
transformName Changes to this property will trigger replacement. String
The Transform name.

Outputs

All input properties are implicitly available as output properties. Additionally, the Transform resource produces the following output properties:

AzureApiVersion string
The Azure API version of the resource.
Created string
The UTC date and time when the Transform was created, in 'YYYY-MM-DDThh:mm:ssZ' format.
Id string
The provider-assigned unique ID for this managed resource.
LastModified string
The UTC date and time when the Transform was last updated, in 'YYYY-MM-DDThh:mm:ssZ' format.
Name string
The name of the resource
SystemData Pulumi.AzureNative.Media.Outputs.SystemDataResponse
The system metadata relating to this resource.
Type string
The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or "Microsoft.Storage/storageAccounts"
AzureApiVersion string
The Azure API version of the resource.
Created string
The UTC date and time when the Transform was created, in 'YYYY-MM-DDThh:mm:ssZ' format.
Id string
The provider-assigned unique ID for this managed resource.
LastModified string
The UTC date and time when the Transform was last updated, in 'YYYY-MM-DDThh:mm:ssZ' format.
Name string
The name of the resource
SystemData SystemDataResponse
The system metadata relating to this resource.
Type string
The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or "Microsoft.Storage/storageAccounts"
azureApiVersion String
The Azure API version of the resource.
created String
The UTC date and time when the Transform was created, in 'YYYY-MM-DDThh:mm:ssZ' format.
id String
The provider-assigned unique ID for this managed resource.
lastModified String
The UTC date and time when the Transform was last updated, in 'YYYY-MM-DDThh:mm:ssZ' format.
name String
The name of the resource
systemData SystemDataResponse
The system metadata relating to this resource.
type String
The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or "Microsoft.Storage/storageAccounts"
azureApiVersion string
The Azure API version of the resource.
created string
The UTC date and time when the Transform was created, in 'YYYY-MM-DDThh:mm:ssZ' format.
id string
The provider-assigned unique ID for this managed resource.
lastModified string
The UTC date and time when the Transform was last updated, in 'YYYY-MM-DDThh:mm:ssZ' format.
name string
The name of the resource
systemData SystemDataResponse
The system metadata relating to this resource.
type string
The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or "Microsoft.Storage/storageAccounts"
azure_api_version str
The Azure API version of the resource.
created str
The UTC date and time when the Transform was created, in 'YYYY-MM-DDThh:mm:ssZ' format.
id str
The provider-assigned unique ID for this managed resource.
last_modified str
The UTC date and time when the Transform was last updated, in 'YYYY-MM-DDThh:mm:ssZ' format.
name str
The name of the resource
system_data SystemDataResponse
The system metadata relating to this resource.
type str
The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or "Microsoft.Storage/storageAccounts"
azureApiVersion String
The Azure API version of the resource.
created String
The UTC date and time when the Transform was created, in 'YYYY-MM-DDThh:mm:ssZ' format.
id String
The provider-assigned unique ID for this managed resource.
lastModified String
The UTC date and time when the Transform was last updated, in 'YYYY-MM-DDThh:mm:ssZ' format.
name String
The name of the resource
systemData Property Map
The system metadata relating to this resource.
type String
The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or "Microsoft.Storage/storageAccounts"

Supporting Types

AacAudio
, AacAudioArgs

Bitrate int
The bitrate, in bits per second, of the output encoded audio.
Channels int
The number of channels in the audio.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
Profile string | Pulumi.AzureNative.Media.AacAudioProfile
The encoding profile to be used when encoding audio with AAC.
SamplingRate int
The sampling rate to use for encoding in hertz.
Bitrate int
The bitrate, in bits per second, of the output encoded audio.
Channels int
The number of channels in the audio.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
Profile string | AacAudioProfile
The encoding profile to be used when encoding audio with AAC.
SamplingRate int
The sampling rate to use for encoding in hertz.
bitrate Integer
The bitrate, in bits per second, of the output encoded audio.
channels Integer
The number of channels in the audio.
label String
An optional label for the codec. The label can be used to control muxing behavior.
profile String | AacAudioProfile
The encoding profile to be used when encoding audio with AAC.
samplingRate Integer
The sampling rate to use for encoding in hertz.
bitrate number
The bitrate, in bits per second, of the output encoded audio.
channels number
The number of channels in the audio.
label string
An optional label for the codec. The label can be used to control muxing behavior.
profile string | AacAudioProfile
The encoding profile to be used when encoding audio with AAC.
samplingRate number
The sampling rate to use for encoding in hertz.
bitrate int
The bitrate, in bits per second, of the output encoded audio.
channels int
The number of channels in the audio.
label str
An optional label for the codec. The label can be used to control muxing behavior.
profile str | AacAudioProfile
The encoding profile to be used when encoding audio with AAC.
sampling_rate int
The sampling rate to use for encoding in hertz.
bitrate Number
The bitrate, in bits per second, of the output encoded audio.
channels Number
The number of channels in the audio.
label String
An optional label for the codec. The label can be used to control muxing behavior.
profile String | "AacLc" | "HeAacV1" | "HeAacV2"
The encoding profile to be used when encoding audio with AAC.
samplingRate Number
The sampling rate to use for encoding in hertz.

AacAudioProfile
, AacAudioProfileArgs

AacLc
AacLcSpecifies that the output audio is to be encoded into AAC Low Complexity profile (AAC-LC).
HeAacV1
HeAacV1Specifies that the output audio is to be encoded into HE-AAC v1 profile.
HeAacV2
HeAacV2Specifies that the output audio is to be encoded into HE-AAC v2 profile.
AacAudioProfileAacLc
AacLcSpecifies that the output audio is to be encoded into AAC Low Complexity profile (AAC-LC).
AacAudioProfileHeAacV1
HeAacV1Specifies that the output audio is to be encoded into HE-AAC v1 profile.
AacAudioProfileHeAacV2
HeAacV2Specifies that the output audio is to be encoded into HE-AAC v2 profile.
AacLc
AacLcSpecifies that the output audio is to be encoded into AAC Low Complexity profile (AAC-LC).
HeAacV1
HeAacV1Specifies that the output audio is to be encoded into HE-AAC v1 profile.
HeAacV2
HeAacV2Specifies that the output audio is to be encoded into HE-AAC v2 profile.
AacLc
AacLcSpecifies that the output audio is to be encoded into AAC Low Complexity profile (AAC-LC).
HeAacV1
HeAacV1Specifies that the output audio is to be encoded into HE-AAC v1 profile.
HeAacV2
HeAacV2Specifies that the output audio is to be encoded into HE-AAC v2 profile.
AAC_LC
AacLcSpecifies that the output audio is to be encoded into AAC Low Complexity profile (AAC-LC).
HE_AAC_V1
HeAacV1Specifies that the output audio is to be encoded into HE-AAC v1 profile.
HE_AAC_V2
HeAacV2Specifies that the output audio is to be encoded into HE-AAC v2 profile.
"AacLc"
AacLcSpecifies that the output audio is to be encoded into AAC Low Complexity profile (AAC-LC).
"HeAacV1"
HeAacV1Specifies that the output audio is to be encoded into HE-AAC v1 profile.
"HeAacV2"
HeAacV2Specifies that the output audio is to be encoded into HE-AAC v2 profile.

AacAudioResponse
, AacAudioResponseArgs

Bitrate int
The bitrate, in bits per second, of the output encoded audio.
Channels int
The number of channels in the audio.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
Profile string
The encoding profile to be used when encoding audio with AAC.
SamplingRate int
The sampling rate to use for encoding in hertz.
Bitrate int
The bitrate, in bits per second, of the output encoded audio.
Channels int
The number of channels in the audio.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
Profile string
The encoding profile to be used when encoding audio with AAC.
SamplingRate int
The sampling rate to use for encoding in hertz.
bitrate Integer
The bitrate, in bits per second, of the output encoded audio.
channels Integer
The number of channels in the audio.
label String
An optional label for the codec. The label can be used to control muxing behavior.
profile String
The encoding profile to be used when encoding audio with AAC.
samplingRate Integer
The sampling rate to use for encoding in hertz.
bitrate number
The bitrate, in bits per second, of the output encoded audio.
channels number
The number of channels in the audio.
label string
An optional label for the codec. The label can be used to control muxing behavior.
profile string
The encoding profile to be used when encoding audio with AAC.
samplingRate number
The sampling rate to use for encoding in hertz.
bitrate int
The bitrate, in bits per second, of the output encoded audio.
channels int
The number of channels in the audio.
label str
An optional label for the codec. The label can be used to control muxing behavior.
profile str
The encoding profile to be used when encoding audio with AAC.
sampling_rate int
The sampling rate to use for encoding in hertz.
bitrate Number
The bitrate, in bits per second, of the output encoded audio.
channels Number
The number of channels in the audio.
label String
An optional label for the codec. The label can be used to control muxing behavior.
profile String
The encoding profile to be used when encoding audio with AAC.
samplingRate Number
The sampling rate to use for encoding in hertz.

AnalysisResolution
, AnalysisResolutionArgs

SourceResolution
SourceResolution
StandardDefinition
StandardDefinition
AnalysisResolutionSourceResolution
SourceResolution
AnalysisResolutionStandardDefinition
StandardDefinition
SourceResolution
SourceResolution
StandardDefinition
StandardDefinition
SourceResolution
SourceResolution
StandardDefinition
StandardDefinition
SOURCE_RESOLUTION
SourceResolution
STANDARD_DEFINITION
StandardDefinition
"SourceResolution"
SourceResolution
"StandardDefinition"
StandardDefinition

Audio
, AudioArgs

Bitrate int
The bitrate, in bits per second, of the output encoded audio.
Channels int
The number of channels in the audio.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
SamplingRate int
The sampling rate to use for encoding in hertz.
Bitrate int
The bitrate, in bits per second, of the output encoded audio.
Channels int
The number of channels in the audio.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
SamplingRate int
The sampling rate to use for encoding in hertz.
bitrate Integer
The bitrate, in bits per second, of the output encoded audio.
channels Integer
The number of channels in the audio.
label String
An optional label for the codec. The label can be used to control muxing behavior.
samplingRate Integer
The sampling rate to use for encoding in hertz.
bitrate number
The bitrate, in bits per second, of the output encoded audio.
channels number
The number of channels in the audio.
label string
An optional label for the codec. The label can be used to control muxing behavior.
samplingRate number
The sampling rate to use for encoding in hertz.
bitrate int
The bitrate, in bits per second, of the output encoded audio.
channels int
The number of channels in the audio.
label str
An optional label for the codec. The label can be used to control muxing behavior.
sampling_rate int
The sampling rate to use for encoding in hertz.
bitrate Number
The bitrate, in bits per second, of the output encoded audio.
channels Number
The number of channels in the audio.
label String
An optional label for the codec. The label can be used to control muxing behavior.
samplingRate Number
The sampling rate to use for encoding in hertz.

AudioAnalysisMode
, AudioAnalysisModeArgs

Standard
StandardPerforms all operations included in the Basic mode, additionally performing language detection and speaker diarization.
Basic
BasicThis mode performs speech-to-text transcription and generation of a VTT subtitle/caption file. The output of this mode includes an Insights JSON file including only the keywords, transcription,and timing information. Automatic language detection and speaker diarization are not included in this mode.
AudioAnalysisModeStandard
StandardPerforms all operations included in the Basic mode, additionally performing language detection and speaker diarization.
AudioAnalysisModeBasic
BasicThis mode performs speech-to-text transcription and generation of a VTT subtitle/caption file. The output of this mode includes an Insights JSON file including only the keywords, transcription,and timing information. Automatic language detection and speaker diarization are not included in this mode.
Standard
StandardPerforms all operations included in the Basic mode, additionally performing language detection and speaker diarization.
Basic
BasicThis mode performs speech-to-text transcription and generation of a VTT subtitle/caption file. The output of this mode includes an Insights JSON file including only the keywords, transcription,and timing information. Automatic language detection and speaker diarization are not included in this mode.
Standard
StandardPerforms all operations included in the Basic mode, additionally performing language detection and speaker diarization.
Basic
BasicThis mode performs speech-to-text transcription and generation of a VTT subtitle/caption file. The output of this mode includes an Insights JSON file including only the keywords, transcription,and timing information. Automatic language detection and speaker diarization are not included in this mode.
STANDARD
StandardPerforms all operations included in the Basic mode, additionally performing language detection and speaker diarization.
BASIC
BasicThis mode performs speech-to-text transcription and generation of a VTT subtitle/caption file. The output of this mode includes an Insights JSON file including only the keywords, transcription,and timing information. Automatic language detection and speaker diarization are not included in this mode.
"Standard"
StandardPerforms all operations included in the Basic mode, additionally performing language detection and speaker diarization.
"Basic"
BasicThis mode performs speech-to-text transcription and generation of a VTT subtitle/caption file. The output of this mode includes an Insights JSON file including only the keywords, transcription,and timing information. Automatic language detection and speaker diarization are not included in this mode.

AudioAnalyzerPreset
, AudioAnalyzerPresetArgs

AudioLanguage string
The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
ExperimentalOptions Dictionary<string, string>
Dictionary containing key value pairs for parameters not exposed in the preset itself
Mode string | Pulumi.AzureNative.Media.AudioAnalysisMode
Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
AudioLanguage string
The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
ExperimentalOptions map[string]string
Dictionary containing key value pairs for parameters not exposed in the preset itself
Mode string | AudioAnalysisMode
Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
audioLanguage String
The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
experimentalOptions Map<String,String>
Dictionary containing key value pairs for parameters not exposed in the preset itself
mode String | AudioAnalysisMode
Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
audioLanguage string
The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
experimentalOptions {[key: string]: string}
Dictionary containing key value pairs for parameters not exposed in the preset itself
mode string | AudioAnalysisMode
Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
audio_language str
The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
experimental_options Mapping[str, str]
Dictionary containing key value pairs for parameters not exposed in the preset itself
mode str | AudioAnalysisMode
Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
audioLanguage String
The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
experimentalOptions Map<String>
Dictionary containing key value pairs for parameters not exposed in the preset itself
mode String | "Standard" | "Basic"
Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.

AudioAnalyzerPresetResponse
, AudioAnalyzerPresetResponseArgs

AudioLanguage string
The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
ExperimentalOptions Dictionary<string, string>
Dictionary containing key value pairs for parameters not exposed in the preset itself
Mode string
Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
AudioLanguage string
The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
ExperimentalOptions map[string]string
Dictionary containing key value pairs for parameters not exposed in the preset itself
Mode string
Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
audioLanguage String
The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
experimentalOptions Map<String,String>
Dictionary containing key value pairs for parameters not exposed in the preset itself
mode String
Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
audioLanguage string
The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
experimentalOptions {[key: string]: string}
Dictionary containing key value pairs for parameters not exposed in the preset itself
mode string
Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
audio_language str
The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
experimental_options Mapping[str, str]
Dictionary containing key value pairs for parameters not exposed in the preset itself
mode str
Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
audioLanguage String
The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
experimentalOptions Map<String>
Dictionary containing key value pairs for parameters not exposed in the preset itself
mode String
Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.

AudioOverlay
, AudioOverlayArgs

InputLabel This property is required. string
The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
AudioGainLevel double
The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
End string
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
FadeInDuration string
The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
FadeOutDuration string
The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
Start string
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
InputLabel This property is required. string
The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
AudioGainLevel float64
The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
End string
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
FadeInDuration string
The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
FadeOutDuration string
The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
Start string
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
inputLabel This property is required. String
The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
audioGainLevel Double
The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
end String
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
fadeInDuration String
The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
fadeOutDuration String
The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
start String
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
inputLabel This property is required. string
The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
audioGainLevel number
The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
end string
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
fadeInDuration string
The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
fadeOutDuration string
The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
start string
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
input_label This property is required. str
The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
audio_gain_level float
The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
end str
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
fade_in_duration str
The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
fade_out_duration str
The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
start str
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
inputLabel This property is required. String
The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
audioGainLevel Number
The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
end String
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
fadeInDuration String
The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
fadeOutDuration String
The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
start String
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.

AudioOverlayResponse
, AudioOverlayResponseArgs

InputLabel This property is required. string
The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
AudioGainLevel double
The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
End string
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
FadeInDuration string
The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
FadeOutDuration string
The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
Start string
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
InputLabel This property is required. string
The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
AudioGainLevel float64
The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
End string
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
FadeInDuration string
The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
FadeOutDuration string
The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
Start string
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
inputLabel This property is required. String
The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
audioGainLevel Double
The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
end String
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
fadeInDuration String
The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
fadeOutDuration String
The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
start String
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
inputLabel This property is required. string
The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
audioGainLevel number
The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
end string
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
fadeInDuration string
The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
fadeOutDuration string
The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
start string
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
input_label This property is required. str
The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
audio_gain_level float
The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
end str
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
fade_in_duration str
The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
fade_out_duration str
The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
start str
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
inputLabel This property is required. String
The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
audioGainLevel Number
The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
end String
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
fadeInDuration String
The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
fadeOutDuration String
The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
start String
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.

AudioResponse
, AudioResponseArgs

Bitrate int
The bitrate, in bits per second, of the output encoded audio.
Channels int
The number of channels in the audio.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
SamplingRate int
The sampling rate to use for encoding in hertz.
Bitrate int
The bitrate, in bits per second, of the output encoded audio.
Channels int
The number of channels in the audio.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
SamplingRate int
The sampling rate to use for encoding in hertz.
bitrate Integer
The bitrate, in bits per second, of the output encoded audio.
channels Integer
The number of channels in the audio.
label String
An optional label for the codec. The label can be used to control muxing behavior.
samplingRate Integer
The sampling rate to use for encoding in hertz.
bitrate number
The bitrate, in bits per second, of the output encoded audio.
channels number
The number of channels in the audio.
label string
An optional label for the codec. The label can be used to control muxing behavior.
samplingRate number
The sampling rate to use for encoding in hertz.
bitrate int
The bitrate, in bits per second, of the output encoded audio.
channels int
The number of channels in the audio.
label str
An optional label for the codec. The label can be used to control muxing behavior.
sampling_rate int
The sampling rate to use for encoding in hertz.
bitrate Number
The bitrate, in bits per second, of the output encoded audio.
channels Number
The number of channels in the audio.
label String
An optional label for the codec. The label can be used to control muxing behavior.
samplingRate Number
The sampling rate to use for encoding in hertz.

BlurType
, BlurTypeArgs

Box
BoxBox: debug filter, bounding box only
Low
LowLow: box-car blur filter
Med
MedMed: Gaussian blur filter
High
HighHigh: Confuse blur filter
Black
BlackBlack: Black out filter
BlurTypeBox
BoxBox: debug filter, bounding box only
BlurTypeLow
LowLow: box-car blur filter
BlurTypeMed
MedMed: Gaussian blur filter
BlurTypeHigh
HighHigh: Confuse blur filter
BlurTypeBlack
BlackBlack: Black out filter
Box
BoxBox: debug filter, bounding box only
Low
LowLow: box-car blur filter
Med
MedMed: Gaussian blur filter
High
HighHigh: Confuse blur filter
Black
BlackBlack: Black out filter
Box
BoxBox: debug filter, bounding box only
Low
LowLow: box-car blur filter
Med
MedMed: Gaussian blur filter
High
HighHigh: Confuse blur filter
Black
BlackBlack: Black out filter
BOX
BoxBox: debug filter, bounding box only
LOW
LowLow: box-car blur filter
MED
MedMed: Gaussian blur filter
HIGH
HighHigh: Confuse blur filter
BLACK
BlackBlack: Black out filter
"Box"
BoxBox: debug filter, bounding box only
"Low"
LowLow: box-car blur filter
"Med"
MedMed: Gaussian blur filter
"High"
HighHigh: Confuse blur filter
"Black"
BlackBlack: Black out filter

BuiltInStandardEncoderPreset
, BuiltInStandardEncoderPresetArgs

PresetName This property is required. string | Pulumi.AzureNative.Media.EncoderNamedPreset
The built-in preset to be used for encoding videos.
Configurations Pulumi.AzureNative.Media.Inputs.PresetConfigurations
Optional configuration settings for encoder. Configurations is only supported for ContentAwareEncoding and H265ContentAwareEncoding BuiltInStandardEncoderPreset.
PresetName This property is required. string | EncoderNamedPreset
The built-in preset to be used for encoding videos.
Configurations PresetConfigurations
Optional configuration settings for encoder. Configurations is only supported for ContentAwareEncoding and H265ContentAwareEncoding BuiltInStandardEncoderPreset.
presetName This property is required. String | EncoderNamedPreset
The built-in preset to be used for encoding videos.
configurations PresetConfigurations
Optional configuration settings for encoder. Configurations is only supported for ContentAwareEncoding and H265ContentAwareEncoding BuiltInStandardEncoderPreset.
presetName This property is required. string | EncoderNamedPreset
The built-in preset to be used for encoding videos.
configurations PresetConfigurations
Optional configuration settings for encoder. Configurations is only supported for ContentAwareEncoding and H265ContentAwareEncoding BuiltInStandardEncoderPreset.
preset_name This property is required. str | EncoderNamedPreset
The built-in preset to be used for encoding videos.
configurations PresetConfigurations
Optional configuration settings for encoder. Configurations is only supported for ContentAwareEncoding and H265ContentAwareEncoding BuiltInStandardEncoderPreset.

BuiltInStandardEncoderPresetResponse
, BuiltInStandardEncoderPresetResponseArgs

PresetName This property is required. string
The built-in preset to be used for encoding videos.
Configurations Pulumi.AzureNative.Media.Inputs.PresetConfigurationsResponse
Optional configuration settings for encoder. Configurations is only supported for ContentAwareEncoding and H265ContentAwareEncoding BuiltInStandardEncoderPreset.
PresetName This property is required. string
The built-in preset to be used for encoding videos.
Configurations PresetConfigurationsResponse
Optional configuration settings for encoder. Configurations is only supported for ContentAwareEncoding and H265ContentAwareEncoding BuiltInStandardEncoderPreset.
presetName This property is required. String
The built-in preset to be used for encoding videos.
configurations PresetConfigurationsResponse
Optional configuration settings for encoder. Configurations is only supported for ContentAwareEncoding and H265ContentAwareEncoding BuiltInStandardEncoderPreset.
presetName This property is required. string
The built-in preset to be used for encoding videos.
configurations PresetConfigurationsResponse
Optional configuration settings for encoder. Configurations is only supported for ContentAwareEncoding and H265ContentAwareEncoding BuiltInStandardEncoderPreset.
preset_name This property is required. str
The built-in preset to be used for encoding videos.
configurations PresetConfigurationsResponse
Optional configuration settings for encoder. Configurations is only supported for ContentAwareEncoding and H265ContentAwareEncoding BuiltInStandardEncoderPreset.
presetName This property is required. String
The built-in preset to be used for encoding videos.
configurations Property Map
Optional configuration settings for encoder. Configurations is only supported for ContentAwareEncoding and H265ContentAwareEncoding BuiltInStandardEncoderPreset.

Complexity
, ComplexityArgs

Speed
SpeedConfigures the encoder to use settings optimized for faster encoding. Quality is sacrificed to decrease encoding time.
Balanced
BalancedConfigures the encoder to use settings that achieve a balance between speed and quality.
Quality
QualityConfigures the encoder to use settings optimized to produce higher quality output at the expense of slower overall encode time.
ComplexitySpeed
SpeedConfigures the encoder to use settings optimized for faster encoding. Quality is sacrificed to decrease encoding time.
ComplexityBalanced
BalancedConfigures the encoder to use settings that achieve a balance between speed and quality.
ComplexityQuality
QualityConfigures the encoder to use settings optimized to produce higher quality output at the expense of slower overall encode time.
Speed
SpeedConfigures the encoder to use settings optimized for faster encoding. Quality is sacrificed to decrease encoding time.
Balanced
BalancedConfigures the encoder to use settings that achieve a balance between speed and quality.
Quality
QualityConfigures the encoder to use settings optimized to produce higher quality output at the expense of slower overall encode time.
Speed
SpeedConfigures the encoder to use settings optimized for faster encoding. Quality is sacrificed to decrease encoding time.
Balanced
BalancedConfigures the encoder to use settings that achieve a balance between speed and quality.
Quality
QualityConfigures the encoder to use settings optimized to produce higher quality output at the expense of slower overall encode time.
SPEED
SpeedConfigures the encoder to use settings optimized for faster encoding. Quality is sacrificed to decrease encoding time.
BALANCED
BalancedConfigures the encoder to use settings that achieve a balance between speed and quality.
QUALITY
QualityConfigures the encoder to use settings optimized to produce higher quality output at the expense of slower overall encode time.
"Speed"
SpeedConfigures the encoder to use settings optimized for faster encoding. Quality is sacrificed to decrease encoding time.
"Balanced"
BalancedConfigures the encoder to use settings that achieve a balance between speed and quality.
"Quality"
QualityConfigures the encoder to use settings optimized to produce higher quality output at the expense of slower overall encode time.

CopyAudio
, CopyAudioArgs

Label string
An optional label for the codec. The label can be used to control muxing behavior.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
label String
An optional label for the codec. The label can be used to control muxing behavior.
label string
An optional label for the codec. The label can be used to control muxing behavior.
label str
An optional label for the codec. The label can be used to control muxing behavior.
label String
An optional label for the codec. The label can be used to control muxing behavior.

CopyAudioResponse
, CopyAudioResponseArgs

Label string
An optional label for the codec. The label can be used to control muxing behavior.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
label String
An optional label for the codec. The label can be used to control muxing behavior.
label string
An optional label for the codec. The label can be used to control muxing behavior.
label str
An optional label for the codec. The label can be used to control muxing behavior.
label String
An optional label for the codec. The label can be used to control muxing behavior.

CopyVideo
, CopyVideoArgs

Label string
An optional label for the codec. The label can be used to control muxing behavior.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
label String
An optional label for the codec. The label can be used to control muxing behavior.
label string
An optional label for the codec. The label can be used to control muxing behavior.
label str
An optional label for the codec. The label can be used to control muxing behavior.
label String
An optional label for the codec. The label can be used to control muxing behavior.

CopyVideoResponse
, CopyVideoResponseArgs

Label string
An optional label for the codec. The label can be used to control muxing behavior.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
label String
An optional label for the codec. The label can be used to control muxing behavior.
label string
An optional label for the codec. The label can be used to control muxing behavior.
label str
An optional label for the codec. The label can be used to control muxing behavior.
label String
An optional label for the codec. The label can be used to control muxing behavior.

DDAudio
, DDAudioArgs

Bitrate int
The bitrate, in bits per second, of the output encoded audio.
Channels int
The number of channels in the audio.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
SamplingRate int
The sampling rate to use for encoding in hertz.
Bitrate int
The bitrate, in bits per second, of the output encoded audio.
Channels int
The number of channels in the audio.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
SamplingRate int
The sampling rate to use for encoding in hertz.
bitrate Integer
The bitrate, in bits per second, of the output encoded audio.
channels Integer
The number of channels in the audio.
label String
An optional label for the codec. The label can be used to control muxing behavior.
samplingRate Integer
The sampling rate to use for encoding in hertz.
bitrate number
The bitrate, in bits per second, of the output encoded audio.
channels number
The number of channels in the audio.
label string
An optional label for the codec. The label can be used to control muxing behavior.
samplingRate number
The sampling rate to use for encoding in hertz.
bitrate int
The bitrate, in bits per second, of the output encoded audio.
channels int
The number of channels in the audio.
label str
An optional label for the codec. The label can be used to control muxing behavior.
sampling_rate int
The sampling rate to use for encoding in hertz.
bitrate Number
The bitrate, in bits per second, of the output encoded audio.
channels Number
The number of channels in the audio.
label String
An optional label for the codec. The label can be used to control muxing behavior.
samplingRate Number
The sampling rate to use for encoding in hertz.

DDAudioResponse
, DDAudioResponseArgs

Bitrate int
The bitrate, in bits per second, of the output encoded audio.
Channels int
The number of channels in the audio.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
SamplingRate int
The sampling rate to use for encoding in hertz.
Bitrate int
The bitrate, in bits per second, of the output encoded audio.
Channels int
The number of channels in the audio.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
SamplingRate int
The sampling rate to use for encoding in hertz.
bitrate Integer
The bitrate, in bits per second, of the output encoded audio.
channels Integer
The number of channels in the audio.
label String
An optional label for the codec. The label can be used to control muxing behavior.
samplingRate Integer
The sampling rate to use for encoding in hertz.
bitrate number
The bitrate, in bits per second, of the output encoded audio.
channels number
The number of channels in the audio.
label string
An optional label for the codec. The label can be used to control muxing behavior.
samplingRate number
The sampling rate to use for encoding in hertz.
bitrate int
The bitrate, in bits per second, of the output encoded audio.
channels int
The number of channels in the audio.
label str
An optional label for the codec. The label can be used to control muxing behavior.
sampling_rate int
The sampling rate to use for encoding in hertz.
bitrate Number
The bitrate, in bits per second, of the output encoded audio.
channels Number
The number of channels in the audio.
label String
An optional label for the codec. The label can be used to control muxing behavior.
samplingRate Number
The sampling rate to use for encoding in hertz.

Deinterlace
, DeinterlaceArgs

Mode string | Pulumi.AzureNative.Media.DeinterlaceMode
The deinterlacing mode. Defaults to AutoPixelAdaptive.
Parity string | Pulumi.AzureNative.Media.DeinterlaceParity
The field parity for de-interlacing, defaults to Auto.
Mode string | DeinterlaceMode
The deinterlacing mode. Defaults to AutoPixelAdaptive.
Parity string | DeinterlaceParity
The field parity for de-interlacing, defaults to Auto.
mode String | DeinterlaceMode
The deinterlacing mode. Defaults to AutoPixelAdaptive.
parity String | DeinterlaceParity
The field parity for de-interlacing, defaults to Auto.
mode string | DeinterlaceMode
The deinterlacing mode. Defaults to AutoPixelAdaptive.
parity string | DeinterlaceParity
The field parity for de-interlacing, defaults to Auto.
mode str | DeinterlaceMode
The deinterlacing mode. Defaults to AutoPixelAdaptive.
parity str | DeinterlaceParity
The field parity for de-interlacing, defaults to Auto.
mode String | "Off" | "AutoPixelAdaptive"
The deinterlacing mode. Defaults to AutoPixelAdaptive.
parity String | "Auto" | "TopFieldFirst" | "BottomFieldFirst"
The field parity for de-interlacing, defaults to Auto.

DeinterlaceMode
, DeinterlaceModeArgs

Off
OffDisables de-interlacing of the source video.
AutoPixelAdaptive
AutoPixelAdaptiveApply automatic pixel adaptive de-interlacing on each frame in the input video.
DeinterlaceModeOff
OffDisables de-interlacing of the source video.
DeinterlaceModeAutoPixelAdaptive
AutoPixelAdaptiveApply automatic pixel adaptive de-interlacing on each frame in the input video.
Off
OffDisables de-interlacing of the source video.
AutoPixelAdaptive
AutoPixelAdaptiveApply automatic pixel adaptive de-interlacing on each frame in the input video.
Off
OffDisables de-interlacing of the source video.
AutoPixelAdaptive
AutoPixelAdaptiveApply automatic pixel adaptive de-interlacing on each frame in the input video.
OFF
OffDisables de-interlacing of the source video.
AUTO_PIXEL_ADAPTIVE
AutoPixelAdaptiveApply automatic pixel adaptive de-interlacing on each frame in the input video.
"Off"
OffDisables de-interlacing of the source video.
"AutoPixelAdaptive"
AutoPixelAdaptiveApply automatic pixel adaptive de-interlacing on each frame in the input video.

DeinterlaceParity
, DeinterlaceParityArgs

Auto
AutoAutomatically detect the order of fields
TopFieldFirst
TopFieldFirstApply top field first processing of input video.
BottomFieldFirst
BottomFieldFirstApply bottom field first processing of input video.
DeinterlaceParityAuto
AutoAutomatically detect the order of fields
DeinterlaceParityTopFieldFirst
TopFieldFirstApply top field first processing of input video.
DeinterlaceParityBottomFieldFirst
BottomFieldFirstApply bottom field first processing of input video.
Auto
AutoAutomatically detect the order of fields
TopFieldFirst
TopFieldFirstApply top field first processing of input video.
BottomFieldFirst
BottomFieldFirstApply bottom field first processing of input video.
Auto
AutoAutomatically detect the order of fields
TopFieldFirst
TopFieldFirstApply top field first processing of input video.
BottomFieldFirst
BottomFieldFirstApply bottom field first processing of input video.
AUTO
AutoAutomatically detect the order of fields
TOP_FIELD_FIRST
TopFieldFirstApply top field first processing of input video.
BOTTOM_FIELD_FIRST
BottomFieldFirstApply bottom field first processing of input video.
"Auto"
AutoAutomatically detect the order of fields
"TopFieldFirst"
TopFieldFirstApply top field first processing of input video.
"BottomFieldFirst"
BottomFieldFirstApply bottom field first processing of input video.

DeinterlaceResponse
, DeinterlaceResponseArgs

Mode string
The deinterlacing mode. Defaults to AutoPixelAdaptive.
Parity string
The field parity for de-interlacing, defaults to Auto.
Mode string
The deinterlacing mode. Defaults to AutoPixelAdaptive.
Parity string
The field parity for de-interlacing, defaults to Auto.
mode String
The deinterlacing mode. Defaults to AutoPixelAdaptive.
parity String
The field parity for de-interlacing, defaults to Auto.
mode string
The deinterlacing mode. Defaults to AutoPixelAdaptive.
parity string
The field parity for de-interlacing, defaults to Auto.
mode str
The deinterlacing mode. Defaults to AutoPixelAdaptive.
parity str
The field parity for de-interlacing, defaults to Auto.
mode String
The deinterlacing mode. Defaults to AutoPixelAdaptive.
parity String
The field parity for de-interlacing, defaults to Auto.

EncoderNamedPreset
, EncoderNamedPresetArgs

H264SingleBitrateSD
H264SingleBitrateSDProduces an MP4 file where the video is encoded with H.264 codec at 2200 kbps and a picture height of 480 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
H264SingleBitrate720p
H264SingleBitrate720pProduces an MP4 file where the video is encoded with H.264 codec at 4500 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
H264SingleBitrate1080p
H264SingleBitrate1080pProduces an MP4 file where the video is encoded with H.264 codec at 6750 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
AdaptiveStreaming
AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.264 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
AACGoodQualityAudio
AACGoodQualityAudioProduces a single MP4 file containing only AAC stereo audio encoded at 192 kbps.
DDGoodQualityAudio
DDGoodQualityAudioProduces a single MP4 file containing only DD(Digital Dolby) stereo audio encoded at 192 kbps.
ContentAwareEncodingExperimental
ContentAwareEncodingExperimentalExposes an experimental preset for content-aware encoding. Given any input content, the service attempts to automatically determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. The underlying algorithms will continue to evolve over time. The output will contain MP4 files with video and audio interleaved.
ContentAwareEncoding
ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
CopyAllBitrateNonInterleaved
CopyAllBitrateNonInterleavedCopy all video and audio streams from the input asset as non-interleaved video and audio output files. This preset can be used to clip an existing asset or convert a group of key frame (GOP) aligned MP4 files as an asset that can be streamed.
H264MultipleBitrate1080p
H264MultipleBitrate1080pProduces a set of 8 GOP-aligned MP4 files, ranging from 6000 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 1080p and goes down to 180p.
H264MultipleBitrate720p
H264MultipleBitrate720pProduces a set of 6 GOP-aligned MP4 files, ranging from 3400 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 720p and goes down to 180p.
H264MultipleBitrateSD
H264MultipleBitrateSDProduces a set of 5 GOP-aligned MP4 files, ranging from 1900kbps to 400 kbps, and stereo AAC audio. Resolution starts at 480p and goes down to 240p.
H265ContentAwareEncoding
H265ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
H265AdaptiveStreaming
H265AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.265 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
H265SingleBitrate720p
H265SingleBitrate720pProduces an MP4 file where the video is encoded with H.265 codec at 1800 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
H265SingleBitrate1080p
H265SingleBitrate1080pProduces an MP4 file where the video is encoded with H.265 codec at 3500 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
H265SingleBitrate4K
H265SingleBitrate4KProduces an MP4 file where the video is encoded with H.265 codec at 9500 kbps and a picture height of 2160 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
EncoderNamedPresetH264SingleBitrateSD
H264SingleBitrateSDProduces an MP4 file where the video is encoded with H.264 codec at 2200 kbps and a picture height of 480 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
EncoderNamedPresetH264SingleBitrate720p
H264SingleBitrate720pProduces an MP4 file where the video is encoded with H.264 codec at 4500 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
EncoderNamedPresetH264SingleBitrate1080p
H264SingleBitrate1080pProduces an MP4 file where the video is encoded with H.264 codec at 6750 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
EncoderNamedPresetAdaptiveStreaming
AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.264 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
EncoderNamedPresetAACGoodQualityAudio
AACGoodQualityAudioProduces a single MP4 file containing only AAC stereo audio encoded at 192 kbps.
EncoderNamedPresetDDGoodQualityAudio
DDGoodQualityAudioProduces a single MP4 file containing only DD(Digital Dolby) stereo audio encoded at 192 kbps.
EncoderNamedPresetContentAwareEncodingExperimental
ContentAwareEncodingExperimentalExposes an experimental preset for content-aware encoding. Given any input content, the service attempts to automatically determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. The underlying algorithms will continue to evolve over time. The output will contain MP4 files with video and audio interleaved.
EncoderNamedPresetContentAwareEncoding
ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
EncoderNamedPresetCopyAllBitrateNonInterleaved
CopyAllBitrateNonInterleavedCopy all video and audio streams from the input asset as non-interleaved video and audio output files. This preset can be used to clip an existing asset or convert a group of key frame (GOP) aligned MP4 files as an asset that can be streamed.
EncoderNamedPresetH264MultipleBitrate1080p
H264MultipleBitrate1080pProduces a set of 8 GOP-aligned MP4 files, ranging from 6000 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 1080p and goes down to 180p.
EncoderNamedPresetH264MultipleBitrate720p
H264MultipleBitrate720pProduces a set of 6 GOP-aligned MP4 files, ranging from 3400 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 720p and goes down to 180p.
EncoderNamedPresetH264MultipleBitrateSD
H264MultipleBitrateSDProduces a set of 5 GOP-aligned MP4 files, ranging from 1900kbps to 400 kbps, and stereo AAC audio. Resolution starts at 480p and goes down to 240p.
EncoderNamedPresetH265ContentAwareEncoding
H265ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
EncoderNamedPresetH265AdaptiveStreaming
H265AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.265 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
EncoderNamedPresetH265SingleBitrate720p
H265SingleBitrate720pProduces an MP4 file where the video is encoded with H.265 codec at 1800 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
EncoderNamedPresetH265SingleBitrate1080p
H265SingleBitrate1080pProduces an MP4 file where the video is encoded with H.265 codec at 3500 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
EncoderNamedPresetH265SingleBitrate4K
H265SingleBitrate4KProduces an MP4 file where the video is encoded with H.265 codec at 9500 kbps and a picture height of 2160 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
H264SingleBitrateSD
H264SingleBitrateSDProduces an MP4 file where the video is encoded with H.264 codec at 2200 kbps and a picture height of 480 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
H264SingleBitrate720p
H264SingleBitrate720pProduces an MP4 file where the video is encoded with H.264 codec at 4500 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
H264SingleBitrate1080p
H264SingleBitrate1080pProduces an MP4 file where the video is encoded with H.264 codec at 6750 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
AdaptiveStreaming
AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.264 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
AACGoodQualityAudio
AACGoodQualityAudioProduces a single MP4 file containing only AAC stereo audio encoded at 192 kbps.
DDGoodQualityAudio
DDGoodQualityAudioProduces a single MP4 file containing only DD(Digital Dolby) stereo audio encoded at 192 kbps.
ContentAwareEncodingExperimental
ContentAwareEncodingExperimentalExposes an experimental preset for content-aware encoding. Given any input content, the service attempts to automatically determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. The underlying algorithms will continue to evolve over time. The output will contain MP4 files with video and audio interleaved.
ContentAwareEncoding
ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
CopyAllBitrateNonInterleaved
CopyAllBitrateNonInterleavedCopy all video and audio streams from the input asset as non-interleaved video and audio output files. This preset can be used to clip an existing asset or convert a group of key frame (GOP) aligned MP4 files as an asset that can be streamed.
H264MultipleBitrate1080p
H264MultipleBitrate1080pProduces a set of 8 GOP-aligned MP4 files, ranging from 6000 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 1080p and goes down to 180p.
H264MultipleBitrate720p
H264MultipleBitrate720pProduces a set of 6 GOP-aligned MP4 files, ranging from 3400 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 720p and goes down to 180p.
H264MultipleBitrateSD
H264MultipleBitrateSDProduces a set of 5 GOP-aligned MP4 files, ranging from 1900kbps to 400 kbps, and stereo AAC audio. Resolution starts at 480p and goes down to 240p.
H265ContentAwareEncoding
H265ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
H265AdaptiveStreaming
H265AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.265 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
H265SingleBitrate720p
H265SingleBitrate720pProduces an MP4 file where the video is encoded with H.265 codec at 1800 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
H265SingleBitrate1080p
H265SingleBitrate1080pProduces an MP4 file where the video is encoded with H.265 codec at 3500 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
H265SingleBitrate4K
H265SingleBitrate4KProduces an MP4 file where the video is encoded with H.265 codec at 9500 kbps and a picture height of 2160 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
H264SingleBitrateSD
H264SingleBitrateSDProduces an MP4 file where the video is encoded with H.264 codec at 2200 kbps and a picture height of 480 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
H264SingleBitrate720p
H264SingleBitrate720pProduces an MP4 file where the video is encoded with H.264 codec at 4500 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
H264SingleBitrate1080p
H264SingleBitrate1080pProduces an MP4 file where the video is encoded with H.264 codec at 6750 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
AdaptiveStreaming
AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.264 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
AACGoodQualityAudio
AACGoodQualityAudioProduces a single MP4 file containing only AAC stereo audio encoded at 192 kbps.
DDGoodQualityAudio
DDGoodQualityAudioProduces a single MP4 file containing only DD(Digital Dolby) stereo audio encoded at 192 kbps.
ContentAwareEncodingExperimental
ContentAwareEncodingExperimentalExposes an experimental preset for content-aware encoding. Given any input content, the service attempts to automatically determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. The underlying algorithms will continue to evolve over time. The output will contain MP4 files with video and audio interleaved.
ContentAwareEncoding
ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
CopyAllBitrateNonInterleaved
CopyAllBitrateNonInterleavedCopy all video and audio streams from the input asset as non-interleaved video and audio output files. This preset can be used to clip an existing asset or convert a group of key frame (GOP) aligned MP4 files as an asset that can be streamed.
H264MultipleBitrate1080p
H264MultipleBitrate1080pProduces a set of 8 GOP-aligned MP4 files, ranging from 6000 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 1080p and goes down to 180p.
H264MultipleBitrate720p
H264MultipleBitrate720pProduces a set of 6 GOP-aligned MP4 files, ranging from 3400 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 720p and goes down to 180p.
H264MultipleBitrateSD
H264MultipleBitrateSDProduces a set of 5 GOP-aligned MP4 files, ranging from 1900kbps to 400 kbps, and stereo AAC audio. Resolution starts at 480p and goes down to 240p.
H265ContentAwareEncoding
H265ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
H265AdaptiveStreaming
H265AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.265 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
H265SingleBitrate720p
H265SingleBitrate720pProduces an MP4 file where the video is encoded with H.265 codec at 1800 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
H265SingleBitrate1080p
H265SingleBitrate1080pProduces an MP4 file where the video is encoded with H.265 codec at 3500 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
H265SingleBitrate4K
H265SingleBitrate4KProduces an MP4 file where the video is encoded with H.265 codec at 9500 kbps and a picture height of 2160 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
H264_SINGLE_BITRATE_SD
H264SingleBitrateSDProduces an MP4 file where the video is encoded with H.264 codec at 2200 kbps and a picture height of 480 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
H264_SINGLE_BITRATE720P
H264SingleBitrate720pProduces an MP4 file where the video is encoded with H.264 codec at 4500 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
H264_SINGLE_BITRATE1080P
H264SingleBitrate1080pProduces an MP4 file where the video is encoded with H.264 codec at 6750 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
ADAPTIVE_STREAMING
AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.264 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
AAC_GOOD_QUALITY_AUDIO
AACGoodQualityAudioProduces a single MP4 file containing only AAC stereo audio encoded at 192 kbps.
DD_GOOD_QUALITY_AUDIO
DDGoodQualityAudioProduces a single MP4 file containing only DD(Digital Dolby) stereo audio encoded at 192 kbps.
CONTENT_AWARE_ENCODING_EXPERIMENTAL
ContentAwareEncodingExperimentalExposes an experimental preset for content-aware encoding. Given any input content, the service attempts to automatically determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. The underlying algorithms will continue to evolve over time. The output will contain MP4 files with video and audio interleaved.
CONTENT_AWARE_ENCODING
ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
COPY_ALL_BITRATE_NON_INTERLEAVED
CopyAllBitrateNonInterleavedCopy all video and audio streams from the input asset as non-interleaved video and audio output files. This preset can be used to clip an existing asset or convert a group of key frame (GOP) aligned MP4 files as an asset that can be streamed.
H264_MULTIPLE_BITRATE1080P
H264MultipleBitrate1080pProduces a set of 8 GOP-aligned MP4 files, ranging from 6000 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 1080p and goes down to 180p.
H264_MULTIPLE_BITRATE720P
H264MultipleBitrate720pProduces a set of 6 GOP-aligned MP4 files, ranging from 3400 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 720p and goes down to 180p.
H264_MULTIPLE_BITRATE_SD
H264MultipleBitrateSDProduces a set of 5 GOP-aligned MP4 files, ranging from 1900kbps to 400 kbps, and stereo AAC audio. Resolution starts at 480p and goes down to 240p.
H265_CONTENT_AWARE_ENCODING
H265ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
H265_ADAPTIVE_STREAMING
H265AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.265 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
H265_SINGLE_BITRATE720P
H265SingleBitrate720pProduces an MP4 file where the video is encoded with H.265 codec at 1800 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
H265_SINGLE_BITRATE1080P
H265SingleBitrate1080pProduces an MP4 file where the video is encoded with H.265 codec at 3500 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
H265_SINGLE_BITRATE4_K
H265SingleBitrate4KProduces an MP4 file where the video is encoded with H.265 codec at 9500 kbps and a picture height of 2160 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
"H264SingleBitrateSD"
H264SingleBitrateSDProduces an MP4 file where the video is encoded with H.264 codec at 2200 kbps and a picture height of 480 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
"H264SingleBitrate720p"
H264SingleBitrate720pProduces an MP4 file where the video is encoded with H.264 codec at 4500 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
"H264SingleBitrate1080p"
H264SingleBitrate1080pProduces an MP4 file where the video is encoded with H.264 codec at 6750 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
"AdaptiveStreaming"
AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.264 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
"AACGoodQualityAudio"
AACGoodQualityAudioProduces a single MP4 file containing only AAC stereo audio encoded at 192 kbps.
"DDGoodQualityAudio"
DDGoodQualityAudioProduces a single MP4 file containing only DD(Digital Dolby) stereo audio encoded at 192 kbps.
"ContentAwareEncodingExperimental"
ContentAwareEncodingExperimentalExposes an experimental preset for content-aware encoding. Given any input content, the service attempts to automatically determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. The underlying algorithms will continue to evolve over time. The output will contain MP4 files with video and audio interleaved.
"ContentAwareEncoding"
ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
"CopyAllBitrateNonInterleaved"
CopyAllBitrateNonInterleavedCopy all video and audio streams from the input asset as non-interleaved video and audio output files. This preset can be used to clip an existing asset or convert a group of key frame (GOP) aligned MP4 files as an asset that can be streamed.
"H264MultipleBitrate1080p"
H264MultipleBitrate1080pProduces a set of 8 GOP-aligned MP4 files, ranging from 6000 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 1080p and goes down to 180p.
"H264MultipleBitrate720p"
H264MultipleBitrate720pProduces a set of 6 GOP-aligned MP4 files, ranging from 3400 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 720p and goes down to 180p.
"H264MultipleBitrateSD"
H264MultipleBitrateSDProduces a set of 5 GOP-aligned MP4 files, ranging from 1900kbps to 400 kbps, and stereo AAC audio. Resolution starts at 480p and goes down to 240p.
"H265ContentAwareEncoding"
H265ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
"H265AdaptiveStreaming"
H265AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.265 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
"H265SingleBitrate720p"
H265SingleBitrate720pProduces an MP4 file where the video is encoded with H.265 codec at 1800 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
"H265SingleBitrate1080p"
H265SingleBitrate1080pProduces an MP4 file where the video is encoded with H.265 codec at 3500 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
"H265SingleBitrate4K"
H265SingleBitrate4KProduces an MP4 file where the video is encoded with H.265 codec at 9500 kbps and a picture height of 2160 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

EntropyMode
, EntropyModeArgs

Cabac
CabacContext Adaptive Binary Arithmetic Coder (CABAC) entropy encoding.
Cavlc
CavlcContext Adaptive Variable Length Coder (CAVLC) entropy encoding.
EntropyModeCabac
CabacContext Adaptive Binary Arithmetic Coder (CABAC) entropy encoding.
EntropyModeCavlc
CavlcContext Adaptive Variable Length Coder (CAVLC) entropy encoding.
Cabac
CabacContext Adaptive Binary Arithmetic Coder (CABAC) entropy encoding.
Cavlc
CavlcContext Adaptive Variable Length Coder (CAVLC) entropy encoding.
Cabac
CabacContext Adaptive Binary Arithmetic Coder (CABAC) entropy encoding.
Cavlc
CavlcContext Adaptive Variable Length Coder (CAVLC) entropy encoding.
CABAC
CabacContext Adaptive Binary Arithmetic Coder (CABAC) entropy encoding.
CAVLC
CavlcContext Adaptive Variable Length Coder (CAVLC) entropy encoding.
"Cabac"
CabacContext Adaptive Binary Arithmetic Coder (CABAC) entropy encoding.
"Cavlc"
CavlcContext Adaptive Variable Length Coder (CAVLC) entropy encoding.

FaceDetectorPreset
, FaceDetectorPresetArgs

BlurType string | Pulumi.AzureNative.Media.BlurType
Blur type
ExperimentalOptions Dictionary<string, string>
Dictionary containing key value pairs for parameters not exposed in the preset itself
Mode string | Pulumi.AzureNative.Media.FaceRedactorMode
This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
Resolution string | Pulumi.AzureNative.Media.AnalysisResolution
Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
BlurType string | BlurType
Blur type
ExperimentalOptions map[string]string
Dictionary containing key value pairs for parameters not exposed in the preset itself
Mode string | FaceRedactorMode
This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
Resolution string | AnalysisResolution
Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
blurType String | BlurType
Blur type
experimentalOptions Map<String,String>
Dictionary containing key value pairs for parameters not exposed in the preset itself
mode String | FaceRedactorMode
This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
resolution String | AnalysisResolution
Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
blurType string | BlurType
Blur type
experimentalOptions {[key: string]: string}
Dictionary containing key value pairs for parameters not exposed in the preset itself
mode string | FaceRedactorMode
This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
resolution string | AnalysisResolution
Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
blur_type str | BlurType
Blur type
experimental_options Mapping[str, str]
Dictionary containing key value pairs for parameters not exposed in the preset itself
mode str | FaceRedactorMode
This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
resolution str | AnalysisResolution
Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
blurType String | "Box" | "Low" | "Med" | "High" | "Black"
Blur type
experimentalOptions Map<String>
Dictionary containing key value pairs for parameters not exposed in the preset itself
mode String | "Analyze" | "Redact" | "Combined"
This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
resolution String | "SourceResolution" | "StandardDefinition"
Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.

FaceDetectorPresetResponse
, FaceDetectorPresetResponseArgs

BlurType string
Blur type
ExperimentalOptions Dictionary<string, string>
Dictionary containing key value pairs for parameters not exposed in the preset itself
Mode string
This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
Resolution string
Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
BlurType string
Blur type
ExperimentalOptions map[string]string
Dictionary containing key value pairs for parameters not exposed in the preset itself
Mode string
This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
Resolution string
Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
blurType String
Blur type
experimentalOptions Map<String,String>
Dictionary containing key value pairs for parameters not exposed in the preset itself
mode String
This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
resolution String
Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
blurType string
Blur type
experimentalOptions {[key: string]: string}
Dictionary containing key value pairs for parameters not exposed in the preset itself
mode string
This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
resolution string
Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
blur_type str
Blur type
experimental_options Mapping[str, str]
Dictionary containing key value pairs for parameters not exposed in the preset itself
mode str
This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
resolution str
Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
blurType String
Blur type
experimentalOptions Map<String>
Dictionary containing key value pairs for parameters not exposed in the preset itself
mode String
This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
resolution String
Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.

FaceRedactorMode
, FaceRedactorModeArgs

Analyze
AnalyzeAnalyze mode detects faces and outputs a metadata file with the results. Allows editing of the metadata file before faces are blurred with Redact mode.
Redact
RedactRedact mode consumes the metadata file from Analyze mode and redacts the faces found.
Combined
CombinedCombined mode does the Analyze and Redact steps in one pass when editing the analyzed faces is not desired.
FaceRedactorModeAnalyze
AnalyzeAnalyze mode detects faces and outputs a metadata file with the results. Allows editing of the metadata file before faces are blurred with Redact mode.
FaceRedactorModeRedact
RedactRedact mode consumes the metadata file from Analyze mode and redacts the faces found.
FaceRedactorModeCombined
CombinedCombined mode does the Analyze and Redact steps in one pass when editing the analyzed faces is not desired.
Analyze
AnalyzeAnalyze mode detects faces and outputs a metadata file with the results. Allows editing of the metadata file before faces are blurred with Redact mode.
Redact
RedactRedact mode consumes the metadata file from Analyze mode and redacts the faces found.
Combined
CombinedCombined mode does the Analyze and Redact steps in one pass when editing the analyzed faces is not desired.
Analyze
AnalyzeAnalyze mode detects faces and outputs a metadata file with the results. Allows editing of the metadata file before faces are blurred with Redact mode.
Redact
RedactRedact mode consumes the metadata file from Analyze mode and redacts the faces found.
Combined
CombinedCombined mode does the Analyze and Redact steps in one pass when editing the analyzed faces is not desired.
ANALYZE
AnalyzeAnalyze mode detects faces and outputs a metadata file with the results. Allows editing of the metadata file before faces are blurred with Redact mode.
REDACT
RedactRedact mode consumes the metadata file from Analyze mode and redacts the faces found.
COMBINED
CombinedCombined mode does the Analyze and Redact steps in one pass when editing the analyzed faces is not desired.
"Analyze"
AnalyzeAnalyze mode detects faces and outputs a metadata file with the results. Allows editing of the metadata file before faces are blurred with Redact mode.
"Redact"
RedactRedact mode consumes the metadata file from Analyze mode and redacts the faces found.
"Combined"
CombinedCombined mode does the Analyze and Redact steps in one pass when editing the analyzed faces is not desired.

Fade
, FadeArgs

Duration This property is required. string
The Duration of the fade effect in the video. The value can be in ISO 8601 format (For example, PT05S to fade In/Out a color during 5 seconds), or a frame count (For example, 10 to fade 10 frames from the start time), or a relative value to stream duration (For example, 10% to fade 10% of stream duration)
FadeColor This property is required. string
The Color for the fade In/Out. it can be on the CSS Level1 colors https://developer.mozilla.org/en-US/docs/Web/CSS/color_value/color_keywords or an RGB/hex value: e.g: rgb(255,0,0), 0xFF0000 or #FF0000
Start string
The position in the input video from where to start fade. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Default is 0
Duration This property is required. string
The Duration of the fade effect in the video. The value can be in ISO 8601 format (For example, PT05S to fade In/Out a color during 5 seconds), or a frame count (For example, 10 to fade 10 frames from the start time), or a relative value to stream duration (For example, 10% to fade 10% of stream duration)
FadeColor This property is required. string
The Color for the fade In/Out. it can be on the CSS Level1 colors https://developer.mozilla.org/en-US/docs/Web/CSS/color_value/color_keywords or an RGB/hex value: e.g: rgb(255,0,0), 0xFF0000 or #FF0000
Start string
The position in the input video from where to start fade. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Default is 0
duration This property is required. String
The Duration of the fade effect in the video. The value can be in ISO 8601 format (For example, PT05S to fade In/Out a color during 5 seconds), or a frame count (For example, 10 to fade 10 frames from the start time), or a relative value to stream duration (For example, 10% to fade 10% of stream duration)
fadeColor This property is required. String
The Color for the fade In/Out. it can be on the CSS Level1 colors https://developer.mozilla.org/en-US/docs/Web/CSS/color_value/color_keywords or an RGB/hex value: e.g: rgb(255,0,0), 0xFF0000 or #FF0000
start String
The position in the input video from where to start fade. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Default is 0
duration This property is required. string
The Duration of the fade effect in the video. The value can be in ISO 8601 format (For example, PT05S to fade In/Out a color during 5 seconds), or a frame count (For example, 10 to fade 10 frames from the start time), or a relative value to stream duration (For example, 10% to fade 10% of stream duration)
fadeColor This property is required. string
The Color for the fade In/Out. it can be on the CSS Level1 colors https://developer.mozilla.org/en-US/docs/Web/CSS/color_value/color_keywords or an RGB/hex value: e.g: rgb(255,0,0), 0xFF0000 or #FF0000
start string
The position in the input video from where to start fade. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Default is 0
duration This property is required. str
The Duration of the fade effect in the video. The value can be in ISO 8601 format (For example, PT05S to fade In/Out a color during 5 seconds), or a frame count (For example, 10 to fade 10 frames from the start time), or a relative value to stream duration (For example, 10% to fade 10% of stream duration)
fade_color This property is required. str
The Color for the fade In/Out. it can be on the CSS Level1 colors https://developer.mozilla.org/en-US/docs/Web/CSS/color_value/color_keywords or an RGB/hex value: e.g: rgb(255,0,0), 0xFF0000 or #FF0000
start str
The position in the input video from where to start fade. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Default is 0
duration This property is required. String
The Duration of the fade effect in the video. The value can be in ISO 8601 format (For example, PT05S to fade In/Out a color during 5 seconds), or a frame count (For example, 10 to fade 10 frames from the start time), or a relative value to stream duration (For example, 10% to fade 10% of stream duration)
fadeColor This property is required. String
The Color for the fade In/Out. it can be on the CSS Level1 colors https://developer.mozilla.org/en-US/docs/Web/CSS/color_value/color_keywords or an RGB/hex value: e.g: rgb(255,0,0), 0xFF0000 or #FF0000
start String
The position in the input video from where to start fade. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Default is 0

FadeResponse
, FadeResponseArgs

Duration This property is required. string
The Duration of the fade effect in the video. The value can be in ISO 8601 format (For example, PT05S to fade In/Out a color during 5 seconds), or a frame count (For example, 10 to fade 10 frames from the start time), or a relative value to stream duration (For example, 10% to fade 10% of stream duration)
FadeColor This property is required. string
The Color for the fade In/Out. it can be on the CSS Level1 colors https://developer.mozilla.org/en-US/docs/Web/CSS/color_value/color_keywords or an RGB/hex value: e.g: rgb(255,0,0), 0xFF0000 or #FF0000
Start string
The position in the input video from where to start fade. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Default is 0
Duration This property is required. string
The Duration of the fade effect in the video. The value can be in ISO 8601 format (For example, PT05S to fade In/Out a color during 5 seconds), or a frame count (For example, 10 to fade 10 frames from the start time), or a relative value to stream duration (For example, 10% to fade 10% of stream duration)
FadeColor This property is required. string
The Color for the fade In/Out. it can be on the CSS Level1 colors https://developer.mozilla.org/en-US/docs/Web/CSS/color_value/color_keywords or an RGB/hex value: e.g: rgb(255,0,0), 0xFF0000 or #FF0000
Start string
The position in the input video from where to start fade. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Default is 0
duration This property is required. String
The Duration of the fade effect in the video. The value can be in ISO 8601 format (For example, PT05S to fade In/Out a color during 5 seconds), or a frame count (For example, 10 to fade 10 frames from the start time), or a relative value to stream duration (For example, 10% to fade 10% of stream duration)
fadeColor This property is required. String
The Color for the fade In/Out. it can be on the CSS Level1 colors https://developer.mozilla.org/en-US/docs/Web/CSS/color_value/color_keywords or an RGB/hex value: e.g: rgb(255,0,0), 0xFF0000 or #FF0000
start String
The position in the input video from where to start fade. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Default is 0
duration This property is required. string
The Duration of the fade effect in the video. The value can be in ISO 8601 format (For example, PT05S to fade In/Out a color during 5 seconds), or a frame count (For example, 10 to fade 10 frames from the start time), or a relative value to stream duration (For example, 10% to fade 10% of stream duration)
fadeColor This property is required. string
The Color for the fade In/Out. it can be on the CSS Level1 colors https://developer.mozilla.org/en-US/docs/Web/CSS/color_value/color_keywords or an RGB/hex value: e.g: rgb(255,0,0), 0xFF0000 or #FF0000
start string
The position in the input video from where to start fade. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Default is 0
duration This property is required. str
The Duration of the fade effect in the video. The value can be in ISO 8601 format (For example, PT05S to fade In/Out a color during 5 seconds), or a frame count (For example, 10 to fade 10 frames from the start time), or a relative value to stream duration (For example, 10% to fade 10% of stream duration)
fade_color This property is required. str
The Color for the fade In/Out. it can be on the CSS Level1 colors https://developer.mozilla.org/en-US/docs/Web/CSS/color_value/color_keywords or an RGB/hex value: e.g: rgb(255,0,0), 0xFF0000 or #FF0000
start str
The position in the input video from where to start fade. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Default is 0
duration This property is required. String
The Duration of the fade effect in the video. The value can be in ISO 8601 format (For example, PT05S to fade In/Out a color during 5 seconds), or a frame count (For example, 10 to fade 10 frames from the start time), or a relative value to stream duration (For example, 10% to fade 10% of stream duration)
fadeColor This property is required. String
The Color for the fade In/Out. it can be on the CSS Level1 colors https://developer.mozilla.org/en-US/docs/Web/CSS/color_value/color_keywords or an RGB/hex value: e.g: rgb(255,0,0), 0xFF0000 or #FF0000
start String
The position in the input video from where to start fade. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Default is 0

Filters
, FiltersArgs

Crop Pulumi.AzureNative.Media.Inputs.Rectangle
The parameters for the rectangular window with which to crop the input video.
Deinterlace Pulumi.AzureNative.Media.Inputs.Deinterlace
The de-interlacing settings.
FadeIn Pulumi.AzureNative.Media.Inputs.Fade
Describes the properties of a Fade effect applied to the input media.
FadeOut Pulumi.AzureNative.Media.Inputs.Fade
Describes the properties of a Fade effect applied to the input media.
Overlays List<Union<Pulumi.AzureNative.Media.Inputs.AudioOverlay, Pulumi.AzureNative.Media.Inputs.VideoOverlay>>
The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
Rotation string | Pulumi.AzureNative.Media.Rotation
The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
Crop Rectangle
The parameters for the rectangular window with which to crop the input video.
Deinterlace Deinterlace
The de-interlacing settings.
FadeIn Fade
Describes the properties of a Fade effect applied to the input media.
FadeOut Fade
Describes the properties of a Fade effect applied to the input media.
Overlays []interface{}
The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
Rotation string | Rotation
The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
crop Rectangle
The parameters for the rectangular window with which to crop the input video.
deinterlace Deinterlace
The de-interlacing settings.
fadeIn Fade
Describes the properties of a Fade effect applied to the input media.
fadeOut Fade
Describes the properties of a Fade effect applied to the input media.
overlays List<Either<AudioOverlay,VideoOverlay>>
The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
rotation String | Rotation
The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
crop Rectangle
The parameters for the rectangular window with which to crop the input video.
deinterlace Deinterlace
The de-interlacing settings.
fadeIn Fade
Describes the properties of a Fade effect applied to the input media.
fadeOut Fade
Describes the properties of a Fade effect applied to the input media.
overlays (AudioOverlay | VideoOverlay)[]
The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
rotation string | Rotation
The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
crop Rectangle
The parameters for the rectangular window with which to crop the input video.
deinterlace Deinterlace
The de-interlacing settings.
fade_in Fade
Describes the properties of a Fade effect applied to the input media.
fade_out Fade
Describes the properties of a Fade effect applied to the input media.
overlays Sequence[Union[AudioOverlay, VideoOverlay]]
The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
rotation str | Rotation
The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
crop Property Map
The parameters for the rectangular window with which to crop the input video.
deinterlace Property Map
The de-interlacing settings.
fadeIn Property Map
Describes the properties of a Fade effect applied to the input media.
fadeOut Property Map
Describes the properties of a Fade effect applied to the input media.
overlays List<Property Map | Property Map>
The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
rotation String | "Auto" | "None" | "Rotate0" | "Rotate90" | "Rotate180" | "Rotate270"
The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto

FiltersResponse
, FiltersResponseArgs

Crop Pulumi.AzureNative.Media.Inputs.RectangleResponse
The parameters for the rectangular window with which to crop the input video.
Deinterlace Pulumi.AzureNative.Media.Inputs.DeinterlaceResponse
The de-interlacing settings.
FadeIn Pulumi.AzureNative.Media.Inputs.FadeResponse
Describes the properties of a Fade effect applied to the input media.
FadeOut Pulumi.AzureNative.Media.Inputs.FadeResponse
Describes the properties of a Fade effect applied to the input media.
Overlays List<Union<Pulumi.AzureNative.Media.Inputs.AudioOverlayResponse, Pulumi.AzureNative.Media.Inputs.VideoOverlayResponse>>
The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
Rotation string
The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
Crop RectangleResponse
The parameters for the rectangular window with which to crop the input video.
Deinterlace DeinterlaceResponse
The de-interlacing settings.
FadeIn FadeResponse
Describes the properties of a Fade effect applied to the input media.
FadeOut FadeResponse
Describes the properties of a Fade effect applied to the input media.
Overlays []interface{}
The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
Rotation string
The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
crop RectangleResponse
The parameters for the rectangular window with which to crop the input video.
deinterlace DeinterlaceResponse
The de-interlacing settings.
fadeIn FadeResponse
Describes the properties of a Fade effect applied to the input media.
fadeOut FadeResponse
Describes the properties of a Fade effect applied to the input media.
overlays List<Either<AudioOverlayResponse,VideoOverlayResponse>>
The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
rotation String
The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
crop RectangleResponse
The parameters for the rectangular window with which to crop the input video.
deinterlace DeinterlaceResponse
The de-interlacing settings.
fadeIn FadeResponse
Describes the properties of a Fade effect applied to the input media.
fadeOut FadeResponse
Describes the properties of a Fade effect applied to the input media.
overlays (AudioOverlayResponse | VideoOverlayResponse)[]
The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
rotation string
The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
crop RectangleResponse
The parameters for the rectangular window with which to crop the input video.
deinterlace DeinterlaceResponse
The de-interlacing settings.
fade_in FadeResponse
Describes the properties of a Fade effect applied to the input media.
fade_out FadeResponse
Describes the properties of a Fade effect applied to the input media.
overlays Sequence[Union[AudioOverlayResponse, VideoOverlayResponse]]
The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
rotation str
The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
crop Property Map
The parameters for the rectangular window with which to crop the input video.
deinterlace Property Map
The de-interlacing settings.
fadeIn Property Map
Describes the properties of a Fade effect applied to the input media.
fadeOut Property Map
Describes the properties of a Fade effect applied to the input media.
overlays List<Property Map | Property Map>
The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
rotation String
The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto

H264Complexity
, H264ComplexityArgs

Speed
SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
Balanced
BalancedTells the encoder to use settings that achieve a balance between speed and quality.
Quality
QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
H264ComplexitySpeed
SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
H264ComplexityBalanced
BalancedTells the encoder to use settings that achieve a balance between speed and quality.
H264ComplexityQuality
QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
Speed
SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
Balanced
BalancedTells the encoder to use settings that achieve a balance between speed and quality.
Quality
QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
Speed
SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
Balanced
BalancedTells the encoder to use settings that achieve a balance between speed and quality.
Quality
QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
SPEED
SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
BALANCED
BalancedTells the encoder to use settings that achieve a balance between speed and quality.
QUALITY
QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
"Speed"
SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
"Balanced"
BalancedTells the encoder to use settings that achieve a balance between speed and quality.
"Quality"
QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.

H264Layer
, H264LayerArgs

Bitrate This property is required. int
The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
AdaptiveBFrame bool
Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
BFrames int
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
BufferWindow string
The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
Crf double
The value of CRF to be used when encoding this layer. This setting takes effect when RateControlMode of video codec is set at CRF mode. The range of CRF value is between 0 and 51, where lower values would result in better quality, at the expense of higher file sizes. Higher values mean more compression, but at some point quality degradation will be noticed. Default value is 23.
EntropyMode string | Pulumi.AzureNative.Media.EntropyMode
The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
FrameRate string
The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
Level string
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
MaxBitrate int
The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
Profile string | Pulumi.AzureNative.Media.H264VideoProfile
We currently support Baseline, Main, High, High422, High444. Default is Auto.
ReferenceFrames int
The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
Slices int
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
Bitrate This property is required. int
The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
AdaptiveBFrame bool
Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
BFrames int
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
BufferWindow string
The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
Crf float64
The value of CRF to be used when encoding this layer. This setting takes effect when RateControlMode of video codec is set at CRF mode. The range of CRF value is between 0 and 51, where lower values would result in better quality, at the expense of higher file sizes. Higher values mean more compression, but at some point quality degradation will be noticed. Default value is 23.
EntropyMode string | EntropyMode
The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
FrameRate string
The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
Level string
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
MaxBitrate int
The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
Profile string | H264VideoProfile
We currently support Baseline, Main, High, High422, High444. Default is Auto.
ReferenceFrames int
The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
Slices int
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
bitrate This property is required. Integer
The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
adaptiveBFrame Boolean
Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
bFrames Integer
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
bufferWindow String
The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
crf Double
The value of CRF to be used when encoding this layer. This setting takes effect when RateControlMode of video codec is set at CRF mode. The range of CRF value is between 0 and 51, where lower values would result in better quality, at the expense of higher file sizes. Higher values mean more compression, but at some point quality degradation will be noticed. Default value is 23.
entropyMode String | EntropyMode
The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
frameRate String
The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
level String
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
maxBitrate Integer
The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
profile String | H264VideoProfile
We currently support Baseline, Main, High, High422, High444. Default is Auto.
referenceFrames Integer
The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
slices Integer
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
bitrate This property is required. number
The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
adaptiveBFrame boolean
Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
bFrames number
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
bufferWindow string
The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
crf number
The value of CRF to be used when encoding this layer. This setting takes effect when RateControlMode of video codec is set at CRF mode. The range of CRF value is between 0 and 51, where lower values would result in better quality, at the expense of higher file sizes. Higher values mean more compression, but at some point quality degradation will be noticed. Default value is 23.
entropyMode string | EntropyMode
The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
frameRate string
The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
level string
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
maxBitrate number
The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
profile string | H264VideoProfile
We currently support Baseline, Main, High, High422, High444. Default is Auto.
referenceFrames number
The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
slices number
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
bitrate This property is required. int
The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
adaptive_b_frame bool
Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
b_frames int
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
buffer_window str
The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
crf float
The value of CRF to be used when encoding this layer. This setting takes effect when RateControlMode of video codec is set at CRF mode. The range of CRF value is between 0 and 51, where lower values would result in better quality, at the expense of higher file sizes. Higher values mean more compression, but at some point quality degradation will be noticed. Default value is 23.
entropy_mode str | EntropyMode
The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
frame_rate str
The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
height str
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label str
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
level str
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
max_bitrate int
The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
profile str | H264VideoProfile
We currently support Baseline, Main, High, High422, High444. Default is Auto.
reference_frames int
The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
slices int
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
width str
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
bitrate This property is required. Number
The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
adaptiveBFrame Boolean
Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
bFrames Number
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
bufferWindow String
The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
crf Number
The value of CRF to be used when encoding this layer. This setting takes effect when RateControlMode of video codec is set at CRF mode. The range of CRF value is between 0 and 51, where lower values would result in better quality, at the expense of higher file sizes. Higher values mean more compression, but at some point quality degradation will be noticed. Default value is 23.
entropyMode String | "Cabac" | "Cavlc"
The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
frameRate String
The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
level String
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
maxBitrate Number
The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
profile String | "Auto" | "Baseline" | "Main" | "High" | "High422" | "High444"
We currently support Baseline, Main, High, High422, High444. Default is Auto.
referenceFrames Number
The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
slices Number
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

H264LayerResponse
, H264LayerResponseArgs

Bitrate This property is required. int
The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
AdaptiveBFrame bool
Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
BFrames int
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
BufferWindow string
The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
Crf double
The value of CRF to be used when encoding this layer. This setting takes effect when RateControlMode of video codec is set at CRF mode. The range of CRF value is between 0 and 51, where lower values would result in better quality, at the expense of higher file sizes. Higher values mean more compression, but at some point quality degradation will be noticed. Default value is 23.
EntropyMode string
The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
FrameRate string
The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
Level string
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
MaxBitrate int
The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
Profile string
We currently support Baseline, Main, High, High422, High444. Default is Auto.
ReferenceFrames int
The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
Slices int
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
Bitrate This property is required. int
The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
AdaptiveBFrame bool
Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
BFrames int
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
BufferWindow string
The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
Crf float64
The value of CRF to be used when encoding this layer. This setting takes effect when RateControlMode of video codec is set at CRF mode. The range of CRF value is between 0 and 51, where lower values would result in better quality, at the expense of higher file sizes. Higher values mean more compression, but at some point quality degradation will be noticed. Default value is 23.
EntropyMode string
The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
FrameRate string
The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
Level string
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
MaxBitrate int
The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
Profile string
We currently support Baseline, Main, High, High422, High444. Default is Auto.
ReferenceFrames int
The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
Slices int
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
bitrate This property is required. Integer
The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
adaptiveBFrame Boolean
Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
bFrames Integer
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
bufferWindow String
The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
crf Double
The value of CRF to be used when encoding this layer. This setting takes effect when RateControlMode of video codec is set at CRF mode. The range of CRF value is between 0 and 51, where lower values would result in better quality, at the expense of higher file sizes. Higher values mean more compression, but at some point quality degradation will be noticed. Default value is 23.
entropyMode String
The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
frameRate String
The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
level String
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
maxBitrate Integer
The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
profile String
We currently support Baseline, Main, High, High422, High444. Default is Auto.
referenceFrames Integer
The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
slices Integer
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
bitrate This property is required. number
The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
adaptiveBFrame boolean
Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
bFrames number
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
bufferWindow string
The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
crf number
The value of CRF to be used when encoding this layer. This setting takes effect when RateControlMode of video codec is set at CRF mode. The range of CRF value is between 0 and 51, where lower values would result in better quality, at the expense of higher file sizes. Higher values mean more compression, but at some point quality degradation will be noticed. Default value is 23.
entropyMode string
The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
frameRate string
The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
level string
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
maxBitrate number
The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
profile string
We currently support Baseline, Main, High, High422, High444. Default is Auto.
referenceFrames number
The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
slices number
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
bitrate This property is required. int
The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
adaptive_b_frame bool
Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
b_frames int
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
buffer_window str
The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
crf float
The value of CRF to be used when encoding this layer. This setting takes effect when RateControlMode of video codec is set at CRF mode. The range of CRF value is between 0 and 51, where lower values would result in better quality, at the expense of higher file sizes. Higher values mean more compression, but at some point quality degradation will be noticed. Default value is 23.
entropy_mode str
The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
frame_rate str
The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
height str
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label str
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
level str
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
max_bitrate int
The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
profile str
We currently support Baseline, Main, High, High422, High444. Default is Auto.
reference_frames int
The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
slices int
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
width str
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
bitrate This property is required. Number
The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
adaptiveBFrame Boolean
Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
bFrames Number
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
bufferWindow String
The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
crf Number
The value of CRF to be used when encoding this layer. This setting takes effect when RateControlMode of video codec is set at CRF mode. The range of CRF value is between 0 and 51, where lower values would result in better quality, at the expense of higher file sizes. Higher values mean more compression, but at some point quality degradation will be noticed. Default value is 23.
entropyMode String
The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
frameRate String
The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
level String
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
maxBitrate Number
The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
profile String
We currently support Baseline, Main, High, High422, High444. Default is Auto.
referenceFrames Number
The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
slices Number
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

H264RateControlMode
, H264RateControlModeArgs

ABR
ABRAverage Bitrate (ABR) mode that hits the target bitrate: Default mode.
CBR
CBRConstant Bitrate (CBR) mode that tightens bitrate variations around target bitrate.
CRF
CRFConstant Rate Factor (CRF) mode that targets at constant subjective quality.
H264RateControlModeABR
ABRAverage Bitrate (ABR) mode that hits the target bitrate: Default mode.
H264RateControlModeCBR
CBRConstant Bitrate (CBR) mode that tightens bitrate variations around target bitrate.
H264RateControlModeCRF
CRFConstant Rate Factor (CRF) mode that targets at constant subjective quality.
ABR
ABRAverage Bitrate (ABR) mode that hits the target bitrate: Default mode.
CBR
CBRConstant Bitrate (CBR) mode that tightens bitrate variations around target bitrate.
CRF
CRFConstant Rate Factor (CRF) mode that targets at constant subjective quality.
ABR
ABRAverage Bitrate (ABR) mode that hits the target bitrate: Default mode.
CBR
CBRConstant Bitrate (CBR) mode that tightens bitrate variations around target bitrate.
CRF
CRFConstant Rate Factor (CRF) mode that targets at constant subjective quality.
ABR
ABRAverage Bitrate (ABR) mode that hits the target bitrate: Default mode.
CBR
CBRConstant Bitrate (CBR) mode that tightens bitrate variations around target bitrate.
CRF
CRFConstant Rate Factor (CRF) mode that targets at constant subjective quality.
"ABR"
ABRAverage Bitrate (ABR) mode that hits the target bitrate: Default mode.
"CBR"
CBRConstant Bitrate (CBR) mode that tightens bitrate variations around target bitrate.
"CRF"
CRFConstant Rate Factor (CRF) mode that targets at constant subjective quality.

H264Video
, H264VideoArgs

Complexity string | Pulumi.AzureNative.Media.H264Complexity
Tells the encoder how to choose its encoding settings. The default value is Balanced.
KeyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
Layers List<Pulumi.AzureNative.Media.Inputs.H264Layer>
The collection of output H.264 layers to be produced by the encoder.
RateControlMode string | Pulumi.AzureNative.Media.H264RateControlMode
The video rate control mode
SceneChangeDetection bool
Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
StretchMode string | Pulumi.AzureNative.Media.StretchMode
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
SyncMode string | Pulumi.AzureNative.Media.VideoSyncMode
The Video Sync Mode
Complexity string | H264Complexity
Tells the encoder how to choose its encoding settings. The default value is Balanced.
KeyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
Layers []H264Layer
The collection of output H.264 layers to be produced by the encoder.
RateControlMode string | H264RateControlMode
The video rate control mode
SceneChangeDetection bool
Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
StretchMode string | StretchMode
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
SyncMode string | VideoSyncMode
The Video Sync Mode
complexity String | H264Complexity
Tells the encoder how to choose its encoding settings. The default value is Balanced.
keyFrameInterval String
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label String
An optional label for the codec. The label can be used to control muxing behavior.
layers List<H264Layer>
The collection of output H.264 layers to be produced by the encoder.
rateControlMode String | H264RateControlMode
The video rate control mode
sceneChangeDetection Boolean
Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
stretchMode String | StretchMode
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode String | VideoSyncMode
The Video Sync Mode
complexity string | H264Complexity
Tells the encoder how to choose its encoding settings. The default value is Balanced.
keyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label string
An optional label for the codec. The label can be used to control muxing behavior.
layers H264Layer[]
The collection of output H.264 layers to be produced by the encoder.
rateControlMode string | H264RateControlMode
The video rate control mode
sceneChangeDetection boolean
Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
stretchMode string | StretchMode
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode string | VideoSyncMode
The Video Sync Mode
complexity str | H264Complexity
Tells the encoder how to choose its encoding settings. The default value is Balanced.
key_frame_interval str
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label str
An optional label for the codec. The label can be used to control muxing behavior.
layers Sequence[H264Layer]
The collection of output H.264 layers to be produced by the encoder.
rate_control_mode str | H264RateControlMode
The video rate control mode
scene_change_detection bool
Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
stretch_mode str | StretchMode
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
sync_mode str | VideoSyncMode
The Video Sync Mode
complexity String | "Speed" | "Balanced" | "Quality"
Tells the encoder how to choose its encoding settings. The default value is Balanced.
keyFrameInterval String
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label String
An optional label for the codec. The label can be used to control muxing behavior.
layers List<Property Map>
The collection of output H.264 layers to be produced by the encoder.
rateControlMode String | "ABR" | "CBR" | "CRF"
The video rate control mode
sceneChangeDetection Boolean
Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
stretchMode String | "None" | "AutoSize" | "AutoFit"
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode String | "Auto" | "Passthrough" | "Cfr" | "Vfr"
The Video Sync Mode

H264VideoProfile
, H264VideoProfileArgs

Auto
AutoTells the encoder to automatically determine the appropriate H.264 profile.
Baseline
BaselineBaseline profile
Main
MainMain profile
High
HighHigh profile.
High422
High422High 4:2:2 profile.
High444
High444High 4:4:4 predictive profile.
H264VideoProfileAuto
AutoTells the encoder to automatically determine the appropriate H.264 profile.
H264VideoProfileBaseline
BaselineBaseline profile
H264VideoProfileMain
MainMain profile
H264VideoProfileHigh
HighHigh profile.
H264VideoProfileHigh422
High422High 4:2:2 profile.
H264VideoProfileHigh444
High444High 4:4:4 predictive profile.
Auto
AutoTells the encoder to automatically determine the appropriate H.264 profile.
Baseline
BaselineBaseline profile
Main
MainMain profile
High
HighHigh profile.
High422
High422High 4:2:2 profile.
High444
High444High 4:4:4 predictive profile.
Auto
AutoTells the encoder to automatically determine the appropriate H.264 profile.
Baseline
BaselineBaseline profile
Main
MainMain profile
High
HighHigh profile.
High422
High422High 4:2:2 profile.
High444
High444High 4:4:4 predictive profile.
AUTO
AutoTells the encoder to automatically determine the appropriate H.264 profile.
BASELINE
BaselineBaseline profile
MAIN
MainMain profile
HIGH
HighHigh profile.
HIGH422
High422High 4:2:2 profile.
HIGH444
High444High 4:4:4 predictive profile.
"Auto"
AutoTells the encoder to automatically determine the appropriate H.264 profile.
"Baseline"
BaselineBaseline profile
"Main"
MainMain profile
"High"
HighHigh profile.
"High422"
High422High 4:2:2 profile.
"High444"
High444High 4:4:4 predictive profile.

H264VideoResponse
, H264VideoResponseArgs

Complexity string
Tells the encoder how to choose its encoding settings. The default value is Balanced.
KeyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
Layers List<Pulumi.AzureNative.Media.Inputs.H264LayerResponse>
The collection of output H.264 layers to be produced by the encoder.
RateControlMode string
The video rate control mode
SceneChangeDetection bool
Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
StretchMode string
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
SyncMode string
The Video Sync Mode
Complexity string
Tells the encoder how to choose its encoding settings. The default value is Balanced.
KeyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
Layers []H264LayerResponse
The collection of output H.264 layers to be produced by the encoder.
RateControlMode string
The video rate control mode
SceneChangeDetection bool
Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
StretchMode string
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
SyncMode string
The Video Sync Mode
complexity String
Tells the encoder how to choose its encoding settings. The default value is Balanced.
keyFrameInterval String
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label String
An optional label for the codec. The label can be used to control muxing behavior.
layers List<H264LayerResponse>
The collection of output H.264 layers to be produced by the encoder.
rateControlMode String
The video rate control mode
sceneChangeDetection Boolean
Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
stretchMode String
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode String
The Video Sync Mode
complexity string
Tells the encoder how to choose its encoding settings. The default value is Balanced.
keyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label string
An optional label for the codec. The label can be used to control muxing behavior.
layers H264LayerResponse[]
The collection of output H.264 layers to be produced by the encoder.
rateControlMode string
The video rate control mode
sceneChangeDetection boolean
Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
stretchMode string
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode string
The Video Sync Mode
complexity str
Tells the encoder how to choose its encoding settings. The default value is Balanced.
key_frame_interval str
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label str
An optional label for the codec. The label can be used to control muxing behavior.
layers Sequence[H264LayerResponse]
The collection of output H.264 layers to be produced by the encoder.
rate_control_mode str
The video rate control mode
scene_change_detection bool
Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
stretch_mode str
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
sync_mode str
The Video Sync Mode
complexity String
Tells the encoder how to choose its encoding settings. The default value is Balanced.
keyFrameInterval String
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label String
An optional label for the codec. The label can be used to control muxing behavior.
layers List<Property Map>
The collection of output H.264 layers to be produced by the encoder.
rateControlMode String
The video rate control mode
sceneChangeDetection Boolean
Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
stretchMode String
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode String
The Video Sync Mode

H265Complexity
, H265ComplexityArgs

Speed
SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
Balanced
BalancedTells the encoder to use settings that achieve a balance between speed and quality.
Quality
QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
H265ComplexitySpeed
SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
H265ComplexityBalanced
BalancedTells the encoder to use settings that achieve a balance between speed and quality.
H265ComplexityQuality
QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
Speed
SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
Balanced
BalancedTells the encoder to use settings that achieve a balance between speed and quality.
Quality
QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
Speed
SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
Balanced
BalancedTells the encoder to use settings that achieve a balance between speed and quality.
Quality
QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
SPEED
SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
BALANCED
BalancedTells the encoder to use settings that achieve a balance between speed and quality.
QUALITY
QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
"Speed"
SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
"Balanced"
BalancedTells the encoder to use settings that achieve a balance between speed and quality.
"Quality"
QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.

H265Layer
, H265LayerArgs

Bitrate This property is required. int
The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
AdaptiveBFrame bool
Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
BFrames int
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
BufferWindow string
The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
Crf double
The value of CRF to be used when encoding this layer. This setting takes effect when RateControlMode of video codec is set at CRF mode. The range of CRF value is between 0 and 51, where lower values would result in better quality, at the expense of higher file sizes. Higher values mean more compression, but at some point quality degradation will be noticed. Default value is 28.
FrameRate string
The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
Level string
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
MaxBitrate int
The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
Profile string | Pulumi.AzureNative.Media.H265VideoProfile
We currently support Main. Default is Auto.
ReferenceFrames int
The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
Slices int
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
Bitrate This property is required. int
The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
AdaptiveBFrame bool
Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
BFrames int
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
BufferWindow string
The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
Crf float64
The value of CRF to be used when encoding this layer. This setting takes effect when RateControlMode of video codec is set at CRF mode. The range of CRF value is between 0 and 51, where lower values would result in better quality, at the expense of higher file sizes. Higher values mean more compression, but at some point quality degradation will be noticed. Default value is 28.
FrameRate string
The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
Level string
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
MaxBitrate int
The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
Profile string | H265VideoProfile
We currently support Main. Default is Auto.
ReferenceFrames int
The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
Slices int
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
bitrate This property is required. Integer
The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
adaptiveBFrame Boolean
Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
bFrames Integer
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
bufferWindow String
The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
crf Double
The value of CRF to be used when encoding this layer. This setting takes effect when RateControlMode of video codec is set at CRF mode. The range of CRF value is between 0 and 51, where lower values would result in better quality, at the expense of higher file sizes. Higher values mean more compression, but at some point quality degradation will be noticed. Default value is 28.
frameRate String
The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
level String
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
maxBitrate Integer
The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
profile String | H265VideoProfile
We currently support Main. Default is Auto.
referenceFrames Integer
The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
slices Integer
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
bitrate This property is required. number
The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
adaptiveBFrame boolean
Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
bFrames number
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
bufferWindow string
The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
crf number
The value of CRF to be used when encoding this layer. This setting takes effect when RateControlMode of video codec is set at CRF mode. The range of CRF value is between 0 and 51, where lower values would result in better quality, at the expense of higher file sizes. Higher values mean more compression, but at some point quality degradation will be noticed. Default value is 28.
frameRate string
The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
level string
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
maxBitrate number
The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
profile string | H265VideoProfile
We currently support Main. Default is Auto.
referenceFrames number
The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
slices number
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
bitrate This property is required. int
The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
adaptive_b_frame bool
Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
b_frames int
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
buffer_window str
The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
crf float
The value of CRF to be used when encoding this layer. This setting takes effect when RateControlMode of video codec is set at CRF mode. The range of CRF value is between 0 and 51, where lower values would result in better quality, at the expense of higher file sizes. Higher values mean more compression, but at some point quality degradation will be noticed. Default value is 28.
frame_rate str
The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
height str
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label str
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
level str
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
max_bitrate int
The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
profile str | H265VideoProfile
We currently support Main. Default is Auto.
reference_frames int
The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
slices int
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
width str
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
bitrate This property is required. Number
The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
adaptiveBFrame Boolean
Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
bFrames Number
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
bufferWindow String
The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
crf Number
The value of CRF to be used when encoding this layer. This setting takes effect when RateControlMode of video codec is set at CRF mode. The range of CRF value is between 0 and 51, where lower values would result in better quality, at the expense of higher file sizes. Higher values mean more compression, but at some point quality degradation will be noticed. Default value is 28.
frameRate String
The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
level String
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
maxBitrate Number
The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
profile String | "Auto" | "Main" | "Main10"
We currently support Main. Default is Auto.
referenceFrames Number
The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
slices Number
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

H265LayerResponse
, H265LayerResponseArgs

Bitrate This property is required. int
The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
AdaptiveBFrame bool
Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
BFrames int
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
BufferWindow string
The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
Crf double
The value of CRF to be used when encoding this layer. This setting takes effect when RateControlMode of video codec is set at CRF mode. The range of CRF value is between 0 and 51, where lower values would result in better quality, at the expense of higher file sizes. Higher values mean more compression, but at some point quality degradation will be noticed. Default value is 28.
FrameRate string
The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
Level string
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
MaxBitrate int
The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
Profile string
We currently support Main. Default is Auto.
ReferenceFrames int
The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
Slices int
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
Bitrate This property is required. int
The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
AdaptiveBFrame bool
Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
BFrames int
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
BufferWindow string
The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
Crf float64
The value of CRF to be used when encoding this layer. This setting takes effect when RateControlMode of video codec is set at CRF mode. The range of CRF value is between 0 and 51, where lower values would result in better quality, at the expense of higher file sizes. Higher values mean more compression, but at some point quality degradation will be noticed. Default value is 28.
FrameRate string
The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
Level string
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
MaxBitrate int
The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
Profile string
We currently support Main. Default is Auto.
ReferenceFrames int
The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
Slices int
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
bitrate This property is required. Integer
The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
adaptiveBFrame Boolean
Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
bFrames Integer
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
bufferWindow String
The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
crf Double
The value of CRF to be used when encoding this layer. This setting takes effect when RateControlMode of video codec is set at CRF mode. The range of CRF value is between 0 and 51, where lower values would result in better quality, at the expense of higher file sizes. Higher values mean more compression, but at some point quality degradation will be noticed. Default value is 28.
frameRate String
The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
level String
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
maxBitrate Integer
The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
profile String
We currently support Main. Default is Auto.
referenceFrames Integer
The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
slices Integer
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
bitrate This property is required. number
The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
adaptiveBFrame boolean
Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
bFrames number
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
bufferWindow string
The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
crf number
The value of CRF to be used when encoding this layer. This setting takes effect when RateControlMode of video codec is set at CRF mode. The range of CRF value is between 0 and 51, where lower values would result in better quality, at the expense of higher file sizes. Higher values mean more compression, but at some point quality degradation will be noticed. Default value is 28.
frameRate string
The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
level string
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
maxBitrate number
The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
profile string
We currently support Main. Default is Auto.
referenceFrames number
The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
slices number
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
bitrate This property is required. int
The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
adaptive_b_frame bool
Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
b_frames int
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
buffer_window str
The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
crf float
The value of CRF to be used when encoding this layer. This setting takes effect when RateControlMode of video codec is set at CRF mode. The range of CRF value is between 0 and 51, where lower values would result in better quality, at the expense of higher file sizes. Higher values mean more compression, but at some point quality degradation will be noticed. Default value is 28.
frame_rate str
The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
height str
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label str
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
level str
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
max_bitrate int
The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
profile str
We currently support Main. Default is Auto.
reference_frames int
The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
slices int
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
width str
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
bitrate This property is required. Number
The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
adaptiveBFrame Boolean
Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
bFrames Number
The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
bufferWindow String
The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
crf Number
The value of CRF to be used when encoding this layer. This setting takes effect when RateControlMode of video codec is set at CRF mode. The range of CRF value is between 0 and 51, where lower values would result in better quality, at the expense of higher file sizes. Higher values mean more compression, but at some point quality degradation will be noticed. Default value is 28.
frameRate String
The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
level String
We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
maxBitrate Number
The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
profile String
We currently support Main. Default is Auto.
referenceFrames Number
The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
slices Number
The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

H265Video
, H265VideoArgs

Complexity string | Pulumi.AzureNative.Media.H265Complexity
Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
KeyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
Layers List<Pulumi.AzureNative.Media.Inputs.H265Layer>
The collection of output H.265 layers to be produced by the encoder.
SceneChangeDetection bool
Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
StretchMode string | Pulumi.AzureNative.Media.StretchMode
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
SyncMode string | Pulumi.AzureNative.Media.VideoSyncMode
The Video Sync Mode
Complexity string | H265Complexity
Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
KeyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
Layers []H265Layer
The collection of output H.265 layers to be produced by the encoder.
SceneChangeDetection bool
Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
StretchMode string | StretchMode
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
SyncMode string | VideoSyncMode
The Video Sync Mode
complexity String | H265Complexity
Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
keyFrameInterval String
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label String
An optional label for the codec. The label can be used to control muxing behavior.
layers List<H265Layer>
The collection of output H.265 layers to be produced by the encoder.
sceneChangeDetection Boolean
Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
stretchMode String | StretchMode
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode String | VideoSyncMode
The Video Sync Mode
complexity string | H265Complexity
Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
keyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label string
An optional label for the codec. The label can be used to control muxing behavior.
layers H265Layer[]
The collection of output H.265 layers to be produced by the encoder.
sceneChangeDetection boolean
Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
stretchMode string | StretchMode
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode string | VideoSyncMode
The Video Sync Mode
complexity str | H265Complexity
Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
key_frame_interval str
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label str
An optional label for the codec. The label can be used to control muxing behavior.
layers Sequence[H265Layer]
The collection of output H.265 layers to be produced by the encoder.
scene_change_detection bool
Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
stretch_mode str | StretchMode
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
sync_mode str | VideoSyncMode
The Video Sync Mode
complexity String | "Speed" | "Balanced" | "Quality"
Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
keyFrameInterval String
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label String
An optional label for the codec. The label can be used to control muxing behavior.
layers List<Property Map>
The collection of output H.265 layers to be produced by the encoder.
sceneChangeDetection Boolean
Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
stretchMode String | "None" | "AutoSize" | "AutoFit"
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode String | "Auto" | "Passthrough" | "Cfr" | "Vfr"
The Video Sync Mode

H265VideoProfile
, H265VideoProfileArgs

Auto
AutoTells the encoder to automatically determine the appropriate H.265 profile.
Main
MainMain profile (https://x265.readthedocs.io/en/default/cli.html?highlight=profile#profile-level-tier)
Main10
Main10Main 10 profile (https://en.wikipedia.org/wiki/High_Efficiency_Video_Coding#Main_10)
H265VideoProfileAuto
AutoTells the encoder to automatically determine the appropriate H.265 profile.
H265VideoProfileMain
MainMain profile (https://x265.readthedocs.io/en/default/cli.html?highlight=profile#profile-level-tier)
H265VideoProfileMain10
Main10Main 10 profile (https://en.wikipedia.org/wiki/High_Efficiency_Video_Coding#Main_10)
Auto
AutoTells the encoder to automatically determine the appropriate H.265 profile.
Main
MainMain profile (https://x265.readthedocs.io/en/default/cli.html?highlight=profile#profile-level-tier)
Main10
Main10Main 10 profile (https://en.wikipedia.org/wiki/High_Efficiency_Video_Coding#Main_10)
Auto
AutoTells the encoder to automatically determine the appropriate H.265 profile.
Main
MainMain profile (https://x265.readthedocs.io/en/default/cli.html?highlight=profile#profile-level-tier)
Main10
Main10Main 10 profile (https://en.wikipedia.org/wiki/High_Efficiency_Video_Coding#Main_10)
AUTO
AutoTells the encoder to automatically determine the appropriate H.265 profile.
MAIN
MainMain profile (https://x265.readthedocs.io/en/default/cli.html?highlight=profile#profile-level-tier)
MAIN10
Main10Main 10 profile (https://en.wikipedia.org/wiki/High_Efficiency_Video_Coding#Main_10)
"Auto"
AutoTells the encoder to automatically determine the appropriate H.265 profile.
"Main"
MainMain profile (https://x265.readthedocs.io/en/default/cli.html?highlight=profile#profile-level-tier)
"Main10"
Main10Main 10 profile (https://en.wikipedia.org/wiki/High_Efficiency_Video_Coding#Main_10)

H265VideoResponse
, H265VideoResponseArgs

Complexity string
Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
KeyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
Layers List<Pulumi.AzureNative.Media.Inputs.H265LayerResponse>
The collection of output H.265 layers to be produced by the encoder.
SceneChangeDetection bool
Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
StretchMode string
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
SyncMode string
The Video Sync Mode
Complexity string
Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
KeyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
Layers []H265LayerResponse
The collection of output H.265 layers to be produced by the encoder.
SceneChangeDetection bool
Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
StretchMode string
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
SyncMode string
The Video Sync Mode
complexity String
Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
keyFrameInterval String
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label String
An optional label for the codec. The label can be used to control muxing behavior.
layers List<H265LayerResponse>
The collection of output H.265 layers to be produced by the encoder.
sceneChangeDetection Boolean
Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
stretchMode String
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode String
The Video Sync Mode
complexity string
Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
keyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label string
An optional label for the codec. The label can be used to control muxing behavior.
layers H265LayerResponse[]
The collection of output H.265 layers to be produced by the encoder.
sceneChangeDetection boolean
Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
stretchMode string
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode string
The Video Sync Mode
complexity str
Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
key_frame_interval str
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label str
An optional label for the codec. The label can be used to control muxing behavior.
layers Sequence[H265LayerResponse]
The collection of output H.265 layers to be produced by the encoder.
scene_change_detection bool
Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
stretch_mode str
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
sync_mode str
The Video Sync Mode
complexity String
Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
keyFrameInterval String
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label String
An optional label for the codec. The label can be used to control muxing behavior.
layers List<Property Map>
The collection of output H.265 layers to be produced by the encoder.
sceneChangeDetection Boolean
Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
stretchMode String
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode String
The Video Sync Mode

Image
, ImageArgs

Start This property is required. string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
KeyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
Range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
Step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
StretchMode string | Pulumi.AzureNative.Media.StretchMode
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
SyncMode string | Pulumi.AzureNative.Media.VideoSyncMode
The Video Sync Mode
Start This property is required. string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
KeyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
Range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
Step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
StretchMode string | StretchMode
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
SyncMode string | VideoSyncMode
The Video Sync Mode
start This property is required. String
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
keyFrameInterval String
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label String
An optional label for the codec. The label can be used to control muxing behavior.
range String
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
step String
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
stretchMode String | StretchMode
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode String | VideoSyncMode
The Video Sync Mode
start This property is required. string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
keyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label string
An optional label for the codec. The label can be used to control muxing behavior.
range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
stretchMode string | StretchMode
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode string | VideoSyncMode
The Video Sync Mode
start This property is required. str
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
key_frame_interval str
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label str
An optional label for the codec. The label can be used to control muxing behavior.
range str
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
step str
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
stretch_mode str | StretchMode
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
sync_mode str | VideoSyncMode
The Video Sync Mode
start This property is required. String
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
keyFrameInterval String
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label String
An optional label for the codec. The label can be used to control muxing behavior.
range String
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
step String
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
stretchMode String | "None" | "AutoSize" | "AutoFit"
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode String | "Auto" | "Passthrough" | "Cfr" | "Vfr"
The Video Sync Mode

ImageFormat
, ImageFormatArgs

FilenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
FilenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
filenamePattern This property is required. String
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
filenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
filename_pattern This property is required. str
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
filenamePattern This property is required. String
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

ImageFormatResponse
, ImageFormatResponseArgs

FilenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
FilenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
filenamePattern This property is required. String
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
filenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
filename_pattern This property is required. str
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
filenamePattern This property is required. String
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

ImageResponse
, ImageResponseArgs

Start This property is required. string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
KeyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
Range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
Step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
StretchMode string
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
SyncMode string
The Video Sync Mode
Start This property is required. string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
KeyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
Range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
Step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
StretchMode string
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
SyncMode string
The Video Sync Mode
start This property is required. String
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
keyFrameInterval String
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label String
An optional label for the codec. The label can be used to control muxing behavior.
range String
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
step String
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
stretchMode String
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode String
The Video Sync Mode
start This property is required. string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
keyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label string
An optional label for the codec. The label can be used to control muxing behavior.
range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
stretchMode string
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode string
The Video Sync Mode
start This property is required. str
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
key_frame_interval str
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label str
An optional label for the codec. The label can be used to control muxing behavior.
range str
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
step str
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
stretch_mode str
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
sync_mode str
The Video Sync Mode
start This property is required. String
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
keyFrameInterval String
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label String
An optional label for the codec. The label can be used to control muxing behavior.
range String
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
step String
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
stretchMode String
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode String
The Video Sync Mode

InsightsType
, InsightsTypeArgs

AudioInsightsOnly
AudioInsightsOnlyGenerate audio only insights. Ignore video even if present. Fails if no audio is present.
VideoInsightsOnly
VideoInsightsOnlyGenerate video only insights. Ignore audio if present. Fails if no video is present.
AllInsights
AllInsightsGenerate both audio and video insights. Fails if either audio or video Insights fail.
InsightsTypeAudioInsightsOnly
AudioInsightsOnlyGenerate audio only insights. Ignore video even if present. Fails if no audio is present.
InsightsTypeVideoInsightsOnly
VideoInsightsOnlyGenerate video only insights. Ignore audio if present. Fails if no video is present.
InsightsTypeAllInsights
AllInsightsGenerate both audio and video insights. Fails if either audio or video Insights fail.
AudioInsightsOnly
AudioInsightsOnlyGenerate audio only insights. Ignore video even if present. Fails if no audio is present.
VideoInsightsOnly
VideoInsightsOnlyGenerate video only insights. Ignore audio if present. Fails if no video is present.
AllInsights
AllInsightsGenerate both audio and video insights. Fails if either audio or video Insights fail.
AudioInsightsOnly
AudioInsightsOnlyGenerate audio only insights. Ignore video even if present. Fails if no audio is present.
VideoInsightsOnly
VideoInsightsOnlyGenerate video only insights. Ignore audio if present. Fails if no video is present.
AllInsights
AllInsightsGenerate both audio and video insights. Fails if either audio or video Insights fail.
AUDIO_INSIGHTS_ONLY
AudioInsightsOnlyGenerate audio only insights. Ignore video even if present. Fails if no audio is present.
VIDEO_INSIGHTS_ONLY
VideoInsightsOnlyGenerate video only insights. Ignore audio if present. Fails if no video is present.
ALL_INSIGHTS
AllInsightsGenerate both audio and video insights. Fails if either audio or video Insights fail.
"AudioInsightsOnly"
AudioInsightsOnlyGenerate audio only insights. Ignore video even if present. Fails if no audio is present.
"VideoInsightsOnly"
VideoInsightsOnlyGenerate video only insights. Ignore audio if present. Fails if no video is present.
"AllInsights"
AllInsightsGenerate both audio and video insights. Fails if either audio or video Insights fail.

InterleaveOutput
, InterleaveOutputArgs

NonInterleavedOutput
NonInterleavedOutputThe output is video-only or audio-only.
InterleavedOutput
InterleavedOutputThe output includes both audio and video.
InterleaveOutputNonInterleavedOutput
NonInterleavedOutputThe output is video-only or audio-only.
InterleaveOutputInterleavedOutput
InterleavedOutputThe output includes both audio and video.
NonInterleavedOutput
NonInterleavedOutputThe output is video-only or audio-only.
InterleavedOutput
InterleavedOutputThe output includes both audio and video.
NonInterleavedOutput
NonInterleavedOutputThe output is video-only or audio-only.
InterleavedOutput
InterleavedOutputThe output includes both audio and video.
NON_INTERLEAVED_OUTPUT
NonInterleavedOutputThe output is video-only or audio-only.
INTERLEAVED_OUTPUT
InterleavedOutputThe output includes both audio and video.
"NonInterleavedOutput"
NonInterleavedOutputThe output is video-only or audio-only.
"InterleavedOutput"
InterleavedOutputThe output includes both audio and video.

JpgFormat
, JpgFormatArgs

FilenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
FilenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
filenamePattern This property is required. String
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
filenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
filename_pattern This property is required. str
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
filenamePattern This property is required. String
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

JpgFormatResponse
, JpgFormatResponseArgs

FilenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
FilenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
filenamePattern This property is required. String
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
filenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
filename_pattern This property is required. str
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
filenamePattern This property is required. String
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

JpgImage
, JpgImageArgs

Start This property is required. string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
KeyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
Layers List<Pulumi.AzureNative.Media.Inputs.JpgLayer>
A collection of output JPEG image layers to be produced by the encoder.
Range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
SpriteColumn int
Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
Step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
StretchMode string | Pulumi.AzureNative.Media.StretchMode
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
SyncMode string | Pulumi.AzureNative.Media.VideoSyncMode
The Video Sync Mode
Start This property is required. string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
KeyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
Layers []JpgLayer
A collection of output JPEG image layers to be produced by the encoder.
Range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
SpriteColumn int
Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
Step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
StretchMode string | StretchMode
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
SyncMode string | VideoSyncMode
The Video Sync Mode
start This property is required. String
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
keyFrameInterval String
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label String
An optional label for the codec. The label can be used to control muxing behavior.
layers List<JpgLayer>
A collection of output JPEG image layers to be produced by the encoder.
range String
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
spriteColumn Integer
Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
step String
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
stretchMode String | StretchMode
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode String | VideoSyncMode
The Video Sync Mode
start This property is required. string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
keyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label string
An optional label for the codec. The label can be used to control muxing behavior.
layers JpgLayer[]
A collection of output JPEG image layers to be produced by the encoder.
range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
spriteColumn number
Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
stretchMode string | StretchMode
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode string | VideoSyncMode
The Video Sync Mode
start This property is required. str
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
key_frame_interval str
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label str
An optional label for the codec. The label can be used to control muxing behavior.
layers Sequence[JpgLayer]
A collection of output JPEG image layers to be produced by the encoder.
range str
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
sprite_column int
Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
step str
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
stretch_mode str | StretchMode
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
sync_mode str | VideoSyncMode
The Video Sync Mode
start This property is required. String
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
keyFrameInterval String
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label String
An optional label for the codec. The label can be used to control muxing behavior.
layers List<Property Map>
A collection of output JPEG image layers to be produced by the encoder.
range String
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
spriteColumn Number
Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
step String
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
stretchMode String | "None" | "AutoSize" | "AutoFit"
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode String | "Auto" | "Passthrough" | "Cfr" | "Vfr"
The Video Sync Mode

JpgImageResponse
, JpgImageResponseArgs

Start This property is required. string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
KeyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
Layers List<Pulumi.AzureNative.Media.Inputs.JpgLayerResponse>
A collection of output JPEG image layers to be produced by the encoder.
Range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
SpriteColumn int
Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
Step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
StretchMode string
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
SyncMode string
The Video Sync Mode
Start This property is required. string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
KeyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
Layers []JpgLayerResponse
A collection of output JPEG image layers to be produced by the encoder.
Range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
SpriteColumn int
Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
Step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
StretchMode string
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
SyncMode string
The Video Sync Mode
start This property is required. String
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
keyFrameInterval String
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label String
An optional label for the codec. The label can be used to control muxing behavior.
layers List<JpgLayerResponse>
A collection of output JPEG image layers to be produced by the encoder.
range String
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
spriteColumn Integer
Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
step String
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
stretchMode String
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode String
The Video Sync Mode
start This property is required. string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
keyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label string
An optional label for the codec. The label can be used to control muxing behavior.
layers JpgLayerResponse[]
A collection of output JPEG image layers to be produced by the encoder.
range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
spriteColumn number
Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
stretchMode string
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode string
The Video Sync Mode
start This property is required. str
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
key_frame_interval str
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label str
An optional label for the codec. The label can be used to control muxing behavior.
layers Sequence[JpgLayerResponse]
A collection of output JPEG image layers to be produced by the encoder.
range str
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
sprite_column int
Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
step str
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
stretch_mode str
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
sync_mode str
The Video Sync Mode
start This property is required. String
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
keyFrameInterval String
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label String
An optional label for the codec. The label can be used to control muxing behavior.
layers List<Property Map>
A collection of output JPEG image layers to be produced by the encoder.
range String
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
spriteColumn Number
Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
step String
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
stretchMode String
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode String
The Video Sync Mode

JpgLayer
, JpgLayerArgs

Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
Quality int
The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
Quality int
The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
quality Integer
The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
quality number
The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
height str
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label str
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
quality int
The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
width str
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
quality Number
The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

JpgLayerResponse
, JpgLayerResponseArgs

Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
Quality int
The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
Quality int
The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
quality Integer
The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
quality number
The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
height str
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label str
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
quality int
The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
width str
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
quality Number
The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

Mp4Format
, Mp4FormatArgs

FilenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
OutputFiles List<Pulumi.AzureNative.Media.Inputs.OutputFile>
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
FilenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
OutputFiles []OutputFile
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
filenamePattern This property is required. String
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
outputFiles List<OutputFile>
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
filenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
outputFiles OutputFile[]
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
filename_pattern This property is required. str
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
output_files Sequence[OutputFile]
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
filenamePattern This property is required. String
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
outputFiles List<Property Map>
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

Mp4FormatResponse
, Mp4FormatResponseArgs

FilenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
OutputFiles List<Pulumi.AzureNative.Media.Inputs.OutputFileResponse>
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
FilenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
OutputFiles []OutputFileResponse
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
filenamePattern This property is required. String
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
outputFiles List<OutputFileResponse>
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
filenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
outputFiles OutputFileResponse[]
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
filename_pattern This property is required. str
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
output_files Sequence[OutputFileResponse]
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
filenamePattern This property is required. String
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
outputFiles List<Property Map>
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

MultiBitrateFormat
, MultiBitrateFormatArgs

FilenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
OutputFiles List<Pulumi.AzureNative.Media.Inputs.OutputFile>
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
FilenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
OutputFiles []OutputFile
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
filenamePattern This property is required. String
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
outputFiles List<OutputFile>
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
filenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
outputFiles OutputFile[]
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
filename_pattern This property is required. str
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
output_files Sequence[OutputFile]
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
filenamePattern This property is required. String
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
outputFiles List<Property Map>
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

MultiBitrateFormatResponse
, MultiBitrateFormatResponseArgs

FilenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
OutputFiles List<Pulumi.AzureNative.Media.Inputs.OutputFileResponse>
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
FilenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
OutputFiles []OutputFileResponse
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
filenamePattern This property is required. String
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
outputFiles List<OutputFileResponse>
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
filenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
outputFiles OutputFileResponse[]
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
filename_pattern This property is required. str
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
output_files Sequence[OutputFileResponse]
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
filenamePattern This property is required. String
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
outputFiles List<Property Map>
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

OnErrorType
, OnErrorTypeArgs

StopProcessingJob
StopProcessingJobTells the service that if this TransformOutput fails, then any other incomplete TransformOutputs can be stopped.
ContinueJob
ContinueJobTells the service that if this TransformOutput fails, then allow any other TransformOutput to continue.
OnErrorTypeStopProcessingJob
StopProcessingJobTells the service that if this TransformOutput fails, then any other incomplete TransformOutputs can be stopped.
OnErrorTypeContinueJob
ContinueJobTells the service that if this TransformOutput fails, then allow any other TransformOutput to continue.
StopProcessingJob
StopProcessingJobTells the service that if this TransformOutput fails, then any other incomplete TransformOutputs can be stopped.
ContinueJob
ContinueJobTells the service that if this TransformOutput fails, then allow any other TransformOutput to continue.
StopProcessingJob
StopProcessingJobTells the service that if this TransformOutput fails, then any other incomplete TransformOutputs can be stopped.
ContinueJob
ContinueJobTells the service that if this TransformOutput fails, then allow any other TransformOutput to continue.
STOP_PROCESSING_JOB
StopProcessingJobTells the service that if this TransformOutput fails, then any other incomplete TransformOutputs can be stopped.
CONTINUE_JOB
ContinueJobTells the service that if this TransformOutput fails, then allow any other TransformOutput to continue.
"StopProcessingJob"
StopProcessingJobTells the service that if this TransformOutput fails, then any other incomplete TransformOutputs can be stopped.
"ContinueJob"
ContinueJobTells the service that if this TransformOutput fails, then allow any other TransformOutput to continue.

OutputFile
, OutputFileArgs

Labels This property is required. List<string>
The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
Labels This property is required. []string
The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
labels This property is required. List<String>
The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
labels This property is required. string[]
The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
labels This property is required. Sequence[str]
The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
labels This property is required. List<String>
The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.

OutputFileResponse
, OutputFileResponseArgs

Labels This property is required. List<string>
The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
Labels This property is required. []string
The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
labels This property is required. List<String>
The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
labels This property is required. string[]
The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
labels This property is required. Sequence[str]
The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
labels This property is required. List<String>
The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.

PngFormat
, PngFormatArgs

FilenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
FilenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
filenamePattern This property is required. String
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
filenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
filename_pattern This property is required. str
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
filenamePattern This property is required. String
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

PngFormatResponse
, PngFormatResponseArgs

FilenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
FilenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
filenamePattern This property is required. String
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
filenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
filename_pattern This property is required. str
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
filenamePattern This property is required. String
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

PngImage
, PngImageArgs

Start This property is required. string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
KeyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
Layers List<Pulumi.AzureNative.Media.Inputs.PngLayer>
A collection of output PNG image layers to be produced by the encoder.
Range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
Step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
StretchMode string | Pulumi.AzureNative.Media.StretchMode
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
SyncMode string | Pulumi.AzureNative.Media.VideoSyncMode
The Video Sync Mode
Start This property is required. string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
KeyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
Layers []PngLayer
A collection of output PNG image layers to be produced by the encoder.
Range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
Step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
StretchMode string | StretchMode
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
SyncMode string | VideoSyncMode
The Video Sync Mode
start This property is required. String
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
keyFrameInterval String
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label String
An optional label for the codec. The label can be used to control muxing behavior.
layers List<PngLayer>
A collection of output PNG image layers to be produced by the encoder.
range String
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
step String
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
stretchMode String | StretchMode
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode String | VideoSyncMode
The Video Sync Mode
start This property is required. string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
keyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label string
An optional label for the codec. The label can be used to control muxing behavior.
layers PngLayer[]
A collection of output PNG image layers to be produced by the encoder.
range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
stretchMode string | StretchMode
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode string | VideoSyncMode
The Video Sync Mode
start This property is required. str
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
key_frame_interval str
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label str
An optional label for the codec. The label can be used to control muxing behavior.
layers Sequence[PngLayer]
A collection of output PNG image layers to be produced by the encoder.
range str
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
step str
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
stretch_mode str | StretchMode
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
sync_mode str | VideoSyncMode
The Video Sync Mode
start This property is required. String
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
keyFrameInterval String
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label String
An optional label for the codec. The label can be used to control muxing behavior.
layers List<Property Map>
A collection of output PNG image layers to be produced by the encoder.
range String
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
step String
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
stretchMode String | "None" | "AutoSize" | "AutoFit"
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode String | "Auto" | "Passthrough" | "Cfr" | "Vfr"
The Video Sync Mode

PngImageResponse
, PngImageResponseArgs

Start This property is required. string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
KeyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
Layers List<Pulumi.AzureNative.Media.Inputs.PngLayerResponse>
A collection of output PNG image layers to be produced by the encoder.
Range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
Step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
StretchMode string
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
SyncMode string
The Video Sync Mode
Start This property is required. string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
KeyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
Layers []PngLayerResponse
A collection of output PNG image layers to be produced by the encoder.
Range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
Step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
StretchMode string
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
SyncMode string
The Video Sync Mode
start This property is required. String
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
keyFrameInterval String
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label String
An optional label for the codec. The label can be used to control muxing behavior.
layers List<PngLayerResponse>
A collection of output PNG image layers to be produced by the encoder.
range String
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
step String
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
stretchMode String
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode String
The Video Sync Mode
start This property is required. string
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
keyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label string
An optional label for the codec. The label can be used to control muxing behavior.
layers PngLayerResponse[]
A collection of output PNG image layers to be produced by the encoder.
range string
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
step string
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
stretchMode string
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode string
The Video Sync Mode
start This property is required. str
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
key_frame_interval str
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label str
An optional label for the codec. The label can be used to control muxing behavior.
layers Sequence[PngLayerResponse]
A collection of output PNG image layers to be produced by the encoder.
range str
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
step str
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
stretch_mode str
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
sync_mode str
The Video Sync Mode
start This property is required. String
The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
keyFrameInterval String
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label String
An optional label for the codec. The label can be used to control muxing behavior.
layers List<Property Map>
A collection of output PNG image layers to be produced by the encoder.
range String
The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
step String
The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
stretchMode String
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode String
The Video Sync Mode

PngLayer
, PngLayerArgs

Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
height str
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label str
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
width str
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

PngLayerResponse
, PngLayerResponseArgs

Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
Height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
Label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
Width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
height string
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label string
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
width string
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
height str
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label str
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
width str
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
height String
The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
label String
The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
width String
The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

PresetConfigurations
, PresetConfigurationsArgs

Complexity string | Pulumi.AzureNative.Media.Complexity
Allows you to configure the encoder settings to control the balance between speed and quality. Example: set Complexity as Speed for faster encoding but less compression efficiency.
InterleaveOutput string | Pulumi.AzureNative.Media.InterleaveOutput
Sets the interleave mode of the output to control how audio and video are stored in the container format. Example: set InterleavedOutput as NonInterleavedOutput to produce audio-only and video-only outputs in separate MP4 files.
KeyFrameIntervalInSeconds double
The key frame interval in seconds. Example: set KeyFrameIntervalInSeconds as 2 to reduce the playback buffering for some players.
MaxBitrateBps int
The maximum bitrate in bits per second (threshold for the top video layer). Example: set MaxBitrateBps as 6000000 to avoid producing very high bitrate outputs for contents with high complexity.
MaxHeight int
The maximum height of output video layers. Example: set MaxHeight as 720 to produce output layers up to 720P even if the input is 4K.
MaxLayers int
The maximum number of output video layers. Example: set MaxLayers as 4 to make sure at most 4 output layers are produced to control the overall cost of the encoding job.
MinBitrateBps int
The minimum bitrate in bits per second (threshold for the bottom video layer). Example: set MinBitrateBps as 200000 to have a bottom layer that covers users with low network bandwidth.
MinHeight int
The minimum height of output video layers. Example: set MinHeight as 360 to avoid output layers of smaller resolutions like 180P.
Complexity string | Complexity
Allows you to configure the encoder settings to control the balance between speed and quality. Example: set Complexity as Speed for faster encoding but less compression efficiency.
InterleaveOutput string | InterleaveOutput
Sets the interleave mode of the output to control how audio and video are stored in the container format. Example: set InterleavedOutput as NonInterleavedOutput to produce audio-only and video-only outputs in separate MP4 files.
KeyFrameIntervalInSeconds float64
The key frame interval in seconds. Example: set KeyFrameIntervalInSeconds as 2 to reduce the playback buffering for some players.
MaxBitrateBps int
The maximum bitrate in bits per second (threshold for the top video layer). Example: set MaxBitrateBps as 6000000 to avoid producing very high bitrate outputs for contents with high complexity.
MaxHeight int
The maximum height of output video layers. Example: set MaxHeight as 720 to produce output layers up to 720P even if the input is 4K.
MaxLayers int
The maximum number of output video layers. Example: set MaxLayers as 4 to make sure at most 4 output layers are produced to control the overall cost of the encoding job.
MinBitrateBps int
The minimum bitrate in bits per second (threshold for the bottom video layer). Example: set MinBitrateBps as 200000 to have a bottom layer that covers users with low network bandwidth.
MinHeight int
The minimum height of output video layers. Example: set MinHeight as 360 to avoid output layers of smaller resolutions like 180P.
complexity String | Complexity
Allows you to configure the encoder settings to control the balance between speed and quality. Example: set Complexity as Speed for faster encoding but less compression efficiency.
interleaveOutput String | InterleaveOutput
Sets the interleave mode of the output to control how audio and video are stored in the container format. Example: set InterleavedOutput as NonInterleavedOutput to produce audio-only and video-only outputs in separate MP4 files.
keyFrameIntervalInSeconds Double
The key frame interval in seconds. Example: set KeyFrameIntervalInSeconds as 2 to reduce the playback buffering for some players.
maxBitrateBps Integer
The maximum bitrate in bits per second (threshold for the top video layer). Example: set MaxBitrateBps as 6000000 to avoid producing very high bitrate outputs for contents with high complexity.
maxHeight Integer
The maximum height of output video layers. Example: set MaxHeight as 720 to produce output layers up to 720P even if the input is 4K.
maxLayers Integer
The maximum number of output video layers. Example: set MaxLayers as 4 to make sure at most 4 output layers are produced to control the overall cost of the encoding job.
minBitrateBps Integer
The minimum bitrate in bits per second (threshold for the bottom video layer). Example: set MinBitrateBps as 200000 to have a bottom layer that covers users with low network bandwidth.
minHeight Integer
The minimum height of output video layers. Example: set MinHeight as 360 to avoid output layers of smaller resolutions like 180P.
complexity string | Complexity
Allows you to configure the encoder settings to control the balance between speed and quality. Example: set Complexity as Speed for faster encoding but less compression efficiency.
interleaveOutput string | InterleaveOutput
Sets the interleave mode of the output to control how audio and video are stored in the container format. Example: set InterleavedOutput as NonInterleavedOutput to produce audio-only and video-only outputs in separate MP4 files.
keyFrameIntervalInSeconds number
The key frame interval in seconds. Example: set KeyFrameIntervalInSeconds as 2 to reduce the playback buffering for some players.
maxBitrateBps number
The maximum bitrate in bits per second (threshold for the top video layer). Example: set MaxBitrateBps as 6000000 to avoid producing very high bitrate outputs for contents with high complexity.
maxHeight number
The maximum height of output video layers. Example: set MaxHeight as 720 to produce output layers up to 720P even if the input is 4K.
maxLayers number
The maximum number of output video layers. Example: set MaxLayers as 4 to make sure at most 4 output layers are produced to control the overall cost of the encoding job.
minBitrateBps number
The minimum bitrate in bits per second (threshold for the bottom video layer). Example: set MinBitrateBps as 200000 to have a bottom layer that covers users with low network bandwidth.
minHeight number
The minimum height of output video layers. Example: set MinHeight as 360 to avoid output layers of smaller resolutions like 180P.
complexity str | Complexity
Allows you to configure the encoder settings to control the balance between speed and quality. Example: set Complexity as Speed for faster encoding but less compression efficiency.
interleave_output str | InterleaveOutput
Sets the interleave mode of the output to control how audio and video are stored in the container format. Example: set InterleavedOutput as NonInterleavedOutput to produce audio-only and video-only outputs in separate MP4 files.
key_frame_interval_in_seconds float
The key frame interval in seconds. Example: set KeyFrameIntervalInSeconds as 2 to reduce the playback buffering for some players.
max_bitrate_bps int
The maximum bitrate in bits per second (threshold for the top video layer). Example: set MaxBitrateBps as 6000000 to avoid producing very high bitrate outputs for contents with high complexity.
max_height int
The maximum height of output video layers. Example: set MaxHeight as 720 to produce output layers up to 720P even if the input is 4K.
max_layers int
The maximum number of output video layers. Example: set MaxLayers as 4 to make sure at most 4 output layers are produced to control the overall cost of the encoding job.
min_bitrate_bps int
The minimum bitrate in bits per second (threshold for the bottom video layer). Example: set MinBitrateBps as 200000 to have a bottom layer that covers users with low network bandwidth.
min_height int
The minimum height of output video layers. Example: set MinHeight as 360 to avoid output layers of smaller resolutions like 180P.
complexity String | "Speed" | "Balanced" | "Quality"
Allows you to configure the encoder settings to control the balance between speed and quality. Example: set Complexity as Speed for faster encoding but less compression efficiency.
interleaveOutput String | "NonInterleavedOutput" | "InterleavedOutput"
Sets the interleave mode of the output to control how audio and video are stored in the container format. Example: set InterleavedOutput as NonInterleavedOutput to produce audio-only and video-only outputs in separate MP4 files.
keyFrameIntervalInSeconds Number
The key frame interval in seconds. Example: set KeyFrameIntervalInSeconds as 2 to reduce the playback buffering for some players.
maxBitrateBps Number
The maximum bitrate in bits per second (threshold for the top video layer). Example: set MaxBitrateBps as 6000000 to avoid producing very high bitrate outputs for contents with high complexity.
maxHeight Number
The maximum height of output video layers. Example: set MaxHeight as 720 to produce output layers up to 720P even if the input is 4K.
maxLayers Number
The maximum number of output video layers. Example: set MaxLayers as 4 to make sure at most 4 output layers are produced to control the overall cost of the encoding job.
minBitrateBps Number
The minimum bitrate in bits per second (threshold for the bottom video layer). Example: set MinBitrateBps as 200000 to have a bottom layer that covers users with low network bandwidth.
minHeight Number
The minimum height of output video layers. Example: set MinHeight as 360 to avoid output layers of smaller resolutions like 180P.

PresetConfigurationsResponse
, PresetConfigurationsResponseArgs

Complexity string
Allows you to configure the encoder settings to control the balance between speed and quality. Example: set Complexity as Speed for faster encoding but less compression efficiency.
InterleaveOutput string
Sets the interleave mode of the output to control how audio and video are stored in the container format. Example: set InterleavedOutput as NonInterleavedOutput to produce audio-only and video-only outputs in separate MP4 files.
KeyFrameIntervalInSeconds double
The key frame interval in seconds. Example: set KeyFrameIntervalInSeconds as 2 to reduce the playback buffering for some players.
MaxBitrateBps int
The maximum bitrate in bits per second (threshold for the top video layer). Example: set MaxBitrateBps as 6000000 to avoid producing very high bitrate outputs for contents with high complexity.
MaxHeight int
The maximum height of output video layers. Example: set MaxHeight as 720 to produce output layers up to 720P even if the input is 4K.
MaxLayers int
The maximum number of output video layers. Example: set MaxLayers as 4 to make sure at most 4 output layers are produced to control the overall cost of the encoding job.
MinBitrateBps int
The minimum bitrate in bits per second (threshold for the bottom video layer). Example: set MinBitrateBps as 200000 to have a bottom layer that covers users with low network bandwidth.
MinHeight int
The minimum height of output video layers. Example: set MinHeight as 360 to avoid output layers of smaller resolutions like 180P.
Complexity string
Allows you to configure the encoder settings to control the balance between speed and quality. Example: set Complexity as Speed for faster encoding but less compression efficiency.
InterleaveOutput string
Sets the interleave mode of the output to control how audio and video are stored in the container format. Example: set InterleavedOutput as NonInterleavedOutput to produce audio-only and video-only outputs in separate MP4 files.
KeyFrameIntervalInSeconds float64
The key frame interval in seconds. Example: set KeyFrameIntervalInSeconds as 2 to reduce the playback buffering for some players.
MaxBitrateBps int
The maximum bitrate in bits per second (threshold for the top video layer). Example: set MaxBitrateBps as 6000000 to avoid producing very high bitrate outputs for contents with high complexity.
MaxHeight int
The maximum height of output video layers. Example: set MaxHeight as 720 to produce output layers up to 720P even if the input is 4K.
MaxLayers int
The maximum number of output video layers. Example: set MaxLayers as 4 to make sure at most 4 output layers are produced to control the overall cost of the encoding job.
MinBitrateBps int
The minimum bitrate in bits per second (threshold for the bottom video layer). Example: set MinBitrateBps as 200000 to have a bottom layer that covers users with low network bandwidth.
MinHeight int
The minimum height of output video layers. Example: set MinHeight as 360 to avoid output layers of smaller resolutions like 180P.
complexity String
Allows you to configure the encoder settings to control the balance between speed and quality. Example: set Complexity as Speed for faster encoding but less compression efficiency.
interleaveOutput String
Sets the interleave mode of the output to control how audio and video are stored in the container format. Example: set InterleavedOutput as NonInterleavedOutput to produce audio-only and video-only outputs in separate MP4 files.
keyFrameIntervalInSeconds Double
The key frame interval in seconds. Example: set KeyFrameIntervalInSeconds as 2 to reduce the playback buffering for some players.
maxBitrateBps Integer
The maximum bitrate in bits per second (threshold for the top video layer). Example: set MaxBitrateBps as 6000000 to avoid producing very high bitrate outputs for contents with high complexity.
maxHeight Integer
The maximum height of output video layers. Example: set MaxHeight as 720 to produce output layers up to 720P even if the input is 4K.
maxLayers Integer
The maximum number of output video layers. Example: set MaxLayers as 4 to make sure at most 4 output layers are produced to control the overall cost of the encoding job.
minBitrateBps Integer
The minimum bitrate in bits per second (threshold for the bottom video layer). Example: set MinBitrateBps as 200000 to have a bottom layer that covers users with low network bandwidth.
minHeight Integer
The minimum height of output video layers. Example: set MinHeight as 360 to avoid output layers of smaller resolutions like 180P.
complexity string
Allows you to configure the encoder settings to control the balance between speed and quality. Example: set Complexity as Speed for faster encoding but less compression efficiency.
interleaveOutput string
Sets the interleave mode of the output to control how audio and video are stored in the container format. Example: set InterleavedOutput as NonInterleavedOutput to produce audio-only and video-only outputs in separate MP4 files.
keyFrameIntervalInSeconds number
The key frame interval in seconds. Example: set KeyFrameIntervalInSeconds as 2 to reduce the playback buffering for some players.
maxBitrateBps number
The maximum bitrate in bits per second (threshold for the top video layer). Example: set MaxBitrateBps as 6000000 to avoid producing very high bitrate outputs for contents with high complexity.
maxHeight number
The maximum height of output video layers. Example: set MaxHeight as 720 to produce output layers up to 720P even if the input is 4K.
maxLayers number
The maximum number of output video layers. Example: set MaxLayers as 4 to make sure at most 4 output layers are produced to control the overall cost of the encoding job.
minBitrateBps number
The minimum bitrate in bits per second (threshold for the bottom video layer). Example: set MinBitrateBps as 200000 to have a bottom layer that covers users with low network bandwidth.
minHeight number
The minimum height of output video layers. Example: set MinHeight as 360 to avoid output layers of smaller resolutions like 180P.
complexity str
Allows you to configure the encoder settings to control the balance between speed and quality. Example: set Complexity as Speed for faster encoding but less compression efficiency.
interleave_output str
Sets the interleave mode of the output to control how audio and video are stored in the container format. Example: set InterleavedOutput as NonInterleavedOutput to produce audio-only and video-only outputs in separate MP4 files.
key_frame_interval_in_seconds float
The key frame interval in seconds. Example: set KeyFrameIntervalInSeconds as 2 to reduce the playback buffering for some players.
max_bitrate_bps int
The maximum bitrate in bits per second (threshold for the top video layer). Example: set MaxBitrateBps as 6000000 to avoid producing very high bitrate outputs for contents with high complexity.
max_height int
The maximum height of output video layers. Example: set MaxHeight as 720 to produce output layers up to 720P even if the input is 4K.
max_layers int
The maximum number of output video layers. Example: set MaxLayers as 4 to make sure at most 4 output layers are produced to control the overall cost of the encoding job.
min_bitrate_bps int
The minimum bitrate in bits per second (threshold for the bottom video layer). Example: set MinBitrateBps as 200000 to have a bottom layer that covers users with low network bandwidth.
min_height int
The minimum height of output video layers. Example: set MinHeight as 360 to avoid output layers of smaller resolutions like 180P.
complexity String
Allows you to configure the encoder settings to control the balance between speed and quality. Example: set Complexity as Speed for faster encoding but less compression efficiency.
interleaveOutput String
Sets the interleave mode of the output to control how audio and video are stored in the container format. Example: set InterleavedOutput as NonInterleavedOutput to produce audio-only and video-only outputs in separate MP4 files.
keyFrameIntervalInSeconds Number
The key frame interval in seconds. Example: set KeyFrameIntervalInSeconds as 2 to reduce the playback buffering for some players.
maxBitrateBps Number
The maximum bitrate in bits per second (threshold for the top video layer). Example: set MaxBitrateBps as 6000000 to avoid producing very high bitrate outputs for contents with high complexity.
maxHeight Number
The maximum height of output video layers. Example: set MaxHeight as 720 to produce output layers up to 720P even if the input is 4K.
maxLayers Number
The maximum number of output video layers. Example: set MaxLayers as 4 to make sure at most 4 output layers are produced to control the overall cost of the encoding job.
minBitrateBps Number
The minimum bitrate in bits per second (threshold for the bottom video layer). Example: set MinBitrateBps as 200000 to have a bottom layer that covers users with low network bandwidth.
minHeight Number
The minimum height of output video layers. Example: set MinHeight as 360 to avoid output layers of smaller resolutions like 180P.

Priority
, PriorityArgs

Low
LowUsed for TransformOutputs that can be generated after Normal and High priority TransformOutputs.
Normal
NormalUsed for TransformOutputs that can be generated at Normal priority.
High
HighUsed for TransformOutputs that should take precedence over others.
PriorityLow
LowUsed for TransformOutputs that can be generated after Normal and High priority TransformOutputs.
PriorityNormal
NormalUsed for TransformOutputs that can be generated at Normal priority.
PriorityHigh
HighUsed for TransformOutputs that should take precedence over others.
Low
LowUsed for TransformOutputs that can be generated after Normal and High priority TransformOutputs.
Normal
NormalUsed for TransformOutputs that can be generated at Normal priority.
High
HighUsed for TransformOutputs that should take precedence over others.
Low
LowUsed for TransformOutputs that can be generated after Normal and High priority TransformOutputs.
Normal
NormalUsed for TransformOutputs that can be generated at Normal priority.
High
HighUsed for TransformOutputs that should take precedence over others.
LOW
LowUsed for TransformOutputs that can be generated after Normal and High priority TransformOutputs.
NORMAL
NormalUsed for TransformOutputs that can be generated at Normal priority.
HIGH
HighUsed for TransformOutputs that should take precedence over others.
"Low"
LowUsed for TransformOutputs that can be generated after Normal and High priority TransformOutputs.
"Normal"
NormalUsed for TransformOutputs that can be generated at Normal priority.
"High"
HighUsed for TransformOutputs that should take precedence over others.

Rectangle
, RectangleArgs

Height string
The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
Left string
The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
Top string
The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
Width string
The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
Height string
The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
Left string
The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
Top string
The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
Width string
The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
height String
The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
left String
The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
top String
The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
width String
The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
height string
The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
left string
The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
top string
The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
width string
The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
height str
The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
left str
The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
top str
The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
width str
The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
height String
The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
left String
The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
top String
The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
width String
The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

RectangleResponse
, RectangleResponseArgs

Height string
The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
Left string
The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
Top string
The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
Width string
The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
Height string
The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
Left string
The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
Top string
The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
Width string
The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
height String
The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
left String
The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
top String
The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
width String
The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
height string
The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
left string
The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
top string
The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
width string
The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
height str
The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
left str
The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
top str
The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
width str
The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
height String
The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
left String
The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
top String
The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
width String
The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

Rotation
, RotationArgs

Auto
AutoAutomatically detect and rotate as needed.
None
NoneDo not rotate the video. If the output format supports it, any metadata about rotation is kept intact.
Rotate0
Rotate0Do not rotate the video but remove any metadata about the rotation.
Rotate90
Rotate90Rotate 90 degrees clockwise.
Rotate180
Rotate180Rotate 180 degrees clockwise.
Rotate270
Rotate270Rotate 270 degrees clockwise.
RotationAuto
AutoAutomatically detect and rotate as needed.
RotationNone
NoneDo not rotate the video. If the output format supports it, any metadata about rotation is kept intact.
RotationRotate0
Rotate0Do not rotate the video but remove any metadata about the rotation.
RotationRotate90
Rotate90Rotate 90 degrees clockwise.
RotationRotate180
Rotate180Rotate 180 degrees clockwise.
RotationRotate270
Rotate270Rotate 270 degrees clockwise.
Auto
AutoAutomatically detect and rotate as needed.
None
NoneDo not rotate the video. If the output format supports it, any metadata about rotation is kept intact.
Rotate0
Rotate0Do not rotate the video but remove any metadata about the rotation.
Rotate90
Rotate90Rotate 90 degrees clockwise.
Rotate180
Rotate180Rotate 180 degrees clockwise.
Rotate270
Rotate270Rotate 270 degrees clockwise.
Auto
AutoAutomatically detect and rotate as needed.
None
NoneDo not rotate the video. If the output format supports it, any metadata about rotation is kept intact.
Rotate0
Rotate0Do not rotate the video but remove any metadata about the rotation.
Rotate90
Rotate90Rotate 90 degrees clockwise.
Rotate180
Rotate180Rotate 180 degrees clockwise.
Rotate270
Rotate270Rotate 270 degrees clockwise.
AUTO
AutoAutomatically detect and rotate as needed.
NONE
NoneDo not rotate the video. If the output format supports it, any metadata about rotation is kept intact.
ROTATE0
Rotate0Do not rotate the video but remove any metadata about the rotation.
ROTATE90
Rotate90Rotate 90 degrees clockwise.
ROTATE180
Rotate180Rotate 180 degrees clockwise.
ROTATE270
Rotate270Rotate 270 degrees clockwise.
"Auto"
AutoAutomatically detect and rotate as needed.
"None"
NoneDo not rotate the video. If the output format supports it, any metadata about rotation is kept intact.
"Rotate0"
Rotate0Do not rotate the video but remove any metadata about the rotation.
"Rotate90"
Rotate90Rotate 90 degrees clockwise.
"Rotate180"
Rotate180Rotate 180 degrees clockwise.
"Rotate270"
Rotate270Rotate 270 degrees clockwise.

StandardEncoderPreset
, StandardEncoderPresetArgs

Codecs This property is required. List<object>
The list of codecs to be used when encoding the input video.
Formats This property is required. List<object>
The list of outputs to be produced by the encoder.
ExperimentalOptions Dictionary<string, string>
Dictionary containing key value pairs for parameters not exposed in the preset itself
Filters Pulumi.AzureNative.Media.Inputs.Filters
One or more filtering operations that are applied to the input media before encoding.
Codecs This property is required. []interface{}
The list of codecs to be used when encoding the input video.
Formats This property is required. []interface{}
The list of outputs to be produced by the encoder.
ExperimentalOptions map[string]string
Dictionary containing key value pairs for parameters not exposed in the preset itself
Filters Filters
One or more filtering operations that are applied to the input media before encoding.
codecs This property is required. List<Object>
The list of codecs to be used when encoding the input video.
formats This property is required. List<Object>
The list of outputs to be produced by the encoder.
experimentalOptions Map<String,String>
Dictionary containing key value pairs for parameters not exposed in the preset itself
filters Filters
One or more filtering operations that are applied to the input media before encoding.
codecs This property is required. (AacAudio | Audio | CopyAudio | CopyVideo | DDAudio | H264Video | H265Video | Image | JpgImage | PngImage | Video)[]
The list of codecs to be used when encoding the input video.
formats This property is required. (ImageFormat | JpgFormat | Mp4Format | MultiBitrateFormat | PngFormat | TransportStreamFormat)[]
The list of outputs to be produced by the encoder.
experimentalOptions {[key: string]: string}
Dictionary containing key value pairs for parameters not exposed in the preset itself
filters Filters
One or more filtering operations that are applied to the input media before encoding.
codecs This property is required. Sequence[Union[AacAudio, Audio, CopyAudio, CopyVideo, DDAudio, H264Video, H265Video, Image, JpgImage, PngImage, Video]]
The list of codecs to be used when encoding the input video.
formats This property is required. Sequence[Union[ImageFormat, JpgFormat, Mp4Format, MultiBitrateFormat, PngFormat, TransportStreamFormat]]
The list of outputs to be produced by the encoder.
experimental_options Mapping[str, str]
Dictionary containing key value pairs for parameters not exposed in the preset itself
filters Filters
One or more filtering operations that are applied to the input media before encoding.
codecs This property is required. List<Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map>
The list of codecs to be used when encoding the input video.
formats This property is required. List<Property Map | Property Map | Property Map | Property Map | Property Map | Property Map>
The list of outputs to be produced by the encoder.
experimentalOptions Map<String>
Dictionary containing key value pairs for parameters not exposed in the preset itself
filters Property Map
One or more filtering operations that are applied to the input media before encoding.

StandardEncoderPresetResponse
, StandardEncoderPresetResponseArgs

Codecs This property is required. List<object>
The list of codecs to be used when encoding the input video.
Formats This property is required. List<object>
The list of outputs to be produced by the encoder.
ExperimentalOptions Dictionary<string, string>
Dictionary containing key value pairs for parameters not exposed in the preset itself
Filters Pulumi.AzureNative.Media.Inputs.FiltersResponse
One or more filtering operations that are applied to the input media before encoding.
Codecs This property is required. []interface{}
The list of codecs to be used when encoding the input video.
Formats This property is required. []interface{}
The list of outputs to be produced by the encoder.
ExperimentalOptions map[string]string
Dictionary containing key value pairs for parameters not exposed in the preset itself
Filters FiltersResponse
One or more filtering operations that are applied to the input media before encoding.
codecs This property is required. List<Object>
The list of codecs to be used when encoding the input video.
formats This property is required. List<Object>
The list of outputs to be produced by the encoder.
experimentalOptions Map<String,String>
Dictionary containing key value pairs for parameters not exposed in the preset itself
filters FiltersResponse
One or more filtering operations that are applied to the input media before encoding.
codecs This property is required. (AacAudioResponse | AudioResponse | CopyAudioResponse | CopyVideoResponse | DDAudioResponse | H264VideoResponse | H265VideoResponse | ImageResponse | JpgImageResponse | PngImageResponse | VideoResponse)[]
The list of codecs to be used when encoding the input video.
formats This property is required. (ImageFormatResponse | JpgFormatResponse | Mp4FormatResponse | MultiBitrateFormatResponse | PngFormatResponse | TransportStreamFormatResponse)[]
The list of outputs to be produced by the encoder.
experimentalOptions {[key: string]: string}
Dictionary containing key value pairs for parameters not exposed in the preset itself
filters FiltersResponse
One or more filtering operations that are applied to the input media before encoding.
codecs This property is required. Sequence[Union[AacAudioResponse, AudioResponse, CopyAudioResponse, CopyVideoResponse, DDAudioResponse, H264VideoResponse, H265VideoResponse, ImageResponse, JpgImageResponse, PngImageResponse, VideoResponse]]
The list of codecs to be used when encoding the input video.
formats This property is required. Sequence[Union[ImageFormatResponse, JpgFormatResponse, Mp4FormatResponse, MultiBitrateFormatResponse, PngFormatResponse, TransportStreamFormatResponse]]
The list of outputs to be produced by the encoder.
experimental_options Mapping[str, str]
Dictionary containing key value pairs for parameters not exposed in the preset itself
filters FiltersResponse
One or more filtering operations that are applied to the input media before encoding.
codecs This property is required. List<Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map>
The list of codecs to be used when encoding the input video.
formats This property is required. List<Property Map | Property Map | Property Map | Property Map | Property Map | Property Map>
The list of outputs to be produced by the encoder.
experimentalOptions Map<String>
Dictionary containing key value pairs for parameters not exposed in the preset itself
filters Property Map
One or more filtering operations that are applied to the input media before encoding.

StretchMode
, StretchModeArgs

None
NoneStrictly respect the output resolution without considering the pixel aspect ratio or display aspect ratio of the input video.
AutoSize
AutoSizeOverride the output resolution, and change it to match the display aspect ratio of the input, without padding. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the value in the preset is overridden, and the output will be at 1280x720, which maintains the input aspect ratio of 16:9.
AutoFit
AutoFitPad the output (with either letterbox or pillar box) to honor the output resolution, while ensuring that the active video region in the output has the same aspect ratio as the input. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the output will be at 1280x1280, which contains an inner rectangle of 1280x720 at aspect ratio of 16:9, and pillar box regions 280 pixels wide at the left and right.
StretchModeNone
NoneStrictly respect the output resolution without considering the pixel aspect ratio or display aspect ratio of the input video.
StretchModeAutoSize
AutoSizeOverride the output resolution, and change it to match the display aspect ratio of the input, without padding. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the value in the preset is overridden, and the output will be at 1280x720, which maintains the input aspect ratio of 16:9.
StretchModeAutoFit
AutoFitPad the output (with either letterbox or pillar box) to honor the output resolution, while ensuring that the active video region in the output has the same aspect ratio as the input. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the output will be at 1280x1280, which contains an inner rectangle of 1280x720 at aspect ratio of 16:9, and pillar box regions 280 pixels wide at the left and right.
None
NoneStrictly respect the output resolution without considering the pixel aspect ratio or display aspect ratio of the input video.
AutoSize
AutoSizeOverride the output resolution, and change it to match the display aspect ratio of the input, without padding. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the value in the preset is overridden, and the output will be at 1280x720, which maintains the input aspect ratio of 16:9.
AutoFit
AutoFitPad the output (with either letterbox or pillar box) to honor the output resolution, while ensuring that the active video region in the output has the same aspect ratio as the input. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the output will be at 1280x1280, which contains an inner rectangle of 1280x720 at aspect ratio of 16:9, and pillar box regions 280 pixels wide at the left and right.
None
NoneStrictly respect the output resolution without considering the pixel aspect ratio or display aspect ratio of the input video.
AutoSize
AutoSizeOverride the output resolution, and change it to match the display aspect ratio of the input, without padding. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the value in the preset is overridden, and the output will be at 1280x720, which maintains the input aspect ratio of 16:9.
AutoFit
AutoFitPad the output (with either letterbox or pillar box) to honor the output resolution, while ensuring that the active video region in the output has the same aspect ratio as the input. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the output will be at 1280x1280, which contains an inner rectangle of 1280x720 at aspect ratio of 16:9, and pillar box regions 280 pixels wide at the left and right.
NONE
NoneStrictly respect the output resolution without considering the pixel aspect ratio or display aspect ratio of the input video.
AUTO_SIZE
AutoSizeOverride the output resolution, and change it to match the display aspect ratio of the input, without padding. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the value in the preset is overridden, and the output will be at 1280x720, which maintains the input aspect ratio of 16:9.
AUTO_FIT
AutoFitPad the output (with either letterbox or pillar box) to honor the output resolution, while ensuring that the active video region in the output has the same aspect ratio as the input. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the output will be at 1280x1280, which contains an inner rectangle of 1280x720 at aspect ratio of 16:9, and pillar box regions 280 pixels wide at the left and right.
"None"
NoneStrictly respect the output resolution without considering the pixel aspect ratio or display aspect ratio of the input video.
"AutoSize"
AutoSizeOverride the output resolution, and change it to match the display aspect ratio of the input, without padding. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the value in the preset is overridden, and the output will be at 1280x720, which maintains the input aspect ratio of 16:9.
"AutoFit"
AutoFitPad the output (with either letterbox or pillar box) to honor the output resolution, while ensuring that the active video region in the output has the same aspect ratio as the input. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the output will be at 1280x1280, which contains an inner rectangle of 1280x720 at aspect ratio of 16:9, and pillar box regions 280 pixels wide at the left and right.

SystemDataResponse
, SystemDataResponseArgs

CreatedAt string
The timestamp of resource creation (UTC).
CreatedBy string
The identity that created the resource.
CreatedByType string
The type of identity that created the resource.
LastModifiedAt string
The timestamp of resource last modification (UTC)
LastModifiedBy string
The identity that last modified the resource.
LastModifiedByType string
The type of identity that last modified the resource.
CreatedAt string
The timestamp of resource creation (UTC).
CreatedBy string
The identity that created the resource.
CreatedByType string
The type of identity that created the resource.
LastModifiedAt string
The timestamp of resource last modification (UTC)
LastModifiedBy string
The identity that last modified the resource.
LastModifiedByType string
The type of identity that last modified the resource.
createdAt String
The timestamp of resource creation (UTC).
createdBy String
The identity that created the resource.
createdByType String
The type of identity that created the resource.
lastModifiedAt String
The timestamp of resource last modification (UTC)
lastModifiedBy String
The identity that last modified the resource.
lastModifiedByType String
The type of identity that last modified the resource.
createdAt string
The timestamp of resource creation (UTC).
createdBy string
The identity that created the resource.
createdByType string
The type of identity that created the resource.
lastModifiedAt string
The timestamp of resource last modification (UTC)
lastModifiedBy string
The identity that last modified the resource.
lastModifiedByType string
The type of identity that last modified the resource.
created_at str
The timestamp of resource creation (UTC).
created_by str
The identity that created the resource.
created_by_type str
The type of identity that created the resource.
last_modified_at str
The timestamp of resource last modification (UTC)
last_modified_by str
The identity that last modified the resource.
last_modified_by_type str
The type of identity that last modified the resource.
createdAt String
The timestamp of resource creation (UTC).
createdBy String
The identity that created the resource.
createdByType String
The type of identity that created the resource.
lastModifiedAt String
The timestamp of resource last modification (UTC)
lastModifiedBy String
The identity that last modified the resource.
lastModifiedByType String
The type of identity that last modified the resource.

TransformOutput
, TransformOutputArgs

Preset This property is required. Pulumi.AzureNative.Media.Inputs.AudioAnalyzerPreset | Pulumi.AzureNative.Media.Inputs.BuiltInStandardEncoderPreset | Pulumi.AzureNative.Media.Inputs.FaceDetectorPreset | Pulumi.AzureNative.Media.Inputs.StandardEncoderPreset | Pulumi.AzureNative.Media.Inputs.VideoAnalyzerPreset
Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
OnError string | Pulumi.AzureNative.Media.OnErrorType
A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
RelativePriority string | Pulumi.AzureNative.Media.Priority
Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
Preset This property is required. AudioAnalyzerPreset | BuiltInStandardEncoderPreset | FaceDetectorPreset | StandardEncoderPreset | VideoAnalyzerPreset
Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
OnError string | OnErrorType
A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
RelativePriority string | Priority
Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
preset This property is required. AudioAnalyzerPreset | BuiltInStandardEncoderPreset | FaceDetectorPreset | StandardEncoderPreset | VideoAnalyzerPreset
Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
onError String | OnErrorType
A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
relativePriority String | Priority
Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
preset This property is required. AudioAnalyzerPreset | BuiltInStandardEncoderPreset | FaceDetectorPreset | StandardEncoderPreset | VideoAnalyzerPreset
Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
onError string | OnErrorType
A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
relativePriority string | Priority
Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
preset This property is required. AudioAnalyzerPreset | BuiltInStandardEncoderPreset | FaceDetectorPreset | StandardEncoderPreset | VideoAnalyzerPreset
Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
on_error str | OnErrorType
A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
relative_priority str | Priority
Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
preset This property is required. Property Map | Property Map | Property Map | Property Map | Property Map
Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
onError String | "StopProcessingJob" | "ContinueJob"
A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
relativePriority String | "Low" | "Normal" | "High"
Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.

TransformOutputResponse
, TransformOutputResponseArgs

Preset This property is required. Pulumi.AzureNative.Media.Inputs.AudioAnalyzerPresetResponse | Pulumi.AzureNative.Media.Inputs.BuiltInStandardEncoderPresetResponse | Pulumi.AzureNative.Media.Inputs.FaceDetectorPresetResponse | Pulumi.AzureNative.Media.Inputs.StandardEncoderPresetResponse | Pulumi.AzureNative.Media.Inputs.VideoAnalyzerPresetResponse
Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
OnError string
A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
RelativePriority string
Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
Preset This property is required. AudioAnalyzerPresetResponse | BuiltInStandardEncoderPresetResponse | FaceDetectorPresetResponse | StandardEncoderPresetResponse | VideoAnalyzerPresetResponse
Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
OnError string
A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
RelativePriority string
Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
preset This property is required. AudioAnalyzerPresetResponse | BuiltInStandardEncoderPresetResponse | FaceDetectorPresetResponse | StandardEncoderPresetResponse | VideoAnalyzerPresetResponse
Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
onError String
A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
relativePriority String
Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
preset This property is required. AudioAnalyzerPresetResponse | BuiltInStandardEncoderPresetResponse | FaceDetectorPresetResponse | StandardEncoderPresetResponse | VideoAnalyzerPresetResponse
Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
onError string
A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
relativePriority string
Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
preset This property is required. AudioAnalyzerPresetResponse | BuiltInStandardEncoderPresetResponse | FaceDetectorPresetResponse | StandardEncoderPresetResponse | VideoAnalyzerPresetResponse
Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
on_error str
A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
relative_priority str
Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
preset This property is required. Property Map | Property Map | Property Map | Property Map | Property Map
Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
onError String
A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
relativePriority String
Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.

TransportStreamFormat
, TransportStreamFormatArgs

FilenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
OutputFiles List<Pulumi.AzureNative.Media.Inputs.OutputFile>
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
FilenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
OutputFiles []OutputFile
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
filenamePattern This property is required. String
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
outputFiles List<OutputFile>
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
filenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
outputFiles OutputFile[]
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
filename_pattern This property is required. str
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
output_files Sequence[OutputFile]
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
filenamePattern This property is required. String
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
outputFiles List<Property Map>
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

TransportStreamFormatResponse
, TransportStreamFormatResponseArgs

FilenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
OutputFiles List<Pulumi.AzureNative.Media.Inputs.OutputFileResponse>
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
FilenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
OutputFiles []OutputFileResponse
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
filenamePattern This property is required. String
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
outputFiles List<OutputFileResponse>
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
filenamePattern This property is required. string
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
outputFiles OutputFileResponse[]
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
filename_pattern This property is required. str
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
output_files Sequence[OutputFileResponse]
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
filenamePattern This property is required. String
The file naming pattern used for the creation of output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {AudioStream} - string "Audio" plus audio stream number(start from 1). {Bitrate} - The audio/video bitrate in kbps. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
outputFiles List<Property Map>
The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

Video
, VideoArgs

KeyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
StretchMode string | Pulumi.AzureNative.Media.StretchMode
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
SyncMode string | Pulumi.AzureNative.Media.VideoSyncMode
The Video Sync Mode
KeyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
StretchMode string | StretchMode
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
SyncMode string | VideoSyncMode
The Video Sync Mode
keyFrameInterval String
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label String
An optional label for the codec. The label can be used to control muxing behavior.
stretchMode String | StretchMode
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode String | VideoSyncMode
The Video Sync Mode
keyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label string
An optional label for the codec. The label can be used to control muxing behavior.
stretchMode string | StretchMode
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode string | VideoSyncMode
The Video Sync Mode
key_frame_interval str
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label str
An optional label for the codec. The label can be used to control muxing behavior.
stretch_mode str | StretchMode
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
sync_mode str | VideoSyncMode
The Video Sync Mode
keyFrameInterval String
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label String
An optional label for the codec. The label can be used to control muxing behavior.
stretchMode String | "None" | "AutoSize" | "AutoFit"
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode String | "Auto" | "Passthrough" | "Cfr" | "Vfr"
The Video Sync Mode

VideoAnalyzerPreset
, VideoAnalyzerPresetArgs

AudioLanguage string
The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
ExperimentalOptions Dictionary<string, string>
Dictionary containing key value pairs for parameters not exposed in the preset itself
InsightsToExtract string | Pulumi.AzureNative.Media.InsightsType
Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
Mode string | Pulumi.AzureNative.Media.AudioAnalysisMode
Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
AudioLanguage string
The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
ExperimentalOptions map[string]string
Dictionary containing key value pairs for parameters not exposed in the preset itself
InsightsToExtract string | InsightsType
Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
Mode string | AudioAnalysisMode
Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
audioLanguage String
The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
experimentalOptions Map<String,String>
Dictionary containing key value pairs for parameters not exposed in the preset itself
insightsToExtract String | InsightsType
Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
mode String | AudioAnalysisMode
Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
audioLanguage string
The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
experimentalOptions {[key: string]: string}
Dictionary containing key value pairs for parameters not exposed in the preset itself
insightsToExtract string | InsightsType
Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
mode string | AudioAnalysisMode
Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
audio_language str
The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
experimental_options Mapping[str, str]
Dictionary containing key value pairs for parameters not exposed in the preset itself
insights_to_extract str | InsightsType
Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
mode str | AudioAnalysisMode
Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
audioLanguage String
The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
experimentalOptions Map<String>
Dictionary containing key value pairs for parameters not exposed in the preset itself
insightsToExtract String | "AudioInsightsOnly" | "VideoInsightsOnly" | "AllInsights"
Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
mode String | "Standard" | "Basic"
Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.

VideoAnalyzerPresetResponse
, VideoAnalyzerPresetResponseArgs

AudioLanguage string
The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
ExperimentalOptions Dictionary<string, string>
Dictionary containing key value pairs for parameters not exposed in the preset itself
InsightsToExtract string
Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
Mode string
Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
AudioLanguage string
The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
ExperimentalOptions map[string]string
Dictionary containing key value pairs for parameters not exposed in the preset itself
InsightsToExtract string
Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
Mode string
Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
audioLanguage String
The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
experimentalOptions Map<String,String>
Dictionary containing key value pairs for parameters not exposed in the preset itself
insightsToExtract String
Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
mode String
Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
audioLanguage string
The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
experimentalOptions {[key: string]: string}
Dictionary containing key value pairs for parameters not exposed in the preset itself
insightsToExtract string
Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
mode string
Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
audio_language str
The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
experimental_options Mapping[str, str]
Dictionary containing key value pairs for parameters not exposed in the preset itself
insights_to_extract str
Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
mode str
Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
audioLanguage String
The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
experimentalOptions Map<String>
Dictionary containing key value pairs for parameters not exposed in the preset itself
insightsToExtract String
Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
mode String
Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.

VideoOverlay
, VideoOverlayArgs

InputLabel This property is required. string
The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
AudioGainLevel double
The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
CropRectangle Pulumi.AzureNative.Media.Inputs.Rectangle
An optional rectangular window used to crop the overlay image or video.
End string
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
FadeInDuration string
The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
FadeOutDuration string
The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
Opacity double
The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
Position Pulumi.AzureNative.Media.Inputs.Rectangle
The location in the input video where the overlay is applied.
Start string
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
InputLabel This property is required. string
The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
AudioGainLevel float64
The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
CropRectangle Rectangle
An optional rectangular window used to crop the overlay image or video.
End string
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
FadeInDuration string
The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
FadeOutDuration string
The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
Opacity float64
The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
Position Rectangle
The location in the input video where the overlay is applied.
Start string
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
inputLabel This property is required. String
The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
audioGainLevel Double
The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
cropRectangle Rectangle
An optional rectangular window used to crop the overlay image or video.
end String
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
fadeInDuration String
The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
fadeOutDuration String
The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
opacity Double
The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
position Rectangle
The location in the input video where the overlay is applied.
start String
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
inputLabel This property is required. string
The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
audioGainLevel number
The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
cropRectangle Rectangle
An optional rectangular window used to crop the overlay image or video.
end string
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
fadeInDuration string
The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
fadeOutDuration string
The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
opacity number
The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
position Rectangle
The location in the input video where the overlay is applied.
start string
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
input_label This property is required. str
The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
audio_gain_level float
The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
crop_rectangle Rectangle
An optional rectangular window used to crop the overlay image or video.
end str
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
fade_in_duration str
The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
fade_out_duration str
The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
opacity float
The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
position Rectangle
The location in the input video where the overlay is applied.
start str
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
inputLabel This property is required. String
The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
audioGainLevel Number
The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
cropRectangle Property Map
An optional rectangular window used to crop the overlay image or video.
end String
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
fadeInDuration String
The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
fadeOutDuration String
The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
opacity Number
The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
position Property Map
The location in the input video where the overlay is applied.
start String
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.

VideoOverlayResponse
, VideoOverlayResponseArgs

InputLabel This property is required. string
The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
AudioGainLevel double
The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
CropRectangle Pulumi.AzureNative.Media.Inputs.RectangleResponse
An optional rectangular window used to crop the overlay image or video.
End string
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
FadeInDuration string
The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
FadeOutDuration string
The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
Opacity double
The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
Position Pulumi.AzureNative.Media.Inputs.RectangleResponse
The location in the input video where the overlay is applied.
Start string
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
InputLabel This property is required. string
The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
AudioGainLevel float64
The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
CropRectangle RectangleResponse
An optional rectangular window used to crop the overlay image or video.
End string
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
FadeInDuration string
The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
FadeOutDuration string
The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
Opacity float64
The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
Position RectangleResponse
The location in the input video where the overlay is applied.
Start string
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
inputLabel This property is required. String
The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
audioGainLevel Double
The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
cropRectangle RectangleResponse
An optional rectangular window used to crop the overlay image or video.
end String
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
fadeInDuration String
The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
fadeOutDuration String
The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
opacity Double
The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
position RectangleResponse
The location in the input video where the overlay is applied.
start String
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
inputLabel This property is required. string
The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
audioGainLevel number
The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
cropRectangle RectangleResponse
An optional rectangular window used to crop the overlay image or video.
end string
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
fadeInDuration string
The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
fadeOutDuration string
The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
opacity number
The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
position RectangleResponse
The location in the input video where the overlay is applied.
start string
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
input_label This property is required. str
The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
audio_gain_level float
The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
crop_rectangle RectangleResponse
An optional rectangular window used to crop the overlay image or video.
end str
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
fade_in_duration str
The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
fade_out_duration str
The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
opacity float
The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
position RectangleResponse
The location in the input video where the overlay is applied.
start str
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
inputLabel This property is required. String
The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
audioGainLevel Number
The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
cropRectangle Property Map
An optional rectangular window used to crop the overlay image or video.
end String
The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
fadeInDuration String
The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
fadeOutDuration String
The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
opacity Number
The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
position Property Map
The location in the input video where the overlay is applied.
start String
The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.

VideoResponse
, VideoResponseArgs

KeyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
StretchMode string
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
SyncMode string
The Video Sync Mode
KeyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
Label string
An optional label for the codec. The label can be used to control muxing behavior.
StretchMode string
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
SyncMode string
The Video Sync Mode
keyFrameInterval String
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label String
An optional label for the codec. The label can be used to control muxing behavior.
stretchMode String
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode String
The Video Sync Mode
keyFrameInterval string
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label string
An optional label for the codec. The label can be used to control muxing behavior.
stretchMode string
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode string
The Video Sync Mode
key_frame_interval str
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label str
An optional label for the codec. The label can be used to control muxing behavior.
stretch_mode str
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
sync_mode str
The Video Sync Mode
keyFrameInterval String
The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
label String
An optional label for the codec. The label can be used to control muxing behavior.
stretchMode String
The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
syncMode String
The Video Sync Mode

VideoSyncMode
, VideoSyncModeArgs

Auto
AutoThis is the default method. Chooses between Cfr and Vfr depending on muxer capabilities. For output format MP4, the default mode is Cfr.
Passthrough
PassthroughThe presentation timestamps on frames are passed through from the input file to the output file writer. Recommended when the input source has variable frame rate, and are attempting to produce multiple layers for adaptive streaming in the output which have aligned GOP boundaries. Note: if two or more frames in the input have duplicate timestamps, then the output will also have the same behavior
Cfr
CfrInput frames will be repeated and/or dropped as needed to achieve exactly the requested constant frame rate. Recommended when the output frame rate is explicitly set at a specified value
Vfr
VfrSimilar to the Passthrough mode, but if the input has frames that have duplicate timestamps, then only one frame is passed through to the output, and others are dropped. Recommended when the number of output frames is expected to be equal to the number of input frames. For example, the output is used to calculate a quality metric like PSNR against the input
VideoSyncModeAuto
AutoThis is the default method. Chooses between Cfr and Vfr depending on muxer capabilities. For output format MP4, the default mode is Cfr.
VideoSyncModePassthrough
PassthroughThe presentation timestamps on frames are passed through from the input file to the output file writer. Recommended when the input source has variable frame rate, and are attempting to produce multiple layers for adaptive streaming in the output which have aligned GOP boundaries. Note: if two or more frames in the input have duplicate timestamps, then the output will also have the same behavior
VideoSyncModeCfr
CfrInput frames will be repeated and/or dropped as needed to achieve exactly the requested constant frame rate. Recommended when the output frame rate is explicitly set at a specified value
VideoSyncModeVfr
VfrSimilar to the Passthrough mode, but if the input has frames that have duplicate timestamps, then only one frame is passed through to the output, and others are dropped. Recommended when the number of output frames is expected to be equal to the number of input frames. For example, the output is used to calculate a quality metric like PSNR against the input
Auto
AutoThis is the default method. Chooses between Cfr and Vfr depending on muxer capabilities. For output format MP4, the default mode is Cfr.
Passthrough
PassthroughThe presentation timestamps on frames are passed through from the input file to the output file writer. Recommended when the input source has variable frame rate, and are attempting to produce multiple layers for adaptive streaming in the output which have aligned GOP boundaries. Note: if two or more frames in the input have duplicate timestamps, then the output will also have the same behavior
Cfr
CfrInput frames will be repeated and/or dropped as needed to achieve exactly the requested constant frame rate. Recommended when the output frame rate is explicitly set at a specified value
Vfr
VfrSimilar to the Passthrough mode, but if the input has frames that have duplicate timestamps, then only one frame is passed through to the output, and others are dropped. Recommended when the number of output frames is expected to be equal to the number of input frames. For example, the output is used to calculate a quality metric like PSNR against the input
Auto
AutoThis is the default method. Chooses between Cfr and Vfr depending on muxer capabilities. For output format MP4, the default mode is Cfr.
Passthrough
PassthroughThe presentation timestamps on frames are passed through from the input file to the output file writer. Recommended when the input source has variable frame rate, and are attempting to produce multiple layers for adaptive streaming in the output which have aligned GOP boundaries. Note: if two or more frames in the input have duplicate timestamps, then the output will also have the same behavior
Cfr
CfrInput frames will be repeated and/or dropped as needed to achieve exactly the requested constant frame rate. Recommended when the output frame rate is explicitly set at a specified value
Vfr
VfrSimilar to the Passthrough mode, but if the input has frames that have duplicate timestamps, then only one frame is passed through to the output, and others are dropped. Recommended when the number of output frames is expected to be equal to the number of input frames. For example, the output is used to calculate a quality metric like PSNR against the input
AUTO
AutoThis is the default method. Chooses between Cfr and Vfr depending on muxer capabilities. For output format MP4, the default mode is Cfr.
PASSTHROUGH
PassthroughThe presentation timestamps on frames are passed through from the input file to the output file writer. Recommended when the input source has variable frame rate, and are attempting to produce multiple layers for adaptive streaming in the output which have aligned GOP boundaries. Note: if two or more frames in the input have duplicate timestamps, then the output will also have the same behavior
CFR
CfrInput frames will be repeated and/or dropped as needed to achieve exactly the requested constant frame rate. Recommended when the output frame rate is explicitly set at a specified value
VFR
VfrSimilar to the Passthrough mode, but if the input has frames that have duplicate timestamps, then only one frame is passed through to the output, and others are dropped. Recommended when the number of output frames is expected to be equal to the number of input frames. For example, the output is used to calculate a quality metric like PSNR against the input
"Auto"
AutoThis is the default method. Chooses between Cfr and Vfr depending on muxer capabilities. For output format MP4, the default mode is Cfr.
"Passthrough"
PassthroughThe presentation timestamps on frames are passed through from the input file to the output file writer. Recommended when the input source has variable frame rate, and are attempting to produce multiple layers for adaptive streaming in the output which have aligned GOP boundaries. Note: if two or more frames in the input have duplicate timestamps, then the output will also have the same behavior
"Cfr"
CfrInput frames will be repeated and/or dropped as needed to achieve exactly the requested constant frame rate. Recommended when the output frame rate is explicitly set at a specified value
"Vfr"
VfrSimilar to the Passthrough mode, but if the input has frames that have duplicate timestamps, then only one frame is passed through to the output, and others are dropped. Recommended when the number of output frames is expected to be equal to the number of input frames. For example, the output is used to calculate a quality metric like PSNR against the input

Import

An existing resource can be imported using its type token, name, and identifier, e.g.

$ pulumi import azure-native:media:Transform createdTransform /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Media/mediaServices/{accountName}/transforms/{transformName} 
Copy

To learn more about importing existing cloud resources, see Importing resources.

Package Details

Repository
Azure Native pulumi/pulumi-azure-native
License
Apache-2.0
This is the latest version of Azure Native. Use the Azure Native v2 docs if using the v2 version of this package.
Azure Native v3.2.0 published on Monday, Apr 14, 2025 by Pulumi