Documentation ¶
Index ¶
- func CfnChannel_CFN_RESOURCE_TYPE_NAME() *string
- func CfnChannel_IsCfnElement(x interface{}) *bool
- func CfnChannel_IsCfnResource(construct constructs.IConstruct) *bool
- func CfnChannel_IsConstruct(x interface{}) *bool
- func CfnInputSecurityGroup_CFN_RESOURCE_TYPE_NAME() *string
- func CfnInputSecurityGroup_IsCfnElement(x interface{}) *bool
- func CfnInputSecurityGroup_IsCfnResource(construct constructs.IConstruct) *bool
- func CfnInputSecurityGroup_IsConstruct(x interface{}) *bool
- func CfnInput_CFN_RESOURCE_TYPE_NAME() *string
- func CfnInput_IsCfnElement(x interface{}) *bool
- func CfnInput_IsCfnResource(construct constructs.IConstruct) *bool
- func CfnInput_IsConstruct(x interface{}) *bool
- func NewCfnChannel_Override(c CfnChannel, scope awscdk.Construct, id *string, props *CfnChannelProps)
- func NewCfnInputSecurityGroup_Override(c CfnInputSecurityGroup, scope awscdk.Construct, id *string, ...)
- func NewCfnInput_Override(c CfnInput, scope awscdk.Construct, id *string, props *CfnInputProps)
- type CfnChannel
- type CfnChannelProps
- type CfnChannel_AacSettingsProperty
- type CfnChannel_Ac3SettingsProperty
- type CfnChannel_AncillarySourceSettingsProperty
- type CfnChannel_ArchiveCdnSettingsProperty
- type CfnChannel_ArchiveContainerSettingsProperty
- type CfnChannel_ArchiveGroupSettingsProperty
- type CfnChannel_ArchiveOutputSettingsProperty
- type CfnChannel_ArchiveS3SettingsProperty
- type CfnChannel_AribDestinationSettingsProperty
- type CfnChannel_AribSourceSettingsProperty
- type CfnChannel_AudioChannelMappingProperty
- type CfnChannel_AudioCodecSettingsProperty
- type CfnChannel_AudioDescriptionProperty
- type CfnChannel_AudioHlsRenditionSelectionProperty
- type CfnChannel_AudioLanguageSelectionProperty
- type CfnChannel_AudioNormalizationSettingsProperty
- type CfnChannel_AudioOnlyHlsSettingsProperty
- type CfnChannel_AudioPidSelectionProperty
- type CfnChannel_AudioSelectorProperty
- type CfnChannel_AudioSelectorSettingsProperty
- type CfnChannel_AudioSilenceFailoverSettingsProperty
- type CfnChannel_AudioTrackProperty
- type CfnChannel_AudioTrackSelectionProperty
- type CfnChannel_AudioWatermarkSettingsProperty
- type CfnChannel_AutomaticInputFailoverSettingsProperty
- type CfnChannel_AvailBlankingProperty
- type CfnChannel_AvailConfigurationProperty
- type CfnChannel_AvailSettingsProperty
- type CfnChannel_BlackoutSlateProperty
- type CfnChannel_BurnInDestinationSettingsProperty
- type CfnChannel_CaptionDescriptionProperty
- type CfnChannel_CaptionDestinationSettingsProperty
- type CfnChannel_CaptionLanguageMappingProperty
- type CfnChannel_CaptionRectangleProperty
- type CfnChannel_CaptionSelectorProperty
- type CfnChannel_CaptionSelectorSettingsProperty
- type CfnChannel_CdiInputSpecificationProperty
- type CfnChannel_ColorSpacePassthroughSettingsProperty
- type CfnChannel_DvbNitSettingsProperty
- type CfnChannel_DvbSdtSettingsProperty
- type CfnChannel_DvbSubDestinationSettingsProperty
- type CfnChannel_DvbSubSourceSettingsProperty
- type CfnChannel_DvbTdtSettingsProperty
- type CfnChannel_Eac3SettingsProperty
- type CfnChannel_EbuTtDDestinationSettingsProperty
- type CfnChannel_EmbeddedDestinationSettingsProperty
- type CfnChannel_EmbeddedPlusScte20DestinationSettingsProperty
- type CfnChannel_EmbeddedSourceSettingsProperty
- type CfnChannel_EncoderSettingsProperty
- type CfnChannel_FailoverConditionProperty
- type CfnChannel_FailoverConditionSettingsProperty
- type CfnChannel_FeatureActivationsProperty
- type CfnChannel_FecOutputSettingsProperty
- type CfnChannel_Fmp4HlsSettingsProperty
- type CfnChannel_FrameCaptureCdnSettingsProperty
- type CfnChannel_FrameCaptureGroupSettingsProperty
- type CfnChannel_FrameCaptureHlsSettingsProperty
- type CfnChannel_FrameCaptureOutputSettingsProperty
- type CfnChannel_FrameCaptureS3SettingsProperty
- type CfnChannel_FrameCaptureSettingsProperty
- type CfnChannel_GlobalConfigurationProperty
- type CfnChannel_H264ColorSpaceSettingsProperty
- type CfnChannel_H264FilterSettingsProperty
- type CfnChannel_H264SettingsProperty
- type CfnChannel_H265ColorSpaceSettingsProperty
- type CfnChannel_H265FilterSettingsProperty
- type CfnChannel_H265SettingsProperty
- type CfnChannel_Hdr10SettingsProperty
- type CfnChannel_HlsAkamaiSettingsProperty
- type CfnChannel_HlsBasicPutSettingsProperty
- type CfnChannel_HlsCdnSettingsProperty
- type CfnChannel_HlsGroupSettingsProperty
- type CfnChannel_HlsInputSettingsProperty
- type CfnChannel_HlsMediaStoreSettingsProperty
- type CfnChannel_HlsOutputSettingsProperty
- type CfnChannel_HlsS3SettingsProperty
- type CfnChannel_HlsSettingsProperty
- type CfnChannel_HlsWebdavSettingsProperty
- type CfnChannel_HtmlMotionGraphicsSettingsProperty
- type CfnChannel_InputAttachmentProperty
- type CfnChannel_InputChannelLevelProperty
- type CfnChannel_InputLocationProperty
- type CfnChannel_InputLossBehaviorProperty
- type CfnChannel_InputLossFailoverSettingsProperty
- type CfnChannel_InputSettingsProperty
- type CfnChannel_InputSpecificationProperty
- type CfnChannel_KeyProviderSettingsProperty
- type CfnChannel_M2tsSettingsProperty
- type CfnChannel_M3u8SettingsProperty
- type CfnChannel_MediaPackageGroupSettingsProperty
- type CfnChannel_MediaPackageOutputDestinationSettingsProperty
- type CfnChannel_MediaPackageOutputSettingsProperty
- type CfnChannel_MotionGraphicsConfigurationProperty
- type CfnChannel_MotionGraphicsSettingsProperty
- type CfnChannel_Mp2SettingsProperty
- type CfnChannel_Mpeg2FilterSettingsProperty
- type CfnChannel_Mpeg2SettingsProperty
- type CfnChannel_MsSmoothGroupSettingsProperty
- type CfnChannel_MsSmoothOutputSettingsProperty
- type CfnChannel_MultiplexGroupSettingsProperty
- type CfnChannel_MultiplexOutputSettingsProperty
- type CfnChannel_MultiplexProgramChannelDestinationSettingsProperty
- type CfnChannel_NetworkInputSettingsProperty
- type CfnChannel_NielsenCBETProperty
- type CfnChannel_NielsenConfigurationProperty
- type CfnChannel_NielsenNaesIiNwProperty
- type CfnChannel_NielsenWatermarksSettingsProperty
- type CfnChannel_OutputDestinationProperty
- type CfnChannel_OutputDestinationSettingsProperty
- type CfnChannel_OutputGroupProperty
- type CfnChannel_OutputGroupSettingsProperty
- type CfnChannel_OutputLocationRefProperty
- type CfnChannel_OutputProperty
- type CfnChannel_OutputSettingsProperty
- type CfnChannel_PassThroughSettingsProperty
- type CfnChannel_RawSettingsProperty
- type CfnChannel_Rec601SettingsProperty
- type CfnChannel_Rec709SettingsProperty
- type CfnChannel_RemixSettingsProperty
- type CfnChannel_RtmpCaptionInfoDestinationSettingsProperty
- type CfnChannel_RtmpGroupSettingsProperty
- type CfnChannel_RtmpOutputSettingsProperty
- type CfnChannel_Scte20PlusEmbeddedDestinationSettingsProperty
- type CfnChannel_Scte20SourceSettingsProperty
- type CfnChannel_Scte27DestinationSettingsProperty
- type CfnChannel_Scte27SourceSettingsProperty
- type CfnChannel_Scte35SpliceInsertProperty
- type CfnChannel_Scte35TimeSignalAposProperty
- type CfnChannel_SmpteTtDestinationSettingsProperty
- type CfnChannel_StandardHlsSettingsProperty
- type CfnChannel_StaticKeySettingsProperty
- type CfnChannel_TeletextDestinationSettingsProperty
- type CfnChannel_TeletextSourceSettingsProperty
- type CfnChannel_TemporalFilterSettingsProperty
- type CfnChannel_TimecodeConfigProperty
- type CfnChannel_TtmlDestinationSettingsProperty
- type CfnChannel_UdpContainerSettingsProperty
- type CfnChannel_UdpGroupSettingsProperty
- type CfnChannel_UdpOutputSettingsProperty
- type CfnChannel_VideoBlackFailoverSettingsProperty
- type CfnChannel_VideoCodecSettingsProperty
- type CfnChannel_VideoDescriptionProperty
- type CfnChannel_VideoSelectorColorSpaceSettingsProperty
- type CfnChannel_VideoSelectorPidProperty
- type CfnChannel_VideoSelectorProgramIdProperty
- type CfnChannel_VideoSelectorProperty
- type CfnChannel_VideoSelectorSettingsProperty
- type CfnChannel_VpcOutputSettingsProperty
- type CfnChannel_WavSettingsProperty
- type CfnChannel_WebvttDestinationSettingsProperty
- type CfnInput
- type CfnInputProps
- type CfnInputSecurityGroup
- type CfnInputSecurityGroupProps
- type CfnInputSecurityGroup_InputWhitelistRuleCidrProperty
- type CfnInput_InputDestinationRequestProperty
- type CfnInput_InputDeviceRequestProperty
- type CfnInput_InputDeviceSettingsProperty
- type CfnInput_InputSourceRequestProperty
- type CfnInput_InputVpcRequestProperty
- type CfnInput_MediaConnectFlowRequestProperty
Constants ¶
This section is empty.
Variables ¶
This section is empty.
Functions ¶
func CfnChannel_CFN_RESOURCE_TYPE_NAME ¶
func CfnChannel_CFN_RESOURCE_TYPE_NAME() *string
func CfnChannel_IsCfnElement ¶
func CfnChannel_IsCfnElement(x interface{}) *bool
Returns `true` if a construct is a stack element (i.e. part of the synthesized cloudformation template).
Uses duck-typing instead of `instanceof` to allow stack elements from different versions of this library to be included in the same stack.
Returns: The construct as a stack element or undefined if it is not a stack element. Experimental.
func CfnChannel_IsCfnResource ¶
func CfnChannel_IsCfnResource(construct constructs.IConstruct) *bool
Check whether the given construct is a CfnResource. Experimental.
func CfnChannel_IsConstruct ¶
func CfnChannel_IsConstruct(x interface{}) *bool
Return whether the given object is a Construct. Experimental.
func CfnInputSecurityGroup_CFN_RESOURCE_TYPE_NAME ¶
func CfnInputSecurityGroup_CFN_RESOURCE_TYPE_NAME() *string
func CfnInputSecurityGroup_IsCfnElement ¶
func CfnInputSecurityGroup_IsCfnElement(x interface{}) *bool
Returns `true` if a construct is a stack element (i.e. part of the synthesized cloudformation template).
Uses duck-typing instead of `instanceof` to allow stack elements from different versions of this library to be included in the same stack.
Returns: The construct as a stack element or undefined if it is not a stack element. Experimental.
func CfnInputSecurityGroup_IsCfnResource ¶
func CfnInputSecurityGroup_IsCfnResource(construct constructs.IConstruct) *bool
Check whether the given construct is a CfnResource. Experimental.
func CfnInputSecurityGroup_IsConstruct ¶
func CfnInputSecurityGroup_IsConstruct(x interface{}) *bool
Return whether the given object is a Construct. Experimental.
func CfnInput_CFN_RESOURCE_TYPE_NAME ¶
func CfnInput_CFN_RESOURCE_TYPE_NAME() *string
func CfnInput_IsCfnElement ¶
func CfnInput_IsCfnElement(x interface{}) *bool
Returns `true` if a construct is a stack element (i.e. part of the synthesized cloudformation template).
Uses duck-typing instead of `instanceof` to allow stack elements from different versions of this library to be included in the same stack.
Returns: The construct as a stack element or undefined if it is not a stack element. Experimental.
func CfnInput_IsCfnResource ¶
func CfnInput_IsCfnResource(construct constructs.IConstruct) *bool
Check whether the given construct is a CfnResource. Experimental.
func CfnInput_IsConstruct ¶
func CfnInput_IsConstruct(x interface{}) *bool
Return whether the given object is a Construct. Experimental.
func NewCfnChannel_Override ¶
func NewCfnChannel_Override(c CfnChannel, scope awscdk.Construct, id *string, props *CfnChannelProps)
Create a new `AWS::MediaLive::Channel`.
func NewCfnInputSecurityGroup_Override ¶
func NewCfnInputSecurityGroup_Override(c CfnInputSecurityGroup, scope awscdk.Construct, id *string, props *CfnInputSecurityGroupProps)
Create a new `AWS::MediaLive::InputSecurityGroup`.
func NewCfnInput_Override ¶
func NewCfnInput_Override(c CfnInput, scope awscdk.Construct, id *string, props *CfnInputProps)
Create a new `AWS::MediaLive::Input`.
Types ¶
type CfnChannel ¶
type CfnChannel interface { awscdk.CfnResource awscdk.IInspectable AttrArn() *string AttrInputs() *[]*string CdiInputSpecification() interface{} SetCdiInputSpecification(val interface{}) CfnOptions() awscdk.ICfnResourceOptions CfnProperties() *map[string]interface{} CfnResourceType() *string ChannelClass() *string SetChannelClass(val *string) CreationStack() *[]*string Destinations() interface{} SetDestinations(val interface{}) EncoderSettings() interface{} SetEncoderSettings(val interface{}) InputAttachments() interface{} SetInputAttachments(val interface{}) InputSpecification() interface{} SetInputSpecification(val interface{}) LogicalId() *string LogLevel() *string SetLogLevel(val *string) Name() *string SetName(val *string) Node() awscdk.ConstructNode Ref() *string RoleArn() *string SetRoleArn(val *string) Stack() awscdk.Stack Tags() awscdk.TagManager UpdatedProperites() *map[string]interface{} Vpc() interface{} SetVpc(val interface{}) AddDeletionOverride(path *string) AddDependsOn(target awscdk.CfnResource) AddMetadata(key *string, value interface{}) AddOverride(path *string, value interface{}) AddPropertyDeletionOverride(propertyPath *string) AddPropertyOverride(propertyPath *string, value interface{}) ApplyRemovalPolicy(policy awscdk.RemovalPolicy, options *awscdk.RemovalPolicyOptions) GetAtt(attributeName *string) awscdk.Reference GetMetadata(key *string) interface{} Inspect(inspector awscdk.TreeInspector) OnPrepare() OnSynthesize(session constructs.ISynthesisSession) OnValidate() *[]*string OverrideLogicalId(newLogicalId *string) Prepare() RenderProperties(props *map[string]interface{}) *map[string]interface{} ShouldSynthesize() *bool Synthesize(session awscdk.ISynthesisSession) ToString() *string Validate() *[]*string ValidateProperties(_properties interface{}) }
A CloudFormation `AWS::MediaLive::Channel`.
The AWS::MediaLive::Channel resource is a MediaLive resource type that creates a channel.
A MediaLive channel ingests and transcodes (decodes and encodes) source content from the inputs that are attached to that channel, and packages the new content into outputs.
TODO: EXAMPLE
func NewCfnChannel ¶
func NewCfnChannel(scope awscdk.Construct, id *string, props *CfnChannelProps) CfnChannel
Create a new `AWS::MediaLive::Channel`.
type CfnChannelProps ¶
type CfnChannelProps struct { // Specification of CDI inputs for this channel. CdiInputSpecification interface{} `json:"cdiInputSpecification" yaml:"cdiInputSpecification"` // The class for this channel. // // For a channel with two pipelines, the class is STANDARD. For a channel with one pipeline, the class is SINGLE_PIPELINE. ChannelClass *string `json:"channelClass" yaml:"channelClass"` // The settings that identify the destination for the outputs in this MediaLive output package. Destinations interface{} `json:"destinations" yaml:"destinations"` // The encoding configuration for the output content. EncoderSettings interface{} `json:"encoderSettings" yaml:"encoderSettings"` // The list of input attachments for the channel. InputAttachments interface{} `json:"inputAttachments" yaml:"inputAttachments"` // The input specification for this channel. // // It specifies the key characteristics of the inputs for this channel: the maximum bitrate, the resolution, and the codec. InputSpecification interface{} `json:"inputSpecification" yaml:"inputSpecification"` // The verbosity for logging activity for this channel. // // Charges for logging (which are generated through Amazon CloudWatch Logging) are higher for higher verbosities. LogLevel *string `json:"logLevel" yaml:"logLevel"` // A name for this audio selector. // // The AudioDescription (in an output) references this name in order to identify a specific input audio to include in that output. Name *string `json:"name" yaml:"name"` // The IAM role for MediaLive to assume when running this channel. // // The role is identified by its ARN. RoleArn *string `json:"roleArn" yaml:"roleArn"` // A collection of tags for this channel. // // Each tag is a key-value pair. Tags interface{} `json:"tags" yaml:"tags"` // Settings to enable VPC mode in the channel, so that the endpoints for all outputs are in your VPC. Vpc interface{} `json:"vpc" yaml:"vpc"` }
Properties for defining a `CfnChannel`.
TODO: EXAMPLE
type CfnChannel_AacSettingsProperty ¶
type CfnChannel_AacSettingsProperty struct { // The average bitrate in bits/second. // // Valid values depend on the rate control mode and profile. Bitrate *float64 `json:"bitrate" yaml:"bitrate"` // Mono, stereo, or 5.1 channel layout. Valid values depend on the rate control mode and profile. The adReceiverMix setting receives a stereo description plus control track, and emits a mono AAC encode of the description track, with control data emitted in the PES header as per ETSI TS 101 154 Annex E. CodingMode *string `json:"codingMode" yaml:"codingMode"` // Set to broadcasterMixedAd when the input contains pre-mixed main audio + AD (narration) as a stereo pair. // // The Audio Type field (audioType) will be set to 3, which signals to downstream systems that this stream contains broadcaster mixed AD. Note that the input received by the encoder must contain pre-mixed audio; MediaLive does not perform the mixing. The values in audioTypeControl and audioType (in AudioDescription) are ignored when set to broadcasterMixedAd. Leave this set to normal when the input does not contain pre-mixed audio + AD. InputType *string `json:"inputType" yaml:"inputType"` // The AAC profile. Profile *string `json:"profile" yaml:"profile"` // The rate control mode. RateControlMode *string `json:"rateControlMode" yaml:"rateControlMode"` // Sets the LATM/LOAS AAC output for raw containers. RawFormat *string `json:"rawFormat" yaml:"rawFormat"` // The sample rate in Hz. // // Valid values depend on the rate control mode and profile. SampleRate *float64 `json:"sampleRate" yaml:"sampleRate"` // Uses MPEG-2 AAC audio instead of MPEG-4 AAC audio for raw or MPEG-2 Transport Stream containers. Spec *string `json:"spec" yaml:"spec"` // The VBR quality level. // // This is used only if rateControlMode is VBR. VbrQuality *string `json:"vbrQuality" yaml:"vbrQuality"` }
The settings for an AAC audio encode in the output.
The parent of this entity is AudioCodecSettings.
TODO: EXAMPLE
type CfnChannel_Ac3SettingsProperty ¶
type CfnChannel_Ac3SettingsProperty struct { // The average bitrate in bits/second. // // Valid bitrates depend on the coding mode. Bitrate *float64 `json:"bitrate" yaml:"bitrate"` // Specifies the bitstream mode (bsmod) for the emitted AC-3 stream. // // For more information about these values, see ATSC A/52-2012. BitstreamMode *string `json:"bitstreamMode" yaml:"bitstreamMode"` // The Dolby Digital coding mode. // // This determines the number of channels. CodingMode *string `json:"codingMode" yaml:"codingMode"` // Sets the dialnorm for the output. // // If excluded and the input audio is Dolby Digital, dialnorm is passed through. Dialnorm *float64 `json:"dialnorm" yaml:"dialnorm"` // If set to filmStandard, adds dynamic range compression signaling to the output bitstream as defined in the Dolby Digital specification. DrcProfile *string `json:"drcProfile" yaml:"drcProfile"` // When set to enabled, applies a 120Hz lowpass filter to the LFE channel prior to encoding. // // This is valid only in codingMode32Lfe mode. LfeFilter *string `json:"lfeFilter" yaml:"lfeFilter"` // When set to followInput, encoder metadata is sourced from the DD, DD+, or DolbyE decoder that supplies this audio data. // // If the audio is supplied from one of these streams, the static metadata settings are used. MetadataControl *string `json:"metadataControl" yaml:"metadataControl"` }
The settings for an AC3 audio encode in the output.
The parent of this entity is AudioCodecSettings.
TODO: EXAMPLE
type CfnChannel_AncillarySourceSettingsProperty ¶
type CfnChannel_AncillarySourceSettingsProperty struct { // Specifies the number (1 to 4) of the captions channel you want to extract from the ancillary captions. // // If you plan to convert the ancillary captions to another format, complete this field. If you plan to choose Embedded as the captions destination in the output (to pass through all the channels in the ancillary captions), leave this field blank because MediaLive ignores the field. SourceAncillaryChannelNumber *float64 `json:"sourceAncillaryChannelNumber" yaml:"sourceAncillaryChannelNumber"` }
Information about the ancillary captions to extract from the input.
The parent of this entity is CaptionSelectorSettings.
TODO: EXAMPLE
type CfnChannel_ArchiveCdnSettingsProperty ¶
type CfnChannel_ArchiveCdnSettingsProperty struct {
// Sets up Amazon S3 as the destination for this Archive output.
ArchiveS3Settings interface{} `json:"archiveS3Settings" yaml:"archiveS3Settings"`
}
Settings to configure the destination of an Archive output.
The parent of this entity is ArchiveGroupSettings.
TODO: EXAMPLE
type CfnChannel_ArchiveContainerSettingsProperty ¶
type CfnChannel_ArchiveContainerSettingsProperty struct { // The settings for the M2TS in the archive output. M2TsSettings interface{} `json:"m2TsSettings" yaml:"m2TsSettings"` // The settings for Raw archive output type. RawSettings interface{} `json:"rawSettings" yaml:"rawSettings"` }
The archive container settings.
The parent of this entity is ArchiveOutputSettings.
TODO: EXAMPLE
type CfnChannel_ArchiveGroupSettingsProperty ¶
type CfnChannel_ArchiveGroupSettingsProperty struct { // Settings to configure the destination of an Archive output. ArchiveCdnSettings interface{} `json:"archiveCdnSettings" yaml:"archiveCdnSettings"` // A directory and base file name where archive files should be written. Destination interface{} `json:"destination" yaml:"destination"` // The number of seconds to write to an archive file before closing and starting a new one. RolloverInterval *float64 `json:"rolloverInterval" yaml:"rolloverInterval"` }
The settings for an archive output group.
The parent of this entity is OutputGroupSettings.
TODO: EXAMPLE
type CfnChannel_ArchiveOutputSettingsProperty ¶
type CfnChannel_ArchiveOutputSettingsProperty struct { // The settings that are specific to the container type of the file. ContainerSettings interface{} `json:"containerSettings" yaml:"containerSettings"` // The output file extension. // // If excluded, this is auto-selected from the container type. Extension *string `json:"extension" yaml:"extension"` // A string that is concatenated to the end of the destination file name. // // The string is required for multiple outputs of the same type. NameModifier *string `json:"nameModifier" yaml:"nameModifier"` }
The archive output settings.
The parent of this entity is OutputSettings.
TODO: EXAMPLE
type CfnChannel_ArchiveS3SettingsProperty ¶
type CfnChannel_ArchiveS3SettingsProperty struct { // Specify the canned ACL to apply to each S3 request. // // Defaults to none. CannedAcl *string `json:"cannedAcl" yaml:"cannedAcl"` }
Sets up Amazon S3 as the destination for this Archive output.
The parent of this entity is ArchiveCdnSettings.
TODO: EXAMPLE
type CfnChannel_AribDestinationSettingsProperty ¶
type CfnChannel_AribDestinationSettingsProperty struct { }
The configuration of ARIB captions in the output.
The parent of this entity is CaptionDestinationSettings.
TODO: EXAMPLE
type CfnChannel_AribSourceSettingsProperty ¶
type CfnChannel_AribSourceSettingsProperty struct { }
Information about the ARIB captions to extract from the input.
The parent of this entity is CaptionSelectorSettings.
TODO: EXAMPLE
type CfnChannel_AudioChannelMappingProperty ¶
type CfnChannel_AudioChannelMappingProperty struct { // The indices and gain values for each input channel that should be remixed into this output channel. InputChannelLevels interface{} `json:"inputChannelLevels" yaml:"inputChannelLevels"` // The index of the output channel that is being produced. OutputChannel *float64 `json:"outputChannel" yaml:"outputChannel"` }
The settings for remixing audio.
The parent of this entity is RemixSettings.
TODO: EXAMPLE
type CfnChannel_AudioCodecSettingsProperty ¶
type CfnChannel_AudioCodecSettingsProperty struct { // The setup of the AAC audio codec in the output. AacSettings interface{} `json:"aacSettings" yaml:"aacSettings"` // The setup of an AC3 audio codec in the output. Ac3Settings interface{} `json:"ac3Settings" yaml:"ac3Settings"` // The setup of an EAC3 audio codec in the output. Eac3Settings interface{} `json:"eac3Settings" yaml:"eac3Settings"` // The setup of an MP2 audio codec in the output. Mp2Settings interface{} `json:"mp2Settings" yaml:"mp2Settings"` // The setup to pass through the Dolby audio codec to the output. PassThroughSettings interface{} `json:"passThroughSettings" yaml:"passThroughSettings"` // Settings for audio encoded with the WAV codec. WavSettings interface{} `json:"wavSettings" yaml:"wavSettings"` }
The configuration of the audio codec in the audio output.
The parent of this entity is AudioDescription.
TODO: EXAMPLE
type CfnChannel_AudioDescriptionProperty ¶
type CfnChannel_AudioDescriptionProperty struct { // The advanced audio normalization settings. AudioNormalizationSettings interface{} `json:"audioNormalizationSettings" yaml:"audioNormalizationSettings"` // The name of the AudioSelector that is used as the source for this AudioDescription. AudioSelectorName *string `json:"audioSelectorName" yaml:"audioSelectorName"` // Applies only if audioTypeControl is useConfigured. // // The values for audioType are defined in ISO-IEC 13818-1. AudioType *string `json:"audioType" yaml:"audioType"` // Determines how audio type is determined. // // followInput: If the input contains an ISO 639 audioType, then that value is passed through to the output. If the input contains no ISO 639 audioType, the value in Audio Type is included in the output. useConfigured: The value in Audio Type is included in the output. Note that this field and audioType are both ignored if inputType is broadcasterMixedAd. AudioTypeControl *string `json:"audioTypeControl" yaml:"audioTypeControl"` // Settings to configure one or more solutions that insert audio watermarks in the audio encode. AudioWatermarkingSettings interface{} `json:"audioWatermarkingSettings" yaml:"audioWatermarkingSettings"` // The audio codec settings. CodecSettings interface{} `json:"codecSettings" yaml:"codecSettings"` // Indicates the language of the audio output track. // // Used only if languageControlMode is useConfigured, or there is no ISO 639 language code specified in the input. LanguageCode *string `json:"languageCode" yaml:"languageCode"` // Choosing followInput causes the ISO 639 language code of the output to follow the ISO 639 language code of the input. // // The languageCode setting is used when useConfigured is set, or when followInput is selected but there is no ISO 639 language code specified by the input. LanguageCodeControl *string `json:"languageCodeControl" yaml:"languageCodeControl"` // The name of this AudioDescription. // // Outputs use this name to uniquely identify this AudioDescription. Description names should be unique within this channel. Name *string `json:"name" yaml:"name"` // The settings that control how input audio channels are remixed into the output audio channels. RemixSettings interface{} `json:"remixSettings" yaml:"remixSettings"` // Used for Microsoft Smooth and Apple HLS outputs. // // Indicates the name displayed by the player (for example, English or Director Commentary). StreamName *string `json:"streamName" yaml:"streamName"` }
The encoding information for one output audio.
The parent of this entity is EncoderSettings.
TODO: EXAMPLE
type CfnChannel_AudioHlsRenditionSelectionProperty ¶
type CfnChannel_AudioHlsRenditionSelectionProperty struct { // Specifies the GROUP-ID in the #EXT-X-MEDIA tag of the target HLS audio rendition. GroupId *string `json:"groupId" yaml:"groupId"` // Specifies the NAME in the #EXT-X-MEDIA tag of the target HLS audio rendition. Name *string `json:"name" yaml:"name"` }
Selector for HLS audio rendition.
The parent of this entity is AudioSelectorSettings.
TODO: EXAMPLE
type CfnChannel_AudioLanguageSelectionProperty ¶
type CfnChannel_AudioLanguageSelectionProperty struct { // Selects a specific three-letter language code from within an audio source. LanguageCode *string `json:"languageCode" yaml:"languageCode"` // When set to "strict," the transport stream demux strictly identifies audio streams by their language descriptor. // // If a PMT update occurs such that an audio stream matching the initially selected language is no longer present, then mute is encoded until the language returns. If set to "loose," then on a PMT update the demux chooses another audio stream in the program with the same stream type if it can't find one with the same language. LanguageSelectionPolicy *string `json:"languageSelectionPolicy" yaml:"languageSelectionPolicy"` }
Information about the audio language to extract.
The parent of this entity is AudioSelectorSettings.
TODO: EXAMPLE
type CfnChannel_AudioNormalizationSettingsProperty ¶
type CfnChannel_AudioNormalizationSettingsProperty struct { // The audio normalization algorithm to use. // // itu17701 conforms to the CALM Act specification. itu17702 conforms to the EBU R-128 specification. Algorithm *string `json:"algorithm" yaml:"algorithm"` // When set to correctAudio, the output audio is corrected using the chosen algorithm. // // If set to measureOnly, the audio is measured but not adjusted. AlgorithmControl *string `json:"algorithmControl" yaml:"algorithmControl"` // The Target LKFS(loudness) to adjust volume to. // // If no value is entered, a default value is used according to the chosen algorithm. The CALM Act (1770-1) recommends a target of -24 LKFS. The EBU R-128 specification (1770-2) recommends a target of -23 LKFS. TargetLkfs *float64 `json:"targetLkfs" yaml:"targetLkfs"` }
The settings for normalizing video.
The parent of this entity is AudioDescription.
TODO: EXAMPLE
type CfnChannel_AudioOnlyHlsSettingsProperty ¶
type CfnChannel_AudioOnlyHlsSettingsProperty struct { // Specifies the group that the audio rendition belongs to. AudioGroupId *string `json:"audioGroupId" yaml:"audioGroupId"` // Used with an audio-only stream. // // It must be a .jpg or .png file. If given, this image is used as the cover art for the audio-only output. Ideally, it should be formatted for an iPhone screen for two reasons. The iPhone does not resize the image; instead, it crops a centered image on the top/bottom and left/right. Additionally, this image file gets saved bit-for-bit into every 10-second segment file, so it increases bandwidth by {image file size} * {segment count} * {user count.}. AudioOnlyImage interface{} `json:"audioOnlyImage" yaml:"audioOnlyImage"` // Four types of audio-only tracks are supported: Audio-Only Variant Stream The client can play back this audio-only stream instead of video in low-bandwidth scenarios. // // Represented as an EXT-X-STREAM-INF in the HLS manifest. Alternate Audio, Auto Select, Default Alternate rendition that the client should try to play back by default. Represented as an EXT-X-MEDIA in the HLS manifest with DEFAULT=YES, AUTOSELECT=YES Alternate Audio, Auto Select, Not Default Alternate rendition that the client might try to play back by default. Represented as an EXT-X-MEDIA in the HLS manifest with DEFAULT=NO, AUTOSELECT=YES Alternate Audio, not Auto Select Alternate rendition that the client will not try to play back by default. Represented as an EXT-X-MEDIA in the HLS manifest with DEFAULT=NO, AUTOSELECT=NO. AudioTrackType *string `json:"audioTrackType" yaml:"audioTrackType"` // Specifies the segment type. SegmentType *string `json:"segmentType" yaml:"segmentType"` }
The configuration of an audio-only HLS output.
The parent of this entity is HlsSettings.
TODO: EXAMPLE
type CfnChannel_AudioPidSelectionProperty ¶
type CfnChannel_AudioPidSelectionProperty struct { // Select the audio by this PID. Pid *float64 `json:"pid" yaml:"pid"` }
Used to extract audio by The PID.
The parent of this entity is AudioSelectorSettings.
TODO: EXAMPLE
type CfnChannel_AudioSelectorProperty ¶
type CfnChannel_AudioSelectorProperty struct { // A name for this AudioSelector. Name *string `json:"name" yaml:"name"` // Information about the specific audio to extract from the input. SelectorSettings interface{} `json:"selectorSettings" yaml:"selectorSettings"` }
Information about one audio to extract from the input.
The parent of this entity is InputSettings.
TODO: EXAMPLE
type CfnChannel_AudioSelectorSettingsProperty ¶
type CfnChannel_AudioSelectorSettingsProperty struct { // Selector for HLS audio rendition. AudioHlsRenditionSelection interface{} `json:"audioHlsRenditionSelection" yaml:"audioHlsRenditionSelection"` // The language code of the audio to select. AudioLanguageSelection interface{} `json:"audioLanguageSelection" yaml:"audioLanguageSelection"` // The PID of the audio to select. AudioPidSelection interface{} `json:"audioPidSelection" yaml:"audioPidSelection"` // Information about the audio track to extract. AudioTrackSelection interface{} `json:"audioTrackSelection" yaml:"audioTrackSelection"` }
Information about the audio to extract from the input.
The parent of this entity is AudioSelector.
TODO: EXAMPLE
type CfnChannel_AudioSilenceFailoverSettingsProperty ¶
type CfnChannel_AudioSilenceFailoverSettingsProperty struct { // The name of the audio selector in the input that MediaLive should monitor to detect silence. // // Select your most important rendition. If you didn't create an audio selector in this input, leave blank. AudioSelectorName *string `json:"audioSelectorName" yaml:"audioSelectorName"` // The amount of time (in milliseconds) that the active input must be silent before automatic input failover occurs. // // Silence is defined as audio loss or audio quieter than -50 dBFS. AudioSilenceThresholdMsec *float64 `json:"audioSilenceThresholdMsec" yaml:"audioSilenceThresholdMsec"` }
MediaLive will perform a failover if audio is not detected in this input for the specified period.
The parent of this entity is FailoverConditionSettings.
TODO: EXAMPLE
type CfnChannel_AudioTrackProperty ¶
type CfnChannel_AudioTrackProperty struct { // 1-based integer value that maps to a specific audio track. Track *float64 `json:"track" yaml:"track"` }
Information about one audio track to extract. You can select multiple tracks.
The parent of this entity is AudioTrackSelection.
TODO: EXAMPLE
type CfnChannel_AudioTrackSelectionProperty ¶
type CfnChannel_AudioTrackSelectionProperty struct {
// Selects one or more unique audio tracks from within a source.
Tracks interface{} `json:"tracks" yaml:"tracks"`
}
Information about the audio track to extract.
The parent of this entity is AudioSelectorSettings.
TODO: EXAMPLE
type CfnChannel_AudioWatermarkSettingsProperty ¶
type CfnChannel_AudioWatermarkSettingsProperty struct {
// Settings to configure Nielsen Watermarks in the audio encode.
NielsenWatermarksSettings interface{} `json:"nielsenWatermarksSettings" yaml:"nielsenWatermarksSettings"`
}
Audio Watermark Settings.
The parent of this entity is AudioDescription.
TODO: EXAMPLE
type CfnChannel_AutomaticInputFailoverSettingsProperty ¶
type CfnChannel_AutomaticInputFailoverSettingsProperty struct { // This clear time defines the requirement a recovered input must meet to be considered healthy. // // The input must have no failover conditions for this length of time. Enter a time in milliseconds. This value is particularly important if the input_preference for the failover pair is set to PRIMARY_INPUT_PREFERRED, because after this time, MediaLive will switch back to the primary input. ErrorClearTimeMsec *float64 `json:"errorClearTimeMsec" yaml:"errorClearTimeMsec"` // A list of failover conditions. // // If any of these conditions occur, MediaLive will perform a failover to the other input. FailoverConditions interface{} `json:"failoverConditions" yaml:"failoverConditions"` // Input preference when deciding which input to make active when a previously failed input has recovered. InputPreference *string `json:"inputPreference" yaml:"inputPreference"` // The input ID of the secondary input in the automatic input failover pair. SecondaryInputId *string `json:"secondaryInputId" yaml:"secondaryInputId"` }
Settings to configure the conditions that will define the input as unhealthy and that will make MediaLive fail over to the other input in the input failover pair.
The parent of this entity is InputAttachment.
TODO: EXAMPLE
type CfnChannel_AvailBlankingProperty ¶
type CfnChannel_AvailBlankingProperty struct { // The blanking image to be used. // // Keep empty for solid black. Only .bmp and .png images are supported. AvailBlankingImage interface{} `json:"availBlankingImage" yaml:"availBlankingImage"` // When set to enabled, the video, audio, and captions are blanked when insertion metadata is added. State *string `json:"state" yaml:"state"` }
The configuration of ad avail blanking in the output.
The parent of this entity is EncoderSettings.
TODO: EXAMPLE
type CfnChannel_AvailConfigurationProperty ¶
type CfnChannel_AvailConfigurationProperty struct {
// The setup of ad avail handling in the output.
AvailSettings interface{} `json:"availSettings" yaml:"availSettings"`
}
The setup of ad avail handling in the output.
The parent of this entity is EncoderSettings.
TODO: EXAMPLE
type CfnChannel_AvailSettingsProperty ¶
type CfnChannel_AvailSettingsProperty struct { // The setup for SCTE-35 splice insert handling. Scte35SpliceInsert interface{} `json:"scte35SpliceInsert" yaml:"scte35SpliceInsert"` // The setup for SCTE-35 time signal APOS handling. Scte35TimeSignalApos interface{} `json:"scte35TimeSignalApos" yaml:"scte35TimeSignalApos"` }
The settings for the ad avail setup in the output.
The parent of this entity is AvailConfiguration.
TODO: EXAMPLE
type CfnChannel_BlackoutSlateProperty ¶
type CfnChannel_BlackoutSlateProperty struct { // The blackout slate image to be used. // // Keep empty for solid black. Only .bmp and .png images are supported. BlackoutSlateImage interface{} `json:"blackoutSlateImage" yaml:"blackoutSlateImage"` // Setting to enabled causes MediaLive to blackout the video, audio, and captions, and raise the "Network Blackout Image" slate when an SCTE104/35 Network End Segmentation Descriptor is encountered. // // The blackout is lifted when the Network Start Segmentation Descriptor is encountered. The Network End and Network Start descriptors must contain a network ID that matches the value entered in Network ID. NetworkEndBlackout *string `json:"networkEndBlackout" yaml:"networkEndBlackout"` // The path to the local file to use as the Network End Blackout image. // // The image is scaled to fill the entire output raster. NetworkEndBlackoutImage interface{} `json:"networkEndBlackoutImage" yaml:"networkEndBlackoutImage"` // Provides a Network ID that matches EIDR ID format (for example, "10.XXXX/XXXX-XXXX-XXXX-XXXX-XXXX-C"). NetworkId *string `json:"networkId" yaml:"networkId"` // When set to enabled, this causes video, audio, and captions to be blanked when indicated by program metadata. State *string `json:"state" yaml:"state"` }
The settings for a blackout slate.
The parent of this entity is EncoderSettings.
TODO: EXAMPLE
type CfnChannel_BurnInDestinationSettingsProperty ¶
type CfnChannel_BurnInDestinationSettingsProperty struct { // If no explicit xPosition or yPosition is provided, setting alignment to centered places the captions at the bottom center of the output. // // Similarly, setting a left alignment aligns captions to the bottom left of the output. If x and y positions are specified in conjunction with the alignment parameter, the font is justified (either left or centered) relative to those coordinates. Selecting "smart" justification left-justifies live subtitles and center-justifies pre-recorded subtitles. All burn-in and DVB-Sub font settings must match. Alignment *string `json:"alignment" yaml:"alignment"` // Specifies the color of the rectangle behind the captions. // // All burn-in and DVB-Sub font settings must match. BackgroundColor *string `json:"backgroundColor" yaml:"backgroundColor"` // Specifies the opacity of the background rectangle. // // 255 is opaque; 0 is transparent. Keeping this parameter blank is equivalent to setting it to 0 (transparent). All burn-in and DVB-Sub font settings must match. BackgroundOpacity *float64 `json:"backgroundOpacity" yaml:"backgroundOpacity"` // The external font file that is used for captions burn-in. // // The file extension must be .ttf or .tte. Although you can select output fonts for many different types of input captions, embedded, STL, and Teletext sources use a strict grid system. Using external fonts with these captions sources could cause an unexpected display of proportional fonts. All burn-in and DVB-Sub font settings must match. Font interface{} `json:"font" yaml:"font"` // Specifies the color of the burned-in captions. // // This option is not valid for source captions that are STL, 608/embedded, or Teletext. These source settings are already pre-defined by the captions stream. All burn-in and DVB-Sub font settings must match. FontColor *string `json:"fontColor" yaml:"fontColor"` // Specifies the opacity of the burned-in captions. // // 255 is opaque; 0 is transparent. All burn-in and DVB-Sub font settings must match. FontOpacity *float64 `json:"fontOpacity" yaml:"fontOpacity"` // The font resolution in DPI (dots per inch). // // The default is 96 dpi. All burn-in and DVB-Sub font settings must match. FontResolution *float64 `json:"fontResolution" yaml:"fontResolution"` // When set to auto, fontSize scales depending on the size of the output. // // Providing a positive integer specifies the exact font size in points. All burn-in and DVB-Sub font settings must match. FontSize *string `json:"fontSize" yaml:"fontSize"` // Specifies the font outline color. // // This option is not valid for source captions that are either 608/embedded or Teletext. These source settings are already pre-defined by the captions stream. All burn-in and DVB-Sub font settings must match. OutlineColor *string `json:"outlineColor" yaml:"outlineColor"` // Specifies font outline size in pixels. // // This option is not valid for source captions that are either 608/embedded or Teletext. These source settings are already pre-defined by the captions stream. All burn-in and DVB-Sub font settings must match. OutlineSize *float64 `json:"outlineSize" yaml:"outlineSize"` // Specifies the color of the shadow cast by the captions. // // All burn-in and DVB-Sub font settings must match. ShadowColor *string `json:"shadowColor" yaml:"shadowColor"` // Specifies the opacity of the shadow. // // 255 is opaque; 0 is transparent. Keeping this parameter blank is equivalent to setting it to 0 (transparent). All burn-in and DVB-Sub font settings must match. ShadowOpacity *float64 `json:"shadowOpacity" yaml:"shadowOpacity"` // Specifies the horizontal offset of the shadow that is relative to the captions in pixels. // // A value of -2 would result in a shadow offset 2 pixels to the left. All burn-in and DVB-Sub font settings must match. ShadowXOffset *float64 `json:"shadowXOffset" yaml:"shadowXOffset"` // Specifies the vertical offset of the shadow that is relative to the captions in pixels. // // A value of -2 would result in a shadow offset 2 pixels above the text. All burn-in and DVB-Sub font settings must match. ShadowYOffset *float64 `json:"shadowYOffset" yaml:"shadowYOffset"` // Controls whether a fixed grid size is used to generate the output subtitles bitmap. // // This applies only to Teletext inputs and DVB-Sub/Burn-in outputs. TeletextGridControl *string `json:"teletextGridControl" yaml:"teletextGridControl"` // Specifies the horizontal position of the captions relative to the left side of the output in pixels. // // A value of 10 would result in the captions starting 10 pixels from the left of the output. If no explicit xPosition is provided, the horizontal captions position is determined by the alignment parameter. All burn-in and DVB-Sub font settings must match. XPosition *float64 `json:"xPosition" yaml:"xPosition"` // Specifies the vertical position of the captions relative to the top of the output in pixels. // // A value of 10 would result in the captions starting 10 pixels from the top of the output. If no explicit yPosition is provided, the captions are positioned towards the bottom of the output. All burn-in and DVB-Sub font settings must match. YPosition *float64 `json:"yPosition" yaml:"yPosition"` }
The settings for burn-in captions in the output.
The parent of this entity is CaptionDestinationSettings.
TODO: EXAMPLE
type CfnChannel_CaptionDescriptionProperty ¶
type CfnChannel_CaptionDescriptionProperty struct { // Specifies which input captions selector to use as a captions source when generating output captions. // // This field should match a captionSelector name. CaptionSelectorName *string `json:"captionSelectorName" yaml:"captionSelectorName"` // Additional settings for a captions destination that depend on the destination type. DestinationSettings interface{} `json:"destinationSettings" yaml:"destinationSettings"` // An ISO 639-2 three-digit code. // // For more information, see http://www.loc.gov/standards/iso639-2/. LanguageCode *string `json:"languageCode" yaml:"languageCode"` // Human-readable information to indicate the captions that are available for players (for example, English or Spanish). LanguageDescription *string `json:"languageDescription" yaml:"languageDescription"` // The name of the captions description. // // The name is used to associate a captions description with an output. Names must be unique within a channel. Name *string `json:"name" yaml:"name"` }
The encoding information for output captions.
The parent of this entity is EncoderSettings.
TODO: EXAMPLE
type CfnChannel_CaptionDestinationSettingsProperty ¶
type CfnChannel_CaptionDestinationSettingsProperty struct { // The configuration of one ARIB captions encode in the output. AribDestinationSettings interface{} `json:"aribDestinationSettings" yaml:"aribDestinationSettings"` // The configuration of one burn-in captions encode in the output. BurnInDestinationSettings interface{} `json:"burnInDestinationSettings" yaml:"burnInDestinationSettings"` // The configuration of one DVB Sub captions encode in the output. DvbSubDestinationSettings interface{} `json:"dvbSubDestinationSettings" yaml:"dvbSubDestinationSettings"` // Settings for EBU-TT captions in the output. EbuTtDDestinationSettings interface{} `json:"ebuTtDDestinationSettings" yaml:"ebuTtDDestinationSettings"` // The configuration of one embedded captions encode in the output. EmbeddedDestinationSettings interface{} `json:"embeddedDestinationSettings" yaml:"embeddedDestinationSettings"` // The configuration of one embedded plus SCTE-20 captions encode in the output. EmbeddedPlusScte20DestinationSettings interface{} `json:"embeddedPlusScte20DestinationSettings" yaml:"embeddedPlusScte20DestinationSettings"` // The configuration of one RTMPCaptionInfo captions encode in the output. RtmpCaptionInfoDestinationSettings interface{} `json:"rtmpCaptionInfoDestinationSettings" yaml:"rtmpCaptionInfoDestinationSettings"` // The configuration of one SCTE20 plus embedded captions encode in the output. Scte20PlusEmbeddedDestinationSettings interface{} `json:"scte20PlusEmbeddedDestinationSettings" yaml:"scte20PlusEmbeddedDestinationSettings"` // The configuration of one SCTE-27 captions encode in the output. Scte27DestinationSettings interface{} `json:"scte27DestinationSettings" yaml:"scte27DestinationSettings"` // The configuration of one SMPTE-TT captions encode in the output. SmpteTtDestinationSettings interface{} `json:"smpteTtDestinationSettings" yaml:"smpteTtDestinationSettings"` // The configuration of one Teletext captions encode in the output. TeletextDestinationSettings interface{} `json:"teletextDestinationSettings" yaml:"teletextDestinationSettings"` // The configuration of one TTML captions encode in the output. TtmlDestinationSettings interface{} `json:"ttmlDestinationSettings" yaml:"ttmlDestinationSettings"` // The configuration of one WebVTT captions encode in the output. WebvttDestinationSettings interface{} `json:"webvttDestinationSettings" yaml:"webvttDestinationSettings"` }
The configuration of one captions encode in the output.
The parent of this entity is CaptionDescription.
TODO: EXAMPLE
type CfnChannel_CaptionLanguageMappingProperty ¶
type CfnChannel_CaptionLanguageMappingProperty struct { // The closed caption channel being described by this CaptionLanguageMapping. // // Each channel mapping must have a unique channel number (maximum of 4). CaptionChannel *float64 `json:"captionChannel" yaml:"captionChannel"` // A three-character ISO 639-2 language code (see http://www.loc.gov/standards/iso639-2). LanguageCode *string `json:"languageCode" yaml:"languageCode"` // The textual description of language. LanguageDescription *string `json:"languageDescription" yaml:"languageDescription"` }
Maps a captions channel to an ISO 693-2 language code (http://www.loc.gov/standards/iso639-2), with an optional description.
The parent of this entity is HlsGroupSettings.
TODO: EXAMPLE
type CfnChannel_CaptionRectangleProperty ¶
type CfnChannel_CaptionRectangleProperty struct { // See the description in leftOffset. // // For height, specify the entire height of the rectangle as a percentage of the underlying frame height. For example, \"80\" means the rectangle height is 80% of the underlying frame height. The topOffset and rectangleHeight must add up to 100% or less. This field corresponds to tts:extent - Y in the TTML standard. Height *float64 `json:"height" yaml:"height"` // Applies only if you plan to convert these source captions to EBU-TT-D or TTML in an output. // // (Make sure to leave the default if you don't have either of these formats in the output.) You can define a display rectangle for the captions that is smaller than the underlying video frame. You define the rectangle by specifying the position of the left edge, top edge, bottom edge, and right edge of the rectangle, all within the underlying video frame. The units for the measurements are percentages. If you specify a value for one of these fields, you must specify a value for all of them. // // For leftOffset, specify the position of the left edge of the rectangle, as a percentage of the underlying frame width, and relative to the left edge of the frame. For example, \"10\" means the measurement is 10% of the underlying frame width. The rectangle left edge starts at that position from the left edge of the frame. This field corresponds to tts:origin - X in the TTML standard. LeftOffset *float64 `json:"leftOffset" yaml:"leftOffset"` // See the description in leftOffset. // // For topOffset, specify the position of the top edge of the rectangle, as a percentage of the underlying frame height, and relative to the top edge of the frame. For example, \"10\" means the measurement is 10% of the underlying frame height. The rectangle top edge starts at that position from the top edge of the frame. This field corresponds to tts:origin - Y in the TTML standard. TopOffset *float64 `json:"topOffset" yaml:"topOffset"` // See the description in leftOffset. // // For width, specify the entire width of the rectangle as a percentage of the underlying frame width. For example, \"80\" means the rectangle width is 80% of the underlying frame width. The leftOffset and rectangleWidth must add up to 100% or less. This field corresponds to tts:extent - X in the TTML standard. Width *float64 `json:"width" yaml:"width"` }
Settings to configure the caption rectangle for an output captions that will be created using this Teletext source captions.
The parent of this entity is TeletextSourceSettings.
TODO: EXAMPLE
type CfnChannel_CaptionSelectorProperty ¶
type CfnChannel_CaptionSelectorProperty struct { // When specified, this field indicates the three-letter language code of the captions track to extract from the source. LanguageCode *string `json:"languageCode" yaml:"languageCode"` // The name identifier for a captions selector. // // This name is used to associate this captions selector with one or more captions descriptions. Names must be unique within a channel. Name *string `json:"name" yaml:"name"` // Information about the specific audio to extract from the input. SelectorSettings interface{} `json:"selectorSettings" yaml:"selectorSettings"` }
Information about one caption to extract from the input.
The parent of this entity is InputSettings.
TODO: EXAMPLE
type CfnChannel_CaptionSelectorSettingsProperty ¶
type CfnChannel_CaptionSelectorSettingsProperty struct { // Information about the ancillary captions to extract from the input. AncillarySourceSettings interface{} `json:"ancillarySourceSettings" yaml:"ancillarySourceSettings"` // Information about the ARIB captions to extract from the input. AribSourceSettings interface{} `json:"aribSourceSettings" yaml:"aribSourceSettings"` // Information about the DVB Sub captions to extract from the input. DvbSubSourceSettings interface{} `json:"dvbSubSourceSettings" yaml:"dvbSubSourceSettings"` // Information about the embedded captions to extract from the input. EmbeddedSourceSettings interface{} `json:"embeddedSourceSettings" yaml:"embeddedSourceSettings"` // Information about the SCTE-20 captions to extract from the input. Scte20SourceSettings interface{} `json:"scte20SourceSettings" yaml:"scte20SourceSettings"` // Information about the SCTE-27 captions to extract from the input. Scte27SourceSettings interface{} `json:"scte27SourceSettings" yaml:"scte27SourceSettings"` // Information about the Teletext captions to extract from the input. TeletextSourceSettings interface{} `json:"teletextSourceSettings" yaml:"teletextSourceSettings"` }
Captions Selector Settings.
The parent of this entity is CaptionSelector.
TODO: EXAMPLE
type CfnChannel_CdiInputSpecificationProperty ¶
type CfnChannel_CdiInputSpecificationProperty struct { // Maximum CDI input resolution. Resolution *string `json:"resolution" yaml:"resolution"` }
The input specification for this channel.
It specifies the key characteristics of CDI inputs for this channel, when those characteristics are different from other inputs.
This entity is at the top level in the channel.
TODO: EXAMPLE
type CfnChannel_ColorSpacePassthroughSettingsProperty ¶
type CfnChannel_ColorSpacePassthroughSettingsProperty struct { }
Passthrough applies no color space conversion to the output.
The parents of this entity are H264ColorSpaceSettings and H265ColorSpaceSettings.
TODO: EXAMPLE
type CfnChannel_DvbNitSettingsProperty ¶
type CfnChannel_DvbNitSettingsProperty struct { // The numeric value placed in the Network Information Table (NIT). NetworkId *float64 `json:"networkId" yaml:"networkId"` // The network name text placed in the networkNameDescriptor inside the Network Information Table (NIT). // // The maximum length is 256 characters. NetworkName *string `json:"networkName" yaml:"networkName"` // The number of milliseconds between instances of this table in the output transport stream. RepInterval *float64 `json:"repInterval" yaml:"repInterval"` }
The configuration of DVB NIT.
The parent of this entity is M2tsSettings.
TODO: EXAMPLE
type CfnChannel_DvbSdtSettingsProperty ¶
type CfnChannel_DvbSdtSettingsProperty struct { // Selects a method of inserting SDT information into an output stream. // // The sdtFollow setting copies SDT information from input stream to output stream. The sdtFollowIfPresent setting copies SDT information from input stream to output stream if SDT information is present in the input. Otherwise, it falls back on the user-defined values. The sdtManual setting means that the user will enter the SDT information. The sdtNone setting means that the output stream will not contain SDT information. OutputSdt *string `json:"outputSdt" yaml:"outputSdt"` // The number of milliseconds between instances of this table in the output transport stream. RepInterval *float64 `json:"repInterval" yaml:"repInterval"` // The service name placed in the serviceDescriptor in the Service Description Table (SDT). // // The maximum length is 256 characters. ServiceName *string `json:"serviceName" yaml:"serviceName"` // The service provider name placed in the serviceDescriptor in the Service Description Table (SDT). // // The maximum length is 256 characters. ServiceProviderName *string `json:"serviceProviderName" yaml:"serviceProviderName"` }
A DVB Service Description Table (SDT).
The parent of this entity is M2tsSettings.
TODO: EXAMPLE
type CfnChannel_DvbSubDestinationSettingsProperty ¶
type CfnChannel_DvbSubDestinationSettingsProperty struct { // If no explicit xPosition or yPosition is provided, setting the alignment to centered places the captions at the bottom center of the output. // // Similarly, setting a left alignment aligns captions to the bottom left of the output. If x and y positions are specified in conjunction with the alignment parameter, the font is justified (either left or centered) relative to those coordinates. Selecting "smart" justification left-justifies live subtitles and center-justifies pre-recorded subtitles. This option is not valid for source captions that are STL or 608/embedded. These source settings are already pre-defined by the captions stream. All burn-in and DVB-Sub font settings must match. Alignment *string `json:"alignment" yaml:"alignment"` // Specifies the color of the rectangle behind the captions. // // All burn-in and DVB-Sub font settings must match. BackgroundColor *string `json:"backgroundColor" yaml:"backgroundColor"` // Specifies the opacity of the background rectangle. // // 255 is opaque; 0 is transparent. Keeping this parameter blank is equivalent to setting it to 0 (transparent). All burn-in and DVB-Sub font settings must match. BackgroundOpacity *float64 `json:"backgroundOpacity" yaml:"backgroundOpacity"` // The external font file that is used for captions burn-in. // // The file extension must be .ttf or .tte. Although you can select output fonts for many different types of input captions, embedded, STL, and Teletext sources use a strict grid system. Using external fonts with these captions sources could cause an unexpected display of proportional fonts. All burn-in and DVB-Sub font settings must match. Font interface{} `json:"font" yaml:"font"` // Specifies the color of the burned-in captions. // // This option is not valid for source captions that are STL, 608/embedded, or Teletext. These source settings are already pre-defined by the captions stream. All burn-in and DVB-Sub font settings must match. FontColor *string `json:"fontColor" yaml:"fontColor"` // Specifies the opacity of the burned-in captions. // // 255 is opaque; 0 is transparent. All burn-in and DVB-Sub font settings must match. FontOpacity *float64 `json:"fontOpacity" yaml:"fontOpacity"` // The font resolution in DPI (dots per inch). // // The default is 96 dpi. All burn-in and DVB-Sub font settings must match. FontResolution *float64 `json:"fontResolution" yaml:"fontResolution"` // When set to auto, fontSize scales depending on the size of the output. // // Providing a positive integer specifies the exact font size in points. All burn-in and DVB-Sub font settings must match. FontSize *string `json:"fontSize" yaml:"fontSize"` // Specifies the font outline color. // // This option is not valid for source captions that are either 608/embedded or Teletext. These source settings are already pre-defined by the captions stream. All burn-in and DVB-Sub font settings must match. OutlineColor *string `json:"outlineColor" yaml:"outlineColor"` // Specifies the font outline size in pixels. // // This option is not valid for source captions that are either 608/embedded or Teletext. These source settings are already pre-defined by the captions stream. All burn-in and DVB-Sub font settings must match. OutlineSize *float64 `json:"outlineSize" yaml:"outlineSize"` // Specifies the color of the shadow that is cast by the captions. // // All burn-in and DVB-Sub font settings must match. ShadowColor *string `json:"shadowColor" yaml:"shadowColor"` // Specifies the opacity of the shadow. // // 255 is opaque; 0 is transparent. Keeping this parameter blank is equivalent to setting it to 0 (transparent). All burn-in and DVB-Sub font settings must match. ShadowOpacity *float64 `json:"shadowOpacity" yaml:"shadowOpacity"` // Specifies the horizontal offset of the shadow relative to the captions in pixels. // // A value of -2 would result in a shadow offset 2 pixels to the left. All burn-in and DVB-Sub font settings must match. ShadowXOffset *float64 `json:"shadowXOffset" yaml:"shadowXOffset"` // Specifies the vertical offset of the shadow relative to the captions in pixels. // // A value of -2 would result in a shadow offset 2 pixels above the text. All burn-in and DVB-Sub font settings must match. ShadowYOffset *float64 `json:"shadowYOffset" yaml:"shadowYOffset"` // Controls whether a fixed grid size is used to generate the output subtitles bitmap. // // This applies to only Teletext inputs and DVB-Sub/Burn-in outputs. TeletextGridControl *string `json:"teletextGridControl" yaml:"teletextGridControl"` // Specifies the horizontal position of the captions relative to the left side of the output in pixels. // // A value of 10 would result in the captions starting 10 pixels from the left of the output. If no explicit xPosition is provided, the horizontal captions position is determined by the alignment parameter. This option is not valid for source captions that are STL, 608/embedded, or Teletext. These source settings are already pre-defined by the captions stream. All burn-in and DVB-Sub font settings must match. XPosition *float64 `json:"xPosition" yaml:"xPosition"` // Specifies the vertical position of the captions relative to the top of the output in pixels. // // A value of 10 would result in the captions starting 10 pixels from the top of the output. If no explicit yPosition is provided, the captions are positioned towards the bottom of the output. This option is not valid for source captions that are STL, 608/embedded, or Teletext. These source settings are already pre-defined by the captions stream. All burn-in and DVB-Sub font settings must match. YPosition *float64 `json:"yPosition" yaml:"yPosition"` }
The settings for DVB Sub captions in the output.
The parent of this entity is CaptionDestinationSettings.
TODO: EXAMPLE
type CfnChannel_DvbSubSourceSettingsProperty ¶
type CfnChannel_DvbSubSourceSettingsProperty struct { // If you will configure a WebVTT caption description that references this caption selector, use this field to provide the language to consider when translating the image-based source to text. OcrLanguage *string `json:"ocrLanguage" yaml:"ocrLanguage"` // When using DVB-Sub with burn-in or SMPTE-TT, use this PID for the source content. // // It is unused for DVB-Sub passthrough. All DVB-Sub content is passed through, regardless of selectors. Pid *float64 `json:"pid" yaml:"pid"` }
Information about the DVB Sub captions to extract from the input.
The parent of this entity is CaptionSelectorSettings.
TODO: EXAMPLE
type CfnChannel_DvbTdtSettingsProperty ¶
type CfnChannel_DvbTdtSettingsProperty struct { // The number of milliseconds between instances of this table in the output transport stream. RepInterval *float64 `json:"repInterval" yaml:"repInterval"` }
The DVB Time and Date Table (TDT).
The parent of this entity is M2tsSettings.
TODO: EXAMPLE
type CfnChannel_Eac3SettingsProperty ¶
type CfnChannel_Eac3SettingsProperty struct { // When set to attenuate3Db, applies a 3 dB attenuation to the surround channels. // // Used only for the 3/2 coding mode. AttenuationControl *string `json:"attenuationControl" yaml:"attenuationControl"` // The average bitrate in bits/second. // // Valid bitrates depend on the coding mode. Bitrate *float64 `json:"bitrate" yaml:"bitrate"` // Specifies the bitstream mode (bsmod) for the emitted E-AC-3 stream. // // For more information, see ATSC A/52-2012 (Annex E). BitstreamMode *string `json:"bitstreamMode" yaml:"bitstreamMode"` // The Dolby Digital Plus coding mode. // // This mode determines the number of channels. CodingMode *string `json:"codingMode" yaml:"codingMode"` // When set to enabled, activates a DC highpass filter for all input channels. DcFilter *string `json:"dcFilter" yaml:"dcFilter"` // Sets the dialnorm for the output. // // If blank and the input audio is Dolby Digital Plus, dialnorm will be passed through. Dialnorm *float64 `json:"dialnorm" yaml:"dialnorm"` // Sets the Dolby dynamic range compression profile. DrcLine *string `json:"drcLine" yaml:"drcLine"` // Sets the profile for heavy Dolby dynamic range compression, ensuring that the instantaneous signal peaks do not exceed specified levels. DrcRf *string `json:"drcRf" yaml:"drcRf"` // When encoding 3/2 audio, setting to lfe enables the LFE channel. LfeControl *string `json:"lfeControl" yaml:"lfeControl"` // When set to enabled, applies a 120Hz lowpass filter to the LFE channel prior to encoding. // // Valid only with a codingMode32 coding mode. LfeFilter *string `json:"lfeFilter" yaml:"lfeFilter"` // The Left only/Right only center mix level. // // Used only for the 3/2 coding mode. LoRoCenterMixLevel *float64 `json:"loRoCenterMixLevel" yaml:"loRoCenterMixLevel"` // The Left only/Right only surround mix level. // // Used only for a 3/2 coding mode. LoRoSurroundMixLevel *float64 `json:"loRoSurroundMixLevel" yaml:"loRoSurroundMixLevel"` // The Left total/Right total center mix level. // // Used only for a 3/2 coding mode. LtRtCenterMixLevel *float64 `json:"ltRtCenterMixLevel" yaml:"ltRtCenterMixLevel"` // The Left total/Right total surround mix level. // // Used only for the 3/2 coding mode. LtRtSurroundMixLevel *float64 `json:"ltRtSurroundMixLevel" yaml:"ltRtSurroundMixLevel"` // When set to followInput, encoder metadata is sourced from the DD, DD+, or DolbyE decoder that supplies this audio data. // // If the audio is not supplied from one of these streams, then the static metadata settings are used. MetadataControl *string `json:"metadataControl" yaml:"metadataControl"` // When set to whenPossible, input DD+ audio will be passed through if it is present on the input. // // This detection is dynamic over the life of the transcode. Inputs that alternate between DD+ and non-DD+ content will have a consistent DD+ output as the system alternates between passthrough and encoding. PassthroughControl *string `json:"passthroughControl" yaml:"passthroughControl"` // When set to shift90Degrees, applies a 90-degree phase shift to the surround channels. // // Used only for a 3/2 coding mode. PhaseControl *string `json:"phaseControl" yaml:"phaseControl"` // A stereo downmix preference. // // Used only for the 3/2 coding mode. StereoDownmix *string `json:"stereoDownmix" yaml:"stereoDownmix"` // When encoding 3/2 audio, sets whether an extra center back surround channel is matrix encoded into the left and right surround channels. SurroundExMode *string `json:"surroundExMode" yaml:"surroundExMode"` // When encoding 2/0 audio, sets whether Dolby Surround is matrix-encoded into the two channels. SurroundMode *string `json:"surroundMode" yaml:"surroundMode"` }
The settings for an EAC3 audio encode in the output.
The parent of this entity is AudioCodecSettings.
TODO: EXAMPLE
type CfnChannel_EbuTtDDestinationSettingsProperty ¶
type CfnChannel_EbuTtDDestinationSettingsProperty struct { // Applies only if you plan to convert these source captions to EBU-TT-D or TTML in an output. // // Complete this field if you want to include the name of the copyright holder in the copyright metadata tag in the TTML CopyrightHolder *string `json:"copyrightHolder" yaml:"copyrightHolder"` // Specifies how to handle the gap between the lines (in multi-line captions). // // - enabled: Fill with the captions background color (as specified in the input captions). // - disabled: Leave the gap unfilled. FillLineGap *string `json:"fillLineGap" yaml:"fillLineGap"` // Specifies the font family to include in the font data attached to the EBU-TT captions. // // Valid only if styleControl is set to include. If you leave this field empty, the font family is set to "monospaced". (If styleControl is set to exclude, the font family is always set to "monospaced".) You specify only the font family. All other style information (color, bold, position and so on) is copied from the input captions. The size is always set to 100% to allow the downstream player to choose the size. - Enter a list of font families, as a comma-separated list of font names, in order of preference. The name can be a font family (such as “Arial”), or a generic font family (such as “serif”), or “default” (to let the downstream player choose the font). // - Leave blank to set the family to “monospace”. FontFamily *string `json:"fontFamily" yaml:"fontFamily"` // Specifies the style information (font color, font position, and so on) to include in the font data that is attached to the EBU-TT captions. // // - include: Take the style information (font color, font position, and so on) from the source captions and include that information in the font data attached to the EBU-TT captions. This option is valid only if the source captions are Embedded or Teletext. // - exclude: In the font data attached to the EBU-TT captions, set the font family to "monospaced". Do not include any other style information. StyleControl *string `json:"styleControl" yaml:"styleControl"` }
Settings for EBU-TT captions in the output.
The parent of this entity is CaptionDestinationSettings.
TODO: EXAMPLE
type CfnChannel_EmbeddedDestinationSettingsProperty ¶
type CfnChannel_EmbeddedDestinationSettingsProperty struct { }
The configuration of embedded captions in the output.
The parent of this entity is CaptionDestinationSettings.
TODO: EXAMPLE
type CfnChannel_EmbeddedPlusScte20DestinationSettingsProperty ¶
type CfnChannel_EmbeddedPlusScte20DestinationSettingsProperty struct { }
The settings for embedded plus SCTE-20 captions in the output.
The parent of this entity is CaptionDestinationSettings.
TODO: EXAMPLE
type CfnChannel_EmbeddedSourceSettingsProperty ¶
type CfnChannel_EmbeddedSourceSettingsProperty struct { // If this is upconvert, 608 data is both passed through the "608 compatibility bytes" fields of the 708 wrapper as well as translated into 708. // // If 708 data is present in the source content, it is discarded. Convert608To708 *string `json:"convert608To708" yaml:"convert608To708"` // Set to "auto" to handle streams with intermittent or non-aligned SCTE-20 and embedded captions. Scte20Detection *string `json:"scte20Detection" yaml:"scte20Detection"` // Specifies the 608/708 channel number within the video track from which to extract captions. // // This is unused for passthrough. Source608ChannelNumber *float64 `json:"source608ChannelNumber" yaml:"source608ChannelNumber"` // This field is unused and deprecated. Source608TrackNumber *float64 `json:"source608TrackNumber" yaml:"source608TrackNumber"` }
Information about the embedded captions to extract from the input.
The parent of this entity is CaptionSelectorSettings.
TODO: EXAMPLE
type CfnChannel_EncoderSettingsProperty ¶
type CfnChannel_EncoderSettingsProperty struct { // The encoding information for output audio. AudioDescriptions interface{} `json:"audioDescriptions" yaml:"audioDescriptions"` // The settings for ad avail blanking. AvailBlanking interface{} `json:"availBlanking" yaml:"availBlanking"` // The configuration settings for the ad avail handling. AvailConfiguration interface{} `json:"availConfiguration" yaml:"availConfiguration"` // The settings for the blackout slate. BlackoutSlate interface{} `json:"blackoutSlate" yaml:"blackoutSlate"` // The encoding information for output captions. CaptionDescriptions interface{} `json:"captionDescriptions" yaml:"captionDescriptions"` // Settings to enable specific features. FeatureActivations interface{} `json:"featureActivations" yaml:"featureActivations"` // The configuration settings that apply to the entire channel. GlobalConfiguration interface{} `json:"globalConfiguration" yaml:"globalConfiguration"` // Settings to enable and configure the motion graphics overlay feature in the channel. MotionGraphicsConfiguration interface{} `json:"motionGraphicsConfiguration" yaml:"motionGraphicsConfiguration"` // The settings to configure Nielsen watermarks. NielsenConfiguration interface{} `json:"nielsenConfiguration" yaml:"nielsenConfiguration"` // The settings for the output groups in the channel. OutputGroups interface{} `json:"outputGroups" yaml:"outputGroups"` // Contains settings used to acquire and adjust timecode information from the inputs. TimecodeConfig interface{} `json:"timecodeConfig" yaml:"timecodeConfig"` // The encoding information for output videos. VideoDescriptions interface{} `json:"videoDescriptions" yaml:"videoDescriptions"` }
The settings for the encoding of outputs.
This entity is at the top level in the channel.
TODO: EXAMPLE
type CfnChannel_FailoverConditionProperty ¶
type CfnChannel_FailoverConditionProperty struct {
// Settings for a specific failover condition.
FailoverConditionSettings interface{} `json:"failoverConditionSettings" yaml:"failoverConditionSettings"`
}
Failover Condition settings. There can be multiple failover conditions inside AutomaticInputFailoverSettings.
The parent of this entity is AutomaticInputFailoverSettings.
TODO: EXAMPLE
type CfnChannel_FailoverConditionSettingsProperty ¶
type CfnChannel_FailoverConditionSettingsProperty struct { // MediaLive will perform a failover if the specified audio selector is silent for the specified period. AudioSilenceSettings interface{} `json:"audioSilenceSettings" yaml:"audioSilenceSettings"` // MediaLive will perform a failover if content is not detected in this input for the specified period. InputLossSettings interface{} `json:"inputLossSettings" yaml:"inputLossSettings"` // MediaLive will perform a failover if content is considered black for the specified period. VideoBlackSettings interface{} `json:"videoBlackSettings" yaml:"videoBlackSettings"` }
Settings for one failover condition.
The parent of this entity is FailoverCondition.
TODO: EXAMPLE
type CfnChannel_FeatureActivationsProperty ¶
type CfnChannel_FeatureActivationsProperty struct { // Enables the Input Prepare feature. // // You can create Input Prepare actions in the schedule only if this feature is enabled. // If you disable the feature on an existing schedule, make sure that you first delete all input prepare actions from the schedule. InputPrepareScheduleActions *string `json:"inputPrepareScheduleActions" yaml:"inputPrepareScheduleActions"` }
Settings to enable specific features. You can't configure these features until you have enabled them in the channel.
The parent of this entity is EncoderSettings.
TODO: EXAMPLE
type CfnChannel_FecOutputSettingsProperty ¶
type CfnChannel_FecOutputSettingsProperty struct { // The parameter D from SMPTE 2022-1. // // The height of the FEC protection matrix. The number of transport stream packets per column error correction packet. The number must be between 4 and 20, inclusive. ColumnDepth *float64 `json:"columnDepth" yaml:"columnDepth"` // Enables column only or column and row-based FEC. IncludeFec *string `json:"includeFec" yaml:"includeFec"` // The parameter L from SMPTE 2022-1. // // The width of the FEC protection matrix. Must be between 1 and 20, inclusive. If only Column FEC is used, then larger values increase robustness. If Row FEC is used, then this is the number of transport stream packets per row error correction packet, and the value must be between 4 and 20, inclusive, if includeFec is columnAndRow. If includeFec is column, this value must be 1 to 20, inclusive. RowLength *float64 `json:"rowLength" yaml:"rowLength"` }
The settings for FEC.
The parent of this entity is UdpOutputSettings.
TODO: EXAMPLE
type CfnChannel_Fmp4HlsSettingsProperty ¶
type CfnChannel_Fmp4HlsSettingsProperty struct { // List all the audio groups that are used with the video output stream. // // Input all the audio GROUP-IDs that are associated to the video, separate by ','. AudioRenditionSets *string `json:"audioRenditionSets" yaml:"audioRenditionSets"` // If set to passthrough, Nielsen inaudible tones for media tracking will be detected in the input audio and an equivalent ID3 tag will be inserted in the output. NielsenId3Behavior *string `json:"nielsenId3Behavior" yaml:"nielsenId3Behavior"` // When set to passthrough, timed metadata is passed through from input to output. TimedMetadataBehavior *string `json:"timedMetadataBehavior" yaml:"timedMetadataBehavior"` }
Settings for the fMP4 containers.
The parent of this entity is HlsSettings.
TODO: EXAMPLE
type CfnChannel_FrameCaptureCdnSettingsProperty ¶
type CfnChannel_FrameCaptureCdnSettingsProperty struct {
// Sets up Amazon S3 as the destination for this Frame Capture output.
FrameCaptureS3Settings interface{} `json:"frameCaptureS3Settings" yaml:"frameCaptureS3Settings"`
}
Settings to configure the destination of a Frame Capture output.
The parent of this entity is FrameCaptureGroupSettings.
TODO: EXAMPLE
type CfnChannel_FrameCaptureGroupSettingsProperty ¶
type CfnChannel_FrameCaptureGroupSettingsProperty struct { // The destination for the frame capture files. // // The destination is either the URI for an Amazon S3 bucket and object, plus a file name prefix (for example, s3ssl://sportsDelivery/highlights/20180820/curling_) or the URI for a MediaStore container, plus a file name prefix (for example, mediastoressl://sportsDelivery/20180820/curling_). The final file names consist of the prefix from the destination field (for example, "curling_") + name modifier + the counter (5 digits, starting from 00001) + extension (which is always .jpg). For example, curlingLow.00001.jpg. Destination interface{} `json:"destination" yaml:"destination"` // Settings to configure the destination of a Frame Capture output. FrameCaptureCdnSettings interface{} `json:"frameCaptureCdnSettings" yaml:"frameCaptureCdnSettings"` }
The settings for a frame capture output group.
The parent of this entity is OutputGroupSettings.
TODO: EXAMPLE
type CfnChannel_FrameCaptureHlsSettingsProperty ¶
type CfnChannel_FrameCaptureHlsSettingsProperty struct { }
Settings for a frame capture output in an HLS output group.
The parent of this entity is HlsSettings.
TODO: EXAMPLE
type CfnChannel_FrameCaptureOutputSettingsProperty ¶
type CfnChannel_FrameCaptureOutputSettingsProperty struct { // Required if the output group contains more than one output. // // This modifier forms part of the output file name. NameModifier *string `json:"nameModifier" yaml:"nameModifier"` }
The frame capture output settings.
The parent of this entity is OutputSettings.
TODO: EXAMPLE
type CfnChannel_FrameCaptureS3SettingsProperty ¶
type CfnChannel_FrameCaptureS3SettingsProperty struct { // Specify the canned ACL to apply to each S3 request. // // Defaults to none. CannedAcl *string `json:"cannedAcl" yaml:"cannedAcl"` }
Sets up Amazon S3 as the destination for this Frame Capture output.
The parent of this entity is FrameCaptureCdnSettings.
TODO: EXAMPLE
type CfnChannel_FrameCaptureSettingsProperty ¶
type CfnChannel_FrameCaptureSettingsProperty struct { // The frequency, in seconds, for capturing frames for inclusion in the output. // // For example, "10" means capture a frame every 10 seconds. CaptureInterval *float64 `json:"captureInterval" yaml:"captureInterval"` // Unit for the frame capture interval. CaptureIntervalUnits *string `json:"captureIntervalUnits" yaml:"captureIntervalUnits"` }
The frame capture settings.
The parent of this entity is VideoCodecSettings.
TODO: EXAMPLE
type CfnChannel_GlobalConfigurationProperty ¶
type CfnChannel_GlobalConfigurationProperty struct { // The value to set the initial audio gain for the channel. InitialAudioGain *float64 `json:"initialAudioGain" yaml:"initialAudioGain"` // Indicates the action to take when the current input completes (for example, end-of-file). // // When switchAndLoopInputs is configured, MediaLive restarts at the beginning of the first input. When "none" is configured, MediaLive transcodes either black, a solid color, or a user-specified slate images per the "Input Loss Behavior" configuration until the next input switch occurs (which is controlled through the Channel Schedule API). InputEndAction *string `json:"inputEndAction" yaml:"inputEndAction"` // The settings for system actions when the input is lost. InputLossBehavior interface{} `json:"inputLossBehavior" yaml:"inputLossBehavior"` // Indicates how MediaLive pipelines are synchronized. // // PIPELINELOCKING - MediaLive attempts to synchronize the output of each pipeline to the other. EPOCHLOCKING - MediaLive attempts to synchronize the output of each pipeline to the Unix epoch. OutputLockingMode *string `json:"outputLockingMode" yaml:"outputLockingMode"` // Indicates whether the rate of frames emitted by the Live encoder should be paced by its system clock (which optionally might be locked to another source through NTP) or should be locked to the clock of the source that is providing the input stream. OutputTimingSource *string `json:"outputTimingSource" yaml:"outputTimingSource"` // Adjusts the video input buffer for streams with very low video frame rates. // // This is commonly set to enabled for music channels with less than one video frame per second. SupportLowFramerateInputs *string `json:"supportLowFramerateInputs" yaml:"supportLowFramerateInputs"` }
The configuration settings that apply to the entire channel.
The parent of this entity is EncoderSettings.
TODO: EXAMPLE
type CfnChannel_H264ColorSpaceSettingsProperty ¶
type CfnChannel_H264ColorSpaceSettingsProperty struct { // Passthrough applies no color space conversion to the output. ColorSpacePassthroughSettings interface{} `json:"colorSpacePassthroughSettings" yaml:"colorSpacePassthroughSettings"` // Settings to configure the handling of Rec601 color space. Rec601Settings interface{} `json:"rec601Settings" yaml:"rec601Settings"` // Settings to configure the handling of Rec709 color space. Rec709Settings interface{} `json:"rec709Settings" yaml:"rec709Settings"` }
Settings for configuring color space in an H264 video encode.
The parent of this entity is H264Settings.
TODO: EXAMPLE
type CfnChannel_H264FilterSettingsProperty ¶
type CfnChannel_H264FilterSettingsProperty struct {
// Settings for applying the temporal filter to the video.
TemporalFilterSettings interface{} `json:"temporalFilterSettings" yaml:"temporalFilterSettings"`
}
Settings to configure video filters that apply to the H264 codec.
The parent of this entity is H264Settings.
TODO: EXAMPLE
type CfnChannel_H264SettingsProperty ¶
type CfnChannel_H264SettingsProperty struct { // The adaptive quantization. // // This allows intra-frame quantizers to vary to improve visual quality. AdaptiveQuantization *string `json:"adaptiveQuantization" yaml:"adaptiveQuantization"` // Indicates that AFD values will be written into the output stream. // // If afdSignaling is auto, the system tries to preserve the input AFD value (in cases where multiple AFD values are valid). If set to fixed, the AFD value is the value configured in the fixedAfd parameter. AfdSignaling *string `json:"afdSignaling" yaml:"afdSignaling"` // The average bitrate in bits/second. // // This is required when the rate control mode is VBR or CBR. It isn't used for QVBR. In a Microsoft Smooth output group, each output must have a unique value when its bitrate is rounded down to the nearest multiple of 1000. Bitrate *float64 `json:"bitrate" yaml:"bitrate"` // The percentage of the buffer that should initially be filled (HRD buffer model). BufFillPct *float64 `json:"bufFillPct" yaml:"bufFillPct"` // The size of the buffer (HRD buffer model) in bits/second. BufSize *float64 `json:"bufSize" yaml:"bufSize"` // Includes color space metadata in the output. ColorMetadata *string `json:"colorMetadata" yaml:"colorMetadata"` // Settings to configure the color space handling for the video. ColorSpaceSettings interface{} `json:"colorSpaceSettings" yaml:"colorSpaceSettings"` // The entropy encoding mode. // // Use cabac (must be in Main or High profile) or cavlc. EntropyEncoding *string `json:"entropyEncoding" yaml:"entropyEncoding"` // Optional filters that you can apply to an encode. FilterSettings interface{} `json:"filterSettings" yaml:"filterSettings"` // A four-bit AFD value to write on all frames of video in the output stream. // // Valid only when afdSignaling is set to Fixed. FixedAfd *string `json:"fixedAfd" yaml:"fixedAfd"` // If set to enabled, adjusts the quantization within each frame to reduce flicker or pop on I-frames. FlickerAq *string `json:"flickerAq" yaml:"flickerAq"` // This setting applies only when scan type is "interlaced." It controls whether coding is performed on a field basis or on a frame basis. (When the video is progressive, the coding is always performed on a frame basis.) enabled: Force MediaLive to code on a field basis, so that odd and even sets of fields are coded separately. disabled: Code the two sets of fields separately (on a field basis) or together (on a frame basis using PAFF), depending on what is most appropriate for the content. ForceFieldPictures *string `json:"forceFieldPictures" yaml:"forceFieldPictures"` // Indicates how the output video frame rate is specified. // // If you select "specified," the output video frame rate is determined by framerateNumerator and framerateDenominator. If you select "initializeFromSource," the output video frame rate is set equal to the input video frame rate of the first input. FramerateControl *string `json:"framerateControl" yaml:"framerateControl"` // The frame rate denominator. FramerateDenominator *float64 `json:"framerateDenominator" yaml:"framerateDenominator"` // The frame rate numerator. // // The frame rate is a fraction, for example, 24000/1001 = 23.976 fps. FramerateNumerator *float64 `json:"framerateNumerator" yaml:"framerateNumerator"` // If enabled, uses reference B frames for GOP structures that have B frames > 1. GopBReference *string `json:"gopBReference" yaml:"gopBReference"` // The frequency of closed GOPs. // // In streaming applications, we recommend that you set this to 1 so that a decoder joining mid-stream will receive an IDR frame as quickly as possible. Setting this value to 0 will break output segmenting. GopClosedCadence *float64 `json:"gopClosedCadence" yaml:"gopClosedCadence"` // The number of B-frames between reference frames. GopNumBFrames *float64 `json:"gopNumBFrames" yaml:"gopNumBFrames"` // The GOP size (keyframe interval) in units of either frames or seconds per gopSizeUnits. // // The value must be greater than zero. GopSize *float64 `json:"gopSize" yaml:"gopSize"` // Indicates if the gopSize is specified in frames or seconds. // // If seconds, the system converts the gopSize into a frame count at runtime. GopSizeUnits *string `json:"gopSizeUnits" yaml:"gopSizeUnits"` // The H.264 level. Level *string `json:"level" yaml:"level"` // The amount of lookahead. // // A value of low can decrease latency and memory usage, while high can produce better quality for certain content. LookAheadRateControl *string `json:"lookAheadRateControl" yaml:"lookAheadRateControl"` // For QVBR: See the tooltip for Quality level. // // For VBR: Set the maximum bitrate in order to accommodate expected spikes in the complexity of the video. MaxBitrate *float64 `json:"maxBitrate" yaml:"maxBitrate"` // Meaningful only if sceneChangeDetect is set to enabled. // // This setting enforces separation between repeated (cadence) I-frames and I-frames inserted by Scene Change Detection. If a scene change I-frame is within I-interval frames of a cadence I-frame, the GOP is shrunk or stretched to the scene change I-frame. GOP stretch requires enabling lookahead as well as setting the I-interval. The normal cadence resumes for the next GOP. Note that the maximum GOP stretch = GOP size + Min-I-interval - 1. MinIInterval *float64 `json:"minIInterval" yaml:"minIInterval"` // The number of reference frames to use. // // The encoder might use more than requested if you use B-frames or interlaced encoding. NumRefFrames *float64 `json:"numRefFrames" yaml:"numRefFrames"` // Indicates how the output pixel aspect ratio is specified. // // If "specified" is selected, the output video pixel aspect ratio is determined by parNumerator and parDenominator. If "initializeFromSource" is selected, the output pixels aspect ratio will be set equal to the input video pixel aspect ratio of the first input. ParControl *string `json:"parControl" yaml:"parControl"` // The Pixel Aspect Ratio denominator. ParDenominator *float64 `json:"parDenominator" yaml:"parDenominator"` // The Pixel Aspect Ratio numerator. ParNumerator *float64 `json:"parNumerator" yaml:"parNumerator"` // An H.264 profile. Profile *string `json:"profile" yaml:"profile"` // Leave as STANDARD_QUALITY or choose a different value (which might result in additional costs to run the channel). // // - ENHANCED_QUALITY: Produces a slightly better video quality without an increase in the bitrate. Has an effect only when the Rate control mode is QVBR or CBR. If this channel is in a MediaLive multiplex, the value must be ENHANCED_QUALITY. // - STANDARD_QUALITY: Valid for any Rate control mode. QualityLevel *string `json:"qualityLevel" yaml:"qualityLevel"` // Controls the target quality for the video encode. // // This applies only when the rate control mode is QVBR. Set values for the QVBR quality level field and Max bitrate field that suit your most important viewing devices. Recommended values are: - Primary screen: Quality level: 8 to 10. Max bitrate: 4M - PC or tablet: Quality level: 7. Max bitrate: 1.5M to 3M - Smartphone: Quality level: 6. Max bitrate: 1M to 1.5M. QvbrQualityLevel *float64 `json:"qvbrQualityLevel" yaml:"qvbrQualityLevel"` // The rate control mode. // // QVBR: The quality will match the specified quality level except when it is constrained by the maximum bitrate. We recommend this if you or your viewers pay for bandwidth. VBR: The quality and bitrate vary, depending on the video complexity. We recommend this instead of QVBR if you want to maintain a specific average bitrate over the duration of the channel. CBR: The quality varies, depending on the video complexity. We recommend this only if you distribute your assets to devices that can't handle variable bitrates. RateControlMode *string `json:"rateControlMode" yaml:"rateControlMode"` // Sets the scan type of the output to progressive or top-field-first interlaced. ScanType *string `json:"scanType" yaml:"scanType"` // The scene change detection. // // On: inserts I-frames when the scene change is detected. Off: does not force an I-frame when the scene change is detected. SceneChangeDetect *string `json:"sceneChangeDetect" yaml:"sceneChangeDetect"` // The number of slices per picture. // // The number must be less than or equal to the number of macroblock rows for progressive pictures, and less than or equal to half the number of macroblock rows for interlaced pictures. This field is optional. If you don't specify a value, MediaLive chooses the number of slices based on the encode resolution. Slices *float64 `json:"slices" yaml:"slices"` // Softness. // // Selects a quantizer matrix. Larger values reduce high-frequency content in the encoded image. Softness *float64 `json:"softness" yaml:"softness"` // If set to enabled, adjusts quantization within each frame based on the spatial variation of content complexity. SpatialAq *string `json:"spatialAq" yaml:"spatialAq"` // If set to fixed, uses gopNumBFrames B-frames per sub-GOP. // // If set to dynamic, optimizes the number of B-frames used for each sub-GOP to improve visual quality. SubgopLength *string `json:"subgopLength" yaml:"subgopLength"` // Produces a bitstream that is compliant with SMPTE RP-2027. Syntax *string `json:"syntax" yaml:"syntax"` // If set to enabled, adjusts quantization within each frame based on the temporal variation of content complexity. TemporalAq *string `json:"temporalAq" yaml:"temporalAq"` // Determines how timecodes should be inserted into the video elementary stream. // // disabled: don't include timecodes. picTimingSei: pass through picture timing SEI messages from the source specified in Timecode Config. TimecodeInsertion *string `json:"timecodeInsertion" yaml:"timecodeInsertion"` }
The settings for the H.264 codec in the output.
The parent of this entity is VideoCodecSettings.
TODO: EXAMPLE
type CfnChannel_H265ColorSpaceSettingsProperty ¶
type CfnChannel_H265ColorSpaceSettingsProperty struct { // Passthrough applies no color space conversion to the output. ColorSpacePassthroughSettings interface{} `json:"colorSpacePassthroughSettings" yaml:"colorSpacePassthroughSettings"` // Settings to configure the handling of HDR10 color space. Hdr10Settings interface{} `json:"hdr10Settings" yaml:"hdr10Settings"` // Settings to configure the handling of Rec601 color space. Rec601Settings interface{} `json:"rec601Settings" yaml:"rec601Settings"` // Settings to configure the handling of Rec709 color space. Rec709Settings interface{} `json:"rec709Settings" yaml:"rec709Settings"` }
H265 Color Space Settings.
The parent of this entity is H265Settings.
TODO: EXAMPLE
type CfnChannel_H265FilterSettingsProperty ¶
type CfnChannel_H265FilterSettingsProperty struct {
// Settings for applying the temporal filter to the video.
TemporalFilterSettings interface{} `json:"temporalFilterSettings" yaml:"temporalFilterSettings"`
}
Settings to configure video filters that apply to the H265 codec.
The parent of this entity is H265Settings.
TODO: EXAMPLE
type CfnChannel_H265SettingsProperty ¶
type CfnChannel_H265SettingsProperty struct { // Adaptive quantization. // // Allows intra-frame quantizers to vary to improve visual quality. AdaptiveQuantization *string `json:"adaptiveQuantization" yaml:"adaptiveQuantization"` // Indicates that AFD values will be written into the output stream. // // If afdSignaling is "auto", the system will try to preserve the input AFD value (in cases where multiple AFD values are valid). If set to "fixed", the AFD value will be the value configured in the fixedAfd parameter. AfdSignaling *string `json:"afdSignaling" yaml:"afdSignaling"` // Whether or not EML should insert an Alternative Transfer Function SEI message to support backwards compatibility with non-HDR decoders and displays. AlternativeTransferFunction *string `json:"alternativeTransferFunction" yaml:"alternativeTransferFunction"` // Average bitrate in bits/second. // // Required when the rate control mode is VBR or CBR. Not used for QVBR. In an MS Smooth output group, each output must have a unique value when its bitrate is rounded down to the nearest multiple of 1000. Bitrate *float64 `json:"bitrate" yaml:"bitrate"` // Size of buffer (HRD buffer model) in bits. BufSize *float64 `json:"bufSize" yaml:"bufSize"` // Includes colorspace metadata in the output. ColorMetadata *string `json:"colorMetadata" yaml:"colorMetadata"` // Color Space settings. ColorSpaceSettings interface{} `json:"colorSpaceSettings" yaml:"colorSpaceSettings"` // Optional filters that you can apply to an encode. FilterSettings interface{} `json:"filterSettings" yaml:"filterSettings"` // Four bit AFD value to write on all frames of video in the output stream. // // Only valid when afdSignaling is set to 'Fixed'. FixedAfd *string `json:"fixedAfd" yaml:"fixedAfd"` // If set to enabled, adjust quantization within each frame to reduce flicker or 'pop' on I-frames. FlickerAq *string `json:"flickerAq" yaml:"flickerAq"` // Framerate denominator. FramerateDenominator *float64 `json:"framerateDenominator" yaml:"framerateDenominator"` // Framerate numerator - framerate is a fraction, e.g. 24000 / 1001 = 23.976 fps. FramerateNumerator *float64 `json:"framerateNumerator" yaml:"framerateNumerator"` // Frequency of closed GOPs. // // In streaming applications, it is recommended that this be set to 1 so a decoder joining mid-stream will receive an IDR frame as quickly as possible. Setting this value to 0 will break output segmenting. GopClosedCadence *float64 `json:"gopClosedCadence" yaml:"gopClosedCadence"` // GOP size (keyframe interval) in units of either frames or seconds per gopSizeUnits. // // If gopSizeUnits is frames, gopSize must be an integer and must be greater than or equal to 1. // If gopSizeUnits is seconds, gopSize must be greater than 0, but need not be an integer. GopSize *float64 `json:"gopSize" yaml:"gopSize"` // Indicates if the gopSize is specified in frames or seconds. // // If seconds the system will convert the gopSize into a frame count at run time. GopSizeUnits *string `json:"gopSizeUnits" yaml:"gopSizeUnits"` // H.265 Level. Level *string `json:"level" yaml:"level"` // Amount of lookahead. // // A value of low can decrease latency and memory usage, while high can produce better quality for certain content. LookAheadRateControl *string `json:"lookAheadRateControl" yaml:"lookAheadRateControl"` // For QVBR: See the tooltip for Quality level. MaxBitrate *float64 `json:"maxBitrate" yaml:"maxBitrate"` // Only meaningful if sceneChangeDetect is set to enabled. // // Defaults to 5 if multiplex rate control is used. Enforces separation between repeated (cadence) I-frames and I-frames inserted by Scene Change Detection. If a scene change I-frame is within I-interval frames of a cadence I-frame, the GOP is shrunk and/or stretched to the scene change I-frame. GOP stretch requires enabling lookahead as well as setting I-interval. The normal cadence resumes for the next GOP. Note: Maximum GOP stretch = GOP size + Min-I-interval - 1 MinIInterval *float64 `json:"minIInterval" yaml:"minIInterval"` // Pixel Aspect Ratio denominator. ParDenominator *float64 `json:"parDenominator" yaml:"parDenominator"` // Pixel Aspect Ratio numerator. ParNumerator *float64 `json:"parNumerator" yaml:"parNumerator"` // H.265 Profile. Profile *string `json:"profile" yaml:"profile"` // Controls the target quality for the video encode. // // Applies only when the rate control mode is QVBR. Set values for the QVBR quality level field and Max bitrate field that suit your most important viewing devices. Recommended values are: // - Primary screen: Quality level: 8 to 10. Max bitrate: 4M // - PC or tablet: Quality level: 7. Max bitrate: 1.5M to 3M // - Smartphone: Quality level: 6. Max bitrate: 1M to 1.5M QvbrQualityLevel *float64 `json:"qvbrQualityLevel" yaml:"qvbrQualityLevel"` // Rate control mode. // // QVBR: Quality will match the specified quality level except when it is constrained by the // maximum bitrate. Recommended if you or your viewers pay for bandwidth. CBR: Quality varies, depending on the video complexity. Recommended only if you distribute // your assets to devices that cannot handle variable bitrates. Multiplex: This rate control mode is only supported (and is required) when the video is being // delivered to a MediaLive Multiplex in which case the rate control configuration is controlled // by the properties within the Multiplex Program. RateControlMode *string `json:"rateControlMode" yaml:"rateControlMode"` // Sets the scan type of the output to progressive or top-field-first interlaced. ScanType *string `json:"scanType" yaml:"scanType"` // Scene change detection. SceneChangeDetect *string `json:"sceneChangeDetect" yaml:"sceneChangeDetect"` // Number of slices per picture. // // Must be less than or equal to the number of macroblock rows for progressive pictures, and less than or equal to half the number of macroblock rows for interlaced pictures. // This field is optional; when no value is specified the encoder will choose the number of slices based on encode resolution. Slices *float64 `json:"slices" yaml:"slices"` // H.265 Tier. Tier *string `json:"tier" yaml:"tier"` // Determines how timecodes should be inserted into the video elementary stream. // // - 'disabled': Do not include timecodes // - 'picTimingSei': Pass through picture timing SEI messages from the source specified in Timecode Config TimecodeInsertion *string `json:"timecodeInsertion" yaml:"timecodeInsertion"` }
H265 Settings.
The parent of this entity is VideoCodecSettings.
TODO: EXAMPLE
type CfnChannel_Hdr10SettingsProperty ¶
type CfnChannel_Hdr10SettingsProperty struct { // Maximum Content Light Level An integer metadata value defining the maximum light level, in nits, of any single pixel within an encoded HDR video stream or file. MaxCll *float64 `json:"maxCll" yaml:"maxCll"` // Maximum Frame Average Light Level An integer metadata value defining the maximum average light level, in nits, for any single frame within an encoded HDR video stream or file. MaxFall *float64 `json:"maxFall" yaml:"maxFall"` }
Hdr10 Settings.
The parents of this entity are H265ColorSpaceSettings (for color space settings in the output) and VideoSelectorColorSpaceSettings (for color space settings in the input).
TODO: EXAMPLE
type CfnChannel_HlsAkamaiSettingsProperty ¶
type CfnChannel_HlsAkamaiSettingsProperty struct { // The number of seconds to wait before retrying a connection to the CDN if the connection is lost. ConnectionRetryInterval *float64 `json:"connectionRetryInterval" yaml:"connectionRetryInterval"` // The size, in seconds, of the file cache for streaming outputs. FilecacheDuration *float64 `json:"filecacheDuration" yaml:"filecacheDuration"` // Specifies whether to use chunked transfer encoding to Akamai. // // To enable this feature, contact Akamai. HttpTransferMode *string `json:"httpTransferMode" yaml:"httpTransferMode"` // The number of retry attempts that will be made before the channel is put into an error state. NumRetries *float64 `json:"numRetries" yaml:"numRetries"` // If a streaming output fails, the number of seconds to wait until a restart is initiated. // // A value of 0 means never restart. RestartDelay *float64 `json:"restartDelay" yaml:"restartDelay"` // The salt for authenticated Akamai. Salt *string `json:"salt" yaml:"salt"` // The token parameter for authenticated Akamai. // // If this is not specified, _gda_ is used. Token *string `json:"token" yaml:"token"` }
The Akamai settings in an HLS output.
The parent of this entity is HlsCdnSettings.
TODO: EXAMPLE
type CfnChannel_HlsBasicPutSettingsProperty ¶
type CfnChannel_HlsBasicPutSettingsProperty struct { // The number of seconds to wait before retrying a connection to the CDN if the connection is lost. ConnectionRetryInterval *float64 `json:"connectionRetryInterval" yaml:"connectionRetryInterval"` // The size, in seconds, of the file cache for streaming outputs. FilecacheDuration *float64 `json:"filecacheDuration" yaml:"filecacheDuration"` // The number of retry attempts that MediaLive makes before the channel is put into an error state. NumRetries *float64 `json:"numRetries" yaml:"numRetries"` // If a streaming output fails, the number of seconds to wait until a restart is initiated. // // A value of 0 means never restart. RestartDelay *float64 `json:"restartDelay" yaml:"restartDelay"` }
The configuration of HLS Basic Put Settings.
The parent of this entity is HlsCdnSettings.
TODO: EXAMPLE
type CfnChannel_HlsCdnSettingsProperty ¶
type CfnChannel_HlsCdnSettingsProperty struct { // Sets up Akamai as the downstream system for the HLS output group. HlsAkamaiSettings interface{} `json:"hlsAkamaiSettings" yaml:"hlsAkamaiSettings"` // The settings for Basic Put for the HLS output. HlsBasicPutSettings interface{} `json:"hlsBasicPutSettings" yaml:"hlsBasicPutSettings"` // Sets up MediaStore as the destination for the HLS output. HlsMediaStoreSettings interface{} `json:"hlsMediaStoreSettings" yaml:"hlsMediaStoreSettings"` // Sets up Amazon S3 as the destination for this HLS output. HlsS3Settings interface{} `json:"hlsS3Settings" yaml:"hlsS3Settings"` // The settings for Web VTT captions in the HLS output group. // // The parent of this entity is HlsGroupSettings. HlsWebdavSettings interface{} `json:"hlsWebdavSettings" yaml:"hlsWebdavSettings"` }
The settings for the CDN of an HLS output.
The parent of this entity is HlsGroupSettings.
TODO: EXAMPLE
type CfnChannel_HlsGroupSettingsProperty ¶
type CfnChannel_HlsGroupSettingsProperty struct { // Chooses one or more ad marker types to pass SCTE35 signals through to this group of Apple HLS outputs. AdMarkers *[]*string `json:"adMarkers" yaml:"adMarkers"` // A partial URI prefix that will be prepended to each output in the media .m3u8 file. The partial URI prefix can be used if the base manifest is delivered from a different URL than the main .m3u8 file. BaseUrlContent *string `json:"baseUrlContent" yaml:"baseUrlContent"` // Optional. // // One value per output group. This field is required only if you are completing Base URL content A, and the downstream system has notified you that the media files for pipeline 1 of all outputs are in a location different from the media files for pipeline 0. BaseUrlContent1 *string `json:"baseUrlContent1" yaml:"baseUrlContent1"` // A partial URI prefix that will be prepended to each output in the media .m3u8 file. The partial URI prefix can be used if the base manifest is delivered from a different URL than the main .m3u8 file. BaseUrlManifest *string `json:"baseUrlManifest" yaml:"baseUrlManifest"` // Optional. // // One value per output group. Complete this field only if you are completing Base URL manifest A, and the downstream system has notified you that the child manifest files for pipeline 1 of all outputs are in a location different from the child manifest files for pipeline 0. BaseUrlManifest1 *string `json:"baseUrlManifest1" yaml:"baseUrlManifest1"` // A mapping of up to 4 captions channels to captions languages. // // This is meaningful only if captionLanguageSetting is set to "insert." CaptionLanguageMappings interface{} `json:"captionLanguageMappings" yaml:"captionLanguageMappings"` // Applies only to 608 embedded output captions. // // Insert: Include CLOSED-CAPTIONS lines in the manifest. Specify at least one language in the CC1 Language Code field. One CLOSED-CAPTION line is added for each Language Code that you specify. Make sure to specify the languages in the order in which they appear in the original source (if the source is embedded format) or the order of the captions selectors (if the source is other than embedded). Otherwise, languages in the manifest will not match properly with the output captions. None: Include the CLOSED-CAPTIONS=NONE line in the manifest. Omit: Omit any CLOSED-CAPTIONS line from the manifest. CaptionLanguageSetting *string `json:"captionLanguageSetting" yaml:"captionLanguageSetting"` // When set to "disabled," sets the #EXT-X-ALLOW-CACHE:no tag in the manifest, which prevents clients from saving media segments for later replay. ClientCache *string `json:"clientCache" yaml:"clientCache"` // The specification to use (RFC-6381 or the default RFC-4281) during m3u8 playlist generation. CodecSpecification *string `json:"codecSpecification" yaml:"codecSpecification"` // Used with encryptionType. // // This is a 128-bit, 16-byte hex value that is represented by a 32-character text string. If ivSource is set to "explicit," this parameter is required and is used as the IV for encryption. ConstantIv *string `json:"constantIv" yaml:"constantIv"` // A directory or HTTP destination for the HLS segments, manifest files, and encryption keys (if enabled). Destination interface{} `json:"destination" yaml:"destination"` // Places segments in subdirectories. DirectoryStructure *string `json:"directoryStructure" yaml:"directoryStructure"` // Specifies whether to insert EXT-X-DISCONTINUITY tags in the HLS child manifests for this output group. // // Typically, choose Insert because these tags are required in the manifest (according to the HLS specification) and serve an important purpose. // Choose Never Insert only if the downstream system is doing real-time failover (without using the MediaLive automatic failover feature) and only if that downstream system has advised you to exclude the tags. DiscontinuityTags *string `json:"discontinuityTags" yaml:"discontinuityTags"` // Encrypts the segments with the specified encryption scheme. // // Exclude this parameter if you don't want encryption. EncryptionType *string `json:"encryptionType" yaml:"encryptionType"` // The parameters that control interactions with the CDN. HlsCdnSettings interface{} `json:"hlsCdnSettings" yaml:"hlsCdnSettings"` // State of HLS ID3 Segment Tagging. HlsId3SegmentTagging *string `json:"hlsId3SegmentTagging" yaml:"hlsId3SegmentTagging"` // DISABLED: Don't create an I-frame-only manifest, but do create the master and media manifests (according to the Output Selection field). // // STANDARD: Create an I-frame-only manifest for each output that contains video, as well as the other manifests (according to the Output Selection field). The I-frame manifest contains a #EXT-X-I-FRAMES-ONLY tag to indicate it is I-frame only, and one or more #EXT-X-BYTERANGE entries identifying the I-frame position. For example, #EXT-X-BYTERANGE:160364@1461888". IFrameOnlyPlaylists *string `json:"iFrameOnlyPlaylists" yaml:"iFrameOnlyPlaylists"` // Specifies whether to include the final (incomplete) segment in the media output when the pipeline stops producing output because of a channel stop, a channel pause or a loss of input to the pipeline. // // Auto means that MediaLive decides whether to include the final segment, depending on the channel class and the types of output groups. // Suppress means to never include the incomplete segment. We recommend you choose Auto and let MediaLive control the behavior. IncompleteSegmentBehavior *string `json:"incompleteSegmentBehavior" yaml:"incompleteSegmentBehavior"` // Applies only if the Mode field is LIVE. // // Specifies the maximum number of segments in the media manifest file. After this maximum, older segments are removed from the media manifest. This number must be less than or equal to the Keep Segments field. IndexNSegments *float64 `json:"indexNSegments" yaml:"indexNSegments"` // A parameter that controls output group behavior on an input loss. InputLossAction *string `json:"inputLossAction" yaml:"inputLossAction"` // Used with encryptionType. // // The IV (initialization vector) is a 128-bit number used in conjunction with the key for encrypting blocks. If set to "include," the IV is listed in the manifest. Otherwise, the IV is not in the manifest. IvInManifest *string `json:"ivInManifest" yaml:"ivInManifest"` // Used with encryptionType. // // The IV (initialization vector) is a 128-bit number used in conjunction with the key for encrypting blocks. If this setting is "followsSegmentNumber," it causes the IV to change every segment (to match the segment number). If this is set to "explicit," you must enter a constantIv value. IvSource *string `json:"ivSource" yaml:"ivSource"` // Applies only if the Mode field is LIVE. // // Specifies the number of media segments (.ts files) to retain in the destination directory. KeepSegments *float64 `json:"keepSegments" yaml:"keepSegments"` // Specifies how the key is represented in the resource identified by the URI. // // If the parameter is absent, an implicit value of "identity" is used. A reverse DNS string can also be specified. KeyFormat *string `json:"keyFormat" yaml:"keyFormat"` // Either a single positive integer version value or a slash-delimited list of version values (1/2/3). KeyFormatVersions *string `json:"keyFormatVersions" yaml:"keyFormatVersions"` // The key provider settings. KeyProviderSettings interface{} `json:"keyProviderSettings" yaml:"keyProviderSettings"` // When set to gzip, compresses HLS playlist. ManifestCompression *string `json:"manifestCompression" yaml:"manifestCompression"` // Indicates whether the output manifest should use a floating point or integer values for segment duration. ManifestDurationFormat *string `json:"manifestDurationFormat" yaml:"manifestDurationFormat"` // When set, minimumSegmentLength is enforced by looking ahead and back within the specified range for a nearby avail and extending the segment size if needed. MinSegmentLength *float64 `json:"minSegmentLength" yaml:"minSegmentLength"` // If "vod," all segments are indexed and kept permanently in the destination and manifest. // // If "live," only the number segments specified in keepSegments and indexNSegments are kept. Newer segments replace older segments, which might prevent players from rewinding all the way to the beginning of the channel. VOD mode uses HLS EXT-X-PLAYLIST-TYPE of EVENT while the channel is running, converting it to a "VOD" type manifest on completion of the stream. Mode *string `json:"mode" yaml:"mode"` // MANIFESTSANDSEGMENTS: Generates manifests (the master manifest, if applicable, and media manifests) for this output group. // // SEGMENTSONLY: Doesn't generate any manifests for this output group. OutputSelection *string `json:"outputSelection" yaml:"outputSelection"` // Includes or excludes the EXT-X-PROGRAM-DATE-TIME tag in .m3u8 manifest files. The value is calculated as follows: Either the program date and time are initialized using the input timecode source, or the time is initialized using the input timecode source and the date is initialized using the timestampOffset. ProgramDateTime *string `json:"programDateTime" yaml:"programDateTime"` // `CfnChannel.HlsGroupSettingsProperty.ProgramDateTimeClock`. ProgramDateTimeClock *string `json:"programDateTimeClock" yaml:"programDateTimeClock"` // The period of insertion of the EXT-X-PROGRAM-DATE-TIME entry, in seconds. ProgramDateTimePeriod *float64 `json:"programDateTimePeriod" yaml:"programDateTimePeriod"` // ENABLED: The master manifest (.m3u8 file) for each pipeline includes information about both pipelines: first its own media files, then the media files of the other pipeline. This feature allows a playout device that supports stale manifest detection to switch from one manifest to the other, when the current manifest seems to be stale. There are still two destinations and two master manifests, but both master manifests reference the media files from both pipelines. DISABLED: The master manifest (.m3u8 file) for each pipeline includes information about its own pipeline only. For an HLS output group with MediaPackage as the destination, the DISABLED behavior is always followed. MediaPackage regenerates the manifests it serves to players, so a redundant manifest from MediaLive is irrelevant. RedundantManifest *string `json:"redundantManifest" yaml:"redundantManifest"` // useInputSegmentation has been deprecated. // // The configured segment size is always used. SegmentationMode *string `json:"segmentationMode" yaml:"segmentationMode"` // The length of the MPEG-2 Transport Stream segments to create, in seconds. // // Note that segments will end on the next keyframe after this number of seconds, so the actual segment length might be longer. SegmentLength *float64 `json:"segmentLength" yaml:"segmentLength"` // The number of segments to write to a subdirectory before starting a new one. // // For this setting to have an effect, directoryStructure must be subdirectoryPerStream. SegmentsPerSubdirectory *float64 `json:"segmentsPerSubdirectory" yaml:"segmentsPerSubdirectory"` // The include or exclude RESOLUTION attribute for a video in the EXT-X-STREAM-INF tag of a variant manifest. StreamInfResolution *string `json:"streamInfResolution" yaml:"streamInfResolution"` // Indicates the ID3 frame that has the timecode. TimedMetadataId3Frame *string `json:"timedMetadataId3Frame" yaml:"timedMetadataId3Frame"` // The timed metadata interval, in seconds. TimedMetadataId3Period *float64 `json:"timedMetadataId3Period" yaml:"timedMetadataId3Period"` // Provides an extra millisecond delta offset to fine tune the timestamps. TimestampDeltaMilliseconds *float64 `json:"timestampDeltaMilliseconds" yaml:"timestampDeltaMilliseconds"` // SEGMENTEDFILES: Emits the program as segments -multiple .ts media files. SINGLEFILE: Applies only if the Mode field is VOD. Emits the program as a single .ts media file. The media manifest includes #EXT-X-BYTERANGE tags to index segments for playback. A typical use for this value is when sending the output to AWS Elemental MediaConvert, which can accept only a single media file. Playback while the channel is running is not guaranteed due to HTTP server caching. TsFileMode *string `json:"tsFileMode" yaml:"tsFileMode"` }
The settings for an HLS output group.
The parent of this entity is OutputGroupSettings.
TODO: EXAMPLE
type CfnChannel_HlsInputSettingsProperty ¶
type CfnChannel_HlsInputSettingsProperty struct { // When specified, the HLS stream with the m3u8 bandwidth that most closely matches this value is chosen. // // Otherwise, the highest bandwidth stream in the m3u8 is chosen. The bitrate is specified in bits per second, as in an HLS manifest. Bandwidth *float64 `json:"bandwidth" yaml:"bandwidth"` // When specified, reading of the HLS input begins this many buffer segments from the end (most recently written segment). // // When not specified, the HLS input begins with the first segment specified in the m3u8. BufferSegments *float64 `json:"bufferSegments" yaml:"bufferSegments"` // The number of consecutive times that attempts to read a manifest or segment must fail before the input is considered unavailable. Retries *float64 `json:"retries" yaml:"retries"` // The number of seconds between retries when an attempt to read a manifest or segment fails. RetryInterval *float64 `json:"retryInterval" yaml:"retryInterval"` // Identifies the source for the SCTE-35 messages that MediaLive will ingest. // // Messages can be ingested from the content segments (in the stream) or from tags in the playlist (the HLS manifest). MediaLive ignores SCTE-35 information in the source that is not selected. Scte35Source *string `json:"scte35Source" yaml:"scte35Source"` }
Information about how to connect to the upstream system.
The parent of this entity is NetworkInputSettings.
TODO: EXAMPLE
type CfnChannel_HlsMediaStoreSettingsProperty ¶
type CfnChannel_HlsMediaStoreSettingsProperty struct { // The number of seconds to wait before retrying a connection to the CDN if the connection is lost. ConnectionRetryInterval *float64 `json:"connectionRetryInterval" yaml:"connectionRetryInterval"` // The size, in seconds, of the file cache for streaming outputs. FilecacheDuration *float64 `json:"filecacheDuration" yaml:"filecacheDuration"` // When set to temporal, output files are stored in non-persistent memory for faster reading and writing. MediaStoreStorageClass *string `json:"mediaStoreStorageClass" yaml:"mediaStoreStorageClass"` // The number of retry attempts that are made before the channel is put into an error state. NumRetries *float64 `json:"numRetries" yaml:"numRetries"` // If a streaming output fails, the number of seconds to wait until a restart is initiated. // // A value of 0 means never restart. RestartDelay *float64 `json:"restartDelay" yaml:"restartDelay"` }
The configuration of a MediaStore container as the destination for an HLS output.
The parent of this entity is HlsCdnSettings.
TODO: EXAMPLE
type CfnChannel_HlsOutputSettingsProperty ¶
type CfnChannel_HlsOutputSettingsProperty struct { // Only applicable when this output is referencing an H.265 video description. Specifies whether MP4 segments should be packaged as HEV1 or HVC1. H265PackagingType *string `json:"h265PackagingType" yaml:"h265PackagingType"` // The settings regarding the underlying stream. // // These settings are different for audio-only outputs. HlsSettings interface{} `json:"hlsSettings" yaml:"hlsSettings"` // A string that is concatenated to the end of the destination file name. // // Accepts \"Format Identifiers\":#formatIdentifierParameters. NameModifier *string `json:"nameModifier" yaml:"nameModifier"` // A string that is concatenated to the end of segment file names. SegmentModifier *string `json:"segmentModifier" yaml:"segmentModifier"` }
The settings for an HLS output.
The parent of this entity is OutputSettings.
TODO: EXAMPLE
type CfnChannel_HlsS3SettingsProperty ¶
type CfnChannel_HlsS3SettingsProperty struct { // Specify the canned ACL to apply to each S3 request. // // Defaults to none. CannedAcl *string `json:"cannedAcl" yaml:"cannedAcl"` }
Sets up Amazon S3 as the destination for this HLS output.
The parent of this entity is HlsCdnSettings.
TODO: EXAMPLE
type CfnChannel_HlsSettingsProperty ¶
type CfnChannel_HlsSettingsProperty struct { // The settings for an audio-only output. AudioOnlyHlsSettings interface{} `json:"audioOnlyHlsSettings" yaml:"audioOnlyHlsSettings"` // The settings for an fMP4 container. Fmp4HlsSettings interface{} `json:"fmp4HlsSettings" yaml:"fmp4HlsSettings"` // Settings for a frame capture output in an HLS output group. FrameCaptureHlsSettings interface{} `json:"frameCaptureHlsSettings" yaml:"frameCaptureHlsSettings"` // The settings for a standard output (an output that is not audio-only). StandardHlsSettings interface{} `json:"standardHlsSettings" yaml:"standardHlsSettings"` }
The settings for an HLS output.
The parent of this entity is HlsOutputSettings.
TODO: EXAMPLE
type CfnChannel_HlsWebdavSettingsProperty ¶
type CfnChannel_HlsWebdavSettingsProperty struct { // The number of seconds to wait before retrying a connection to the CDN if the connection is lost. ConnectionRetryInterval *float64 `json:"connectionRetryInterval" yaml:"connectionRetryInterval"` // The size, in seconds, of the file cache for streaming outputs. FilecacheDuration *float64 `json:"filecacheDuration" yaml:"filecacheDuration"` // Specifies whether to use chunked transfer encoding to WebDAV. HttpTransferMode *string `json:"httpTransferMode" yaml:"httpTransferMode"` // The number of retry attempts that are made before the channel is put into an error state. NumRetries *float64 `json:"numRetries" yaml:"numRetries"` // If a streaming output fails, the number of seconds to wait until a restart is initiated. // // A value of 0 means never restart. RestartDelay *float64 `json:"restartDelay" yaml:"restartDelay"` }
The configuration of a WebDav server as the downstream system for an HLS output.
The parent of this entity is HlsCdnSettings.
TODO: EXAMPLE
type CfnChannel_HtmlMotionGraphicsSettingsProperty ¶
type CfnChannel_HtmlMotionGraphicsSettingsProperty struct { }
Settings to configure the motion graphics overlay to use an HTML asset.
The parent of this entity is MotionGraphicsSetting.
TODO: EXAMPLE
type CfnChannel_InputAttachmentProperty ¶
type CfnChannel_InputAttachmentProperty struct { // Settings to implement automatic input failover in this input. AutomaticInputFailoverSettings interface{} `json:"automaticInputFailoverSettings" yaml:"automaticInputFailoverSettings"` // A name for the attachment. // // This is required if you want to use this input in an input switch action. InputAttachmentName *string `json:"inputAttachmentName" yaml:"inputAttachmentName"` // The ID of the input to attach. InputId *string `json:"inputId" yaml:"inputId"` // Information about the content to extract from the input and about the general handling of the content. InputSettings interface{} `json:"inputSettings" yaml:"inputSettings"` }
An input to attach to this channel.
This entity is at the top level in the channel.
TODO: EXAMPLE
type CfnChannel_InputChannelLevelProperty ¶
type CfnChannel_InputChannelLevelProperty struct { // The remixing value. // // Units are in dB, and acceptable values are within the range from -60 (mute) to 6 dB. Gain *float64 `json:"gain" yaml:"gain"` // The index of the input channel that is used as a source. InputChannel *float64 `json:"inputChannel" yaml:"inputChannel"` }
The setting to remix the audio.
The parent of this entity is AudioChannelMappings.
TODO: EXAMPLE
type CfnChannel_InputLocationProperty ¶
type CfnChannel_InputLocationProperty struct { // The password parameter that holds the password for accessing the downstream system. // // This applies only if the downstream system requires credentials. PasswordParam *string `json:"passwordParam" yaml:"passwordParam"` // The URI should be a path to a file that is accessible to the Live system (for example, an http:// URI) depending on the output type. // // For example, an RTMP destination should have a URI similar to rtmp://fmsserver/live. Uri *string `json:"uri" yaml:"uri"` // The user name to connect to the downstream system. // // This applies only if the downstream system requires credentials. Username *string `json:"username" yaml:"username"` }
The input location.
The parent of this entity is InputLossBehavior.
TODO: EXAMPLE
type CfnChannel_InputLossBehaviorProperty ¶
type CfnChannel_InputLossBehaviorProperty struct { // On input loss, the number of milliseconds to substitute black into the output before switching to the frame specified by inputLossImageType. // // A value x, where 0 <= x <= 1,000,000 and a value of 1,000,000, is interpreted as infinite. BlackFrameMsec *float64 `json:"blackFrameMsec" yaml:"blackFrameMsec"` // When the input loss image type is "color," this field specifies the color to use. // // Value: 6 hex characters that represent the values of RGB. InputLossImageColor *string `json:"inputLossImageColor" yaml:"inputLossImageColor"` // When the input loss image type is "slate," these fields specify the parameters for accessing the slate. InputLossImageSlate interface{} `json:"inputLossImageSlate" yaml:"inputLossImageSlate"` // Indicates whether to substitute a solid color or a slate into the output after the input loss exceeds blackFrameMsec. InputLossImageType *string `json:"inputLossImageType" yaml:"inputLossImageType"` // On input loss, the number of milliseconds to repeat the previous picture before substituting black into the output. // // A value x, where 0 <= x <= 1,000,000 and a value of 1,000,000, is interpreted as infinite. RepeatFrameMsec *float64 `json:"repeatFrameMsec" yaml:"repeatFrameMsec"` }
The configuration of channel behavior when the input is lost.
The parent of this entity is GlobalConfiguration.
TODO: EXAMPLE
type CfnChannel_InputLossFailoverSettingsProperty ¶
type CfnChannel_InputLossFailoverSettingsProperty struct { // The amount of time (in milliseconds) that no input is detected. // // After that time, an input failover will occur. InputLossThresholdMsec *float64 `json:"inputLossThresholdMsec" yaml:"inputLossThresholdMsec"` }
MediaLive will perform a failover if content is not detected in this input for the specified period.
The parent of this entity is FailoverConditionSettings.
TODO: EXAMPLE
type CfnChannel_InputSettingsProperty ¶
type CfnChannel_InputSettingsProperty struct { // Information about the specific audio to extract from the input. // // The parent of this entity is InputSettings. AudioSelectors interface{} `json:"audioSelectors" yaml:"audioSelectors"` // Information about the specific captions to extract from the input. CaptionSelectors interface{} `json:"captionSelectors" yaml:"captionSelectors"` // Enables or disables the deblock filter when filtering. DeblockFilter *string `json:"deblockFilter" yaml:"deblockFilter"` // Enables or disables the denoise filter when filtering. DenoiseFilter *string `json:"denoiseFilter" yaml:"denoiseFilter"` // Adjusts the magnitude of filtering from 1 (minimal) to 5 (strongest). FilterStrength *float64 `json:"filterStrength" yaml:"filterStrength"` // Turns on the filter for this input. // // MPEG-2 inputs have the deblocking filter enabled by default. 1) auto - filtering is applied depending on input type/quality 2) disabled - no filtering is applied to the input 3) forced - filtering is applied regardless of the input type. InputFilter *string `json:"inputFilter" yaml:"inputFilter"` // Information about how to connect to the upstream system. NetworkInputSettings interface{} `json:"networkInputSettings" yaml:"networkInputSettings"` // `CfnChannel.InputSettingsProperty.Scte35Pid`. Scte35Pid *float64 `json:"scte35Pid" yaml:"scte35Pid"` // Specifies whether to extract applicable ancillary data from a SMPTE-2038 source in this input. // // Applicable data types are captions, timecode, AFD, and SCTE-104 messages. // - PREFER: Extract from SMPTE-2038 if present in this input, otherwise extract from another source (if any). // - IGNORE: Never extract any ancillary data from SMPTE-2038. Smpte2038DataPreference *string `json:"smpte2038DataPreference" yaml:"smpte2038DataPreference"` // The loop input if it is a file. SourceEndBehavior *string `json:"sourceEndBehavior" yaml:"sourceEndBehavior"` // Information about one video to extract from the input. VideoSelector interface{} `json:"videoSelector" yaml:"videoSelector"` }
Information about extracting content from the input and about handling the content.
The parent of this entity is InputAttachment.
TODO: EXAMPLE
type CfnChannel_InputSpecificationProperty ¶
type CfnChannel_InputSpecificationProperty struct { // The codec to include in the input specification for this channel. Codec *string `json:"codec" yaml:"codec"` // The maximum input bitrate for any input attached to this channel. MaximumBitrate *string `json:"maximumBitrate" yaml:"maximumBitrate"` // The resolution for any input attached to this channel. Resolution *string `json:"resolution" yaml:"resolution"` }
The input specification for this channel.
It specifies the key characteristics of the inputs for this channel: the maximum bitrate, the resolution, and the codec.
This entity is at the top level in the channel.
TODO: EXAMPLE
type CfnChannel_KeyProviderSettingsProperty ¶
type CfnChannel_KeyProviderSettingsProperty struct {
// The configuration of static key settings.
StaticKeySettings interface{} `json:"staticKeySettings" yaml:"staticKeySettings"`
}
The configuration of key provider settings.
The parent of this entity is HlsGroupSettings.
TODO: EXAMPLE
type CfnChannel_M2tsSettingsProperty ¶
type CfnChannel_M2tsSettingsProperty struct { // When set to drop, the output audio streams are removed from the program if the selected input audio stream is removed from the input. // // This allows the output audio configuration to dynamically change based on the input configuration. If this is set to encodeSilence, all output audio streams will output encoded silence when not connected to an active input stream. AbsentInputAudioBehavior *string `json:"absentInputAudioBehavior" yaml:"absentInputAudioBehavior"` // When set to enabled, uses ARIB-compliant field muxing and removes video descriptor. Arib *string `json:"arib" yaml:"arib"` // The PID for ARIB Captions in the transport stream. // // You can enter the value as a decimal or hexadecimal value. Valid values are 32 (or 0x20)..8182 (or 0x1ff6). AribCaptionsPid *string `json:"aribCaptionsPid" yaml:"aribCaptionsPid"` // If set to auto, The PID number used for ARIB Captions will be auto-selected from unused PIDs. // // If set to useConfigured, ARIB captions will be on the configured PID number. AribCaptionsPidControl *string `json:"aribCaptionsPidControl" yaml:"aribCaptionsPidControl"` // When set to dvb, uses the DVB buffer model for Dolby Digital audio. // // When set to atsc, the ATSC model is used. AudioBufferModel *string `json:"audioBufferModel" yaml:"audioBufferModel"` // The number of audio frames to insert for each PES packet. AudioFramesPerPes *float64 `json:"audioFramesPerPes" yaml:"audioFramesPerPes"` // The PID of the elementary audio streams in the transport stream. // // Multiple values are accepted, and can be entered in ranges or by comma separation. You can enter the value as a decimal or hexadecimal value. Each PID specified must be in the range of 32 (or 0x20)..8182 (or 0x1ff6). AudioPids *string `json:"audioPids" yaml:"audioPids"` // When set to atsc, uses stream type = 0x81 for AC3 and stream type = 0x87 for EAC3. // // When set to dvb, uses stream type = 0x06. AudioStreamType *string `json:"audioStreamType" yaml:"audioStreamType"` // The output bitrate of the transport stream in bits per second. // // Setting to 0 lets the muxer automatically determine the appropriate bitrate. Bitrate *float64 `json:"bitrate" yaml:"bitrate"` // If set to multiplex, uses the multiplex buffer model for accurate interleaving. // // Setting to bufferModel to none can lead to lower latency, but low-memory devices might not be able to play back the stream without interruptions. BufferModel *string `json:"bufferModel" yaml:"bufferModel"` // When set to enabled, generates captionServiceDescriptor in PMT. CcDescriptor *string `json:"ccDescriptor" yaml:"ccDescriptor"` // Inserts a DVB Network Information Table (NIT) at the specified table repetition interval. DvbNitSettings interface{} `json:"dvbNitSettings" yaml:"dvbNitSettings"` // Inserts a DVB Service Description Table (SDT) at the specified table repetition interval. DvbSdtSettings interface{} `json:"dvbSdtSettings" yaml:"dvbSdtSettings"` // The PID for the input source DVB Subtitle data to this output. // // Multiple values are accepted, and can be entered in ranges and/or by comma separation. You can enter the value as a decimal or hexadecimal value. Each PID specified must be in the range of 32 (or 0x20)..8182 (or 0x1ff6). DvbSubPids *string `json:"dvbSubPids" yaml:"dvbSubPids"` // Inserts DVB Time and Date Table (TDT) at the specified table repetition interval. DvbTdtSettings interface{} `json:"dvbTdtSettings" yaml:"dvbTdtSettings"` // The PID for the input source DVB Teletext data to this output. // // You can enter the value as a decimal or hexadecimal value. Valid values are 32 (or 0x20)..8182 (or 0x1ff6). DvbTeletextPid *string `json:"dvbTeletextPid" yaml:"dvbTeletextPid"` // If set to passthrough, passes any EBIF data from the input source to this output. Ebif *string `json:"ebif" yaml:"ebif"` // When videoAndFixedIntervals is selected, audio EBP markers are added to partitions 3 and 4. // // The interval between these additional markers is fixed, and is slightly shorter than the video EBP marker interval. This is only available when EBP Cablelabs segmentation markers are selected. Partitions 1 and 2 always follow the video interval. EbpAudioInterval *string `json:"ebpAudioInterval" yaml:"ebpAudioInterval"` // When set, enforces that Encoder Boundary Points do not come within the specified time interval of each other by looking ahead at input video. // // If another EBP is going to come in within the specified time interval, the current EBP is not emitted, and the segment is "stretched" to the next marker. The lookahead value does not add latency to the system. The channel must be configured elsewhere to create sufficient latency to make the lookahead accurate. EbpLookaheadMs *float64 `json:"ebpLookaheadMs" yaml:"ebpLookaheadMs"` // Controls placement of EBP on audio PIDs. // // If set to videoAndAudioPids, EBP markers are placed on the video PID and all audio PIDs. If set to videoPid, EBP markers are placed on only the video PID. EbpPlacement *string `json:"ebpPlacement" yaml:"ebpPlacement"` // This field is unused and deprecated. EcmPid *string `json:"ecmPid" yaml:"ecmPid"` // Includes or excludes the ES Rate field in the PES header. EsRateInPes *string `json:"esRateInPes" yaml:"esRateInPes"` // The PID for the input source ETV Platform data to this output. // // You can enter it as a decimal or hexadecimal value. Valid values are 32 (or 0x20) to 8182 (or 0x1ff6). EtvPlatformPid *string `json:"etvPlatformPid" yaml:"etvPlatformPid"` // The PID for input source ETV Signal data to this output. // // You can enter the value as a decimal or hexadecimal value. Valid values are 32 (or 0x20)..8182 (or 0x1ff6). EtvSignalPid *string `json:"etvSignalPid" yaml:"etvSignalPid"` // The length in seconds of each fragment. // // This is used only with EBP markers. FragmentTime *float64 `json:"fragmentTime" yaml:"fragmentTime"` // If set to passthrough, passes any KLV data from the input source to this output. Klv *string `json:"klv" yaml:"klv"` // The PID for the input source KLV data to this output. // // Multiple values are accepted, and can be entered in ranges or by comma separation. You can enter the value as a decimal or hexadecimal value. Each PID specified must be in the range of 32 (or 0x20)..8182 (or 0x1ff6). KlvDataPids *string `json:"klvDataPids" yaml:"klvDataPids"` // If set to passthrough, Nielsen inaudible tones for media tracking will be detected in the input audio and an equivalent ID3 tag will be inserted in the output. NielsenId3Behavior *string `json:"nielsenId3Behavior" yaml:"nielsenId3Behavior"` // The value, in bits per second, of extra null packets to insert into the transport stream. // // This can be used if a downstream encryption system requires periodic null packets. NullPacketBitrate *float64 `json:"nullPacketBitrate" yaml:"nullPacketBitrate"` // The number of milliseconds between instances of this table in the output transport stream. // // Valid values are 0, 10..1000. PatInterval *float64 `json:"patInterval" yaml:"patInterval"` // When set to pcrEveryPesPacket, a Program Clock Reference value is inserted for every Packetized Elementary Stream (PES) header. // // This parameter is effective only when the PCR PID is the same as the video or audio elementary stream. PcrControl *string `json:"pcrControl" yaml:"pcrControl"` // The maximum time, in milliseconds, between Program Clock References (PCRs) inserted into the transport stream. PcrPeriod *float64 `json:"pcrPeriod" yaml:"pcrPeriod"` // The PID of the Program Clock Reference (PCR) in the transport stream. // // When no value is given, MediaLive assigns the same value as the video PID. You can enter the value as a decimal or hexadecimal value. Valid values are 32 (or 0x20)..8182 (or 0x1ff6). PcrPid *string `json:"pcrPid" yaml:"pcrPid"` // The number of milliseconds between instances of this table in the output transport stream. // // Valid values are 0, 10..1000. PmtInterval *float64 `json:"pmtInterval" yaml:"pmtInterval"` // The PID for the Program Map Table (PMT) in the transport stream. // // You can enter the value as a decimal or hexadecimal value. Valid values are 32 (or 0x20)..8182 (or 0x1ff6). PmtPid *string `json:"pmtPid" yaml:"pmtPid"` // The value of the program number field in the Program Map Table (PMT). ProgramNum *float64 `json:"programNum" yaml:"programNum"` // When VBR, does not insert null packets into the transport stream to fill the specified bitrate. // // The bitrate setting acts as the maximum bitrate when VBR is set. RateMode *string `json:"rateMode" yaml:"rateMode"` // The PID for the input source SCTE-27 data to this output. // // Multiple values are accepted, and can be entered in ranges or by comma separation. You can enter the value as a decimal or hexadecimal value. Each PID specified must be in the range of 32 (or 0x20)..8182 (or 0x1ff6). Scte27Pids *string `json:"scte27Pids" yaml:"scte27Pids"` // Optionally passes SCTE-35 signals from the input source to this output. Scte35Control *string `json:"scte35Control" yaml:"scte35Control"` // The PID of the SCTE-35 stream in the transport stream. // // You can enter the value as a decimal or hexadecimal value. Valid values are 32 (or 0x20)..8182 (or 0x1ff6). Scte35Pid *string `json:"scte35Pid" yaml:"scte35Pid"` // Inserts segmentation markers at each segmentationTime period. // // raiSegstart sets the Random Access Indicator bit in the adaptation field. raiAdapt sets the RAI bit and adds the current timecode in the private data bytes. psiSegstart inserts PAT and PMT tables at the start of segments. ebp adds Encoder Boundary Point information to the adaptation field as per OpenCable specification OC-SP-EBP-I01-130118. ebpLegacy adds Encoder Boundary Point information to the adaptation field using a legacy proprietary format. SegmentationMarkers *string `json:"segmentationMarkers" yaml:"segmentationMarkers"` // The segmentation style parameter controls how segmentation markers are inserted into the transport stream. // // With avails, it is possible that segments might be truncated, which can influence where future segmentation markers are inserted. When a segmentation style of resetCadence is selected and a segment is truncated due to an avail, we will reset the segmentation cadence. This means the subsequent segment will have a duration of $segmentationTime seconds. When a segmentation style of maintainCadence is selected and a segment is truncated due to an avail, we will not reset the segmentation cadence. This means the subsequent segment will likely be truncated as well. However, all segments after that will have a duration of $segmentationTime seconds. Note that EBP lookahead is a slight exception to this rule. SegmentationStyle *string `json:"segmentationStyle" yaml:"segmentationStyle"` // The length, in seconds, of each segment. // // This is required unless markers is set to None_. SegmentationTime *float64 `json:"segmentationTime" yaml:"segmentationTime"` // When set to passthrough, timed metadata is passed through from input to output. TimedMetadataBehavior *string `json:"timedMetadataBehavior" yaml:"timedMetadataBehavior"` // The PID of the timed metadata stream in the transport stream. // // You can enter the value as a decimal or hexadecimal value. Valid values are 32 (or 0x20)..8182 (or 0x1ff6). TimedMetadataPid *string `json:"timedMetadataPid" yaml:"timedMetadataPid"` // The value of the transport stream ID field in the Program Map Table (PMT). TransportStreamId *float64 `json:"transportStreamId" yaml:"transportStreamId"` // The PID of the elementary video stream in the transport stream. // // You can enter the value as a decimal or hexadecimal value. Valid values are 32 (or 0x20)..8182 (or 0x1ff6). VideoPid *string `json:"videoPid" yaml:"videoPid"` }
The configuration of the M2TS in the output.
The parents of this entity are ArchiveContainerSettings and UdpContainerSettings.
TODO: EXAMPLE
type CfnChannel_M3u8SettingsProperty ¶
type CfnChannel_M3u8SettingsProperty struct { // The number of audio frames to insert for each PES packet. AudioFramesPerPes *float64 `json:"audioFramesPerPes" yaml:"audioFramesPerPes"` // The PID of the elementary audio streams in the transport stream. // // Multiple values are accepted, and can be entered in ranges or by comma separation. You can enter the value as a decimal or hexadecimal value. AudioPids *string `json:"audioPids" yaml:"audioPids"` // This parameter is unused and deprecated. EcmPid *string `json:"ecmPid" yaml:"ecmPid"` // If set to passthrough, Nielsen inaudible tones for media tracking will be detected in the input audio and an equivalent ID3 tag will be inserted in the output. NielsenId3Behavior *string `json:"nielsenId3Behavior" yaml:"nielsenId3Behavior"` // The number of milliseconds between instances of this table in the output transport stream. // // A value of \"0\" writes out the PMT once per segment file. PatInterval *float64 `json:"patInterval" yaml:"patInterval"` // When set to pcrEveryPesPacket, a Program Clock Reference value is inserted for every Packetized Elementary Stream (PES) header. // // This parameter is effective only when the PCR PID is the same as the video or audio elementary stream. PcrControl *string `json:"pcrControl" yaml:"pcrControl"` // The maximum time, in milliseconds, between Program Clock References (PCRs) inserted into the transport stream. PcrPeriod *float64 `json:"pcrPeriod" yaml:"pcrPeriod"` // The PID of the Program Clock Reference (PCR) in the transport stream. // // When no value is given, MediaLive assigns the same value as the video PID. You can enter the value as a decimal or hexadecimal value. PcrPid *string `json:"pcrPid" yaml:"pcrPid"` // The number of milliseconds between instances of this table in the output transport stream. // // A value of \"0\" writes out the PMT once per segment file. PmtInterval *float64 `json:"pmtInterval" yaml:"pmtInterval"` // The PID for the Program Map Table (PMT) in the transport stream. // // You can enter the value as a decimal or hexadecimal value. PmtPid *string `json:"pmtPid" yaml:"pmtPid"` // The value of the program number field in the Program Map Table (PMT). ProgramNum *float64 `json:"programNum" yaml:"programNum"` // If set to passthrough, passes any SCTE-35 signals from the input source to this output. Scte35Behavior *string `json:"scte35Behavior" yaml:"scte35Behavior"` // The PID of the SCTE-35 stream in the transport stream. // // You can enter the value as a decimal or hexadecimal value. Scte35Pid *string `json:"scte35Pid" yaml:"scte35Pid"` // When set to passthrough, timed metadata is passed through from input to output. TimedMetadataBehavior *string `json:"timedMetadataBehavior" yaml:"timedMetadataBehavior"` // The PID of the timed metadata stream in the transport stream. // // You can enter the value as a decimal or hexadecimal value. Valid values are 32 (or 0x20)..8182 (or 0x1ff6). TimedMetadataPid *string `json:"timedMetadataPid" yaml:"timedMetadataPid"` // The value of the transport stream ID field in the Program Map Table (PMT). TransportStreamId *float64 `json:"transportStreamId" yaml:"transportStreamId"` // The PID of the elementary video stream in the transport stream. // // You can enter the value as a decimal or hexadecimal value. VideoPid *string `json:"videoPid" yaml:"videoPid"` }
Settings for the M3U8 container.
The parent of this entity is StandardHlsSettings.
TODO: EXAMPLE
type CfnChannel_MediaPackageGroupSettingsProperty ¶
type CfnChannel_MediaPackageGroupSettingsProperty struct {
// The MediaPackage channel destination.
Destination interface{} `json:"destination" yaml:"destination"`
}
The settings for the MediaPackage group.
The parent of this entity is OutputGroupSettings.
TODO: EXAMPLE
type CfnChannel_MediaPackageOutputDestinationSettingsProperty ¶
type CfnChannel_MediaPackageOutputDestinationSettingsProperty struct { // The ID of the channel in MediaPackage that is the destination for this output group. // // You don't need to specify the individual inputs in MediaPackage; MediaLive handles the connection of the two MediaLive pipelines to the two MediaPackage inputs. The MediaPackage channel and MediaLive channel must be in the same Region. ChannelId *string `json:"channelId" yaml:"channelId"` }
Destination settings for a MediaPackage output.
The parent of this entity is OutputDestination.
TODO: EXAMPLE
type CfnChannel_MediaPackageOutputSettingsProperty ¶
type CfnChannel_MediaPackageOutputSettingsProperty struct { }
The settings for a MediaPackage output.
The parent of this entity is OutputSettings.
TODO: EXAMPLE
type CfnChannel_MotionGraphicsConfigurationProperty ¶
type CfnChannel_MotionGraphicsConfigurationProperty struct { // Enables or disables the motion graphics overlay feature in the channel. MotionGraphicsInsertion *string `json:"motionGraphicsInsertion" yaml:"motionGraphicsInsertion"` // Settings to enable and configure the motion graphics overlay feature in the channel. MotionGraphicsSettings interface{} `json:"motionGraphicsSettings" yaml:"motionGraphicsSettings"` }
Settings to enable and configure the motion graphics overlay feature in the channel.
The parent of this entity is EncoderSettings.
TODO: EXAMPLE
type CfnChannel_MotionGraphicsSettingsProperty ¶
type CfnChannel_MotionGraphicsSettingsProperty struct {
// Settings to configure the motion graphics overlay to use an HTML asset.
HtmlMotionGraphicsSettings interface{} `json:"htmlMotionGraphicsSettings" yaml:"htmlMotionGraphicsSettings"`
}
Settings to enable and configure the motion graphics overlay feature in the channel.
The parent of this entity is MotionGraphicsConfiguration.
TODO: EXAMPLE
type CfnChannel_Mp2SettingsProperty ¶
type CfnChannel_Mp2SettingsProperty struct { // The average bitrate in bits/second. Bitrate *float64 `json:"bitrate" yaml:"bitrate"` // The MPEG2 Audio coding mode. // // Valid values are codingMode10 (for mono) or codingMode20 (for stereo). CodingMode *string `json:"codingMode" yaml:"codingMode"` // The sample rate in Hz. SampleRate *float64 `json:"sampleRate" yaml:"sampleRate"` }
The configuration for this MP2 audio.
The parent of this entity is AudioCodecSettings.
TODO: EXAMPLE
type CfnChannel_Mpeg2FilterSettingsProperty ¶
type CfnChannel_Mpeg2FilterSettingsProperty struct {
// Settings for applying the temporal filter to the video.
TemporalFilterSettings interface{} `json:"temporalFilterSettings" yaml:"temporalFilterSettings"`
}
Settings to configure video filters that apply to the MPEG-2 codec.
The parent of this entity is Mpeg2FilterSettings.
TODO: EXAMPLE
type CfnChannel_Mpeg2SettingsProperty ¶
type CfnChannel_Mpeg2SettingsProperty struct { // Choose Off to disable adaptive quantization. // // Or choose another value to enable the quantizer and set its strength. The strengths are: Auto, Off, Low, Medium, High. When you enable this field, MediaLive allows intra-frame quantizers to vary, which might improve visual quality. AdaptiveQuantization *string `json:"adaptiveQuantization" yaml:"adaptiveQuantization"` // Indicates the AFD values that MediaLive will write into the video encode. // // If you do not know what AFD signaling is, or if your downstream system has not given you guidance, choose AUTO. // AUTO: MediaLive will try to preserve the input AFD value (in cases where multiple AFD values are valid). // FIXED: MediaLive will use the value you specify in fixedAFD. AfdSignaling *string `json:"afdSignaling" yaml:"afdSignaling"` // Specifies whether to include the color space metadata. // // The metadata describes the color space that applies to the video (the colorSpace field). We recommend that you insert the metadata. ColorMetadata *string `json:"colorMetadata" yaml:"colorMetadata"` // Choose the type of color space conversion to apply to the output. // // For detailed information on setting up both the input and the output to obtain the desired color space in the output, see the section on \"MediaLive Features - Video - color space\" in the MediaLive User Guide. // PASSTHROUGH: Keep the color space of the input content - do not convert it. // AUTO:Convert all content that is SD to rec 601, and convert all content that is HD to rec 709. ColorSpace *string `json:"colorSpace" yaml:"colorSpace"` // Sets the pixel aspect ratio for the encode. DisplayAspectRatio *string `json:"displayAspectRatio" yaml:"displayAspectRatio"` // Optionally specify a noise reduction filter, which can improve quality of compressed content. // // If you do not choose a filter, no filter will be applied. // TEMPORAL: This filter is useful for both source content that is noisy (when it has excessive digital artifacts) and source content that is clean. // When the content is noisy, the filter cleans up the source content before the encoding phase, with these two effects: First, it improves the output video quality because the content has been cleaned up. Secondly, it decreases the bandwidth because MediaLive does not waste bits on encoding noise. // When the content is reasonably clean, the filter tends to decrease the bitrate. FilterSettings interface{} `json:"filterSettings" yaml:"filterSettings"` // Complete this field only when afdSignaling is set to FIXED. // // Enter the AFD value (4 bits) to write on all frames of the video encode. FixedAfd *string `json:"fixedAfd" yaml:"fixedAfd"` // description": "The framerate denominator. // // For example, 1001. The framerate is the numerator divided by the denominator. For example, 24000 / 1001 = 23.976 FPS. FramerateDenominator *float64 `json:"framerateDenominator" yaml:"framerateDenominator"` // The framerate numerator. // // For example, 24000. The framerate is the numerator divided by the denominator. For example, 24000 / 1001 = 23.976 FPS. FramerateNumerator *float64 `json:"framerateNumerator" yaml:"framerateNumerator"` // MPEG2: default is open GOP. GopClosedCadence *float64 `json:"gopClosedCadence" yaml:"gopClosedCadence"` // Relates to the GOP structure. // // The number of B-frames between reference frames. If you do not know what a B-frame is, use the default. GopNumBFrames *float64 `json:"gopNumBFrames" yaml:"gopNumBFrames"` // Relates to the GOP structure. // // The GOP size (keyframe interval) in the units specified in gopSizeUnits. If you do not know what GOP is, use the default. // If gopSizeUnits is frames, then the gopSize must be an integer and must be greater than or equal to 1. // If gopSizeUnits is seconds, the gopSize must be greater than 0, but does not need to be an integer. GopSize *float64 `json:"gopSize" yaml:"gopSize"` // Relates to the GOP structure. // // Specifies whether the gopSize is specified in frames or seconds. If you do not plan to change the default gopSize, leave the default. If you specify SECONDS, MediaLive will internally convert the gop size to a frame count. GopSizeUnits *string `json:"gopSizeUnits" yaml:"gopSizeUnits"` // Set the scan type of the output to PROGRESSIVE or INTERLACED (top field first). ScanType *string `json:"scanType" yaml:"scanType"` // Relates to the GOP structure. // // If you do not know what GOP is, use the default. // FIXED: Set the number of B-frames in each sub-GOP to the value in gopNumBFrames. // DYNAMIC: Let MediaLive optimize the number of B-frames in each sub-GOP, to improve visual quality. SubgopLength *string `json:"subgopLength" yaml:"subgopLength"` // Determines how MediaLive inserts timecodes in the output video. // // For detailed information about setting up the input and the output for a timecode, see the section on \"MediaLive Features - Timecode configuration\" in the MediaLive User Guide. // DISABLED: do not include timecodes. // GOP_TIMECODE: Include timecode metadata in the GOP header. TimecodeInsertion *string `json:"timecodeInsertion" yaml:"timecodeInsertion"` }
The settings for the MPEG-2 codec in the output.
The parent of this entity is VideoCodecSetting.
TODO: EXAMPLE
type CfnChannel_MsSmoothGroupSettingsProperty ¶
type CfnChannel_MsSmoothGroupSettingsProperty struct { // The value of the Acquisition Point Identity element that is used in each message placed in the sparse track. // // Enabled only if sparseTrackType is not "none." AcquisitionPointId *string `json:"acquisitionPointId" yaml:"acquisitionPointId"` // If set to passthrough for an audio-only Microsoft Smooth output, the fragment absolute time is set to the current timecode. // // This option does not write timecodes to the audio elementary stream. AudioOnlyTimecodeControl *string `json:"audioOnlyTimecodeControl" yaml:"audioOnlyTimecodeControl"` // If set to verifyAuthenticity, verifies the HTTPS certificate chain to a trusted certificate authority (CA). // // This causes HTTPS outputs to self-signed certificates to fail. CertificateMode *string `json:"certificateMode" yaml:"certificateMode"` // The number of seconds to wait before retrying the connection to the IIS server if the connection is lost. // // Content is cached during this time, and the cache is delivered to the IIS server after the connection is re-established. ConnectionRetryInterval *float64 `json:"connectionRetryInterval" yaml:"connectionRetryInterval"` // The Smooth Streaming publish point on an IIS server. // // MediaLive acts as a "Push" encoder to IIS. Destination interface{} `json:"destination" yaml:"destination"` // The Microsoft Smooth channel ID that is sent to the IIS server. // // Specify the ID only if eventIdMode is set to useConfigured. EventId *string `json:"eventId" yaml:"eventId"` // Specifies whether to send a channel ID to the IIS server. // // If no channel ID is sent and the same channel is used without changing the publishing point, clients might see cached video from the previous run. Options: - "useConfigured" - use the value provided in eventId - "useTimestamp" - generate and send a channel ID based on the current timestamp - "noEventId" - do not send a channel ID to the IIS server. EventIdMode *string `json:"eventIdMode" yaml:"eventIdMode"` // When set to sendEos, sends an EOS signal to an IIS server when stopping the channel. EventStopBehavior *string `json:"eventStopBehavior" yaml:"eventStopBehavior"` // The size, in seconds, of the file cache for streaming outputs. FilecacheDuration *float64 `json:"filecacheDuration" yaml:"filecacheDuration"` // The length, in seconds, of mp4 fragments to generate. // // The fragment length must be compatible with GOP size and frame rate. FragmentLength *float64 `json:"fragmentLength" yaml:"fragmentLength"` // A parameter that controls output group behavior on an input loss. InputLossAction *string `json:"inputLossAction" yaml:"inputLossAction"` // The number of retry attempts. NumRetries *float64 `json:"numRetries" yaml:"numRetries"` // The number of seconds before initiating a restart due to output failure, due to exhausting the numRetries on one segment, or exceeding filecacheDuration. RestartDelay *float64 `json:"restartDelay" yaml:"restartDelay"` // useInputSegmentation has been deprecated. // // The configured segment size is always used. SegmentationMode *string `json:"segmentationMode" yaml:"segmentationMode"` // The number of milliseconds to delay the output from the second pipeline. SendDelayMs *float64 `json:"sendDelayMs" yaml:"sendDelayMs"` // If set to scte35, uses incoming SCTE-35 messages to generate a sparse track in this group of Microsoft Smooth outputs. SparseTrackType *string `json:"sparseTrackType" yaml:"sparseTrackType"` // When set to send, sends a stream manifest so that the publishing point doesn't start until all streams start. StreamManifestBehavior *string `json:"streamManifestBehavior" yaml:"streamManifestBehavior"` // The timestamp offset for the channel. // // Used only if timestampOffsetMode is set to useConfiguredOffset. TimestampOffset *string `json:"timestampOffset" yaml:"timestampOffset"` // The type of timestamp date offset to use. // // - useEventStartDate: Use the date the channel was started as the offset - useConfiguredOffset: Use an explicitly configured date as the offset. TimestampOffsetMode *string `json:"timestampOffsetMode" yaml:"timestampOffsetMode"` }
The settings for a Microsoft Smooth output group.
The parent of this entity is OutputGroupSettings.
TODO: EXAMPLE
type CfnChannel_MsSmoothOutputSettingsProperty ¶
type CfnChannel_MsSmoothOutputSettingsProperty struct { // Only applicable when this output is referencing an H.265 video description. Specifies whether MP4 segments should be packaged as HEV1 or HVC1. H265PackagingType *string `json:"h265PackagingType" yaml:"h265PackagingType"` // A string that is concatenated to the end of the destination file name. // // This is required for multiple outputs of the same type. NameModifier *string `json:"nameModifier" yaml:"nameModifier"` }
Configuration of a Microsoft Smooth output.
The parent of this entity is OutputSettings.
TODO: EXAMPLE
type CfnChannel_MultiplexGroupSettingsProperty ¶
type CfnChannel_MultiplexGroupSettingsProperty struct { }
The settings for a Multiplex output group.
The parent of this entity is OutputGroupSettings.
TODO: EXAMPLE
type CfnChannel_MultiplexOutputSettingsProperty ¶
type CfnChannel_MultiplexOutputSettingsProperty struct {
// Destination is a Multiplex.
Destination interface{} `json:"destination" yaml:"destination"`
}
Configuration of a Multiplex output.
The parent of this entity is OutputSettings.
TODO: EXAMPLE
type CfnChannel_MultiplexProgramChannelDestinationSettingsProperty ¶
type CfnChannel_MultiplexProgramChannelDestinationSettingsProperty struct { // The ID of the Multiplex that the encoder is providing output to. // // You do not need to specify the individual inputs to the Multiplex; MediaLive will handle the connection of the two MediaLive pipelines to the two Multiplex instances. // The Multiplex must be in the same region as the Channel. MultiplexId *string `json:"multiplexId" yaml:"multiplexId"` // The program name of the Multiplex program that the encoder is providing output to. ProgramName *string `json:"programName" yaml:"programName"` }
Destination settings for a Multiplex output.
The parent of this entity is OutputDestination.
TODO: EXAMPLE
type CfnChannel_NetworkInputSettingsProperty ¶
type CfnChannel_NetworkInputSettingsProperty struct { // Information about how to connect to the upstream system. HlsInputSettings interface{} `json:"hlsInputSettings" yaml:"hlsInputSettings"` // Checks HTTPS server certificates. // // When set to checkCryptographyOnly, cryptography in the certificate is checked, but not the server's name. Certain subdomains (notably S3 buckets that use dots in the bucket name) don't strictly match the corresponding certificate's wildcard pattern and would otherwise cause the channel to error. This setting is ignored for protocols that do not use HTTPS. ServerValidation *string `json:"serverValidation" yaml:"serverValidation"` }
Information about how to connect to the upstream system.
The parent of this entity is InputSettings.
TODO: EXAMPLE
type CfnChannel_NielsenCBETProperty ¶
type CfnChannel_NielsenCBETProperty struct { // Enter the CBET check digits to use in the watermark. CbetCheckDigitString *string `json:"cbetCheckDigitString" yaml:"cbetCheckDigitString"` // Determines the method of CBET insertion mode when prior encoding is detected on the same layer. CbetStepaside *string `json:"cbetStepaside" yaml:"cbetStepaside"` // Enter the CBET Source ID (CSID) to use in the watermark. Csid *string `json:"csid" yaml:"csid"` }
Complete these fields only if you want to insert watermarks of type Nielsen CBET.
The parent of this entity is NielsenWatermarksSettings ¶
TODO: EXAMPLE
type CfnChannel_NielsenConfigurationProperty ¶
type CfnChannel_NielsenConfigurationProperty struct { // Enter the Distributor ID assigned to your organization by Nielsen. DistributorId *string `json:"distributorId" yaml:"distributorId"` // Enables Nielsen PCM to ID3 tagging. NielsenPcmToId3Tagging *string `json:"nielsenPcmToId3Tagging" yaml:"nielsenPcmToId3Tagging"` }
The settings to configure Nielsen watermarks.
The parent of this entity is EncoderSettings.
TODO: EXAMPLE
type CfnChannel_NielsenNaesIiNwProperty ¶
type CfnChannel_NielsenNaesIiNwProperty struct { // Enter the check digit string for the watermark. CheckDigitString *string `json:"checkDigitString" yaml:"checkDigitString"` // Enter the Nielsen Source ID (SID) to include in the watermark. Sid *float64 `json:"sid" yaml:"sid"` }
Complete these fields only if you want to insert watermarks of type Nielsen NAES II (N2) and Nielsen NAES VI (NW).
The parent of this entity is NielsenWatermarksSettings.
TODO: EXAMPLE
type CfnChannel_NielsenWatermarksSettingsProperty ¶
type CfnChannel_NielsenWatermarksSettingsProperty struct { // Complete these fields only if you want to insert watermarks of type Nielsen CBET. NielsenCbetSettings interface{} `json:"nielsenCbetSettings" yaml:"nielsenCbetSettings"` // Choose the distribution types that you want to assign to the watermarks: - PROGRAM_CONTENT - FINAL_DISTRIBUTOR. NielsenDistributionType *string `json:"nielsenDistributionType" yaml:"nielsenDistributionType"` // Complete these fields only if you want to insert watermarks of type Nielsen NAES II (N2) and Nielsen NAES VI (NW). NielsenNaesIiNwSettings interface{} `json:"nielsenNaesIiNwSettings" yaml:"nielsenNaesIiNwSettings"` }
Settings to configure Nielsen Watermarks in the audio encode.
The parent of this entity is AudioWatermarkSettings.
TODO: EXAMPLE
type CfnChannel_OutputDestinationProperty ¶
type CfnChannel_OutputDestinationProperty struct { // The ID for this destination. Id *string `json:"id" yaml:"id"` // The destination settings for a MediaPackage output. MediaPackageSettings interface{} `json:"mediaPackageSettings" yaml:"mediaPackageSettings"` // Destination settings for a Multiplex output; // // one destination for both encoders. MultiplexSettings interface{} `json:"multiplexSettings" yaml:"multiplexSettings"` // The destination settings for an output. Settings interface{} `json:"settings" yaml:"settings"` }
Configuration information for an output.
This entity is at the top level in the channel.
TODO: EXAMPLE
type CfnChannel_OutputDestinationSettingsProperty ¶
type CfnChannel_OutputDestinationSettingsProperty struct { // The password parameter that holds the password for accessing the downstream system. // // This password parameter applies only if the downstream system requires credentials. PasswordParam *string `json:"passwordParam" yaml:"passwordParam"` // The stream name for the content. // // This applies only to RTMP outputs. StreamName *string `json:"streamName" yaml:"streamName"` // The URL for the destination. Url *string `json:"url" yaml:"url"` // The user name to connect to the downstream system. // // This applies only if the downstream system requires credentials. Username *string `json:"username" yaml:"username"` }
The configuration information for this output.
The parent of this entity is OutputDestination.
TODO: EXAMPLE
type CfnChannel_OutputGroupProperty ¶
type CfnChannel_OutputGroupProperty struct { // A custom output group name that you can optionally define. // // Only letters, numbers, and the underscore character are allowed. The maximum length is 32 characters. Name *string `json:"name" yaml:"name"` // The settings associated with the output group. OutputGroupSettings interface{} `json:"outputGroupSettings" yaml:"outputGroupSettings"` // The settings for the outputs in the output group. Outputs interface{} `json:"outputs" yaml:"outputs"` }
The settings for one output group.
The parent of this entity is EncoderSettings.
TODO: EXAMPLE
type CfnChannel_OutputGroupSettingsProperty ¶
type CfnChannel_OutputGroupSettingsProperty struct { // The configuration of an archive output group. // // The parent of this entity is OutputGroupSettings. ArchiveGroupSettings interface{} `json:"archiveGroupSettings" yaml:"archiveGroupSettings"` // The configuration of a frame capture output group. FrameCaptureGroupSettings interface{} `json:"frameCaptureGroupSettings" yaml:"frameCaptureGroupSettings"` // The configuration of an HLS output group. HlsGroupSettings interface{} `json:"hlsGroupSettings" yaml:"hlsGroupSettings"` // The configuration of a MediaPackage output group. MediaPackageGroupSettings interface{} `json:"mediaPackageGroupSettings" yaml:"mediaPackageGroupSettings"` // The configuration of a Microsoft Smooth output group. MsSmoothGroupSettings interface{} `json:"msSmoothGroupSettings" yaml:"msSmoothGroupSettings"` // The settings for a Multiplex output group. MultiplexGroupSettings interface{} `json:"multiplexGroupSettings" yaml:"multiplexGroupSettings"` // The configuration of an RTMP output group. RtmpGroupSettings interface{} `json:"rtmpGroupSettings" yaml:"rtmpGroupSettings"` // The configuration of a UDP output group. UdpGroupSettings interface{} `json:"udpGroupSettings" yaml:"udpGroupSettings"` }
The configuration of the output group.
The parent of this entity is OutputGroup.
TODO: EXAMPLE
type CfnChannel_OutputLocationRefProperty ¶
type CfnChannel_OutputLocationRefProperty struct { // A reference ID for this destination. DestinationRefId *string `json:"destinationRefId" yaml:"destinationRefId"` }
A reference to an OutputDestination ID that is defined in the channel.
This entity is used by ArchiveGroupSettings, FrameCaptureGroupSettings, HlsGroupSettings, MediaPackageGroupSettings, MSSmoothGroupSettings, RtmpOutputSettings, and UdpOutputSettings.
TODO: EXAMPLE
type CfnChannel_OutputProperty ¶
type CfnChannel_OutputProperty struct { // The names of the audio descriptions that are used as audio sources for this output. AudioDescriptionNames *[]*string `json:"audioDescriptionNames" yaml:"audioDescriptionNames"` // The names of the caption descriptions that are used as captions sources for this output. CaptionDescriptionNames *[]*string `json:"captionDescriptionNames" yaml:"captionDescriptionNames"` // The name that is used to identify an output. OutputName *string `json:"outputName" yaml:"outputName"` // The output type-specific settings. OutputSettings interface{} `json:"outputSettings" yaml:"outputSettings"` // The name of the VideoDescription that is used as the source for this output. VideoDescriptionName *string `json:"videoDescriptionName" yaml:"videoDescriptionName"` }
The output settings.
The parent of this entity is OutputGroup.
TODO: EXAMPLE
type CfnChannel_OutputSettingsProperty ¶
type CfnChannel_OutputSettingsProperty struct { // The settings for an archive output. ArchiveOutputSettings interface{} `json:"archiveOutputSettings" yaml:"archiveOutputSettings"` // The settings for a frame capture output. // // The parent of this entity is OutputGroupSettings. FrameCaptureOutputSettings interface{} `json:"frameCaptureOutputSettings" yaml:"frameCaptureOutputSettings"` // The settings for an HLS output. // // The parent of this entity is OutputGroupSettings. HlsOutputSettings interface{} `json:"hlsOutputSettings" yaml:"hlsOutputSettings"` // The settings for a MediaPackage output. // // The parent of this entity is OutputGroupSettings. MediaPackageOutputSettings interface{} `json:"mediaPackageOutputSettings" yaml:"mediaPackageOutputSettings"` // The settings for a Microsoft Smooth output. MsSmoothOutputSettings interface{} `json:"msSmoothOutputSettings" yaml:"msSmoothOutputSettings"` // Configuration of a Multiplex output. MultiplexOutputSettings interface{} `json:"multiplexOutputSettings" yaml:"multiplexOutputSettings"` // The settings for an RTMP output. // // The parent of this entity is OutputGroupSettings. RtmpOutputSettings interface{} `json:"rtmpOutputSettings" yaml:"rtmpOutputSettings"` // The settings for a UDP output. // // The parent of this entity is OutputGroupSettings. UdpOutputSettings interface{} `json:"udpOutputSettings" yaml:"udpOutputSettings"` }
The output settings.
The parent of this entity is Output.
TODO: EXAMPLE
type CfnChannel_PassThroughSettingsProperty ¶
type CfnChannel_PassThroughSettingsProperty struct { }
The settings for passing through audio to the output.
The parent of this entity is AudioCodecSettings.
TODO: EXAMPLE
type CfnChannel_RawSettingsProperty ¶
type CfnChannel_RawSettingsProperty struct { }
The container for WAV audio in the output group.
The parent of this entity is ArchiveContainerSettings.
TODO: EXAMPLE
type CfnChannel_Rec601SettingsProperty ¶
type CfnChannel_Rec601SettingsProperty struct { }
Rec601 Settings.
The parents of this entity are H264ColorSpaceSettings and H265ColorSpaceSettings.
TODO: EXAMPLE
type CfnChannel_Rec709SettingsProperty ¶
type CfnChannel_Rec709SettingsProperty struct { }
Rec709 Settings.
The parents of this entity are H264ColorSpaceSettings and H265ColorSpaceSettings.
TODO: EXAMPLE
type CfnChannel_RemixSettingsProperty ¶
type CfnChannel_RemixSettingsProperty struct { // A mapping of input channels to output channels, with appropriate gain adjustments. ChannelMappings interface{} `json:"channelMappings" yaml:"channelMappings"` // The number of input channels to be used. ChannelsIn *float64 `json:"channelsIn" yaml:"channelsIn"` // The number of output channels to be produced. // // Valid values: 1, 2, 4, 6, 8. ChannelsOut *float64 `json:"channelsOut" yaml:"channelsOut"` }
The settings for remixing audio in the output.
The parent of this entity is AudioDescription.
TODO: EXAMPLE
type CfnChannel_RtmpCaptionInfoDestinationSettingsProperty ¶
type CfnChannel_RtmpCaptionInfoDestinationSettingsProperty struct { }
The settings for RTMPCaptionInfo captions encode in the output.
The parent of this entity is CaptionDestinationSettings.
TODO: EXAMPLE
type CfnChannel_RtmpGroupSettingsProperty ¶
type CfnChannel_RtmpGroupSettingsProperty struct { // Choose the ad marker type for this output group. // // MediaLive will create a message based on the content of each SCTE-35 message, format it for that marker type, and insert it in the datastream. AdMarkers *[]*string `json:"adMarkers" yaml:"adMarkers"` // An authentication scheme to use when connecting with a CDN. AuthenticationScheme *string `json:"authenticationScheme" yaml:"authenticationScheme"` // Controls behavior when the content cache fills up. // // If a remote origin server stalls the RTMP connection and doesn't accept content fast enough, the media cache fills up. When the cache reaches the duration specified by cacheLength, the cache stops accepting new content. If set to disconnectImmediately, the RTMP output forces a disconnect. Clear the media cache, and reconnect after restartDelay seconds. If set to waitForServer, the RTMP output waits up to 5 minutes to allow the origin server to begin accepting data again. CacheFullBehavior *string `json:"cacheFullBehavior" yaml:"cacheFullBehavior"` // The cache length, in seconds, that is used to calculate buffer size. CacheLength *float64 `json:"cacheLength" yaml:"cacheLength"` // Controls the types of data that pass to onCaptionInfo outputs. // // If set to all, 608 and 708 carried DTVCC data is passed. If set to field1AndField2608, DTVCC data is stripped out, but 608 data from both fields is passed. If set to field1608, only the data carried in 608 from field 1 video is passed. CaptionData *string `json:"captionData" yaml:"captionData"` // Controls the behavior of this RTMP group if the input becomes unavailable. // // emitOutput: Emit a slate until the input returns. pauseOutput: Stop transmitting data until the input returns. This does not close the underlying RTMP connection. InputLossAction *string `json:"inputLossAction" yaml:"inputLossAction"` // If a streaming output fails, the number of seconds to wait until a restart is initiated. // // A value of 0 means never restart. RestartDelay *float64 `json:"restartDelay" yaml:"restartDelay"` }
The configuration of an RTMP output group.
The parent of this entity is OutputGroupSettings.
TODO: EXAMPLE
type CfnChannel_RtmpOutputSettingsProperty ¶
type CfnChannel_RtmpOutputSettingsProperty struct { // If set to verifyAuthenticity, verifies the TLS certificate chain to a trusted certificate authority (CA). // // This causes RTMPS outputs with self-signed certificates to fail. CertificateMode *string `json:"certificateMode" yaml:"certificateMode"` // The number of seconds to wait before retrying a connection to the Flash Media server if the connection is lost. ConnectionRetryInterval *float64 `json:"connectionRetryInterval" yaml:"connectionRetryInterval"` // The RTMP endpoint excluding the stream name (for example, rtmp://host/appname). Destination interface{} `json:"destination" yaml:"destination"` // The number of retry attempts. NumRetries *float64 `json:"numRetries" yaml:"numRetries"` }
The settings for one RTMP output.
The parent of this entity is OutputSettings.
TODO: EXAMPLE
type CfnChannel_Scte20PlusEmbeddedDestinationSettingsProperty ¶
type CfnChannel_Scte20PlusEmbeddedDestinationSettingsProperty struct { }
The configuration of SCTE-20 plus embedded captions in the output.
The parent of this entity is CaptionDestinationSettings.
TODO: EXAMPLE
type CfnChannel_Scte20SourceSettingsProperty ¶
type CfnChannel_Scte20SourceSettingsProperty struct { // If upconvert, 608 data is both passed through the "608 compatibility bytes" fields of the 708 wrapper as well as translated into 708. // // Any 708 data present in the source content is discarded. Convert608To708 *string `json:"convert608To708" yaml:"convert608To708"` // Specifies the 608/708 channel number within the video track from which to extract captions. Source608ChannelNumber *float64 `json:"source608ChannelNumber" yaml:"source608ChannelNumber"` }
Information about the SCTE-20 captions to extract from the input.
The parent of this entity is CaptionSelectorSettings.
TODO: EXAMPLE
type CfnChannel_Scte27DestinationSettingsProperty ¶
type CfnChannel_Scte27DestinationSettingsProperty struct { }
The configuration of SCTE-27 captions in the output.
The parent of this entity is CaptionDestinationSettings.
TODO: EXAMPLE
type CfnChannel_Scte27SourceSettingsProperty ¶
type CfnChannel_Scte27SourceSettingsProperty struct { // If you will configure a WebVTT caption description that references this caption selector, use this field to provide the language to consider when translating the image-based source to text. OcrLanguage *string `json:"ocrLanguage" yaml:"ocrLanguage"` // The PID field is used in conjunction with the captions selector languageCode field as follows: Specify PID and Language: Extracts captions from that PID; // // the language is "informational." Specify PID and omit Language: Extracts the specified PID. Omit PID and specify Language: Extracts the specified language, whichever PID that happens to be. Omit PID and omit Language: Valid only if source is DVB-Sub that is being passed through; all languages are passed through. Pid *float64 `json:"pid" yaml:"pid"` }
Information about the SCTE-27 captions to extract from the input.
The parent of this entity is CaptionSelectorSettings.
TODO: EXAMPLE
type CfnChannel_Scte35SpliceInsertProperty ¶
type CfnChannel_Scte35SpliceInsertProperty struct { // When specified, this offset (in milliseconds) is added to the input ad avail PTS time. // // This applies only to embedded SCTE 104/35 messages. It doesn't apply to OOB messages. AdAvailOffset *float64 `json:"adAvailOffset" yaml:"adAvailOffset"` // When set to ignore, segment descriptors with noRegionalBlackoutFlag set to 0 no longer trigger blackouts or ad avail slates. NoRegionalBlackoutFlag *string `json:"noRegionalBlackoutFlag" yaml:"noRegionalBlackoutFlag"` // When set to ignore, segment descriptors with webDeliveryAllowedFlag set to 0 no longer trigger blackouts or ad avail slates. WebDeliveryAllowedFlag *string `json:"webDeliveryAllowedFlag" yaml:"webDeliveryAllowedFlag"` }
The setup of SCTE-35 splice insert handling.
The parent of this entity is AvailSettings.
TODO: EXAMPLE
type CfnChannel_Scte35TimeSignalAposProperty ¶
type CfnChannel_Scte35TimeSignalAposProperty struct { // When specified, this offset (in milliseconds) is added to the input ad avail PTS time. // // This applies only to embedded SCTE 104/35 messages. It doesn't apply to OOB messages. AdAvailOffset *float64 `json:"adAvailOffset" yaml:"adAvailOffset"` // When set to ignore, segment descriptors with noRegionalBlackoutFlag set to 0 no longer trigger blackouts or ad avail slates. NoRegionalBlackoutFlag *string `json:"noRegionalBlackoutFlag" yaml:"noRegionalBlackoutFlag"` // When set to ignore, segment descriptors with webDeliveryAllowedFlag set to 0 no longer trigger blackouts or ad avail slates. WebDeliveryAllowedFlag *string `json:"webDeliveryAllowedFlag" yaml:"webDeliveryAllowedFlag"` }
The settings for the SCTE-35 time signal APOS mode.
The parent of this entity is AvailSettings.
TODO: EXAMPLE
type CfnChannel_SmpteTtDestinationSettingsProperty ¶
type CfnChannel_SmpteTtDestinationSettingsProperty struct { }
The setup of SMPTE-TT captions in the output.
The parent of this entity is CaptionDestinationSettings.
TODO: EXAMPLE
type CfnChannel_StandardHlsSettingsProperty ¶
type CfnChannel_StandardHlsSettingsProperty struct { // Lists all the audio groups that are used with the video output stream. // // This inputs all the audio GROUP-IDs that are associated with the video, separated by a comma (,). AudioRenditionSets *string `json:"audioRenditionSets" yaml:"audioRenditionSets"` // Settings for the M3U8 container. M3U8Settings interface{} `json:"m3U8Settings" yaml:"m3U8Settings"` }
The configuration of an HLS output that is a standard output (not an audio-only output).
The parent of this entity is HlsSettings.
TODO: EXAMPLE
type CfnChannel_StaticKeySettingsProperty ¶
type CfnChannel_StaticKeySettingsProperty struct { // The URL of the license server that is used for protecting content. KeyProviderServer interface{} `json:"keyProviderServer" yaml:"keyProviderServer"` // The static key value as a 32 character hexadecimal string. StaticKeyValue *string `json:"staticKeyValue" yaml:"staticKeyValue"` }
The static key settings.
The parent of this entity is KeyProviderSettings.
TODO: EXAMPLE
type CfnChannel_TeletextDestinationSettingsProperty ¶
type CfnChannel_TeletextDestinationSettingsProperty struct { }
The settings for a Teletext captions output encode.
The parent of this entity is CaptionDestinationSettings.
TODO: EXAMPLE
type CfnChannel_TeletextSourceSettingsProperty ¶
type CfnChannel_TeletextSourceSettingsProperty struct { // Settings to configure the caption rectangle for an output captions that will be created using this Teletext source captions. OutputRectangle interface{} `json:"outputRectangle" yaml:"outputRectangle"` // Specifies the Teletext page number within the data stream from which to extract captions. // // The range is 0x100 (256) to 0x8FF (2303). This is unused for passthrough. It should be specified as a hexadecimal string with no "0x" prefix. PageNumber *string `json:"pageNumber" yaml:"pageNumber"` }
Information about the Teletext captions to extract from the input.
The parent of this entity is CaptionSelectorSettings.
TODO: EXAMPLE
type CfnChannel_TemporalFilterSettingsProperty ¶
type CfnChannel_TemporalFilterSettingsProperty struct { // If you enable this filter, the results are the following: - If the source content is noisy (it contains excessive digital artifacts), the filter cleans up the source. // // - If the source content is already clean, the filter tends to decrease the bitrate, especially when the rate control mode is QVBR. PostFilterSharpening *string `json:"postFilterSharpening" yaml:"postFilterSharpening"` // Choose a filter strength. // // We recommend a strength of 1 or 2. A higher strength might take out good information, resulting in an image that is overly soft. Strength *string `json:"strength" yaml:"strength"` }
Settings for the temporal filter to apply to the video.
The parents of this entity are H264FilterSettings, H265FilterSettings, and Mpeg2FilterSettings.
TODO: EXAMPLE
type CfnChannel_TimecodeConfigProperty ¶
type CfnChannel_TimecodeConfigProperty struct { // Identifies the source for the timecode that will be associated with the channel outputs. // // Embedded (embedded): Initialize the output timecode with timecode from the source. If no embedded timecode is detected in the source, the system falls back to using "Start at 0" (zerobased). System Clock (systemclock): Use the UTC time. Start at 0 (zerobased): The time of the first frame of the channel will be 00:00:00:00. Source *string `json:"source" yaml:"source"` // The threshold in frames beyond which output timecode is resynchronized to the input timecode. // // Discrepancies below this threshold are permitted to avoid unnecessary discontinuities in the output timecode. There is no timecode sync when this is not specified. SyncThreshold *float64 `json:"syncThreshold" yaml:"syncThreshold"` }
The configuration of the timecode in the output.
The parent of this entity is EncoderSettings.
TODO: EXAMPLE
type CfnChannel_TtmlDestinationSettingsProperty ¶
type CfnChannel_TtmlDestinationSettingsProperty struct { // When set to passthrough, passes through style and position information from a TTML-like input source (TTML, SMPTE-TT, CFF-TT) to the CFF-TT output or TTML output. StyleControl *string `json:"styleControl" yaml:"styleControl"` }
The setup of TTML captions in the output.
The parent of this entity is CaptionDestinationSettings.
TODO: EXAMPLE
type CfnChannel_UdpContainerSettingsProperty ¶
type CfnChannel_UdpContainerSettingsProperty struct {
// The M2TS configuration for this UDP output.
M2TsSettings interface{} `json:"m2TsSettings" yaml:"m2TsSettings"`
}
The configuration of a UDP output.
The parent of this entity is UdpOutputSettings.
TODO: EXAMPLE
type CfnChannel_UdpGroupSettingsProperty ¶
type CfnChannel_UdpGroupSettingsProperty struct { // Specifies the behavior of the last resort when the input video is lost, and no more backup inputs are available. // // When dropTs is selected, the entire transport stream stops emitting. When dropProgram is selected, the program can be dropped from the transport stream (and replaced with null packets to meet the TS bitrate requirement). Or when emitProgram is selected, the transport stream continues to be produced normally with repeat frames, black frames, or slate frames substituted for the absent input video. InputLossAction *string `json:"inputLossAction" yaml:"inputLossAction"` // Indicates the ID3 frame that has the timecode. TimedMetadataId3Frame *string `json:"timedMetadataId3Frame" yaml:"timedMetadataId3Frame"` // The timed metadata interval in seconds. TimedMetadataId3Period *float64 `json:"timedMetadataId3Period" yaml:"timedMetadataId3Period"` }
The configuration of a UDP output group.
The parent of this entity is OutputGroupSettings.
TODO: EXAMPLE
type CfnChannel_UdpOutputSettingsProperty ¶
type CfnChannel_UdpOutputSettingsProperty struct { // The UDP output buffering in milliseconds. // // Larger values increase latency through the transcoder but simultaneously assist the transcoder in maintaining a constant, low-jitter UDP/RTP output while accommodating clock recovery, input switching, input disruptions, picture reordering, and so on. BufferMsec *float64 `json:"bufferMsec" yaml:"bufferMsec"` // The settings for the UDP output. ContainerSettings interface{} `json:"containerSettings" yaml:"containerSettings"` // The destination address and port number for RTP or UDP packets. // // These can be unicast or multicast RTP or UDP (for example, rtp://239.10.10.10:5001 or udp://10.100.100.100:5002). Destination interface{} `json:"destination" yaml:"destination"` // The settings for enabling and adjusting Forward Error Correction on UDP outputs. FecOutputSettings interface{} `json:"fecOutputSettings" yaml:"fecOutputSettings"` }
The settings for one UDP output.
The parent of this entity is OutputSettings.
TODO: EXAMPLE
type CfnChannel_VideoBlackFailoverSettingsProperty ¶
type CfnChannel_VideoBlackFailoverSettingsProperty struct { // A value used in calculating the threshold below which MediaLive considers a pixel to be 'black'. // // For the input to be considered black, every pixel in a frame must be below this threshold. The threshold is calculated as a percentage (expressed as a decimal) of white. Therefore .1 means 10% white (or 90% black). Note how the formula works for any color depth. For example, if you set this field to 0.1 in 10-bit color depth: (1023*0.1=102.3), which means a pixel value of 102 or less is 'black'. If you set this field to .1 in an 8-bit color depth: (255*0.1=25.5), which means a pixel value of 25 or less is 'black'. The range is 0.0 to 1.0, with any number of decimal places. BlackDetectThreshold *float64 `json:"blackDetectThreshold" yaml:"blackDetectThreshold"` // The amount of time (in milliseconds) that the active input must be black before automatic input failover occurs. VideoBlackThresholdMsec *float64 `json:"videoBlackThresholdMsec" yaml:"videoBlackThresholdMsec"` }
MediaLive will perform a failover if content is considered black for the specified period.
The parent of this entity is FailoverConditionSettings.
TODO: EXAMPLE
type CfnChannel_VideoCodecSettingsProperty ¶
type CfnChannel_VideoCodecSettingsProperty struct { // The settings for the video codec in a frame capture output. FrameCaptureSettings interface{} `json:"frameCaptureSettings" yaml:"frameCaptureSettings"` // The settings for the H.264 codec in the output. H264Settings interface{} `json:"h264Settings" yaml:"h264Settings"` // Settings for video encoded with the H265 codec. H265Settings interface{} `json:"h265Settings" yaml:"h265Settings"` // Settings for video encoded with the MPEG-2 codec. Mpeg2Settings interface{} `json:"mpeg2Settings" yaml:"mpeg2Settings"` }
The settings for the video codec in the output.
The parent of this entity is VideoDescription.
TODO: EXAMPLE
type CfnChannel_VideoDescriptionProperty ¶
type CfnChannel_VideoDescriptionProperty struct { // The video codec settings. CodecSettings interface{} `json:"codecSettings" yaml:"codecSettings"` // The output video height, in pixels. // // This must be an even number. For most codecs, you can keep this field and width blank in order to use the height and width (resolution) from the source. Note that we don't recommend keeping the field blank. For the Frame Capture codec, height and width are required. Height *float64 `json:"height" yaml:"height"` // The name of this VideoDescription. // // Outputs use this name to uniquely identify this description. Description names should be unique within this channel. Name *string `json:"name" yaml:"name"` // Indicates how to respond to the AFD values in the input stream. // // RESPOND causes input video to be clipped, depending on the AFD value, input display aspect ratio, and output display aspect ratio, and (except for the FRAMECAPTURE codec) includes the values in the output. PASSTHROUGH (does not apply to FRAMECAPTURE codec) ignores the AFD values and includes the values in the output, so input video is not clipped. NONE ignores the AFD values and does not include the values through to the output, so input video is not clipped. RespondToAfd *string `json:"respondToAfd" yaml:"respondToAfd"` // STRETCHTOOUTPUT configures the output position to stretch the video to the specified output resolution (height and width). // // This option overrides any position value. DEFAULT might insert black boxes (pillar boxes or letter boxes) around the video to provide the specified output resolution. ScalingBehavior *string `json:"scalingBehavior" yaml:"scalingBehavior"` // Changes the strength of the anti-alias filter used for scaling. // // 0 is the softest setting, and 100 is the sharpest. We recommend a setting of 50 for most content. Sharpness *float64 `json:"sharpness" yaml:"sharpness"` // The output video width, in pixels. // // It must be an even number. For most codecs, you can keep this field and height blank in order to use the height and width (resolution) from the source. Note that we don't recommend keeping the field blank. For the Frame Capture codec, height and width are required. Width *float64 `json:"width" yaml:"width"` }
Encoding information for one output video.
The parent of this entity is EncoderSettings.
TODO: EXAMPLE
type CfnChannel_VideoSelectorColorSpaceSettingsProperty ¶
type CfnChannel_VideoSelectorColorSpaceSettingsProperty struct {
// Settings to configure color space settings in the incoming video.
Hdr10Settings interface{} `json:"hdr10Settings" yaml:"hdr10Settings"`
}
Settings to configure color space settings in the incoming video.
The parent of this entity is VideoSelector.
TODO: EXAMPLE
type CfnChannel_VideoSelectorPidProperty ¶
type CfnChannel_VideoSelectorPidProperty struct { // Selects a specific PID from within a video source. Pid *float64 `json:"pid" yaml:"pid"` }
Selects a specific PID from within a video source.
The parent of this entity is VideoSelectorSettings.
TODO: EXAMPLE
type CfnChannel_VideoSelectorProgramIdProperty ¶
type CfnChannel_VideoSelectorProgramIdProperty struct { // Selects a specific program from within a multi-program transport stream. // // If the program doesn't exist, MediaLive selects the first program within the transport stream by default. ProgramId *float64 `json:"programId" yaml:"programId"` }
Used to extract video by the program ID.
The parent of this entity is VideoSelectorSettings.
TODO: EXAMPLE
type CfnChannel_VideoSelectorProperty ¶
type CfnChannel_VideoSelectorProperty struct { // Specifies the color space of an input. // // This setting works in tandem with colorSpaceConversion to determine if MediaLive will perform any conversion. ColorSpace *string `json:"colorSpace" yaml:"colorSpace"` // Settings to configure color space settings in the incoming video. ColorSpaceSettings interface{} `json:"colorSpaceSettings" yaml:"colorSpaceSettings"` // Applies only if colorSpace is a value other than Follow. // // This field controls how the value in the colorSpace field is used. Fallback means that when the input does include color space data, that data is used, but when the input has no color space data, the value in colorSpace is used. Choose fallback if your input is sometimes missing color space data, but when it does have color space data, that data is correct. Force means to always use the value in colorSpace. Choose force if your input usually has no color space data or might have unreliable color space data. ColorSpaceUsage *string `json:"colorSpaceUsage" yaml:"colorSpaceUsage"` // Information about the video to select from the content. SelectorSettings interface{} `json:"selectorSettings" yaml:"selectorSettings"` }
Information about the video to extract from the input. An input can contain only one video selector.
The parent of this entity is InputSettings.
TODO: EXAMPLE
type CfnChannel_VideoSelectorSettingsProperty ¶
type CfnChannel_VideoSelectorSettingsProperty struct { // Used to extract video by PID. VideoSelectorPid interface{} `json:"videoSelectorPid" yaml:"videoSelectorPid"` // Used to extract video by program ID. VideoSelectorProgramId interface{} `json:"videoSelectorProgramId" yaml:"videoSelectorProgramId"` }
Information about the video to extract from the input.
The parent of this entity is VideoSelector.
TODO: EXAMPLE
type CfnChannel_VpcOutputSettingsProperty ¶
type CfnChannel_VpcOutputSettingsProperty struct { // List of public address allocation IDs to associate with ENIs that will be created in Output VPC. // // Must specify one for SINGLE_PIPELINE, two for STANDARD channels PublicAddressAllocationIds *[]*string `json:"publicAddressAllocationIds" yaml:"publicAddressAllocationIds"` // A list of up to 5 EC2 VPC security group IDs to attach to the Output VPC network interfaces. // // If none are specified then the VPC default security group will be used SecurityGroupIds *[]*string `json:"securityGroupIds" yaml:"securityGroupIds"` // A list of VPC subnet IDs from the same VPC. // // If STANDARD channel, subnet IDs must be mapped to two unique availability zones (AZ). SubnetIds *[]*string `json:"subnetIds" yaml:"subnetIds"` }
Settings to enable VPC mode in the channel, so that the endpoints for all outputs are in your VPC.
This entity is at the top level in the channel.
TODO: EXAMPLE
type CfnChannel_WavSettingsProperty ¶
type CfnChannel_WavSettingsProperty struct { // Bits per sample. BitDepth *float64 `json:"bitDepth" yaml:"bitDepth"` // The audio coding mode for the WAV audio. // // The mode determines the number of channels in the audio. CodingMode *string `json:"codingMode" yaml:"codingMode"` // Sample rate in Hz. SampleRate *float64 `json:"sampleRate" yaml:"sampleRate"` }
The setup of WAV audio in the output.
The parent of this entity is AudioCodecSettings.
TODO: EXAMPLE
type CfnChannel_WebvttDestinationSettingsProperty ¶
type CfnChannel_WebvttDestinationSettingsProperty struct { // Controls whether the color and position of the source captions is passed through to the WebVTT output captions. // // PASSTHROUGH - Valid only if the source captions are EMBEDDED or TELETEXT. NO_STYLE_DATA - Don't pass through the style. The output captions will not contain any font styling information. StyleControl *string `json:"styleControl" yaml:"styleControl"` }
The configuration of Web VTT captions in the output.
The parent of this entity is CaptionDestinationSettings.
TODO: EXAMPLE
type CfnInput ¶
type CfnInput interface { awscdk.CfnResource awscdk.IInspectable AttrArn() *string AttrDestinations() *[]*string AttrSources() *[]*string CfnOptions() awscdk.ICfnResourceOptions CfnProperties() *map[string]interface{} CfnResourceType() *string CreationStack() *[]*string Destinations() interface{} SetDestinations(val interface{}) InputDevices() interface{} SetInputDevices(val interface{}) InputSecurityGroups() *[]*string SetInputSecurityGroups(val *[]*string) LogicalId() *string MediaConnectFlows() interface{} SetMediaConnectFlows(val interface{}) Name() *string SetName(val *string) Node() awscdk.ConstructNode Ref() *string RoleArn() *string SetRoleArn(val *string) Sources() interface{} SetSources(val interface{}) Stack() awscdk.Stack Tags() awscdk.TagManager Type() *string SetType(val *string) UpdatedProperites() *map[string]interface{} Vpc() interface{} SetVpc(val interface{}) AddDeletionOverride(path *string) AddDependsOn(target awscdk.CfnResource) AddMetadata(key *string, value interface{}) AddOverride(path *string, value interface{}) AddPropertyDeletionOverride(propertyPath *string) AddPropertyOverride(propertyPath *string, value interface{}) ApplyRemovalPolicy(policy awscdk.RemovalPolicy, options *awscdk.RemovalPolicyOptions) GetAtt(attributeName *string) awscdk.Reference GetMetadata(key *string) interface{} Inspect(inspector awscdk.TreeInspector) OnPrepare() OnSynthesize(session constructs.ISynthesisSession) OnValidate() *[]*string OverrideLogicalId(newLogicalId *string) Prepare() RenderProperties(props *map[string]interface{}) *map[string]interface{} ShouldSynthesize() *bool Synthesize(session awscdk.ISynthesisSession) ToString() *string Validate() *[]*string ValidateProperties(_properties interface{}) }
A CloudFormation `AWS::MediaLive::Input`.
The AWS::MediaLive::Input resource is a MediaLive resource type that creates an input.
A MediaLive input holds information that describes how the MediaLive channel is connected to the upstream system that is providing the source content that is to be transcoded.
TODO: EXAMPLE
func NewCfnInput ¶
func NewCfnInput(scope awscdk.Construct, id *string, props *CfnInputProps) CfnInput
Create a new `AWS::MediaLive::Input`.
type CfnInputProps ¶
type CfnInputProps struct { // Settings that apply only if the input is a push type of input. Destinations interface{} `json:"destinations" yaml:"destinations"` // Settings that apply only if the input is an Elemental Link input. InputDevices interface{} `json:"inputDevices" yaml:"inputDevices"` // The list of input security groups (referenced by IDs) to attach to the input if the input is a push type. InputSecurityGroups *[]*string `json:"inputSecurityGroups" yaml:"inputSecurityGroups"` // Settings that apply only if the input is a MediaConnect input. MediaConnectFlows interface{} `json:"mediaConnectFlows" yaml:"mediaConnectFlows"` // A name for the input. Name *string `json:"name" yaml:"name"` // The IAM role for MediaLive to assume when creating a MediaConnect input or Amazon VPC input. // // This doesn't apply to other types of inputs. The role is identified by its ARN. RoleArn *string `json:"roleArn" yaml:"roleArn"` // Settings that apply only if the input is a pull type of input. Sources interface{} `json:"sources" yaml:"sources"` // A collection of tags for this input. // // Each tag is a key-value pair. Tags interface{} `json:"tags" yaml:"tags"` // The type for this input. Type *string `json:"type" yaml:"type"` // Settings that apply only if the input is an push input where the source is on Amazon VPC. Vpc interface{} `json:"vpc" yaml:"vpc"` }
Properties for defining a `CfnInput`.
TODO: EXAMPLE
type CfnInputSecurityGroup ¶
type CfnInputSecurityGroup interface { awscdk.CfnResource awscdk.IInspectable AttrArn() *string CfnOptions() awscdk.ICfnResourceOptions CfnProperties() *map[string]interface{} CfnResourceType() *string CreationStack() *[]*string LogicalId() *string Node() awscdk.ConstructNode Ref() *string Stack() awscdk.Stack Tags() awscdk.TagManager UpdatedProperites() *map[string]interface{} WhitelistRules() interface{} SetWhitelistRules(val interface{}) AddDeletionOverride(path *string) AddDependsOn(target awscdk.CfnResource) AddMetadata(key *string, value interface{}) AddOverride(path *string, value interface{}) AddPropertyDeletionOverride(propertyPath *string) AddPropertyOverride(propertyPath *string, value interface{}) ApplyRemovalPolicy(policy awscdk.RemovalPolicy, options *awscdk.RemovalPolicyOptions) GetAtt(attributeName *string) awscdk.Reference GetMetadata(key *string) interface{} Inspect(inspector awscdk.TreeInspector) OnPrepare() OnSynthesize(session constructs.ISynthesisSession) OnValidate() *[]*string OverrideLogicalId(newLogicalId *string) Prepare() RenderProperties(props *map[string]interface{}) *map[string]interface{} ShouldSynthesize() *bool Synthesize(session awscdk.ISynthesisSession) ToString() *string Validate() *[]*string ValidateProperties(_properties interface{}) }
A CloudFormation `AWS::MediaLive::InputSecurityGroup`.
The AWS::MediaLive::InputSecurityGroup is a MediaLive resource type that creates an input security group.
A MediaLive input security group is associated with a MediaLive input. The input security group is an "allow list" of IP addresses that controls whether an external IP address can push content to the associated MediaLive input.
TODO: EXAMPLE
func NewCfnInputSecurityGroup ¶
func NewCfnInputSecurityGroup(scope awscdk.Construct, id *string, props *CfnInputSecurityGroupProps) CfnInputSecurityGroup
Create a new `AWS::MediaLive::InputSecurityGroup`.
type CfnInputSecurityGroupProps ¶
type CfnInputSecurityGroupProps struct { // A collection of tags for this input security group. // // Each tag is a key-value pair. Tags interface{} `json:"tags" yaml:"tags"` // The list of IPv4 CIDR addresses to include in the input security group as "allowed" addresses. WhitelistRules interface{} `json:"whitelistRules" yaml:"whitelistRules"` }
Properties for defining a `CfnInputSecurityGroup`.
TODO: EXAMPLE
type CfnInputSecurityGroup_InputWhitelistRuleCidrProperty ¶
type CfnInputSecurityGroup_InputWhitelistRuleCidrProperty struct { // An IPv4 CIDR range to include in this input security group. Cidr *string `json:"cidr" yaml:"cidr"` }
An IPv4 CIDR range to include in this input security group.
The parent of this entity is InputSecurityGroup.
TODO: EXAMPLE
type CfnInput_InputDestinationRequestProperty ¶
type CfnInput_InputDestinationRequestProperty struct { // The stream name (application name/application instance) for the location the RTMP source content will be pushed to in MediaLive. StreamName *string `json:"streamName" yaml:"streamName"` }
Settings that apply only if the input is a push type of input.
The parent of this entity is Input.
TODO: EXAMPLE
type CfnInput_InputDeviceRequestProperty ¶
type CfnInput_InputDeviceRequestProperty struct { // This property is not used. // // Ignore it. Id *string `json:"id" yaml:"id"` }
This entity is not used.
Ignore it.
TODO: EXAMPLE
type CfnInput_InputDeviceSettingsProperty ¶
type CfnInput_InputDeviceSettingsProperty struct { // The unique ID for the device. Id *string `json:"id" yaml:"id"` }
Settings that apply only if the input is an Elemental Link input.
The parent of this entity is Input.
TODO: EXAMPLE
type CfnInput_InputSourceRequestProperty ¶
type CfnInput_InputSourceRequestProperty struct { // The password parameter that holds the password for accessing the upstream system. // // The password parameter applies only if the upstream system requires credentials. PasswordParam *string `json:"passwordParam" yaml:"passwordParam"` // For a pull input, the URL where MediaLive pulls the source content from. Url *string `json:"url" yaml:"url"` // The user name to connect to the upstream system. // // The user name applies only if the upstream system requires credentials. Username *string `json:"username" yaml:"username"` }
Settings that apply only if the input is a pull type of input.
The parent of this entity is Input.
TODO: EXAMPLE
type CfnInput_InputVpcRequestProperty ¶
type CfnInput_InputVpcRequestProperty struct { // The list of up to five VPC security group IDs to attach to the input VPC network interfaces. // // The security groups require subnet IDs. If none are specified, MediaLive uses the VPC default security group. SecurityGroupIds *[]*string `json:"securityGroupIds" yaml:"securityGroupIds"` // The list of two VPC subnet IDs from the same VPC. // // You must associate subnet IDs to two unique Availability Zones. SubnetIds *[]*string `json:"subnetIds" yaml:"subnetIds"` }
Settings that apply only if the input is an push input where the source is on Amazon VPC.
The parent of this entity is Input.
TODO: EXAMPLE
type CfnInput_MediaConnectFlowRequestProperty ¶
type CfnInput_MediaConnectFlowRequestProperty struct { // The ARN of one or two MediaConnect flows that are the sources for this MediaConnect input. FlowArn *string `json:"flowArn" yaml:"flowArn"` }
Settings that apply only if the input is a MediaConnect input.
The parent of this entity is Input.
TODO: EXAMPLE