Zynq UltraScale+ MPSoC VCU TRD 2021.2 - PL DDR HLG SDI Audio Video Capture and Display

This page provides detailed information related to Design Module 2 - HLG SDI Video Capture and Display with PLDDR

Table of Contents

1 Overview

The primary goal of this Design is to demonstrate the capabilities of VCU hard block present in Zynq UltraScale+ EV devices. The TRD will serve as a platform to tune the performance parameters of VCU and arrive at optimal configurations for encoder and decoder blocks. It has also added an initial support of 8-channels audio.

This module enables the capture of the Hybrid Log Gamma(HLG) video from an SDI-Rx subsystem implemented in the PL. The Hybrid Log Gamma(HLG) video can be displayed through the SDI-Tx subsystem implemented in the PL. The module can stream-out and stream-in live captured video frames through an Ethernet interface. This module supports single-stream for XV20 pixel format. In this design, PL_DDR is used for decoding and PS_DDR for encoding so that DDR bandwidth would be enough to support high bandwidth VCU applications requiring simultaneous encoder and decoder operations and transcoding at 4k@60 FPS.

This module supports the Encoding-Decoding and Transmission of Hybrid Log Gamma(HLG) video along with backward compatible Standard Dynamic Range(SDR) for SDI. It provides the ability to encode a wide dynamic range, while still being compatible with the existing transmission standards in the standard dynamic range (SDR) region. This HLG format encodes the HDR and SDR information in single signal enabling HDR-compatible TVs to display an enhanced image. Unlike HDR it does not have any metadata, rather it will use the ATC(Alternative transfer characteristics) SEI(supplemental enhanced information) information in the VUI(video usability information) to add extra encoding details.

From VCU point of view, there are two "types" of HLG, which you can enable:

  1. There is a HLG-SDR Backwards Compatible Mode, which uses the BT2020 value in the SPS VUI parameters instead of the HLG transfer characteristics. Then the VCU encoder will insert a 'Alternative Transfer Characteristics' (ATC) SEI with the HLG value. See below video frame snapshot captured in stream-eye:

Depending on version of stream-eye, you may not see SEI message correctly. But if you look at HEX viewer you will see ATC SEI in bit-stream.

0x93 - Payload Type (147 == ATC)
0x01 - Payload Size (1 byte)
0x12 - 18 (HLG EOTF value)
0x80 - payload bits ending

2. There is a HLG only mode. This directly uses the HLG value in the SPS VUI parameters. See below frame snapshot captured in stream-eye:

This design supports the following video interfaces:
Sources:

  • SDI-Rx capture pipeline implemented in the PL.

  • File source (SD card, USB storage, SATA hard disk).

  • Stream-In from network or internet.

Sinks:

  • SDI-Tx display pipeline implemented in the PL.

VCU Codec:

  • Video Encode/Decode capability using VCU hard block in PL 

    • AVC/HEVC encoding.

    • Encoder/decoder parameter configuration.

Streaming Interfaces:

  • 1G Ethernet PS GEM 

Video format:

  • XV20

Audio Configuration:

  • Codec: Opus

  • Format: S24_32LE

  • Channels: 2, 8

  • Sampling rate: 48 kHz

Supported Resolution

The table below provides the supported resolution from GUI and command-line app in this design.

Resolution

GUI

Command Line

Single Stream

Single Stream

Multi Stream

4Kp60/59.94

X

X

4Kp30/29.97

X

X

1080p60/59.94

X

X

√ - Supported
x – Not supported

The below table gives information about the features supported in this design. 

Pipeline

Input Source

Output Type

ALSA Srivers

Resolution

Audio Codec Type

Audio Configuration

Video Codec
Type

 Deliverables

Pipeline

Input Source

Output Type

ALSA Srivers

Resolution

Audio Codec Type

Audio Configuration

Video Codec
Type

 Deliverables

 

 

Record/Stream-Out pipeline

 

 

 

SDI-Rx

 

 

File Sink/ Stream-Out

 

 

 

SDI-Rx ALSA drivers

 

 


4K/1080p

 

 Opus

2 channel @ 48 kHz

 

 

 

HEVC/AVC

SDI-Rx Audio encode with soft codec and video with VCU and store it in a container format.

 Vorbis

8* channel @ 48 kHz

SDI-Rx Audio encode with soft codec and video with VCU and store it in a container format.

 

 

Playback pipeline

 

File Source/ Stream-In

 

 

SDI-Tx

 

 

SDI-Tx ALSA drivers

 

 

 

4K/1080p

Opus

2 channel @ 48 kHz

 

 

 

HEVC/AVC

Playback of the local-file/stream-in with video decoded using VCU and Audio using GStreamer soft codec.

Vorbis

8* channel @ 48 kHz

Playback of the local-file/stream-in with video decoded using VCU and Audio using GStreamer soft codec.

Capture → Display

SDI-Rx

SDI-Tx

SDI-Rx ALSA drivers and SDI-Tx ALSA drivers

4K/1080p

NA

2 channel @ 48 kHz OR

8* channel @ 48 kHz

HEVC/AVC

SDI-Rx Audio /Video pass to SDI-Tx without VCU/Audio-Codec.

Capture → Encode → Decode → Display

 SDI-Rx

 SDI-Tx

 SDI-Rx ALSA drivers and SDI-Tx ALSA drivers

 4K/1080p

 NA

2 channel @ 48 kHz

OR
8* channel @ 48 kHz

 HEVC/AVC

SDI-Rx raw audio and video with VCU encoder and decode to achieve AV sync.

* The 8-channels audio functionality is validated with Phabrix Qx 12G SDI Analyzer/Generator.

The below figure shows the HLG SDI Video Capture and HLG SDI Display with Audio design hardware block diagram.

The below figure shows the HLG SDI Video Capture and HLG SDI Display with Audio design software block diagram.

1.1 Board Setup

Refer below link for board setup

1.2 Run Flow

The TRD package is released with the source code, Vivado project, Petalinux BSP, and SD card image that enables the user to run the demonstration. It also includes the binaries necessary to configure and boot the ZCU106 board. Prior to running the steps mentioned in this wiki page, download the TRD package and extract its contents to a directory referred to as TRD_HOME which is the home directory.

Refer below link to download all TRD contents.

TRD package contents specific to HLG SDI Video Capture and HLG SDI Display with Audio design are placed in the following directory structure. The user needs to copy all the files from the $TRD_HOME/images/vcu_sdi_xv20 to FAT32 formatted SD card directory.

rdf0428-zcu106-vcu-trd-2021-2/ ├── apu │   └── vcu_petalinux_bsp ├── images │   ├── vcu_audio │   ├── vcu_llp2_hdmi_nv12 │   ├── vcu_llp2_plddr_hdmi │   ├── vcu_llp2_sdi_xv20 │   ├── vcu_multistream_nv12 │   ├── vcu_pcie │   ├── vcu_plddrv1_hdr10_hdmi │   ├── vcu_plddrv2_hdr10_hdmi │   └── vcu_sdi_xv20 ├── pcie_host_package │   ├── COPYING │   ├── include │   ├── LICENSE │   ├── readme.txt │   ├── RELEASE │   ├── tests │   ├── tools │   └── xdma ├── pl │   ├── constrs │   ├── designs │   ├── prebuild │   ├── README.md │   └── srcs └── README.txt └── zcu106_vcu_trd_sources_and_licenses.tar.gz 22 directories, 7 files

TRD package contents specific to HLG SDI Video Capture and HLG SDI Display with Audio are placed in the following directory structure.

rdf0428-zcu106-vcu-trd-2021-2 ├── apu │   └── vcu_petalinux_bsp │   └── xilinx-vcu-zcu106-v2021.2-final.bsp ├─images │   └── vcu_sdi_xv20 │   ├── autostart.sh │   ├── BOOT.BIN │ ├── boot.scr │   ├── config │   ├── Image │   ├── rootfs.cpio.gz.u-boot │   ├── system.dtb │   └── vcu ├── pcie_host_package ├── pl │   ├── constrs │   ├── designs │ │ ├── zcu106_picxo_plddr_sdi_8ch │   ├── prebuild │   │   ├── zcu106_picxo_plddr_sdi_8ch │   ├── README.md │   └── srcs │   ├── hdl │   └── ip └── README.txt └── zcu106_vcu_trd_sources_and_licenses.tar.gz

Configuration files (input.cfg) for various Resolutions are placed in the following directory structure in /media/card.

config ├── 1080p60 │ ├── Display │ ├── Record │ ├── Stream-in │ └── Stream-out ├── 4kp30 │ ├── Display │ ├── Record │ ├── Stream-in │ └── Stream-out └── 4kp60 │ ├── Display │ ├── Record │ ├── Stream-in │ └── Stream-out └── input.cfg

1.2.1 GStreamer Application (vcu_gst_app)

The vcu_gst_app is a command-line multi-threaded Linux application. The command-line application requires an input configuration file (input.cfg) to be provided in the plain text.

Execution of the application is shown below:

Example:

4kp60 HEVC_HIGH Display Pipeline execution

4kp60 HEVC_HIGH Record Pipeline execution

4kp60 HEVC_HIGH Stream-out Pipeline execution

4kp60 HEVC_HIGH Stream-in Pipeline execution

Make sure SDI-Rx should be configured to 4kp60 mode.

To measure the latency of the pipeline, run the below command. The latency data is huge, so dump it to a file.

Refer below link for detailed run flow steps

1.3 Build Flow

Refer below link for build flow


2 Other Information

2.1 Known Issues

2.2 Limitations

2.3 Optimum VCU Encoder parameters for use-cases:

Video streaming:

  • Video streaming use-case requires a very stable bitrate graph for all pictures

  • It is good to avoid periodic large Intra pictures during the encoding session

  • Low-latency rate control (hardware RC) is the preferred control-rate for video streaming, it tries to maintain equal amount frame sizes for all pictures.

  • Good to avoid periodic Intra frames instead use low-delay-p (IPPPPP…)

  • VBR is not a preferred mode of streaming

Performance: AVC Encoder settings:

  • It is preferred to use 8 or higher slices for better AVC encoder performance

  • AVC standard does not support Tile mode processing which results in the processing of MB rows sequentially for entropy coding

Quality: Low bitrate AVC encoding:

  • Enable profile=high and use qp-mode=auto for low-bitrate encoding use-cases

  • The high profile enables 8x8 transform which results in better video quality at low bitrate

2.4 Audio-Video Synchronization

Clocks and synchronization in GStreamer

When playing complex media, each sound and video sample must be played in a specific order at a specific time. For this purpose, GStreamer provides a synchronization mechanism.

GStreamer provides support for the following use cases:

  • Non-live sources with access faster than playback rate. This is the case where one is reading media from a file and playing it back in a synchronized fashion. In this case, multiple streams need to be synchronized, like audio, video and subtitles.

  • Capture and synchronized muxing/mixing of media from multiple live sources. This is a typical use case where you record audio and video from a microphone/camera and mux it into a file for storage.

  • Streaming from (slow) network streams with buffering. This is the typical web streaming case where you access content from a streaming server using HTTP.

  • Capture from live source and playback with configurable latency. This is used, for example, when capturing from a camera, applying an effect, and displaying the result. It is also used when streaming low latency content over a network with UDP.

  • Simultaneous live capture and playback from prerecorded content. This is used in audio recording cases where you play a previously recorded audio and record new samples, the purpose is to have the new audio perfectly in sync with the previously recorded data.

GStreamer uses a GstClock object, buffer timestamps and a SEGMENT event to synchronize streams in a pipeline as we will see in the next sections.

Clock running-time

In a typical computer, there are many sources that can be used as a time source, e.g., the system time, soundcards, CPU performance counters, etc. For this reason, GStreamer has many GstClock implementations available. Note that clock time doesn't have to start from 0 or any other known value. Some clocks start counting from particular start date, others from the last reboot, etc.

A GstClock returns the absolute-time according to that clock with gst_clock_get_time (). The absolute-time (or clock time) of a clock is monotonically increasing.

running-time is the difference between a previous snapshot of the absolute-time called the base-time and any other absolute-time.

running-time = absolute-time - base-time

A GStreamer GstPipeline object maintains a GstClock object and a base-time when it goes to the PLAYING state. The pipeline gives a handle to the selected GstClock to each element in the pipeline along with selected base-time. The pipeline will select a base-time in such a way that the running-time reflects the total time spent in the PLAYING state. As a result, when the pipeline is PAUSED, the running-time stands still.

Because all objects in the pipeline have the same clock and base-time, they can thus all calculate the running-time according to the pipeline clock.

Buffer running-time

To calculate a buffer running-time, we need a buffer timestamp and the SEGMENT event that preceded the buffer. First we can convert the SEGMENT event into a GstSegment object and then we can use the gst_segment_to_running_time () function to perform the calculation of the buffer running-time.

Synchronization is now a matter of making sure that a buffer with a certain running-time is played when the clock reaches the same running-time. Usually, this task is performed by sink elements. These elements also have to take into account the configured pipeline's latency and add it to the buffer running-time before synchronizing to the pipeline clock.

Non-live sources timestamp buffers with a running-time starting from 0. After a flushing seek, they will produce buffers again from a running-time of 0.

Live sources need to timestamp buffers with a running-time matching the pipeline running-time when the first byte of the buffer was captured.

Buffer stream-time

The buffer stream-time, also known as the position in the stream, is a value between 0 and the total duration of the media and it's calculated from the buffer timestamps and the preceding SEGMENT event.

The stream-time is used in:

  • Report the current position in the stream with the POSITION query.

  • The position used in the seek events and queries.

  • The position used to synchronize controlled values.

The stream-time is never used to synchronize streams, this is only done with the running-time.

Time overview

Here is an overview of the various timelines used in GStreamer.

The image below represents the different times in the pipeline when playing a 100ms sample and repeating the part between 50ms and 100ms.

You can see how the running-time of a buffer always increments monotonically along with the clock-time. Buffers are played when their running-time is equal to the clock-time - base-time. The stream-time represents the position in the stream and jumps backwards when repeating.

Clock providers

A clock provider is an element in the pipeline that can provide a GstClock object. The clock object needs to report an absolute-time that is monotonically increasing when the element is in the PLAYING state. It is allowed to pause the clock while the element is PAUSED.

Clock providers exist because they play back media at some rate, and this rate is not necessarily the same as the system clock rate. For example, a sound card may play back at 44.1 kHz, but that doesn't mean that after exactly 1 second according to the system clock, the sound card has played back 44100 samples. This is only true by approximation. In fact, the audio device has an internal clock based on the number of samples played that we can expose.

If an element with an internal clock needs to synchronize, it needs to estimate when a time according to the pipeline clock will take place according to the internal clock. To estimate this, it needs to slave its clock to the pipeline clock.

If the pipeline clock is exactly the internal clock of an element, the element can skip the slaving step and directly use the pipeline clock to schedule playback. This can be both faster and more accurate. Therefore, generally, elements with an internal clock like audio input or output devices will be a clock provider for the pipeline.

When the pipeline goes to the PLAYING state, it will go over all elements in the pipeline from sink to source and ask each element if they can provide a clock. The last element that can provide a clock will be used as the clock provider in the pipeline. This algorithm prefers a clock from an audio sink in a typical playback pipeline and a clock from source elements in a typical capture pipeline.

There exist some bus messages to let you know about the clock and clock providers in the pipeline. You can see what clock is selected in the pipeline by looking at the NEW_CLOCK message on the bus. When a clock provider is removed from the pipeline, a CLOCK_LOST message is posted and the application should go to PAUSED and back to PLAYING to select a new clock.

For more detail please refer: https://gstreamer.freedesktop.org/documentation/application-development/advanced/clocks.html?gi-language=c


3 Appendix A - Input Configuration File (input.cfg)

The example configuration files are stored at /media/card/config/ folder.

Configuration Type

Configuration Name

Description

Available Options

Configuration Type

Configuration Name

Description

Available Options

Common

 

Common Configuration

It is the starting point of common configuration

 

Num of Input

Number of input

1

Output

Select the video interface.

SDI or DP

Out Type

Type of output

display, record, stream

Display Rate

Pipeline frame rate

30, 60

Exit

It indicates to the application that the configuration is over

 

 

Input

Input Configuration

It is the starting point of the input configuration

 

Input Num

Starting Nth input configuration

1

Input Type

Input source type

SDI, File, Stream

Uri

File path or Network URL. Applicable for file playback and Stream-in pipeline only. Supported file formats for playback are ts, mp4, and mkv.

file:///media/usb/abc.ts (for file path), udp://192.168.25.89:5004/ (for Network streaming, Here 192.168.25.89 is IP address and 5004 is port number)

Raw

To tell the pipeline is processed or pass-through

TRUE, FALSE

Width

The width of the live source

3840,1920

Height

The height of the live source

2160, 1080

Format

The format of input data

XV20

Exit

It indicates to the application that the configuration is over

 

Encoder

 

Encoder Configuration

It is the starting point of encoder configuration

 

Encoder Num

Starting Nth encoder configuration

1

Encoder Name

Name of the encoder

AVC, HEVC

Profile

Name of the profile

baseline, main or high for AVC. Main for HEVC.

Rate Control

Rate control options

CBR, VBR, and Low_Latency.

Filler Data

Filler Data NAL units for CBR rate control

True, False

QP

QP control mode used by the VCU encoder

Uniform, Auto

L2 Cache

Enable or Disable L2Cache buffer in encoding process.

True, False

Latency Mode

Encoder latency mode.

normal, sub_frame

Low Bandwidth

If enabled, decrease the vertical search range used for P-frame motion estimation to reduce the bandwidth.

True, False

Gop Mode

Group of Pictures mode.

Basic, low_delay_p, low_delay_b

Bitrate

Target bitrate in Kbps

1-60000

B Frames

Number of B-frames between two consecutive P-frames

0-4

Slice

The number of slices produced for each frame. Each slice contains one or more complete macroblock/CTU row(s). Slices are distributed over the frame as regularly as possible. If slice-size is defined as well more slices may be produced to fit the slice-size requirement

4 to 22 : 4Kp resolution with HEVC codec
4 to 32 : 4Kp resolution with AVC codec
4 to 32 : 1080p resolution with HEVC codec
4 to 32 : 1080p resolution with AVC codec

GoP Length

The distance between two consecutive I frames

1-1000

GDR Mode

It specifies which Gradual Decoder Refresh(GDR) scheme should be used when gop-mode = low_delay_p

Horizontal/Vertical/Disabled

Entropy Mode

It specifies the entropy mode for H.264 (AVC) encoding process

CAVLC/CABAC/Default

Max Picture Size

It is used to curtail instantaneous peak in the bit-stream using this parameter. It works in CBR/VBR rate-control only. When it is enabled, max-picture-size value is calculated and set with 10% of AllowedPeakMargin. i.e. max-picture-size =  (TargetBitrate / FrameRate) * 1.1

TRUE/FALSE

HLG_SDR_Compatible

It specifies the whether vcu will use the HLG Only mode or Backward-Compatible-Mode.

TRUE/FALSE

Preset

Encoder configuration Preset

HEVC_HIGH, HEVC_MEDIUM, HEVC_LOW, AVC_HIGH, AVC_MEDIUM, AVC_LOW, Custom

Exit

It indicates to the application that the configuration is over

 

Record

Record Configuration

It is the starting point of record configuration.

 

Record Num

Starting Nth record configuration.

1

Out File Name

Record file path.

e.g. /media/usb/abc.ts

Duration

Duration in minutes.

1-3

Exit

It indicates to the application that the configuration is over.

 

Streaming

Streaming Configuration

It is the starting point of streaming configuration.

 

Streaming Num

Starting Nth Streaming configuration

1

Host IP

The host to send the packets to

192.168.25.89 or Windows PC IP

Port:

The port to send the packets to.

5004, 5008, 5012, 5016

Exit

It indicates to the application that the configuration is over.

 

Audio Configuration

Audio Configuration

It is the starting point of the audio configuration.

 

Audio Enable

Enable or Disable audio in pipeline.

True, False

Audio Format

The format of the audio

S24_32LE

Sampling Rate

To set the audio sampling rate.

48000

Num Of Channel

The number of audio channels.

2

Source

It should be SDI, as currently only SDI audio capture is supported.

 

Renderer

It should be SDI, as currently only SDI audio renderer is supported.

 

Volume

To set the volume level. The default value is 2.0.

0.0 - 10.0

Exit

It indicates to the application that the configuration is over.

 

Trace

Trace Configuration

It is the starting point of trace configuration.

 

FPS Info

To display fps info on the console.

True, False

APM Info

To display APM counter number on the console.

True, False

Pipeline Info

To display pipeline info on console.

True, False

Exit

It indicates to the application that the configuration is over.

 


4 Appendix B - SDI-Rx/Tx Link-up and GStreamer Commands

This section covers configuration of SDI-Rx using media-ctl utility and SDI-Tx using modetest utility, along with demonstrating SDI-Rx/Tx link-up issue and steps to switch resolution. It also contains sample GStreamer SDI XV20 pipelines for Display, Record, Stream-In and Stream-Out use-cases.

Run the below command to check the SDI link status, resolution, video node and output format of the SDI input source. Run the below command for all media nodes to print media device topology where mediaX represents different media nodes. In the topology, log look for the v_smpte_uhdsdi_rx_ss string to identify the SDI input source media node. The media-ctl command generated as part of petalinux bsp will support all the vcu supported formats like NV12, NV16, XV15 and XV20.

  • When HDMI source is connected to 4Kp60 resolution, it shows as below:

  • When the SDI source is not connected, it shows:

Modetest commands:

  • Modetest command for 4Kp60 Display

  • Modetest command for 4Kp60 Display (HLG use-case)

  • Modetest command for 4Kp30 Display

  • Modetest command for 4Kp30 Display (HLG use-case)

  • Modetest command for 1080p60 Display

  • Modetest command for 1080p60 Display (HLG use-case)

  • Modetest command for 4Kp59.94 Display

  • Modetest command for 4Kp29.97 Display

  • Modetest command for 1080p59.94 Display

  • Follow the below steps to switch the SDI-Rx resolution from 1080p60 to 4Kp60.

    • Check current SDI Input Source Resolution (1080p60) by following the above-mentioned steps.

    • Run vcu_gst_app for current SDI resolution (1080p60) by executing the following command.

Below configurations needs to be set in input.cfg for SDI-1080p60.

  • Change Resolution of SDI Input Source from 1080p60 to 4Kp60 by following the below steps.

    • Set the SDI source resolution to 4Kp60 (Home page → settings → display & Sound → Resolution → change to 4Kp60).

    • Save the configuration to take place the change.

    • Verify desired SDI Input Source Resolution (4Kp60) by following the above-mentioned steps.

  • The table below lists the parameters of the pixel format.

Pixel Format

GStreamer Format

Media Bus Format

GStreamer HEVC Profile

GStreamer AVC Profile

Kmssink Plane-id

Pixel Format

GStreamer Format

Media Bus Format

GStreamer HEVC Profile

GStreamer AVC Profile

Kmssink Plane-id

XV20

NV16_10LE32

UYVY10_1X20

main-422-10

high-4:2:2

32

  • Run the following gst-launch-1.0 command to display raw SDI video + 2-channels audio using the GStreamer pipeline.

  • Run the following gst-launch-1.0 command to display raw SDI video + 8-channels audio using the GStreamer pipeline.

  • Run the following gst-launch-1.0 command to display raw HLG SDI video using the GStreamer pipeline.

  • Run the following gst-launch-1.0 command to display processed(capture → encode → decode → display) SDI video + 2-channels audio using the GStreamer pipeline.

  • Run the following gst-launch-1.0 command to display processed SDI video + 8-channels audio using the GStreamer pipeline.

  • Run the following gst-launch-1.0 command to display processed HLG Only SDI video using the GStreamer pipeline.

  • Run the following gst-launch-1.0 command to display processed HLG-SDR-Compatible SDI video using the GStreamer pipeline.

  • Run the following gst-launch-1.0 command to display processed 4Kp59.94 fractional frame-rate SDI video using the GStreamer pipeline.

  • Run the following gst-launch-1.0 command to display processed 4Kp29.97 fractional frame-rate SDI video using the GStreamer pipeline.

  • Run the following gst-launch-1.0 command to record SDI video + 2-channels audio using the GStreamer pipeline.

  • Run the following gst-launch-1.0 command to play the recorded 2-channels audio + video file using the GStreamer pipeline.

  • Run the following gst-launch-1.0 command to record SDI video + 8-channels audio using the GStreamer pipeline.

  • Run the following gst-launch-1.0 command to play the recorded 8-channels audio + video file using the GStreamer pipeline.

  • Run the following gst-launch-1.0 command to record HLG Only SDI video using the GStreamer pipeline.

  • Run the following gst-launch-1.0 command to record HLG-SDR-Compatible SDI video using the GStreamer pipeline.

  • Run the following gst-launch-1.0 command to play the recorded HLG Only / HLG-SDR-Compatible video file using the GStreamer pipeline.

  • Run the following gst-launch-1.0 command to capture, encode and stream-out SDI video and opus encoded SDI 2-channels audio using the GStreamer pipeline, where 192.168.25.89 is host/client IP address and 5004 is port number.

  • Run the following gst-launch-1.0 command to stream-in, decode and play SDI video and opus encoded SDI 2-channels audio using the GStreamer pipeline, where 5004 is port number.

  • Run the following gst-launch-1.0 command to capture, encode and stream-out SDI video and vorbis encoded SDI 8-channels audio using the GStreamer pipeline. It enables transmission of RTP packets. RTP packets are sent on ports 5004(for video) and 5005(for audio).

  • Run the following gst-launch-1.0 command to stream-in, decode and play SDI video and vorbis encoded SDI 8-channels audio using the GStreamer pipeline. It enables reception of RTP packets. RTP packets are received on ports 5004(for video) and 5005(for audio).

  • Run the following gst-launch-1.0 command to capture, encode and stream-out HLG Only SDI video using the GStreamer pipeline, where 192.168.25.89 is host/client IP address and 5004 is port number.

  • Run the following gst-launch-1.0 command to stream-in, decode and play HLG-Only SDI video using the GStreamer pipeline, where 5004 is port number.

  • Run the following gst-launch-1.0 command to capture, encode and stream-out HLG-SDR-Compatible SDI video using the GStreamer pipeline, where 192.168.25.89 is host/client IP address and 5004 is port number.

  • Run the following gst-launch-1.0 command to stream-in, decode and play HLG-SDR-Compatible SDI video using the GStreamer pipeline, where 5004 is port number.

  • Run the following gst-launch-1.0 command to display 1080i60 interlaced pipeline.

  • Run the following gst-launch-1.0 command to display Interlaced PAL pipeline.

  • Run the following gst-launch-1.0 command to display Interlaced NTSC pipeline.

5 References

© Copyright 2019 - 2022 Xilinx Inc. Privacy Policy