Recent from talks
Contribute something
Nothing was collected or created yet.
Serial digital interface
View on WikipediaThis article needs additional citations for verification. (September 2008) |
| SDI | |
|---|---|
| Serial digital interface | |
Serial digital interface uses BNC connectors | |
| Abbreviation | SDI / SD-SDI / HD-SDI / SDI 3G / SDI 6G / SDI 12G / SDI 24G / |
| Status | Active |
| Year started | 1989 |
| Latest version | 24G 2020 |
| Organization | SMPTE (The Society of Motion Picture and Television Engineers) |
| Predecessor | CVBS - Composite Video |
Serial digital interface (SDI) is a family of digital video interfaces first standardized by SMPTE (The Society of Motion Picture and Television Engineers) in 1989.[1][2] For example, ITU-R BT.656 and SMPTE 259M define digital video interfaces used for broadcast-grade video. A related standard, known as high-definition serial digital interface (HD-SDI), is standardized in SMPTE 292M; this provides a nominal data rate of 1.485 Gbit/s.[3]
Additional SDI standards have been introduced to support increasing video resolutions (HD, UHD and beyond), frame rates, stereoscopic (3D) video,[4][5] and color depth.[6] Dual link HD-SDI consists of a pair of SMPTE 292M links, standardized by SMPTE 372M in 1998;[2] this provides a nominal 2.970 Gbit/s interface used in applications (such as digital cinema or HDTV 1080P) that require greater fidelity and resolution than standard HDTV can provide. 3G-SDI (standardized in SMPTE 424M) consists of a single 2.970 Gbit/s serial link that allows replacing dual link HD-SDI. 6G-SDI and 12G-SDI standards were published on March 19, 2015.[7][8]
These standards are used for transmission of uncompressed, unencrypted digital video signals (optionally including embedded audio and time code) within television facilities; they can also be used for packetized data. SDI is used to connect together different pieces of equipment such as recorders, monitors, PCs and vision mixers. Coaxial variants of the specification range in length but are typically less than 300 meters (980 ft). Fiber optic variants of the specification such as 297M allow for long-distance transmission limited only by maximum fiber length or repeaters.
SDI and HD-SDI are usually available only in professional video equipment because various licensing agreements restrict the use of unencrypted digital interfaces, such as SDI, prohibiting their use in consumer equipment. Several professional video and HD-video capable DSLR cameras and all uncompressed video capable consumer cameras use the HDMI interface, often called clean HDMI. There are various mod kits for existing DVD players and other devices such as splitters that ignore HDCP, which allow a user to add a serial digital interface to these devices.[citation needed]
Electrical interface
[edit]The various serial digital interface standards all use (one or more) coaxial cables with BNC connectors, with a nominal impedance of 75 ohms.
This is the same type of cable used in analog composite video setups, potentially allowing for easier "drop-in" equipment upgrades (although, at high bitrates and/or long distances, it may be necessary for older, oxidising, or lower-grade cable to be replaced with optical fibre). The specified signal amplitude at the source is 800 mV (±10%) peak-to-peak; far lower voltages may be measured at the receiver owing to attenuation. Using equalization at the receiver, it is possible to send 270 Mbit/s SDI over 300 meters (980 ft) without use of repeaters, but shorter lengths are preferred. The HD bitrates have a shorter maximum run length, typically 100 meters (330 ft).[9][10]
Uncompressed digital component signals are transmitted. Data is encoded in NRZI format, and a linear feedback shift register is used to scramble the data to reduce the likelihood that long strings of zeroes or ones will be present on the interface. The interface is self-synchronizing and self-clocking. Framing is done by detection of a special synchronization pattern, which appears on the (unscrambled) serial digital signal to be a sequence of ten ones followed by twenty zeroes (twenty ones followed by forty zeroes in HD); this bit pattern is not legal anywhere else within the data payload.
Standards
[edit]| Standard | Name | Introduced | Bitrates (Mbit/s) | Example video formats |
|---|---|---|---|---|
| SMPTE 259M | SD-SDI | 1989[2] | 270, 360, 143, 177 | 480i, 576i |
| SMPTE 344M | ED-SDI | 2000[11] | 540 | 480p, 576p |
| SMPTE 292M | HD-SDI | 1998[2] | 1485 and 1485/1.001 | 720p, 1080i |
| SMPTE 372M | Dual Link HD-SDI | 2002[2] | 2970 and 2970/1.001 | 1080p60 |
| SMPTE 424M | 3G-SDI | 2006[2] | 2970 and 2970/1.001 | 1080p60 |
| SMPTE ST 2081 | 6G-SDI | 2015[7] | 6000 | 1080p120, 2160p30 |
| SMPTE ST 2082 | 12G-SDI | 2015[8] | 12000 | 2160p60 |
| SMPTE ST 2083 | 24G-SDI | 2020[12][13] | 24000 | 2160p120, 4320p30 |
Bit rates
[edit]Several bit rates are used in serial digital video signal:
- For standard-definition applications, as defined by SMPTE 259M, the possible bit rates are 270 Mbit/s, 360 Mbit/s, 143 Mbit/s, and 177 Mbit/s. 270 Mbit/s is by far the most commonly used; though the 360 Mbit/s interface (used for widescreen standard definition) is sometimes encountered. The 143 and 177 Mbit/s interfaces were intended for transmission of composite-encoded (NTSC or PAL) video digitally and are now considered obsolete.
- For enhanced definition applications (mainly 525P), there are several 540 Mbit/s interfaces defined, as well as an interface standard for a dual-link 270 Mbit/s interface. These are rarely encountered.
- For HDTV applications, the serial digital interface is defined by SMPTE 292M. Two bit rates are defined, 1.485 Gbit/s, and 1.485/1.001 Gbit/s. The factor of 1/1.001 is provided to allow SMPTE 292M to support video formats with frame rates of 59.94 Hz, 29.97 Hz, and 23.98 Hz, in order to be compatible with existing NTSC systems. The 1.485 Gbit/s version of the standard supports other frame rates in widespread use, including 60 Hz, 50 Hz, 30 Hz, 25 Hz, and 24 Hz. It is common to collectively refer to both standards as using a nominal bit rate of 1.5 Gbit/s.
- For very high-definition applications, requiring greater resolution, frame rate, or color fidelity than the HD-SDI interface can provide, the SMPTE 372M standard defines the dual link interface. As the name suggests, this interface consists of two SMPTE 292M interconnects operating in parallel. In particular, the dual link interface supports 10-bit, 4:2:2, 1080P formats at frame rates of 60 Hz, 59.94 Hz, and 50 Hz, as well as 12-bit color depth, RGB encoding, and 4:4:4 colour sampling.
- A nominal 3 Gbit/s interface (more accurately, 2.97 Gbit/s, but commonly referred to as "3 gig") was standardized by SMPTE as 424M in 2006. Revised in 2012 as SMPTE ST 424:2012, it supports all of the features supported by the dual 1.485 Gbit/s interface but requires only one cable rather than two.
Other interfaces
[edit]SMPTE 297-2006 defines an optical fiber system for transmitting bit-serial digital signals It is intended for transmitting SMPTE ST 259 signals (143 through 360 Mbit/s), SMPTE ST 344 signals (540 Mbit/s), SMPTE ST 292-1/-2 signals (1.485 Gbit/s and 1.485/1.001 Gbit/s) and SMPTE ST 424 signals (2.970 Gbit/s and 2.970/1.001 Gbit/s). In addition to optical specification, ST 297 also mandates laser safety testing and that all optical interfaces are labelled to indicate safety compliance, application and interoperability.[14]
An 8-bit parallel digital interface is defined by ITU-R Rec. 601; this is obsolete (however, many clauses in the various standards accommodate the possibility of an 8-bit interface).
Data format
[edit]In SD and ED applications, the serial data format is defined to 10 bits wide, whereas in HD applications, it is 20 bits wide, divided into two parallel 10-bit datastreams (known as Y and C). The SD datastream is arranged like this:
- Cb Y Cr Y' Cb Y Cr Y'
whereas the HD datastreams are arranged like this:
- Y
- Y Y' Y Y' Y Y' Y Y'
- C
- Cb Cr Cb Cr Cb Cr Cb Cr
For all serial digital interfaces (excluding the obsolete composite encodings), the native color encoding is 4:2:2 YCbCr format. The luminance channel (Y) is encoded at full bandwidth (13.5 MHz in 270 Mbit/s SD, ~75 MHz in HD), and the two chrominance channels (Cb and Cr) are subsampled horizontally and encoded at half bandwidth (6.75 MHz or 37.5 MHz). The Y, Cr, and Cb samples are co-sited (acquired at the same instance in time), and the Y' sample is acquired at the time halfway between two adjacent Y samples.
In the above, Y refers to luminance samples, and C to chrominance samples. Cr and Cb further refer to the red and blue "color difference" channels; see Component video for more information. This section only discusses the native color encoding of SDI; other color encodings are possible by treating the interface as a generic 10-bit data channel. The use of other colorimetry encodings, and the conversion to and from RGB colorspace, is discussed below.
Video payload (as well as ancillary data payload) may use any 10-bit word in the range 4 to 1,019 (00416 to 3FB16) inclusive; the values 0–3 and 1,020–1,023 (3FC16–3FF16) are reserved and may not appear anywhere in the payload. These reserved words have two purposes; they are used both for Synchronization packets and for Ancillary data headers.
Synchronization packets
[edit]A synchronization packet (commonly known as the timing reference signal or TRS) occurs immediately before the first active sample on every line, and immediately after the last active sample (and before the start of the horizontal blanking region). The synchronization packet consists of four 10-bit words, the first three words are always the same—0x3FF, 0, 0; the fourth consists of 3 flag bits, along with an error correcting code. As a result, there are 8 different synchronization packets possible.
In the HD-SDI and dual link interfaces, synchronization packets must occur simultaneously in both the Y and C datastreams. (Some delay between the two cables in a dual link interface is permissible; equipment which supports dual link is expected to buffer the leading link in order to allow the other link to catch up). In SD-SDI and enhanced definition interfaces, there is only one datastream, and thus only one synchronization packet at a time. Other than the issue of how many packets appear, their format is the same in all versions of the serial-digital interface.
The flags bits found in the fourth word (commonly known as the XYZ word) are known as H, F, and V. The H bit indicates the start of horizontal blank; and synchronization bits immediately preceding the horizontal blanking region must have H set to one. Such packets are commonly referred to as End of Active Video, or EAV packets. Likewise, the packet appearing immediately before the start of the active video has H set to 0; this is the Start of Active Video or SAV packet.
Likewise, the V bit is used to indicate the start of the vertical blanking region; an EAV packet with V=1 indicates the following line (lines are deemed to start at EAV) is part of the vertical interval, an EAV packet with V=0 indicates the following line is part of the active picture.
The F bit is used in interlaced and segmented-frame formats to indicate whether the line comes from the first or second field (or segment). In progressive scan formats, the F bit is always set to zero.
Line counter and CRC
[edit]In the high definition serial digital interface (and in dual-link HD), additional check words are provided to increase the robustness of the interface. In these formats, the four samples immediately following the EAV packets (but not the SAV packets) contain a cyclic redundancy check field, and a line count indicator. The CRC field provides a CRC of the preceding line (CRCs are computed independently for the Y and C streams) and can be used to detect bit errors in the interface. The line count field indicates the line number of the current line.
The CRC and line counts are not provided in the SD and ED interfaces. Instead, a special ancillary data packet known as an EDH packet may be optionally used to provide a CRC check on the data.
Line and sample numbering
[edit]Each sample within a given datastream is assigned a unique line and sample number. In all formats, the first sample immediately following the SAV packet is assigned sample number 0; the next sample is sample 1; all the way up to the XYZ word in the following SAV packet. In SD interfaces, where there is only one datastream, the 0th sample is a Cb sample; the 1st sample a Y sample, the 2nd sample a Cr sample, and the third sample is the Y' sample; the pattern repeats from there. In HD interfaces, each datastream has its own sample numbering—so the 0th sample of the Y datastream is the Y sample, the next sample the Y' sample, etc. Likewise, the first sample in the C datastream is Cb, followed by Cr, followed by Cb again.
Lines are numbered sequentially, starting from 1, up to the number of lines per frame of the indicated format (typically 525, 625, 750, or 1125 (Sony HDVS)). Determination of line 1 is somewhat arbitrary; however, it is unambiguously specified by the relevant standards. In 525-line systems, the first line of vertical blank is line 1, whereas in other interlaced systems (625 and 1125-line), the first line after the F bit transitions to zero is line 1.
Note that lines are deemed to start at EAV, whereas sample zero is the sample following SAV. This produces the somewhat confusing result that the first sample in a given line of 1080i video is sample number 1920 (the first EAV sample in that format), and the line ends at the following sample 1919 (the last active sample in that format). Note that this behavior differs somewhat from analog video interfaces, where the line transition is deemed to occur at the sync pulse, which occurs roughly halfway through the horizontal blanking region.
Link numbering
[edit]Link numbering is only an issue in multi-link interfaces. The first link (the primary link) is assigned a link number of 1, subsequent links are assigned increasing link numbers; so, the second (secondary) link in a dual-link system is link 2. The link number of a given interface is indicated by a VPID packet located in the vertical ancillary data space.
Note that the data layout in dual link is designed so that the primary link can be fed into a single-link interface, and still produce usable (though somewhat degraded) video. The secondary link generally contains things like additional LSBs (in 12-bit formats), non-cosited samples in 4:4:4 sampled video (so that the primary link is still valid 4:2:2), and alpha or data channels. If the second link of a 1080P dual link configuration is absent, the first link still contains a valid 1080i signal.
In the case of 1080p60, 59.94, or 50 Hz video over a dual link; each link contains a valid 1080i signal at the same field rate. The first link contains the 1st, 3rd, and 5th lines of odd fields and the 2nd, 4th, 6th, etc. lines of even fields, and the second link contains the even lines on the odd fields, and the odd lines on the even fields. When the two links are combined, the result is a progressive-scan picture at the higher frame rate.
Ancillary data
[edit]Like SMPTE 259M, SMPTE 292M supports the SMPTE 291M standard for ancillary data. Ancillary data is provided as a standardized transport for non-video payload within a serial digital signal; it is used for things such as embedded audio, closed captions, timecode, and other sorts of metadata. Ancillary data is indicated by a 3-word packet consisting of 0, 3FF, 3FF (the opposite of the synchronization packet header), followed by a two-word identification code, a data count word (indicating 0–255 words of payload), the actual payload, and a one-word checksum. Other than in their use in the header, the codes prohibited to video payload are also prohibited to ancillary data payload.
Specific applications of ancillary data include embedded audio, EDH, VPID and SDTI.
In dual link applications, ancillary data is mostly found on the primary link; the secondary link is to be used for ancillary data only if there is no room on the primary link. One exception to this rule is the VPID packet; both links must have a valid VPID packet present.
Embedded audio
[edit]Both the HD and SD serial interfaces allow for 16 audio channels to be embedded along with the video. The two interfaces use different audio encapsulation methods — SD uses the SMPTE 272M standard, whereas HD uses the SMPTE 299M standard. Typically, 48 kHz, 24-bit (20-bit in SD, but extendable to 24 bit) PCM audio is encoded, in a manner directly compatible with the AES3 digital audio interface. The data is placed in the horizontal blanking periods, when the SDI signal otherwise carries nothing useful (the receiver generates its own blanking signals from the TRS).
In dual-link applications, 32 channels of audio are available, as each link may carry 16 channels.
SMPTE ST 299-2:2010 extends the 3G SDI interface to be able to transmit 32 audio channels (16 pairs) on a single link.
EDH
[edit]As the standard definition interface carries no checksum, CRC, or other data integrity check, an EDH (Error Detection and Handling) packet may be optionally placed in the vertical interval of the video signal. This packet includes CRC values for both the active picture, and the entire field (excluding those lines at which switching may occur, and which should contain no useful data); equipment can compute their own CRC and compare it with the received CRC in order to detect errors.
EDH is typically only used with the standard definition interface; the presence of CRC words in the HD interface make EDH packets unnecessary.
VPID
[edit]VPID (or video payload identifier) packets are increasingly used to describe the video format. In early versions of the serial digital interface, it was always possible to uniquely determine the video format by counting the number of lines and samples between H and V transitions in the TRS. With the introduction of dual link interfaces, and segmented-frame standards, this is no longer possible; thus the VPID standard (defined by SMPTE 352M) provides a way to uniquely and unambiguously identify the format of the video payload.
Video payload and blanking
[edit]The active portion of the video signal is defined to be those samples which follow an SAV packet and precede the next EAV packet; where the corresponding EAV and SAV packets have the V bit set to zero. It is in the active portion that the actual image information is stored.
Color encoding
[edit]Several color encodings are possible in the serial digital interface. The default (and most common case) is 10-bit linearly sampled video data encoded as 4:2:2 YCbCr. (YCbCr is a digital representation of the YPbPr colorspace). Samples of video are stored as described above. Data words correspond to signal levels of the respective video components, as follows:
- The luma (Y) channel is defined such that a signal level of 0 mV is assigned the codeword 64 (40 hex), and 700 millivolts (full scale) is assigned the codeword 940 (3AC hex) .
- For the chroma channels, 0 mV is assigned the code word 512 (200 hex), −350 mV is assigned a code word of 64 (40 hex), and +350 mV is assigned a code word of 960 (3C0 hex).
Note that the scaling of the luma and chroma channels is not identical. The minimum and maximum of these ranges represent the preferred signal limits, though the video payload may venture outside these ranges (providing that the reserved code words of 0–3 and 1020–1023 are never used for video payload). In addition, the corresponding analog signal may have excursions further outside of this range.
Colorimetry
[edit]As YPbPr (and YCbCr) are both derived from the RGB colorspace, a means of converting is required. There are three colorimetries typically used with digital video:
- SD and ED applications typically use a colorimetry matrix specified in ITU-R Rec. 601.
- Most HD, dual link, and 3 Gbit/s applications use a different matrix, specified in ITU-R Rec. 709.
- The 1035-line MUSE HD standards specified by SMPTE 260M (primarily used in Japan and now largely considered obsolete), used a colorimetry matrix specified by SMPTE 240M. This colorimetry is nowadays rarely used, as the 1035-line formats have been superseded by 1080-line formats.
Other color encodings
[edit]The dual-link and 3 Gbit/s interfaces additionally support other color encodings besides 4:2:2 YCbCr, namely:
- 4:2:2 and 4:4:4 YCbCr, with an optional alpha (used for linear keying, a.k.a. alpha compositing) or data (used for non-video payload) channel
- 4:4:4 RGB, also with an optional alpha or data channel
- 4:2:2 YCbCr, 4:4:4 YCbCr, and 4:4:4 RGB, with 12 bits of color information per sample, rather than 10. Note that the interface itself is still 10 bit; the additional 2 bits per channel are multiplexed into an additional 10-bit channel on the second link.
If an RGB encoding is used, the three primaries are all encoded in the same fashion as the Y channel; a value of 64 (40 hex) corresponds to 0 mV, and 940 (3AC hex) corresponds to 700 mV.
12-bit applications are scaled in a similar fashion to their 10-bit counterparts; the additional two bits are considered to be LSBs.
Vertical and horizontal blanking regions
[edit]For portions of the vertical and horizontal blanking regions which are not used for ancillary data, it is recommended that the luma samples be assigned the code word 64 (40 hex), and the chroma samples be assigned 512 (200 hex); both of which correspond to 0 mV. It is permissible to encode analog vertical interval information (such as vertical interval timecode or vertical interval test signals) without breaking the interface, but such usage is nonstandard (and ancillary data is the preferred means for transmitting metadata). Conversion of analog sync and burst signals into digital, however, is not recommended—and neither is necessary in the digital interface.
Different picture formats have different requirements for digital blanking. For example, all so-called 1080-line HD formats have 1080 active lines, but 1125 total lines, with the remainder being vertical blanking.[1]
Supported video formats
[edit]The various versions of the serial digital interface support numerous video formats:
- The 270 Mbit/s interface supports 525-line, interlaced video at a 59.94 Hz field rate (29.97 Hz frame rate), and 625-line, 50 Hz interlaced video. These formats are highly compatible with NTSC and PAL-B/G/D/K/I systems respectively; and the terms NTSC and PAL are often (incorrectly) used to refer to these formats. PAL is a composite color encoding scheme, and the term does not define the line-standard (though it is most usually encountered with 625i), while the serial digital interface—other than the obsolete 143 Mbit/s and 177 Mbit/s forms—is a component standard.
- The 360 Mbit/s interface supports 525i and 625i widescreen. It can also be used to support 525p, if 4:2:0 sampling is used.
- The various 540 Mbit/s interfaces support 525p and 625p formats.
- The nominal 1.49 Gbit/s interfaces support most high-definition video formats. Supported formats include 1080/60i, 1080/59.94i, 1080/50i, 1080/30p, 1080/29.97p, 1080/25p, 1080/24p, 1080/23.98p, 720/60p, 720/59.94p, and 720/50p. In addition, there are several 1035i formats (an obsolete Japanese television standard), half-bandwidth 720p standards such as 720/24p (used in some film conversion applications, and unusual because it has an odd number of samples per line[citation needed]), and various 1080psf (progressive, segmented frame) formats. Progressive Segmented frames formats appear as interlace video but contain video which is progressively scanned. This is done to support analog monitors and televisions, many of which are incapable of locking to low field rates such as 30 Hz and 24 Hz.
- The 2.97 Gbit/s dual link HD interface supports 1080/60p, 1080/59.94p, and 1080/50p, as well as 4:4:4 encoding, greater color depth, RGB encoding, alpha channels, and nonstandard resolutions (often encountered in computer graphics or digital cinema).
- A quad link interface of 3G-SDI supports UHDTV-1 resolution 2160/60p
Related interfaces
[edit]In addition to the regular serial digital interface described here, there are several other similar interfaces which are similar to, or are contained within, a serial digital interface.
SDTI
[edit]There is an expanded specification called SDTI (Serial Data Transport Interface), which allows compressed (i.e. DV, MPEG and others) video streams to be transported over an SDI line. This allows for multiple video streams in one cable or faster-than-realtime (2x, 4x,...) video transmission. A related standard, known as HD-SDTI, provides similar capability over an SMPTE 292M interface.
The SDTI interface is specified by SMPTE 305M. The HD-SDTI interface is specified by SMPTE 348M.
ASI
[edit]The asynchronous serial interface (ASI) specification describes how to transport a MPEG Transport Stream (MPEG-TS), containing multiple MPEG video streams, over 75-ohm copper coaxial cable or multi-mode optical fiber. ASI is popular way to transport broadcast programs from the studio to the final transmission equipment before it reaches viewers sitting at home.
The ASI standard is part of the Digital Video Broadcasting (DVB) standard.
SMPTE 349M
[edit]The standard SMPTE 349M: Transport of Alternate Source Image Formats through SMPTE 292M, specifies a means to encapsulate non-standard and lower-bitrate video formats within an HD-SDI interface. This standard allows, for example, several independent standard-definition video signals to be multiplexed onto an HD-SDI interface and transmitted down one wire. This standard doesn't merely adjust EAV and SAV timing to meet the requirements of the lower-bitrate formats; instead, it provides a means by which an entire SDI format (including synchronization words, ancillary data, and video payload) can be encapsulated and transmitted as ordinary data payload within a 292M stream.
HDMI
[edit]
The HDMI interface is a compact audio/video interface for transferring uncompressed video data and compressed/uncompressed digital audio data from an HDMI-compliant device to a compatible computer monitor, video projector, digital television, or digital audio device. It is mainly used in the consumer area, but increasingly used in professional devices including uncompressed video, often called clean HDMI.
G.703
[edit]The G.703 standard is another high-speed digital interface, originally designed for telephony.
HDcctv
[edit]The HDcctv standard embodies the adaptation of SDI for video surveillance applications, not to be confused with TDI, a similar but different format for video surveillance cameras.
CoaXPress
[edit]The CoaXPress standard is another high-speed digital interface, originally designed for use with industrial cameras. The data rates for CoaXPress go up to 12.5 Gbit/s over a single coaxial cable. A 41 Mbit/s uplink channel and power over coax are also included in the standard.
References
[edit]- ^ a b Charles A. Poynton (2003). Digital Video and HDTV. Morgan Kaufmann. ISBN 978-1-55860-792-7.
- ^ a b c d e f John Hudson (2013). "3Gb/s SDI for Transport of 1080p50/60, 3D, UHDTV1 / 4k and Beyond" (PDF). Archived from the original (PDF) on 2016-03-04. Retrieved 2014-03-13.
- ^ Francis Rumsey, John Watkinson (2004). Digital interface handbook. Taylor & Francis. ISBN 9780240519098.
- ^ "AJA Introduces Hi5-3D Mini-Converter For Stereo 3D Monitoring by Scott Gentry – ProVideo Coalition". 10 September 2010.
- ^ "Archived copy" (PDF). Archived from the original (PDF) on 2021-05-08.
{{cite web}}: CS1 maint: archived copy as title (link) - ^ Signalling and Transport of HDR and Wide Colour Gamut Video over 3G-SDI Interfaces. May 23, 2020 – via tech.ebu.ch.
- ^ a b ST 2081-10:2015 – 2160-Line and 1080-Line Source Image and Ancillary Data Mapping for Single-Link 6G-SDI. IEEE. 2015-03-19. doi:10.5594/SMPTE.ST2081-10.2015. ISBN 978-1-61482-857-0.
- ^ a b ST 2082-10:2015 - 2160-line Source Image and Ancillary Data Mapping for 12G-SDI. IEEE. 2015-03-19. doi:10.5594/SMPTE.ST2082-10.2015. ISBN 978-1-61482-860-0.
- ^ Advice on the use of 3 Gbit/s HD-SDI interfaces (PDF). European Broadcasting Un. July 2011. Retrieved 20 July 2015.
- ^ "Recommended Transmission Distance at Serial Digital Data Rates" (PDF). Belden. Archived from the original (PDF) on 2015-02-26. Retrieved 20 July 2015.
- ^ "Transport of alternate source formats through Recommendation ITU-R BT.1120" (PDF). International Telecommunication Union. Retrieved February 27, 2019.
- ^ [dead link]
- ^ "March 2014 Standards Quarterly Report (page 28)" (PDF). SMPTE. Archived from the original (PDF) on 16 July 2014. Retrieved 19 September 2014.
- ^ SMPTE (2013). "3Gb/s SDI for Transport of 1080p50/60, 3D, UHDTV1 / 4k and Beyond" (PDF). Archived from the original (PDF) on 2016-03-03. Retrieved 2017-03-17.
Sources
[edit]Standards
[edit]- Society of Motion Picture and Television Engineers: SMPTE 274M-2005: Image Sample Structure, Digital Representation and Digital Timing Reference Sequences for Multiple Picture Rates
- Society of Motion Picture and Television Engineers: SMPTE 292M-1998: Bit-Serial Digital Interface for High Definition Television
- Society of Motion Picture and Television Engineers: SMPTE 291M-1998: Ancillary Data Packet and Space Formatting
- Society of Motion Picture and Television Engineers: SMPTE 372M-2002: Dual Link 292M Interface for 1920 x 1080 Picture Raster
External links
[edit]Serial digital interface
View on GrokipediaOverview
Definition and purpose
The Serial Digital Interface (SDI) is a family of standards for the serial transmission of uncompressed digital video signals, embedded audio, and ancillary metadata over coaxial or fiber optic cables in professional video environments. Developed by the Society of Motion Picture and Television Engineers (SMPTE), SDI serializes parallel video data into a single high-speed stream, enabling efficient and robust signal handling without the need for multiple parallel connections. The foundational standard, SMPTE 259M, was first published in 1989 to support standard-definition formats, establishing SDI as a cornerstone for digital video workflows.[11][12] The core purpose of SDI is to facilitate reliable, low-latency transport of high-fidelity video in demanding production settings, such as studios, outside broadcast (OB) vans, and post-production suites, where signal degradation must be minimized. By prioritizing uncompressed transmission and electrical equalization, SDI overcomes the distance and interference limitations inherent in parallel interfaces or consumer-oriented connections like HDMI, allowing signals to travel up to hundreds of meters on coaxial cable or longer distances via fiber optics while preserving quality. This design ensures seamless integration in real-time applications, contrasting with compressed formats that introduce delays or artifacts unsuitable for live or critical workflows.[13][14] In practice, SDI serves key applications in professional video routing, including connections from cameras to switchers in live television production and signal distribution across broadcast facilities and film sets. It enables efficient infrastructure sharing in television broadcasting, where multiple video sources must be switched and monitored without quality loss, and supports post-production pipelines by providing a stable backbone for editing and effects processing. These capabilities have made SDI indispensable for maintaining the integrity of professional-grade content creation and delivery.[15][16]History
The Serial Digital Interface (SDI) originated in the late 1980s as part of the broadcasting industry's shift from analog composite video signals to digital formats, enabling more reliable transmission of standard-definition video. Introduced in 1989 by the Society of Motion Picture and Television Engineers (SMPTE), the initial standard, SMPTE 259M, defined a 10-bit serial interface operating at 270 Mbit/s for 525/60 (NTSC) and 625/50 (PAL) component video signals, replacing analog systems like composite video that suffered from degradation over distance.[17] This development addressed the growing need for digital workflows in professional video production, where analog limitations hindered signal integrity in studio and transmission environments.[1] Subsequent advancements in SDI were driven by the demand for higher resolutions and frame rates, leading to a series of standards that extended bit rates while maintaining compatibility with existing coaxial infrastructure. In 1998, SMPTE 292M established HD-SDI at 1.485 Gbit/s, supporting high-definition formats up to 1080i/60, which became essential for the emerging HDTV era in broadcasting.[17] This was followed by SMPTE 424M in 2006, introducing 3G-SDI at 2.97 Gbit/s to accommodate progressive HD formats like 1080p/60 without requiring dual-link configurations.[18] Further evolution included SMPTE ST 2081 in 2015 for 6G-SDI at 5.94 Gbit/s, enabling single-link 4K/UHD transmission up to 2160p/30, and SMPTE ST 2082 in the same year for 12G-SDI at 11.88 Gbit/s, supporting 2160p/60.[18] These milestones reflected the 2000s-2010s push toward UHD and higher frame rates in film, television, and live events, prioritizing uncompressed video quality over analog's vulnerabilities.[1] The most recent iteration, 24G-SDI under SMPTE ST 2083 published in 2020, operates at 24 Gbit/s to handle 8K resolutions such as 4320p/30, ensuring SDI's scalability for ultra-high-definition production.[19] As of 2025, SDI continues to dominate live broadcast and production environments due to its low-latency, uncompressed nature and robust error handling, particularly in high-stakes scenarios like sports and concerts where reliability is paramount.[20] However, it faces gradual displacement by IP-based workflows, such as SMPTE ST 2110, which offer greater flexibility and scalability in networked systems, though hybrid SDI-IP setups remain common during this transition.[21]Electrical interface
Standards and bit rates
The Serial Digital Interface (SDI) encompasses a series of standards developed by the Society of Motion Picture and Television Engineers (SMPTE) to define bit-serial transmission for uncompressed digital video signals over coaxial or fiber optic cables. These standards specify electrical characteristics, data rates, and mapping methods for various video resolutions and frame rates, evolving from standard-definition to ultra-high-definition formats.[22] The foundational standard, SMPTE 259M, introduced in 1989, defines SD-SDI at a nominal bit rate of 270 Mbit/s, supporting interlaced formats such as 525i (480i) and 625i (576i) in 10-bit YCbCr 4:2:2 color space.[23] Subsequent revisions and related standards like SMPTE 344M extended support to enhanced-definition formats at 540 Mbit/s, but the core 270 Mbit/s rate remains the primary for SD-SDI applications.[24] For high-definition video, SMPTE 292M, published in 1998, establishes HD-SDI with a single-link bit rate of 1.485 Gbit/s (or 1.485/1.001 Gbit/s for certain frame rates), accommodating 1080i and 720p formats in 10-bit YCbCr 4:2:2.[25] This standard introduced parallel-link options for higher bandwidth needs, such as dual-link configurations for 1080p. Advancing to support progressive-scan HD, SMPTE 424M from 2006 defines 3G-SDI at 2.97 Gbit/s (or 2.97/1.001 Gbit/s), enabling single-link transmission for 1080p up to 60 fps or dual-link for deeper color formats, with mapping structures outlined in SMPTE 425M including Level A (progressive) and Level B (segmented frame or dual-link).[26][27] Higher-speed interfaces include 6G-SDI per SMPTE ST 2081 (2015), operating at 5.94 Gbit/s (or 5.94/1.001 Gbit/s) for single-link 1080p60 in 10-bit formats, with a document suite covering electrical, optical, and mapping specifications.[28] 12G-SDI, defined by SMPTE ST 2082 (also 2015), achieves 11.88 Gbit/s (or 11.88/1.001 Gbit/s) to support single-link 2160p60 (4K UHD) in YCbCr 4:2:2 10-bit, reducing cabling complexity for ultra-high-definition workflows.[28] The most recent coaxial standard, 24G-SDI under SMPTE ST 2083 (2020), provides 23.76 Gbit/s for single-link transmission of 2160p120 or 4320p30 (8K) in 10- or 12-bit depths, addressing high-frame-rate and higher-resolution production needs.[29] Additionally, SMPTE ST 297 (2006, revised 2015) specifies fiber optic transmission systems compatible with SDI signals from SMPTE 259M through 424M, enabling longer-distance deployments without electrical limitations.[30]| Standard | Name | Year | Bit Rate (Gbit/s) | Example Supported Formats |
|---|---|---|---|---|
| SMPTE 259M | SD-SDI | 1989 | 0.270 | 525i/625i (480i/576i) |
| SMPTE 292M | HD-SDI | 1998 | 1.485 | 1080i, 720p |
| SMPTE 424M | 3G-SDI | 2006 | 2.97 | 1080p60 (single/dual-link) |
| SMPTE ST 2081 | 6G-SDI | 2015 | 5.94 | 1080p60 |
| SMPTE ST 2082 | 12G-SDI | 2015 | 11.88 | 2160p60 |
| SMPTE ST 2083 | 24G-SDI | 2020 | 23.76 | 2160p120, 4320p30 |
Transmission characteristics
The Serial Digital Interface (SDI) employs a 10-bit parallel-to-serial conversion process, where parallel video data is scrambled using a polynomial-based scrambler to ensure DC balance and self-clocking properties, followed by non-return-to-zero inverted (NRZI) encoding to minimize baseline wander and facilitate clock recovery at the receiver.[31] This encoding scheme, specified in SMPTE ST 259M for standard-definition rates and extended in subsequent standards like SMPTE ST 292M, converts the non-return-to-zero (NRZ) serial stream into NRZI format using the polynomial (X + 1), producing transitions for every bit change to maintain signal integrity over coaxial media.[31] Higher-rate variants, such as those in SMPTE ST 424M, retain NRZI but incorporate enhanced scrambling to handle increased bit rates up to 2.97 Gbit/s.[32] SDI signals are transmitted as differential voltage levels with a peak-to-peak amplitude of 800 mV ±10%, a DC offset of 0.0 V ±0.5 V, and overshoot/undershoot limited to less than 10% to prevent distortion.[33] The interface maintains a characteristic impedance of 75 Ω, with return loss better than 15 dB across the signal bandwidth to minimize reflections and ensure efficient power transfer.[33] These electrical parameters, defined in SMPTE standards such as ST 259M and ST 292M, apply uniformly across SDI variants to support reliable transmission in professional video environments.[34] Connections utilize BNC connectors compliant with IEC 61169-8, which provide a bayonet-style coupling for quick, secure mating on 75 Ω coaxial cables like RG-59 or RG-6/U.[35] Transmission distances vary by bit rate and cable type due to attenuation; for example, at 270 Mb/s (SD-SDI), RG-6 supports up to 300 m, while at 1.485 Gb/s (HD-SDI), distances reduce to approximately 100 m on the same cable.[36] For higher rates like 2.97 Gb/s (3G-SDI), RG-6 limits runs to about 80 m, and 11.88 Gb/s (12G-SDI) further constrains to 30-50 m, necessitating low-loss cables to meet the 20-40 dB loss budgets specified in SMPTE ST 2082-1.[36] To counteract frequency-dependent attenuation in long cable runs, SDI receivers incorporate adaptive equalization circuits that boost high-frequency components, restoring the signal to meet eye pattern requirements with at least 40% eye opening for reliable sampling.[33] Reclocking at intermediate points uses phase-locked loops to extract and regenerate the clock, reducing accumulated jitter; timing jitter must remain below 0.2 unit intervals (UI) for 270 Mb/s links and up to 0.3 UI alignment jitter for rates above 3 Gb/s, as per SMPTE specifications.[33] These measures ensure low bit error rates, typically below 10^{-12}, in cascaded systems. For bandwidth demands exceeding single-link capacities, such as uncompressed 1080p60 video, multi-link configurations aggregate parallel SDI links; dual-link HD-SDI combines two 1.485 Gb/s interfaces per SMPTE ST 372M, while quad-link 3G-SDI uses four 2.97 Gb/s links to achieve 12 Gb/s equivalents for 4K formats under SMPTE ST 425-3. These setups distribute data across links with defined mapping structures to maintain synchronization and simplify cabling in production environments.Data format
Synchronization and framing
The Serial Digital Interface (SDI) structures its data stream into 10-bit words to facilitate reliable transmission of uncompressed digital video. Each word represents a sample or timing element, serialized at high bit rates defined by SMPTE standards such as ST 259 for standard definition and ST 292 for high definition. This word-based format enables deserializers to reconstruct the parallel data bus from the serial stream without additional synchronization lines.[32] Synchronization within SDI relies on Timing Reference Signals (TRS), specifically Start of Active Video (SAV) and End of Active Video (EAV) packets, which delineate the boundaries of active video regions in each line. These packets consist of four consecutive 10-bit words: the first three words are fixed as 3FFh, 000h, and 000h in hexadecimal, forming a unique preamble for detection. The fourth word has bit 9 set to 1 and is denoted in hex as starting with 2 or 3 depending on the F bit (e.g., 200h for SAV in field 1), where bits 8-6 encode the F-bit (bit 8 for field number), V-bit (bit 7 for vertical blanking status), and H-bit (bit 6, set to 0 for SAV and 1 for EAV), with bits 5-0 set to 0. This structure ensures robust word alignment, as the preamble's distinct pattern—reversed from ancillary data flags—allows receivers to identify and lock onto line starts and ends even in noisy environments. For HD formats under SMPTE ST 292, EAV packets are extended with two additional words for line numbering and two for cyclic redundancy check (CRC) values, enhancing framing integrity across the 1125 total lines per frame.[32] Framing in SDI operates on a line-by-line basis, with each horizontal line comprising a fixed number of 10-bit words to maintain constant bit rates across formats. For example, in 1080i/59.94 HD-SDI, each line totals 2200 words, including 1920 active video words multiplexed from Y, Cb, and Cr components in 4:2:2 sampling, plus blanking intervals for SAV, EAV, and ancillary data. The stream is self-clocking, embedding timing information directly in the data to eliminate the need for a separate clock signal; this is achieved through bit scrambling (a pseudo-random polynomial to ensure DC balance and frequent transitions) followed by non-return-to-zero inverted (NRZI) encoding, which toggles the signal on bit changes for reliable clock extraction. Receivers employ phase-locked loops (PLLs) to recover the embedded clock from these transitions, reclocking the data to suppress jitter accumulated over long cable runs. Bit rates, such as 1.485 Gbps for HD-SDI, determine the word rate (e.g., 148.5 MHz), influencing line duration but not the framing structure itself.[32][37] In multi-link configurations, such as dual-link HD-SDI (SMPTE ST 372) or quad-link 12G-SDI (SMPTE ST 2082), synchronization across parallel interfaces requires explicit alignment to reconstruct the full video frame. Link Number (LN) bytes, embedded in payload identification packets per SMPTE ST 352, designate each link (e.g., Link 1 as primary, subsequent as 2–4), enabling receivers to reorder and phase-align streams with timing offsets limited to 40 ns at the source. This ensures seamless deserialization, particularly for ultra-high-definition formats exceeding single-link capacities.[38]Line and sample numbering
In serial digital interfaces (SDI), line numbering provides a structured addressing scheme for video frames, enabling precise synchronization and data placement. For high-definition formats defined in SMPTE ST 292-1, each line is identified by an 11-bit counter embedded in the timing reference signals (TRS) following the end of active video (EAV) and start of active video (SAV) words. These counters, denoted as LN0 and LN1, range from 1 to 1125, starting with the first line of vertical blanking and incrementing sequentially through active video lines. This numbering supports both progressive and interlaced scanning, with line 1 marking the beginning of field 1 in interlaced modes. The horizontal ancillary data (HANC) space occupies the horizontal blanking interval between the EAV and the next SAV on the same line, while vertical ancillary (VANC) data resides in the vertical blanking interval, typically lines 1 through 20 or equivalent, depending on the format. Sample numbering within each line begins at 0 immediately following the SAV word, encompassing the active video region before the EAV. For example, in 1920×1080 progressive formats like 1080p, there are 1920 active luma (Y) samples per line, with chroma (Cb, Cr) subsampled according to the 4:2:2 ratio, resulting in 960 Cb/Cr samples multiplexed pairwise. The total samples per line, including blanking, vary by frame rate to maintain constant bit rates—such as 2200 samples for 59.94/60 Hz or 2750 for 23.98/24 Hz—ensuring consistent data flow. For formats with non-square pixels, such as standard-definition SDI, a multifactor mapping adjusts sample counts to align with square-pixel representation in the interface, preserving aspect ratios without altering the digital stream structure. Field identification is handled by the F bit in the TRS words of SAV and EAV. In interlaced formats, the F bit is set to 0 for field 1 (odd lines) and 1 for field 2 (even lines), while it remains 0 for progressive scan to indicate a single field per frame. This bit, combined with the V bit (1 during vertical blanking, 0 otherwise), allows receivers to distinguish frame structure and reconstruct images accurately. In multi-link SDI configurations, such as dual-link HD-SDI per SMPTE ST 372M or quad-link 3G-SDI per SMPTE ST 425-5, link numbering ensures component mapping across parallel interfaces. Links are designated as 0 through 3 (or A/B for dual), with bytes in the data stream specifying assignments: link 0 typically carries the Y (luma) component, link 1 the Cb (blue-difference), link 2 the Cr (red-difference), and link 3 the alpha channel for transparency in 4:4:4:4 formats. For YCbCr 4:4:4:4 10-bit, even-indexed Cb and Cr samples are mapped to link 0's Cb/Cr space alongside Y, while odd-indexed samples and alpha occupy link 1; similar partitioning applies to RGB variants, distributing bandwidth evenly to support higher resolutions like 1080p at increased bit depths.Error detection
The primary mechanism for error detection in serial digital interfaces (SDI) involves cyclic redundancy checks (CRC) embedded within the timing reference signals and ancillary data packets to identify bit errors in transmitted video lines and frames. In high-definition SDI (HD-SDI), as specified in SMPTE ST 292-1, each line's end-of-active-video (EAV) sequence includes two 10-bit CRC words (one for the Y/luma channel and one for the CbCr/chroma channel) within the extended timing reference signal (TRS), computed separately over the active video samples and horizontal ancillary data (HANC) of the preceding line (from the word following the previous SAV to the word before the line number words). This line CRC enables per-line error detection, allowing receivers to flag transmission errors specific to individual lines without relying solely on frame-level checks.[39] For standard-definition SDI (SD-SDI) per SMPTE ST 259:1, which lacks built-in line CRCs in its three-word TRS structure, error detection relies on the Error Detection and Handling (EDH) system outlined in SMPTE RP 165. EDH inserts ancillary data packets containing two 16-bit CRC checkwords—one for the full field (all active and blanking samples, excluding switching lines) and one for the active picture area only—along with error flags that report line errors (single-line issues), block errors (multiple consecutive line errors), and aggregate errors (cumulative frame issues). These CRCs are generated using the polynomial , providing robust detection of bit flips across the field. EDH packets are placed on designated lines (e.g., line 9 for even fields in NTSC), enabling equipment to monitor and isolate faulty components in the signal chain.[40][41] In HD-SDI and extensions, EDH is adapted with separate 18-bit CRCs for luma (Y) and chroma (CBCR) channels, using the polynomial , inserted in ancillary packets to cover the full frame or active video, complementing the line-level CRCs. The Video Payload Identifier (VPID), defined in SMPTE ST 352, embeds a four-word ancillary packet (using DID and SDID codes) that specifies the video format, bit depth, and mapping structure, providing contextual information to receivers for validating error detection against the expected payload configuration.[40] Higher-rate SDI variants, such as 12G-SDI in SMPTE ST 2082-1, retain similar line CRC and EDH mechanisms scaled to support up to 2160-line formats, with CRC generation and checking integrated into the TRS extensions for per-line integrity. In certain mappings (e.g., ST 2082-10 for 12G Level A), forward error correction (FEC) may be optionally applied using Reed-Solomon codes over the serial stream to not only detect but also correct errors, enhancing reliability over longer cable runs, though this is not mandatory for core video transport. 24G-SDI, as in preliminary extensions, follows analogous CRC-based detection with potential FEC options for ultra-high-resolution payloads.[42]Ancillary data
Embedded audio
Embedded audio in Serial Digital Interface (SDI) transports multi-channel AES3 digital audio signals within the ancillary data spaces of the video stream, allowing synchronized audio and video transmission without separate cables. This embedding follows AES3 formatting, where audio samples are packetized into horizontal ancillary (HANC) and vertical ancillary (VANC) spaces during blanking intervals, ensuring compatibility with video timing.[43][32] For standard-definition SDI (SD-SDI) at 270 Mb/s, SMPTE ST 272 specifies the embedding of up to 16 channels of 48 kHz, 24-bit audio, organized into four groups of four channels each, with each group derived from two AES3 pairs. Audio packets are inserted into HANC spaces, with each packet containing up to 64 audio samples aligned to video lines for synchronous playback at 48 kHz. The packet structure begins with an ancillary data flag (ADF: three words of 0x000, 0x3FF, 0x3FF), followed by a Data Identifier (DID) indicating the audio group—such as 0x61 for group 1 audio data—and a Secondary Data Identifier (SDID) for subgroup details, then Data Block Number (DBN), Data Count (DC), user data words (UDW) holding the audio samples, and a checksum. Each 24-bit AES3 sample (plus validity, user, and channel status bits) is mapped across three 10-bit UDW: the X word carries the Z-bit, channel code, and lower audio bits; X+1 the middle bits; and X+2 the upper bits with auxiliary and parity information.[43][32][44] In high-definition SDI (HD-SDI) at 1.5 Gb/s and 3G-SDI at 3 Gb/s, SMPTE ST 299-1 extends support to 16 channels of 24-bit audio at 48 kHz (or optionally 32 kHz and 44.1 kHz), embedded similarly in HANC/VANC via four groups, but with enhanced packetization for higher data rates—each sample mapped across four 10-bit words including clock phase (CLK) and error correction code (ECC) fields for improved integrity. For ultra-high-definition formats like 6G-SDI and 12G-SDI, the same ST 299-1 framework applies, but with added capacity for 96 kHz sampling rates and up to 32 channels in dual-link configurations to accommodate immersive audio such as 7.1 or 22.2 surround, enabling higher fidelity for cinema and broadcast applications. Audio control packets accompany data packets, carrying metadata like sample alignment and active channel flags.[32][44][45] Channel mapping in embedded audio supports configurations from mono and stereo to multi-channel setups like 5.1 (left, right, center, low-frequency effects, left/rear surround, right/rear surround) and 7.1 (adding left/rear and right/rear), with channels assigned sequentially across groups—for instance, 5.1 occupying channels 1–4 in group 1 and 5–6 in the same group, while metadata in the control packet specifies embedding position, gain levels, and downmix parameters to maintain audio balance during transmission. This mapping ensures interoperability across AES3 sources and SDI receivers, with user bits preserved for additional audio descriptors.[32][43]| Audio Group | DID (Hex) for SD-SDI (ST 272) | DID (Hex) for HD/3G-SDI (ST 299-1) |
|---|---|---|
| Group 1 | 0x61 | 0x47 |
| Group 2 | 0x62 | 0x48 |
| Group 3 | 0x63 | 0x49 |
| Group 4 | 0x64 | 0x4A |
Metadata packets
Metadata packets in the Serial Digital Interface (SDI) refer to non-audio ancillary data packets that convey essential control and descriptive information, such as timecode, captions, and format identifiers, embedded within the horizontal ancillary data (HANC) or vertical ancillary data (VANC) spaces of the video signal. These packets follow the formatting defined in SMPTE ST 291-1, which specifies a structure consisting of an ancillary data flag (ADF) sequence, a data identification word (DID) for packet type recognition, an optional secondary data identification word (SDID) for type 2 packets, a data count (DC) word indicating the number of user data words (up to 255 bytes), the user data words themselves, and a checksum word for verification.[46] This structure allows for flexible embedding of metadata without interfering with the active video payload. Timecode metadata is embedded as ancillary data packets to synchronize video frames, supporting both linear timecode (LTC) and vertical interval timecode (VITC) formats as per SMPTE ST 12-1, with transmission details in SMPTE ST 12-2. These packets use DID = 0x60 and SDID = 0x60 in a type 2 format, placing up to 32 words of timecode data (including hours, minutes, seconds, frames, and user bits) in the VANC space, typically lines 9 through 20 for VITC compatibility in high-definition formats.[47][48] LTC can alternatively be carried in HANC spaces for continuous audio-like synchronization across fields. Captions and subtitles are transported via dedicated ancillary packets, primarily in the HANC space, to ensure compatibility with line 21 data services in legacy systems. For CEA-608 (analog-compatible closed captions), packets follow SMPTE ST 334 with DID = 0x61 and SDID = 0x02, encapsulating two bytes of caption data per packet and requiring sequencing across multiple packets for complete lines. CEA-708 (digital closed captions) uses DID = 0x61 and SDID = 0x01 in a type 2 packet, supporting up to 255 bytes of compressed caption data including multiple services, fonts, and positioning, often sequenced in VANC for HD formats to align with field timing.[49][48] Other metadata includes the Active Format Description (AFD), which describes the active picture aspect ratio and letterbox/pan-scan status, carried in type 2 packets with DID = 0x41 and SDID = 0x05 per SMPTE ST 291-1 registration, typically in HANC line 10 or 13. The Video Payload Identifier (VPID) provides rapid format identification for SDI signals, using a type 1 packet with DID = 0x41 (no SDID) and four user data words encoding details like resolution, frame rate, and scan type as defined in SMPTE ST 352; for example, the payload 0x41 0x04 0x04 0x00 identifies 1080i at 50 Hz.[50][48][51] These packets enable downstream devices to auto-configure without parsing the full video signal.Video payload
Color encoding
In serial digital interface (SDI) transmissions, video samples are primarily encoded using component YCbCr 4:2:2 format, where the luminance (Y) component is sampled at the full pixel rate and the chrominance (Cb and Cr) components are subsampled at half the rate to achieve a 4:2:2 ratio, optimizing bandwidth while preserving perceptual quality. For progressive scan video, the encoding employs Y'CbCr to account for the non-linear gamma correction in the luma signal. This format is standardized for standard-definition (SD) video under SMPTE ST 259 and for high-definition (HD) under SMPTE ST 292-1, ensuring compatibility across broadcast equipment.[52][53] The standard bit depth for SDI video samples is 10 bits per component, providing 1024 quantization levels for enhanced dynamic range and reduced banding compared to 8-bit encoding.[53] In this scheme, the Y component is quantized over digital codes 4 to 1019, corresponding to black at code 4 and peak white at 1019, while Cb and Cr range from 64 to 960, with digital zero (neutral color) at 512. For high dynamic range (HDR) content in higher-speed interfaces like 12G-SDI and beyond, 12-bit depth is supported under SMPTE ST 2082-1, extending the quantization range to 4096 levels per component for greater precision in highlight and shadow details. RGB encoding is available but limited to specific mappings, such as dual-link HD-SDI under SMPTE ST 372, where it supports 10-bit 4:4:4 RGB for applications requiring full chrominance sampling without subsampling. Colorimetry in SDI adheres to ITU-R BT.601 for SD formats, defining primaries and white point (D65) suitable for 525-line systems with a narrower color gamut. For HD and ultra-high-definition (UHD) formats, ITU-R BT.709 is used, expanding the color gamut with updated red, green, and blue primaries to better match modern displays. Transfer functions for standard dynamic range (SDR) content follow a power-law curve approximating gamma 2.4, as specified in BT.709, to compensate for display non-linearities. HDR support in SDI incorporates perceptual quantizer (PQ) from SMPTE ST 2084 or hybrid log-gamma (HLG) from ITU-R BT.2100, enabling absolute or relative luminance mapping up to 10,000 nits for PQ and backward-compatible grading for HLG, respectively. Advanced SDI configurations, such as quad-link 12G-SDI under SMPTE ST 2082-10, support YUV 4:4:4:4 encoding for uncompressed full-bandwidth chrominance in UHD workflows, including an alpha channel for keying and compositing operations in post-production. This allows for RGBA pixel formats at 10- or 12-bit depths, facilitating high-fidelity color grading and visual effects without chroma subsampling artifacts.[54]Blanking regions
In the Serial Digital Interface (SDI), blanking regions refer to the temporal and spatial intervals outside the active video area, originally derived from analog television standards to accommodate synchronization signals and ancillary data without interfering with picture content. These regions include horizontal blanking, which occurs at the end of each line, and vertical blanking, which spans multiple lines between fields or frames, allowing for the insertion of timing information, error detection, and non-video data such as audio or metadata.[27][32] Horizontal blanking in standard-definition SDI (SD-SDI), as defined by SMPTE 259M, consists of 280 pixels per line, encompassing the start-of-active-video (SAV) and end-of-active-video (EAV) synchronization codes along with horizontal ancillary (HANC) space for data packets. This results in a total line length of 858 pixels, with 720 pixels allocated to active video, leaving the remaining space for blanking and overscan allowances that prevent edge artifacts on displays. In high-definition SDI (HD-SDI), per SMPTE 292M and SMPTE 274M, horizontal blanking is similarly 280 timing samples, but the total line comprises 2200 10-bit words for 1920x1080 formats, supporting HANC regions immediately following EAV and preceding SAV. These blanking durations ensure compatibility with legacy equipment while providing bandwidth for embedded services.[27][55][32] Vertical blanking intervals vary by format and scanning method, typically ranging from 20 to 45 lines to separate active video fields or frames. For SD-SDI in 525-line systems, vertical blanking includes approximately 20 lines at the top of each field (e.g., lines 1–20), with additional lines at the bottom, totaling around 39 non-active lines out of 525, enabling vertical ancillary (VANC) data placement. In HD-SDI for 1080i interlaced formats under SMPTE 274M, the total of 1125 lines includes 45 blanking lines, with VANC often occupying lines 1–20 in the full-field blanking region before active video begins on line 21. Progressive formats, such as 1080p, exhibit uniform vertical blanking across the frame without field splits, while interlaced signals divide blanking into odd and even fields for alternating line scans. Full-field vertical blanking utilizes the entire line width for data, contrasting with partial-field approaches that limit ancillary insertion to specific segments. Overscan allowances in active video regions, such as the 720x486 area in SD-SDI, account for display variations by defining safe action and title areas within the digital blanking structure.[27][32][55] Variations in blanking accommodate content types, including film transfers via 3:2 pulldown in interlaced SDI, where cadence flags are embedded in vertical blanking lines to signal the pulldown pattern for 24 fps film to 60-field video conversion, ensuring smooth playback without judder. Ancillary data insertion occurs primarily within these blanking regions, as detailed in related standards.[27][32]Supported formats
The Serial Digital Interface (SDI) supports a range of video formats across its variants, from standard definition (SD) to ultra-high definition (UHD) and beyond, accommodating both interlaced and progressive scan types with square-pixel aspect ratios for HD and higher resolutions.[56] Standard-definition formats, defined under SMPTE ST 259, include 480i at 59.94 fields per second (NTSC) and 576i at 50 fields per second (PAL), both utilizing a 4:3 aspect ratio.[57][58] High-definition formats, standardized in SMPTE ST 292 for HD-SDI, encompass 720p and 1080i/p at frame rates ranging from 23.98 Hz to 60 Hz, with a 16:9 aspect ratio.[59][53] The 3G-SDI extension (SMPTE ST 424) further enables 1080p at up to 60 Hz, supporting progressive formats at higher frame rates while maintaining compatibility with earlier HD timings.[60][27] For UHD and 4K resolutions (2160p), support is provided through higher-speed variants: 6G-SDI (SMPTE ST 2081) handles 2160p at 24–30 Hz in single-link configurations, often using quad-link 3G-SDI for broader compatibility, while 12G-SDI (SMPTE ST 2082) accommodates 2160p at 24–60 Hz via single-link transmission.[56][19] These formats preserve square-pixel mapping and support both progressive and select interlaced modes where applicable.[61] Proposed 24G-SDI (SMPTE ST 2083, in development as of 2025) is intended to support 8K resolutions such as 4320p at 30 Hz in single-link mode and high frame rates like 2160p120 through multi-link setups. As of 2025, 24G-SDI remains in the proposal stage, with multi-link 12G-SDI serving as an interim solution for 8K workflows.[62]| SDI Variant | Key Standards | Supported Resolutions and Frame Rates | Aspect Ratio | Link Configuration |
|---|---|---|---|---|
| SD-SDI | SMPTE ST 259 | 480i59.94, 576i50 | 4:3 | Single-link |
| HD-SDI | SMPTE ST 292 | 720p/1080i/p (23.98–60 Hz) | 16:9 | Single-link |
| 3G-SDI | SMPTE ST 424 | 1080p (up to 60 Hz) | 16:9 | Single-link |
| 6G-SDI | SMPTE ST 2081 | 2160p (24–30 Hz), 1080p120 | 16:9 | Single/quad-link |
| 12G-SDI | SMPTE ST 2082 | 2160p (24–60 Hz) | 16:9 | Single/quad-link |
| 24G-SDI | SMPTE ST 2083 (proposed) | Planned: 4320p30, 2160p120 (multi-link) | 16:9 | Single/multi-link (proposed) |