Ethernet is poised to become the backbone of the live television production industry. Live television broadcasts have for decades been produced using the circuit-switched method, but recent networking advances have enhanced Ethernet to the point where packet-switching can reliably process and deliver multiple uncompressed, high-definition television signals. By moving to near-commodity Ethernet-based technologies the test television industry stands to make significant cost savings over currently-used switching and distribution equipment. This paper examines the current and near-future feasibility and impact of using Ethernet as the sole medium for distributing all signals in a television studio or outside broadcast.
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Ethernet-Based Live Television Production
1. SMPTE AUSTRALIA CONFERENCE 2011
21 JULY 2011, SYDNEY
Ethernet-Based
Live Television
Production
DAN FLETT
BROADCAST ENGINEER
GLOBAL TELEVISION
1
2. The Status Quo
Circuit-switched multicamera production
– Live means live – very low latencies
“Traditional” Ethernet unsuitable
– Drops packets
Broadcast-quality video over IP is possible
– e.g. contribution links or IPTV
– Too much latency
– Latency is unpredictable
2
4. Bandwidth Increases Over Time
HD-SDI 1.5Gbps 100 Gbit Ethernet
SD-SDI 270Mbps
100000
10000
1000
M bps
100 3G-SDI 3Gbps
10
1
1989 1994 1999 2004 2009
Year
VIDEO ETHERNET
Digital Video vs. Ethernet Link Bandwidth
4
5. Live Video over Ethernet –
The Challenges
Why use uncompressed video?
– Doesn’t require codecs everywhere
The big hurdle: Latency
– 1 frame of delay (40ms) is too much
Synchronisation
– Vision mixers have approx. 4 to 20µs
autotiming windows
– Can Ethernet deliver?
5
6. IEEE Audio/Video Bridging
Task Group
“Bounded” latency
– IEEE 802.1Qav Traffic Shaping for AV
Genlock over the network
– IEEE 802.1AS Timing and Synchronisation
– Based on IEEE 1588 Precision Time Protocol
Reserved bandwidth paths
– IEEE 802.1Qat Stream
Reservation Protocol
6
7. Other Helpful Technologies
Encapsulation of SDI for packet-switching
– High Bit Rate Media Transport (HBRMT)
– Adds Forward Error Correction (FEC)
IPv6
– Dynamic addressing - no need manually assign
static addresses
– Merge and separate networks with ease
7
8. Impact on Broadcast Systems
Cameras
– CCU is redundant
Signal Routers and Distribution Amps
– Are redundant!
Vision Mixers
– All signals via 2 100GbE connectors
8
9. Impact on Broadcast Systems
Video Replay Devices
– More flexibility in architecture
Multiviewers
– Auto detection of signals “out of box”
Signal Processors
– 30 processors in a frame, 2 connectors
9
10. Impact on Broadcast Systems
Monitors
– Copper 10 GbE port
– Multitouch screen, replaces router panels
Configuration / Control Server
– Combines network and broadcast systems
management
– Service Oriented Architecture
10
13. Cost Comparison:
Traditional 24 Camera Facility 2011
Video Router (576x576) $ 432,000
Distribution Amps x 50 $ 25,000
CCUs x 24 $ 600,000
Video<>Fibre mux gear $ 40,000
Audio, comms gear $ ?
Cabling, install costs $ ?
TOTAL:~ $1,097,000
13
14. Cost Comparison:
Packet-Switched 24 Camera Facility 2015
Ethernet Port Costs
100GBASE-LR4 x 6 $150,000
100GBASE-SR10 x 72 $360,000
10GBASE-LR x 52 $ 78,000
10GBASE-T x 148 $ 55,000
1000BASE-T x 156 $ 7,800
TOTAL:~ $651,300
14
15. Conclusion
Ethernet opens up significant efficiencies
– price reduction curve
Shorter job-cycle time
= Faster return on investment
AVB standards are open, royalty-free
Ignore Ethernet at your peril!
15