Video Codecs: How Digital Video Is Compressed by Orders of Magnitude
Video codecs compress raw video data by 100-1000× using spatial compression (DCT) and temporal compression (motion prediction), with two main lineages: H.26x and VP/AV.
Video codecs are standards for compressing raw video — which at 4K resolution generates ~12 Gbps uncompressed — into streams small enough for storage and streaming, typically achieving 100–1,000× compression with acceptable quality loss. ## Core Techniques All modern codecs exploit two types of redundancy: - **Spatial** (intra-frame): Within a single frame, adjacent pixels are similar. DCT (discrete cosine transform) converts pixel blocks to frequency components, and high-frequency detail is quantized away. - **Temporal** (inter-frame): Consecutive frames are mostly the same. Motion compensation encodes only the differences, predicting where blocks move between frames. Frames are classified as I-frames (independent, fully encoded), P-frames (predicted from previous frames), and B-frames (bidirectionally predicted from both past and future frames). ## Two Lineages - **ITU H.26x**: H.261 (1988), H.264 (AVC): The Video Codec That Enabled YouTube and Modern Streaming (2003), H.265 (HEVC): The Technically Superior Codec Killed by Patent Licensing (2013), H.266/VVC: The Latest Video Codec Standard Struggling with Patent Licensing (2020). Developed by standards bodies, patent-encumbered. - **Open-source VP/AV**: VP8, VP9: Google's Royalty-Free Video Codec That Powered YouTube's 4K Era (2013), AV1: The Royalty-Free Codec That Won the Web (2018). Royalty-free, backed by the Alliance for Open Media (AOMedia): The Consortium Behind AV1. Each generation roughly halves the bitrate needed for equivalent quality, though at increasing encoder complexity. **See also:** Video Codec History: H.261 to AV2 and the 2026 State of Streaming Compression