The Complete Guide to Live Streaming (Update)

A thumbnail image of a streaming report titled 'The Complete Guide to Live Streaming: What It Is, How It Works, and Why You Need It'

What Live Video Streaming Is, How It Works, and Why You Need It

Live video streaming is skyrocketing in popularity. It will account for approximately 13 percent of all internet traffic by 2022, representing a fifteen-fold growth from 2017.

While streaming technologies have evolved drastically over the years, the basic definitions still apply. In a nutshell, live streaming involves broadcasting video and audio content across the internet to allow for near-simultaneous capture and playback.

But between capturing a live video feed and broadcasting it, quite a bit occurs. The data must be encoded, packaged, and often transcoded for delivery to virtually any screen on the planet.

In this guide, we’ll take a deep dive into the end-to-end workflow. Download the comprehensive PDF or explore our table of contents below.


Table of Contents:

  1. Capture
  2. Video Codecs and Encoding
  3. Packaging and Protocols
  4. Ingest and Transcoding
  5. Delivery
  6. Playback

Video on Demand vs. Live Streaming

Video streaming can take the form of both live and recorded content. With live streaming, the content plays as it’s being captured. Examples of this range from video chats and interactive games to endoscopy cameras and streaming drones.

Video on demand (VOD), on the other hand, describes prerecorded content that internet-connected users can stream by request. Some top players in this space include Netflix, Amazon Prime, Hulu, and Sling. YouTube’s David After Dentist and Netflix’s Stranger Things are both examples of VOD content.

For the purposes of this guide, we’ll be focusing on live video. Let’s start by summarizing the live streaming workflow and then take a closer look at the individual steps.


Live Streaming Workflow

Live video streaming starts with compressing a massive video file for delivery. Content distributors use an encoder to digitally convert the raw video with a codec. These two-part compression tools shrink gigabytes of data into megabytes. The encoder itself might be built into the camera, but it can also be a hardware appliance, a computer software like OBS Studio, or a mobile app.

Once the video stream is compressed, the encoder packages it for delivery across the internet. This involves putting the components of the stream into a file format. These container formats travel according to a protocol, or standardized delivery method. Common protocols include RTMP, HLS, and MPEG-DASH.

The packaged stream is then transported to a media server located either on premises or in the cloud. This is where the magic happens. As the media server ingests the stream, it can transcode the data into a more common codec, transize the video into a lower resolution, transrate the file into a different bitrate, or transmux it into a more scalable format.

This conversion process is critical when streaming to a variety of devices. Without transcoding the original stream, reaching viewers across an array of devices wouldn’t be possible. A streaming server software or cloud streaming service can be employed to accomplish this and more.

The single stream that first entered the media server will likely depart as multiple renditions that accommodate varying bandwidths and devices for large-scale viewing. But distance is also an issue.

The farther viewers are from the media server, the longer it will take to distribute the stream. That’s where a content delivery network (CDN) comes in handy. CDNs use an extensive network of servers placed strategically around the globe to distribute content quickly.

If done right, the live stream will find its way from the CDN to viewers across the world — in a matter of just seconds. The live stream will play back with minimal buffering and in the highest quality possible for spectators across a range of devices and internet speeds

It all starts with employing the right tools along the way. Read on to get the skinny.





Three, Two, One… Action!

Live streaming starts at the camera. Most cameras are digital and can capture images at a stunning 4K resolution (2160p). This resolution requires a very high bitrate to support the raw digital video signal coming out of the camera, so the cables used to transfer this signal must be capable of handling large amounts of data. HDMI or Ethernet cables can be used in some cases. But most often, a 4K signal transferred over long distances requires an SDI cable that can manage the bandwidth requirements.

Wireless cameras can also be used, with portable broadcasting platforms finding their way into the industry. Today’s smartphones are designed for streaming, outperforming digital cameras from ten years prior. The iPhone 12 Pro Max, for example, records HDR video at 60 frames per second.


Multi-Camera Video Production

Multi Camera Production

Some live streaming is done with a smartphone, but more serious live productions employ additional cameras. These multi-camera studio setups, and other video sources, are connected to a switcher that transitions between them. The audio is transferred to a mixer via XLR cables. Generally, the switcher adds the audio from the mixer into the final output signal. The switcher could be hardware, software, or a little of both when capture cards are required.


IP Cameras

Ip Cameras

When production isn’t a priority but speed matters, IP cameras come into play. IP cameras can send live streams directly over Ethernet cables, making them easy to put wherever you want. Most IP cameras use the RTSP protocol, which supports low-latency live streaming. RTSP is pulled to the media server rather than pushed. For that reason, the camera must be on an open, static IP address for the media server to locate it.

From surveillance to conferencing, IP cameras work great when you want to live stream from one location without getting too fancy. These user-friendly streaming devices don’t require a separate encoder, and you can aggregate the content for delivery to any device with a live transcoding solution.


User-Generated Content

User-generated content makes up a significant portion of live streams. In some cases, webcams are used. When it comes to sites like Twitch, users employ a combination of screen-recording software and webcams. But the majority of today’s content creators are on their smartphones. In fact, mobile users account for 58.56 percent of all internet traffic today.

Mobile apps and social media networks leverage video to drive engagement, but the use cases don’t end there. Smartphones can be transformed into everything from bodycams to crime-tracking tools with the addition of live streaming.

While users are responsible for supplying their own recording technology (a.k.a. smartphones or webcams), the live streaming app must have encoding functionality built in.


Video Encoding and Codecs

The second step in any live streaming workflow is video encoding. After capturing the video with your camera(s) of choice — be it a production-quality setup, IP camera, or your end-users’ mobile devices — live video data must be digitized for efficient transportation across the internet. Video encoding is essential to live streaming, helping to ensure quick delivery and playback.


What Is Encoding?

Video encoding refers to the process of converting raw video into a digital format that’s compatible with many devices. Videos are often reduced from gigabytes of data down to megabytes of data. This process involves a two-part compression tool called a codec.


What Is a Codec?

Literally ‘coder-decoder’ or ‘compressor-decompressor,’ codecs apply algorithms to tightly compress a bulky video for delivery. The video is shrunk down for storage and transmission, and later decompressed for viewing.

When it comes to streaming, codecs employ lossy compression by discarding unnecessary data to create a smaller file. Two separate compression processes take place: video and audio. Video codecs act upon the visual data, whereas audio codecs act upon the recorded sound.

H.264, also known as AVC (Advanced Video Coding), is the most common video codec. AAC (Advanced Audio Coding) is the most common audio codec.


What Video Codec Should You Use?

Streaming to a variety of devices starts with supporting a variety of codecs. But to keep the encoding part of the workflow simple and fast, you can always transcode streams later when they’re ingested by the media server.

While industry leaders continue to refine and develop the latest compression tools, many content distributors employ older video codecs like H.264/AVC for delivery to legacy devices. H.264 is your best bet for maximizing compatibility, even though other video codecs are more technologically advanced.

Below are some of the most common video codecs in use today.

Video CodecBenefitsLimitations
H.264/AVCWidely supported.Not the most cutting-edge compression technology.
H.265/HEVCSupports 8K resolution.Takes up to 4x longer to encode than H.264.
VP9Royalty-free.An earlier version of AV1.
AV1Open-source and very advanced.Not yet supported on a large scale.
H.266/VVCIntended to improve upon H.265.Same royalty issues as H.265.

What Audio Codec Should You Use?

AAC takes the cake when balancing quality with compatibility across audio codecs. While open-source alternatives like Opus far outperform AAC, they lack support across as many platforms and devices.

Audio CodecBenefitsLimitations
AACMost common audio codec.Higher-quality alternatives exist.
MP3Also widely supported.Less advanced than AAC.
OpusHighest-quality lossy audio format.Yet to be widely adopted.
VorbisNon-proprietary alternative to AAC.Less advanced than Opus.
SpeexPatent-free speech codec.Also obsoleted by Opus.

Encoding Best Practices

Encoding best practices go way beyond what codec you select. You’ll also want to consider frame rate, keyframe interval, and bitrate.

Luckily, a live stream can always be transcoded into another format once it reaches the server. This can be done using a software and your own servers, or in the cloud for professionally managed delivery.

In addition to the codecs used, it’s important to examine how your stream is packaged and which protocols enter the picture.


Packaging and Protocols

Once compressed, video streams must be packaged for delivery. The distinction between compression and packaging is subtle but relevant. For the sake of explanation, let’s picture the process in terms of residential garbage removal.

We start with the raw video data, which must be compressed down for delivery across the internet. An encoder allows us to do so by compressing gigabytes into megabytes. Think of the encoder as a household trash compactor and the codecs as the bags of compressed trash.

In order to actually transport the compressed trash (audio and video codecs) to the dump (viewer), another step is required. It’s crucial to place the bag of trash, as well as any other odds and ends (such as metadata), into a curbside trash receptacle. File container formats can be thought of in terms of these receptacles. They act as a wrapper for all streaming data so that it’s primed for delivery.

Finally, the contents of the trash bin are transported to the dump via an established route. Think of protocols as the established routes that garbage trucks take.

OK. Enough trash talk. Time to define each term for real.

Visualization: Encoding and Packaging Live Video Streams

What Is a Video Container Format?

Video container formats, also called wrappers, hold all the components of a compressed stream. This could include the audio codec, video codec, closed captioning, and any associated metadata such as subtitles or preview images. Common containers include .mp4, .mov, .ts, and .wmv.


What Is a Streaming Protocol?

A protocol is a set of rules governing how data travels from one device to another. For instance, the Hypertext Transfer Protocol (HTTP) deals with hypertext documents and webpages. Online video delivery uses both streaming protocols and HTTP-based protocols. Streaming protocols like Real-Time Messaging Protocol (RTMP) offer fast video delivery, whereas HTTP-based protocols like HLS can help optimize the viewing experience.

The protocol used can increase streaming latency by to up to 45 seconds.


Traditional Stateful Streaming Protocols

In the early days, traditional protocols such as RTSP (Real-Time Streaming Protocol) and RTMP (Real-Time Messaging Protocol) were the go-to methods for streaming video over the internet and playing it back on home devices. These protocols are stateful, which means they require a dedicated streaming server.

While RTSP and RTMP support lightning-fast video delivery, they aren’t optimized for great viewing experiences at scale. Additionally, fewer players support these protocols than ever before. Many broadcasters choose to transport live streams to their media server using a stateful protocol like RTMP and then transcode it for multi-device delivery.

RTMP and RTSP keep latency at around 5 seconds or less.


HTTP-Based Adaptive Streaming Protocols

The industry eventually shifted in favor of HTTP-based technologies. Streams deployed over HTTP are not technically “streams.” Rather, they’re progressive downloads sent via regular web servers.

Using adaptive bitrate streaming, HTTP-based protocols deliver the best video quality and viewer experience possible — no matter the connection, software, or device. Some of the most common HTTP-based protocols include MPEG-DASH and Apple’s HLS.

HTTP-based protocols are stateless, meaning they can be delivered using a regular old web server. That said, they fall on the high end of the latency spectrum.

HTTP-based protocols can cause 10-45 seconds in latency.

streaming latency and interactivity continuum

Emerging Protocols for Near-Real-Time Delivery

With an increasing number of videos being delivered live, industry leaders continue to improve streaming technology. Emerging standards like WebRTC, SRT, Low-Latency HLS, and low-latency CMAF for DASH (which is a format rather than a protocol) support near-real-time delivery — even over poor connections.

WebRTCReal-time interactivity without a plugin.Only scalable with a live streaming platform like Wowza.
SRTSmooth playback with minimal lag.Playback capabilities are still being developed.
Low-Latency HLSSub-three-second global delivery.Vendors are adding support for this new spec.
Low-Latency CMAFStreamlined workflows and decreased latency.Many organizations are prioritizing other technologies at this time.

These new technology stacks promise to reduce latency to 3 seconds or less!


Video Packaging and Protocols for Every Workflow

Depending on how you set up your streaming workflow, you’re not limited to one protocol from capture to playback. Many broadcasters use RTMP to get from the encoder to server, and then transcode the stream into an adaptive HTTP-based format. The best protocol for your live stream depends entirely on your use case. Let’s take a closer look at how transcoding works.


Ingest and Transcoding

The fourth step in a live streaming workflow is transcoding the stream into a variety of different codecs, bitrates, resolutions, and file containers. While this step can be skipped, it’s essential to most streaming scenarios.

A media server located either on premises or in the cloud ingests the packaged stream, and a powerful conversion process called transcoding ensues. This allows broadcasters to reach almost any device — regardless of the viewer’s connection or hardware. Once transcoding is complete, multiple renditions of the original stream depart for delivery.


Transmuxing, Transcoding, Transizing, Transrating

To optimize the viewing experience across a variety of devices and connection speeds, broadcasters often elect to transmux, transcode, transrate, and transize streams as they pass through the media server.

  • Transmuxing: Taking the compressed audio and video and repackaging it into a different container format. This allows delivery across different protocols without manipulating the actual file. Think of transmuxing like converting a word doc into a pdf and vice versa.
  • Transcoding: An umbrella term for taking a compressed/encoded file and decompressing/decoding it to alter in some way. The manipulated file is then recompressed for delivery. Transrating and transizing are both subcategories of transcoding.
    • Transrating: Changing the bitrate of the decompressed file to accommodate different connection speeds. This can include changing the frame rate or changing the resolution.
    • Transizing: Resizing the video frame — or resolution — to accommodate different screens.

Rather than creating one live stream at one bitrate, transcoding allows you to create multiple streams at different bitrates and resolutions. That way, your live streams can dynamically adapt to fit the screen sizes and internet speeds of all your viewers. This is known as adaptive bitrate (ABR) streaming.


What Is Adaptive Bitrate Streaming?

Adaptive bitrate (ABR) streaming involves outputting multiple renditions of the original video stream to enable playback on a variety of devices and connection speeds. Content distributors use adaptive bitrate streaming to deliver high-quality streams to users with outstanding bandwidth and processing power, while also accommodating those lacking in the speed and power departments.

The result? No buffering or stream interruptions. Plus, as a viewer’s signal strength goes from two bars to three, the stream automatically adjusts to deliver a superior rendition.

Graphic Visualization: Adaptive Bitrate Streaming

How Does Adaptive Bitrate Streaming Work?

The first step to adaptive bitrate streaming is creating multiple renditions of the original stream to provide a variety of resolution and bitrate options. These transcoded files fall on an encoding ladder. At the bottom, a high-bitrate, high-frame-rate, high-resolution stream can be output for viewers with the most high-tech setups. At the top of the ladder, the same video in low quality is available for viewers with small screens and poor service.

During the process of transcoding, these renditions are broken into segments that are 2-10 seconds in length. The video player can then use whichever rendition is best suited for its display, processing power, and connectivity. If power and connectivity change mid-stream, the video will automatically switch to another step on the ladder.

With adaptive bitrate streaming, mobile viewers with poor connections don’t have to wait for the stream to load. And for those plugged into high-speed internet, a higher-resolution alternative can play.



Back to the viewers. Because we don’t know where they’re located, there’s still the issue of distance. The farther your audience is from the media server, the longer it will take to deliver the live stream. This can cause latency and buffering.

To resolve the latency inherent to global delivery, many broadcasters employ a content delivery network (CDN).


What Is a CDN?

As the name suggests, a CDN is a system of geographically distributed servers used to transport media files. This removes the bottleneck of traffic that can result when delivering streams with a single server, as CDNs only require a single stream for each rendition of an outbound video.

These large networks help truncate the time it takes to deliver video streams from origin to end users. Sharing the workload across a network of servers also improves scalability should viewership increase.


Benefits of Using a CDN for Live Streaming

  • Scalability: Employing a CDN is the fastest, most reliable way to get your content in front of numerous viewers — even with viewership spikes.
  • Speed: CDNs use speedy superhighways to deliver content to vast audiences across the globe
  • Quality: Streaming through a CDN allows you to achieve the highest sound quality and video resolution possible, while minimizing buffering and delays.
  • Security: A CDN can help prevent distributed denial of service (DDoS) attacks, which occur when a site or resource is flooded by multiple, simultaneous attempts to breach it.

Even if you forgo a CDN, other strategies can be employed to reach a broad audience. Many content distributors leverage social media platforms to ensure scalability, speed, and quality without breaking the bank. Simulcasting makes its easy.


What Is Simulcasting?

Simulcasting is the ability to take one video stream and broadcast it to multiple destinations at the same time — thereby maximizing your impact. This allows you to reach a broader audience, no matter which platform or service your viewers prefer.

Live streaming social platforms are everywhere, with both wide-ranging and niche applications. While Facebook connects you with the largest general audience, Twitter is a top destinations for news coverage and events. Meanwhile, Twitch is dedicated to gaming. And let’s not forget about YouTube — which has become a search engine in its own right.

While ‘the more the merrier’ might be an intuitive approach, you should only stream to the destinations that make sense for your audience. Broadcasting in the wrong context can result in negative feedback and wasted resources.

It’s also worth noting that the process of simulcasting can be complex. Multi-destination broadcasting requires massive amounts of bandwidth and is prone to errors. That’s why it’s important to start with the right tools.

Whether via a CDN, simulcasting, or both, the last step to consider is your stream’s final destination: playback.



The end-goal of live streaming is playback. And through the process of encoding, transcoding, and globally distributing the video content, your live video stream should do just that — at high quality, with low latency, and at any scale.


Multi-Device Delivery

If every viewer had a 4K home theater plugged into high-speed internet, delivering live streams would be easy. But that’s not the case.

Today’s viewers are out and about. A significant portion of your audience is on mobile devices, relying on depleted batteries and LTE connections. Other viewers stream content to their laptops using public Wi-Fi, or perhaps to their iPads via mobile hotspots.

Different screens and varying internet speeds make transcoding essential. And with adaptive bitrate streaming, you can deliver the highest-quality stream possible to internet-connected TVs and mobile users alike.


Our Tech = Your Superpower

In the content-flooded landscape — where digital interfaces have largely replaced in-person interactions — live streaming is one of the most authentic ways to connect with your audience. While a lot goes into getting that live stream from capture to playback, we designed our software, services, and hardware to help you along the way.

  • Wowza Streaming Engine: Stream on your terms and to any device. Our downloadable media server software powers live and on-demand streaming, on-premises or in the cloud, with fully customizable software.
  • Wowza Video: Quickly build live streaming applications with our fully managed cloud service. The flexible platform gets you up and running quickly, and can be used from capture to playback or as part of a custom solution.
  • Wowza CDN: Reach global audiences of any size with our streaming delivery network. Automatically included with Wowza Video, the Wowza CDN can also be configured as a stream target for Wowza Streaming Engine.

Push the limits of what’s possible with streaming technology. Streamline your streaming workflow with Wowza.


Search Wowza Resources



Follow Us


About Traci Ruether

Traci Ruether is a Colorado-based B2B tech writer with a background in streaming and network infrastructure. Aside from writing, Traci enjoys cooking, gardening, and spending quality time with her kith and kin. Follow her on LinkedIn at or learn more at