const video = await generate(topic)const voice = await tts(script)await render({ scenes, voice })
All articles
🎥Video Creation

Audio Mixing for Video: The Professional Guide

Audio mixing is the invisible skill that separates amateur video from professional production. This guide covers the fundamentals of levels, EQ, and compression, the exact dB and LUFS values for balancing voice, music, and sound effects, step-by-step workflows in Premiere Pro, DaVinci Resolve Fairlight, and CapCut, the measurable impact of audio mixing on video performance metrics, and a pre-export checklist that ensures broadcast-quality sound on every video you publish.

10 min readJanuary 4, 2023

Professional video audio is 90% mixing -- not 90% microphone

Audio levels, EQ, and the mixing checklist that makes every video sound broadcast-ready

Why Audio Mixing Separates Amateur from Pro Video

Audio mixing is the single most underestimated skill in video production. Most creators obsess over camera gear, lighting setups, and color grading while treating audio as an afterthought -- something to fix in post if it sounds "off." This approach produces videos that look cinematic but sound like they were recorded in a bathroom. The result is a disconnect that viewers feel immediately, even if they cannot articulate what is wrong. Research on perceived video quality consistently shows that audiences judge production value primarily by audio, not visuals. A video shot on an iPhone with properly mixed audio feels more professional than a video shot on a RED camera with unbalanced, muddy sound.

The tolerance gap between bad visuals and bad audio is enormous and measurable. Viewers will watch a 480p video with clear, well-balanced audio for its full duration. The same viewers abandon a 4K video with distorted, poorly mixed audio within the first 15 seconds. This asymmetry exists because human hearing is more sensitive to quality degradation than human vision. Clipping audio, uneven volume between speakers, background music drowning out dialogue, and sudden level changes all trigger an unconscious stress response that makes viewers reach for the back button. YouTube analytics confirm this pattern: videos with consistent, properly mixed audio retain viewers 25% longer than videos with identical visual quality but inconsistent audio levels.

Professional audio mixing is not about expensive equipment or years of training. It is about understanding a small set of principles -- levels, equalization, compression, and balance -- and applying them consistently to every video before export. The difference between amateur and professional audio is usually 15 minutes of work in your editing timeline. Those 15 minutes determine whether your video sounds like a broadcast production or a casual voicemail. Every concept in this guide can be learned and applied in a single afternoon, and the improvement in your video quality will be immediately obvious to every viewer.

ℹ️ The Audio Quality Perception Gap

Viewers tolerate bad visuals far longer than bad audio. A video with phone-quality visuals but clean, well-mixed audio feels professional. The same video with great visuals but muddy, unbalanced audio feels amateur. Audio mixing is the invisible skill that determines perceived production quality

Audio Mixing Fundamentals: Levels, EQ, Compression

Audio mixing for video rests on three core concepts: levels, equalization (EQ), and compression. Levels refer to the volume of each audio element in your mix -- how loud the voice is relative to the music, how loud the music is relative to sound effects, and how loud the final combined output is. Levels are measured in decibels (dB), and every digital audio workstation and video editor displays them on a meter that ranges from silence at the bottom to 0 dB at the top. The critical rule of levels is that your audio should never hit 0 dB, because that is the clipping point where digital distortion occurs. Professional mixers target peaks between -6 dB and -3 dB for dialogue, leaving headroom for transient spikes without ever touching the clipping threshold.

Equalization is the process of adjusting specific frequency ranges to improve clarity and reduce muddiness. Human voice occupies a frequency band roughly between 85 Hz and 8,000 Hz, with the critical intelligibility range sitting between 1,000 Hz and 4,000 Hz. A high-pass filter at 80 Hz removes low-frequency rumble from air conditioning, traffic, and handling noise without affecting voice quality. A gentle boost between 2,000 Hz and 5,000 Hz adds presence and clarity to dialogue, making words easier to understand even at lower volumes. These two EQ moves -- cut the lows, boost the presence range -- solve 80% of vocal clarity problems in video production.

Compression reduces the dynamic range of your audio, making quiet parts louder and loud parts quieter so the overall volume stays consistent. Without compression, a speaker who alternates between whispering and projecting will produce audio that forces viewers to constantly adjust their volume. A compressor with a ratio of 3:1 or 4:1, a threshold set just above the average speaking level, and an attack time of 10 to 20 milliseconds tames these fluctuations without making the audio sound artificially squashed. The goal is transparent compression -- the viewer should not be able to hear the compressor working, but they should notice that the dialogue is consistently clear and intelligible from start to finish.

  • Levels: keep dialogue peaks between -6 dB and -3 dB, never let any element hit 0 dB (clipping threshold), and monitor your meters throughout the edit
  • High-pass filter at 80 Hz: removes low-frequency rumble from HVAC, traffic, and handling noise without affecting voice clarity
  • Presence boost between 2,000 Hz and 5,000 Hz: adds vocal clarity and intelligibility, making dialogue cut through background music and effects
  • Compression ratio of 3:1 to 4:1: tames volume fluctuations in dialogue so viewers never need to adjust their volume mid-video
  • Attack time of 10-20 ms on the compressor: fast enough to catch spikes but slow enough to preserve the natural transients of speech
  • Always apply EQ before compression in your effects chain -- shaping the tone first gives the compressor a cleaner signal to work with

The Perfect Audio Balance: Voice, Music, Effects

The most common audio mixing mistake in video production is setting background music too loud relative to dialogue. Creators listen to their mix on studio monitors or headphones in a quiet room and think the balance sounds fine, but viewers watching on laptop speakers, phone speakers, or earbuds in noisy environments cannot separate the voice from the music. The fix is a specific set of level relationships that professional broadcast engineers have used for decades. These numbers are not artistic preferences -- they are engineering standards derived from decades of psychoacoustic research on how human ears perceive layered audio.

Dialogue should sit at -6 dB to -3 dB peak, with an average level (RMS) around -12 dB to -10 dB. Background music should be mixed 12 to 16 dB below the dialogue, putting it in the -18 dB to -22 dB range. This gap feels enormous when you solo the music track -- it will sound almost inaudible -- but in context with voice over it, the music fills the frequency space beneath the dialogue and creates a polished, layered sound without competing for attention. Sound effects sit between the two: -12 dB to -15 dB for ambient effects and room tone, with transient effects like button clicks or whooshes peaking up to -8 dB briefly before returning to the lower level.

The final mix should be normalized to a loudness standard measured in LUFS (Loudness Units Full Scale), which accounts for perceived loudness rather than just peak levels. YouTube recommends -14 LUFS for uploaded video. Podcasts and audio-focused platforms target -16 LUFS. Broadcast television uses -24 LUFS. Normalizing to the correct LUFS target ensures your video sounds the same volume as other content on the platform, preventing the jarring experience of a viewer needing to crank their volume up for your video and then getting blasted by the next video in their feed. Every major NLE includes a loudness meter or supports a free loudness metering plugin.

💡 The Audio Mixing Ratio

The magic audio mixing ratio: voice at -6dB to -3dB, background music at -18dB to -22dB, sound effects at -12dB to -15dB. Normalize your final mix to -14 LUFS for YouTube and -16 LUFS for podcasts. These numbers produce consistent, broadcast-quality audio on every platform

Audio Mixing in Premiere Pro, DaVinci Resolve, CapCut

Adobe Premiere Pro provides audio mixing through its Essential Sound panel and the Audio Track Mixer. Start by assigning each clip a type in the Essential Sound panel: Dialogue for voice tracks, Music for background tracks, and SFX for sound effects. This assignment automatically applies baseline processing suited to each category. For dialogue, enable the Loudness Auto Match feature which normalizes your voice to broadcast standards. Then open the Audio Track Mixer (Window > Audio Track Mixer) to set your fader levels: dialogue track at 0 dB (with the clip itself peaking at -6 dB to -3 dB), music track pulled down to -12 to -16 on the fader, and effects track at -6 to -8. Use the built-in Parametric EQ to apply your high-pass filter and presence boost, then add the Dynamics effect for compression with a 3:1 ratio.

DaVinci Resolve includes Fairlight, a full professional digital audio workstation built directly into the editing software. Switch to the Fairlight page to access the mixer, EQ, dynamics, and loudness metering. Fairlight provides six-band parametric EQ per track, a built-in compressor and limiter, and a loudness meter that reads in real-time LUFS. Set your dialogue track EQ with a high-pass at 80 Hz and a gentle shelf boost at 3 kHz. Apply the compressor with a 3:1 ratio and -20 dB threshold, adjusting the threshold until the gain reduction meter shows 3 to 6 dB of compression on loud passages. Use the Bus system to group all dialogue tracks, all music tracks, and all effects tracks separately, then set the bus levels to achieve the -6/-18/-12 dB relationship. The loudness meter on the master bus shows your integrated LUFS in real time -- export when it reads -14 LUFS.

CapCut has made audio mixing accessible to creators who find Premiere Pro and DaVinci Resolve intimidating. In CapCut desktop or mobile, tap any audio clip to access volume, fade in, fade out, and noise reduction controls. Set your voice track volume so the waveform peaks sit roughly two-thirds up the track height, then reduce music volume until it is barely visible as a thin waveform beneath the dialogue. CapCut includes an auto-adjust volume feature that ducks music automatically when dialogue is detected -- enable this for a quick, automatic mix that gets you 80% of the way to professional results. For finer control, use the volume keyframe feature to manually duck music during dialogue sections and bring it back up during pauses. While CapCut does not offer LUFS metering natively, exporting and checking your loudness in a free tool like Youlean Loudness Meter ensures platform compliance.

  1. Premiere Pro: assign clip types in Essential Sound panel (Dialogue, Music, SFX), enable Loudness Auto Match on dialogue, set fader levels in Audio Track Mixer, apply Parametric EQ and Dynamics effects
  2. DaVinci Resolve Fairlight: switch to Fairlight page, set 6-band EQ with 80 Hz high-pass and 3 kHz presence boost, apply compressor at 3:1 ratio, use Bus system to group tracks, monitor integrated LUFS on master bus
  3. CapCut: adjust clip volumes so dialogue waveform peaks at two-thirds track height, enable auto-adjust volume for automatic music ducking, use volume keyframes for manual control during critical sections
  4. All editors: always monitor on headphones AND speakers before export -- mixes that sound perfect on headphones often reveal problems on laptop or phone speakers
  5. Export a test segment (30 seconds) and check loudness with a free LUFS meter before committing to a full export -- correcting after export wastes render time

Does Better Audio Mixing Improve Video Performance?

The correlation between audio quality and video performance metrics is one of the most consistent findings in creator analytics. YouTube creators who improve their audio mixing without changing any other production element see measurable increases in average view duration, which is the single most important metric for algorithmic recommendation. The mechanism is straightforward: poorly mixed audio causes viewer fatigue. When dialogue levels fluctuate, when background music competes with the voice, or when sudden volume spikes force viewers to grab their volume controls, each of these moments creates a micro-friction that accumulates into an early exit. Properly mixed audio eliminates these friction points entirely.

Completion rates tell the clearest story. A/B testing by multiple YouTube channels has shown that videos with professionally mixed audio (consistent levels, proper EQ, music ducked appropriately) achieve completion rates 15% to 25% higher than the same content with raw, unmixed audio. For a 10-minute video, that translates to viewers watching an additional 90 to 150 seconds on average. In algorithmic terms, this is an enormous difference -- YouTube's recommendation engine weighs average view duration heavily, meaning that better audio mixing directly increases the probability of your video being recommended to new viewers.

The impact extends beyond YouTube. Podcast episodes with properly normalized audio (-16 LUFS) see higher listener retention on Spotify and Apple Podcasts. Instagram Reels and TikTok videos with clear, properly leveled audio get shared more frequently because viewers are more likely to rewatch content that sounds good on phone speakers. Corporate training videos with professional audio mixing receive higher completion scores in LMS platforms. The pattern is universal: every platform and every content type benefits from the 15 minutes of audio mixing work that most creators skip.

The Audio Mixing Performance Boost

Creators who learn basic audio mixing (30 minutes of training) see a 20% increase in average view duration. The improvement comes from reduced viewer fatigue -- properly mixed audio doesn't strain the ears, which means viewers stay longer without consciously knowing why

Quick Audio Mixing Checklist for Every Video

Every video you export should pass through this audio mixing checklist. Memorize it, print it, or pin it next to your editing monitor. These seven checks take less than five minutes and catch the issues that make the difference between amateur and professional audio. Skipping even one of them risks publishing a video with a problem that viewers will notice within the first ten seconds. The checklist works in any editor -- Premiere Pro, DaVinci Resolve, CapCut, Final Cut Pro, or any other NLE.

Before you touch the export button, solo each audio track and listen for problems individually. Check that dialogue has no clipping (peaks never touching 0 dB), that background music sits at -18 dB to -22 dB relative to voice, and that no sound effect is louder than the dialogue at any point in the timeline. Verify that a high-pass filter is applied to every dialogue track to remove low-end rumble. Confirm that your compressor is active on dialogue and showing 3 to 6 dB of gain reduction on loud passages. Then un-solo everything and listen to a 30-second segment on headphones, then again on your laptop or phone speakers. If the dialogue is clear and the music is present but not distracting on both playback systems, your mix is ready.

The final check is loudness normalization. Run your loudness meter and verify the integrated reading hits your target: -14 LUFS for YouTube, -16 LUFS for podcasts, -24 LUFS for broadcast. If you are off target, apply a gain adjustment to the master bus rather than re-adjusting individual tracks, which preserves your carefully set balance relationships. Export, spot-check the first 30 seconds and a random section in the middle of the exported file, and publish with confidence that your audio will sound professional on every device and platform your audience uses.

  • Check 1 -- No clipping: verify all dialogue peaks sit between -6 dB and -3 dB with no element touching 0 dB anywhere in the timeline
  • Check 2 -- Music balance: confirm background music is at -18 dB to -22 dB, sounding barely audible when soloed but full and present beneath dialogue
  • Check 3 -- High-pass filter: ensure every dialogue track has an 80 Hz high-pass filter active to remove rumble, HVAC hum, and handling noise
  • Check 4 -- Compression active: verify the compressor on dialogue tracks is showing 3 to 6 dB of gain reduction on loud passages with a 3:1 or 4:1 ratio
  • Check 5 -- Multi-device test: listen on headphones AND on laptop or phone speakers, confirming dialogue clarity on both playback systems
  • Check 6 -- LUFS normalization: run the loudness meter and verify integrated reading of -14 LUFS (YouTube), -16 LUFS (podcasts), or -24 LUFS (broadcast)
  • Check 7 -- Export spot-check: after rendering, play back the first 30 seconds and one random mid-video section to catch any export artifacts or level shifts
Audio Mixing for Video: The Professional Guide