All articles
💬AI Tools

AI Subtitles vs Captions: The Real Difference

Subtitles and captions are not interchangeable -- they serve different audiences, meet different legal requirements, and are generated differently by AI tools. This guide defines the precise distinction between subtitles, closed captions, and open captions, explains when to use each one based on your platform and audience, compares SRT, VTT, and burned-in formats with platform compatibility details, covers how AI generates subtitles and captions through different pipelines, and provides best practices for timing, styling, and accessibility compliance under WCAG and ADA standards.

11 min readJune 15, 2023

Subtitles and captions serve different audiences -- use the wrong one and you miss both

The definitive guide to subtitles vs captions, file formats, and when to use each

Subtitles and Captions Are Not the Same Thing

Most people use the words subtitles and captions interchangeably, and most of the time nobody corrects them. But in professional video production, accessibility compliance, and broadcast standards, the two terms describe fundamentally different things -- and using one when you need the other means your content fails to serve the audience it was designed for. The distinction is not pedantic. It has legal implications under the Americans with Disabilities Act (ADA) and the Web Content Accessibility Guidelines (WCAG), and it affects how platforms like YouTube, Netflix, and TikTok categorize and deliver your text overlays to viewers.

Subtitles are text translations of spoken dialogue intended for viewers who can hear the audio but do not understand the language being spoken. When you watch a French film with English text at the bottom of the screen, those are subtitles. They assume you can hear everything -- the music, the sound effects, the tone of voice -- and they only provide a written version of the words being said. Subtitles exist to cross a language barrier, not an accessibility barrier. A subtitle track for an English-language video might be a Spanish translation, a Mandarin translation, or a simplified English version for language learners.

Captions, by contrast, are a complete text representation of all meaningful audio in a video. This includes spoken dialogue, but also speaker identification (who is talking when multiple voices are present), sound effects (a door slamming, a phone ringing, glass breaking), music descriptions (upbeat jazz music playing, ominous orchestral score), and non-speech vocalizations (sighing, laughing, clearing throat). Captions are designed for deaf and hard-of-hearing viewers who cannot access any of the audio track. Without those additional descriptions, a deaf viewer watching a horror film would miss every creaking floorboard, every whispered warning, and every sudden musical sting that builds tension -- they would see the visuals but lose half the storytelling.

ℹ️ The Core Distinction

Subtitles translate spoken dialogue for viewers who speak a different language. Captions transcribe all audio -- including dialogue, sound effects, and music -- for deaf and hard-of-hearing viewers. Using one when you need the other means your content fails to serve its intended audience

Subtitles vs Closed Captions vs Open Captions

Once you understand the subtitle-versus-caption distinction, the next layer of complexity is the difference between closed and open formats. This is where the terminology gets confusing because the words closed and open refer to how the text is delivered, not what it contains. A closed caption and an open caption can contain identical text -- the difference is whether the viewer can toggle it on and off.

Closed captions (often abbreviated CC) are separate text tracks that exist alongside the video but are not permanently embedded in the video frames. They are stored as external files -- typically SRT or VTT format -- or as metadata streams within a video container like MP4 or MKV. The viewer chooses whether to display them. On YouTube, you click the CC button. On Netflix, you navigate to the subtitle and caption menu. On a television, you enable closed captioning through your TV settings or the remote. The key characteristic of closed captions is viewer control: they can be turned on, turned off, resized, repositioned, and restyled depending on the platform and device.

Open captions are permanently burned into the video frames themselves. They are part of the image -- you cannot turn them off, resize them, or change their appearance because they are rendered as pixels in every frame, just like any other visual element. When you see a TikTok video with bold white text appearing word-by-word over the footage, those are open captions (sometimes called hardcoded or burned-in captions). The creator chose the font, size, color, and position, and every viewer sees exactly the same text presentation regardless of their device or settings.

Subtitles follow the same open-versus-closed logic. A closed subtitle track is a separate file (like an SRT) that contains translated dialogue and can be toggled by the viewer. An open subtitle is a translation that has been burned directly into the video frames. Foreign films distributed with permanent English text at the bottom use open subtitles. The same film on a streaming service with selectable language options uses closed subtitles. The content is identical -- only the delivery mechanism changes.

  • Closed captions: separate text tracks the viewer can toggle on/off -- stored as SRT, VTT, or embedded metadata -- viewer controls size, position, and styling on most platforms
  • Open captions: permanently burned into the video frames -- cannot be turned off, resized, or restyled -- every viewer sees the identical text presentation
  • Closed subtitles: selectable language tracks the viewer enables through a menu -- standard on streaming platforms like Netflix, YouTube, and Disney+
  • Open subtitles: translated text permanently rendered into the video image -- common in theatrical releases of foreign films and social media content
  • Key trade-off: closed formats give the viewer control and enable searchability and SEO -- open formats guarantee visibility but sacrifice flexibility and accessibility customization

When Should You Use Subtitles vs Captions?

The choice between subtitles and captions depends on three factors: your audience, your platform, and your legal obligations. Getting this decision right is not just a best practice -- for businesses, educational institutions, and government organizations, it is a compliance requirement. The WCAG 2.1 Level AA standard (the benchmark most organizations target) requires captions for all prerecorded audio content and subtitles for prerecorded video content that is in a different language from the page. The ADA has been interpreted by courts to extend these requirements to web video content published by businesses that serve the public.

If your content has a significant deaf or hard-of-hearing audience -- or if you are legally required to be accessible -- you need captions, not subtitles. Educational institutions receiving federal funding must caption all video content. Corporate training videos must be captioned. Government websites must meet WCAG 2.1 AA standards, which mandate captions. Healthcare organizations, financial services companies, and any business operating in a regulated industry should default to full captions for every video they publish. The cost of retrofitting uncaptioned video libraries after an accessibility lawsuit dwarfs the cost of captioning content during production.

If your content targets international audiences who speak different languages, you need subtitle tracks. A YouTube creator with viewers in 30 countries should provide subtitle translations in their top viewer languages. A SaaS company with a global customer base should subtitle their product demo videos. An e-commerce brand selling internationally should subtitle their product videos. In these cases, the viewers can hear the audio perfectly -- they just need a written translation of the dialogue.

For social media content -- TikTok, Instagram Reels, YouTube Shorts, and Facebook videos -- the calculus changes entirely. The primary reason to add text to social media video is not accessibility or translation but silent autoplay. An estimated 85 percent of Facebook video is watched without sound, and similar numbers apply to other feed-based platforms. In this context, you need open captions (burned into the video) because viewers scrolling through their feed will not manually enable a closed caption track. The text needs to be visible by default, styled to match your brand, and positioned to be readable on mobile screens.

💡 Platform Decision Guide

For social media video (TikTok, Reels, Shorts): use open captions (burned into the video) because 85% of viewers watch without sound. For YouTube and website video: use closed captions (SRT/VTT files) because they're searchable, translatable, and viewer-toggleable

How AI Generates Subtitles and Captions Differently

Modern AI tools use automatic speech recognition (ASR) as the foundation for both subtitle and caption generation, but the output pipelines diverge significantly after the initial transcription step. Understanding how these tools work helps you evaluate their accuracy, choose the right tool for your use case, and know where manual review is most critical. The core technology -- converting audio waveforms into text -- is the same, but what happens to that text afterward determines whether you get subtitles or captions.

For subtitle generation, the AI performs speech-to-text transcription, segments the text into timed chunks that match the audio rhythm, and optionally translates the text into one or more target languages. The translation step is where AI subtitle tools vary most dramatically in quality. Some tools use neural machine translation (similar to Google Translate or DeepL) to convert the source-language transcript into the target language. Others use large language models to produce more natural, context-aware translations that account for idioms, cultural references, and sentence structures that do not translate literally. The best subtitle tools preserve the original timing while adapting the translated text to be readable at the same pace -- a non-trivial problem because different languages express the same idea in different numbers of words.

For caption generation, the AI must do significantly more work. Beyond transcribing dialogue, a caption-quality tool needs to identify and label speakers (Speaker 1, Speaker 2, or ideally by name), detect and describe non-speech audio events (sound effects, music, ambient noise), and format the output according to caption standards that specify maximum line length, maximum display duration, minimum gap between captions, and reading speed limits. The FCC standard for broadcast captions, for example, limits display speed to approximately 190 words per minute to ensure readability. Most AI caption tools today handle dialogue transcription well but struggle with sound effect detection -- they might miss a subtle background noise or describe music too generically. This is where manual review adds the most value.

The accuracy gap between AI-generated subtitles and AI-generated captions matters for compliance. If you are producing captions for accessibility purposes under ADA or WCAG requirements, the accuracy standard is typically 99 percent or higher for dialogue and comprehensive coverage of non-speech audio. Most AI tools achieve 90 to 95 percent accuracy on dialogue out of the box, which means a 10-minute video might contain 10 to 20 errors that require manual correction. For subtitles, the accuracy standard depends on the translation quality and is harder to quantify -- a subtitle that is technically correct but awkwardly phrased fails differently than one with a factual error. The practical takeaway is that AI generates an excellent first draft of both subtitles and captions, but human review remains essential for published content.

SRT vs VTT vs Burned-In: Caption File Formats

Caption and subtitle files come in several standard formats, and choosing the right one depends on where your video will be published and what features you need. The three formats you will encounter most often are SRT (SubRip Subtitle), VTT (Web Video Text Tracks), and burned-in (hardcoded into the video). Each has distinct capabilities, platform compatibility, and trade-offs that affect both the viewer experience and your workflow.

SRT is the oldest and most universally supported caption file format. An SRT file is plain text with a simple structure: a sequential number, a timecode range (start time to end time in hours:minutes:seconds,milliseconds format), and the caption text. That is it -- no styling, no positioning, no metadata. This simplicity is both the format's greatest strength and its biggest limitation. Every video platform accepts SRT files: YouTube, Vimeo, Facebook, LinkedIn, Twitter, and virtually every video player and editing application. If you are unsure which format to use, SRT is the safe default. The downside is that SRT cannot specify font color, size, position, or background -- the playback platform or device applies its own default styling, which you cannot control.

VTT (WebVTT) is the modern successor to SRT and the standard for web-based video. VTT supports everything SRT does plus styling (bold, italic, color, font size), positioning (place captions at the top, bottom, or sides of the frame), speaker identification through voice tags, and chapter markers. The VTT format is required by the HTML5 video element, which means any custom video player built for the web uses VTT natively. YouTube accepts both SRT and VTT, but if you upload VTT with styling tags, YouTube will respect some of them. For web accessibility compliance under WCAG, VTT is the recommended format because it supports the richest set of accessibility features including text customization for viewers with low vision.

Burned-in captions are not a file format -- they are captions rendered directly into the video frames during export. You create them in your video editor by placing text elements on the timeline, styling them however you want, and exporting the video with the text permanently visible. The result is a video file (MP4, MOV, etc.) where the captions are part of the image. Burned-in captions give you complete creative control over appearance -- any font, any color, any animation, any position. But they cannot be turned off, searched, translated, or restyled by the viewer. For accessibility purposes, burned-in captions alone do not meet WCAG requirements because users cannot customize text size and color to meet their needs.

  • SRT format: plain text, universally supported, no styling -- best for maximum platform compatibility and simplicity
  • VTT format: supports styling, positioning, speaker tags, and chapter markers -- best for web video and WCAG compliance
  • Burned-in: permanent, fully styled, creative control -- best for social media where captions must be visible by default
  • SRT structure: sequence number, timecode (00:01:15,000 --> 00:01:18,500), caption text -- editable in any text editor
  • VTT structure: WEBVTT header, optional styling block, timecodes (00:01:15.000 --> 00:01:18.500), caption text with optional HTML tags
  • Platform support: YouTube accepts SRT and VTT; TikTok uses burned-in only; Vimeo accepts SRT, VTT, and SCC; HTML5 video requires VTT

The Best Accessibility Strategy

The most accessible approach: provide both closed captions (for hearing-impaired viewers and SEO) AND subtitle tracks (for international audiences). AI tools make generating both from one video trivial -- the transcript is the same base, just formatted differently for each use case

Best Practices for Subtitles and Captions

Effective captions and subtitles are invisible in the best sense -- viewers read them effortlessly without being distracted from the video content. Achieving that invisibility requires attention to placement, timing, line length, reading speed, and styling. The difference between professional captions and amateur ones is not accuracy (both might transcribe the words correctly) but readability: how easily and comfortably a viewer can absorb the text while watching the video.

Timing is the most critical technical element. Each caption should appear at the exact moment the corresponding audio begins and disappear shortly after it ends. The standard minimum display time is 1 second (even for very short phrases) and the maximum is 7 seconds. Captions that flash by too quickly are unreadable; captions that linger too long after the audio has moved on create a disorienting disconnect between what the viewer hears and what they read. For reading speed, the widely accepted maximum is 190 to 200 words per minute for adult viewers, which translates to roughly 15 to 20 words per caption display. If a speaker talks rapidly and you cannot fit their words into readable captions at natural speed, it is acceptable to condense the text slightly -- capturing the meaning rather than every single word.

Line length and line breaks matter more than most creators realize. The standard maximum is two lines per caption, with each line containing no more than 42 characters (including spaces). Line breaks should follow natural linguistic boundaries: break between clauses, between a subject and a long predicate, or before conjunctions -- never in the middle of a word, a proper noun, or a tightly coupled phrase. A caption that reads "The president announced new / climate policy" is harder to parse than one that reads "The president announced / new climate policy" because the second version breaks at a natural grammatical pause.

For styling, high contrast is non-negotiable. White text on a semi-transparent black background is the industry standard because it is readable over any video content -- bright scenes, dark scenes, and everything in between. Avoid decorative fonts, thin typefaces, and low-contrast color combinations. For open captions on social media, bold sans-serif fonts (like Montserrat, Inter, or the platform-native options) at a size large enough to read on a mobile phone screen are the standard. Test your captions on an actual phone screen, not just your computer monitor -- text that looks perfectly readable on a 27-inch display can be illegible on a 6-inch phone.

  1. Start with AI-generated transcription to create your base text -- review and correct errors before formatting into captions or subtitles
  2. Choose your format: SRT or VTT for closed captions on YouTube and web video, burned-in for TikTok, Reels, and Shorts
  3. Set timing to match audio precisely: captions appear when speech begins, disappear 0.5 to 1 second after speech ends, with minimum 1-second display time
  4. Limit each caption to 2 lines maximum, 42 characters per line, and break lines at natural grammatical boundaries
  5. Keep reading speed under 200 words per minute -- condense wordy passages to maintain readability without losing meaning
  6. Use high-contrast styling: white text on a semi-transparent black background for closed captions, bold sans-serif fonts for open captions
  7. Add speaker identification for multi-speaker content and describe sound effects and music for full caption accessibility
  8. Test on mobile devices before publishing -- if captions are unreadable on a phone screen, resize and reposition them