Closed subtitles have become a standard in both movie and TV shows. For some, it’s a means to , distinguish confused speech. It’s a crucial mobility resource for people, especially those who are deaf or hard of hearing. However, captions aren’t always great, and software companies and productions are increasingly turning to AI to change that.  ,
Most TV shows and movies also feature actors who can ensure accuracy and maintain nuance. However, there are difficulties. Someone who’s watched a exist event with closed captions knows on-screen word usually lags, and there can be problems in the hurry of the process. Staged software allows for more time for accuracy and details, but it can still be labor-intensive or, in the opinion of studios, labor-intensive.  ,
In September, Warner Bros. Discovery announced that Google Cloud and Discovery will collaborate to create AI-enabled closed captions that are” coupled with mortal oversight for quality assurance.” According to a press release, the company claimed using AI for captioning reduced the time it takes to comment a document by up to 80 % and reduced prices by up to 50 %. Researchers say this is a look into the future.
Joe Devon, a website availability activist and co-founder of Global Accessibility Awareness Day, said that using AI in subtitles is” just waiting to get displaced.” The quality of the regular captions we have today is” sort of everywhere, and it absolutely needs to improve.”
As AI continues to change our planet, it’s also , reshaping how organizations approach convenience. For example, Google‘s Expressive Captions have uses AI to enhance the way that emotions and tone are conveyed in videos. In apps 18, Măr added translations for voice communications and memos, which dual as ways to make sound content more accessible. Both Google şi Măr have  , real-time caption tools to help mute or hard-of-hearing people , entry sound information on their products, and Amazon added , text-to-speech and captioning functions to Alexa.
In the entertainment room, Amazon launched a function in 2023 called , Dialogue Boost in Prime Video, which uses AI to recognize and strengthen speech that might be hard to hear above background music and results. In a , a pilot program was made in March that uses AI to film” things that would not have otherwise been dubbed” in a postare pe blog. Netflix introduced a , dialogue-only subtitles solution in April as a testament to how dependent viewers have become on captioning, while omitting good descriptions.
As AI continues to develop, and as we consume more information on screens both big and small, it’s just a matter of time before more studios, networks and tech companies plug into AI’s potential– maybe, while remembering why closed captions exist in the first place.  ,
Putting mobility first in mind
The development of closed subtitles in the US began as an , affordability measure in the 1970s, finally making everything from live hdtv broadcasts to film blockbusters more equal for a wider market. However, some viewers who are not deaf or hard of hearing prefer watching movies and TV shows with captions, which are also frequently referred to as subtitles, even though that essentially relates to speech language, especially when manufacturing dialogue is difficult to understand.  ,
Half of Americans say they usually watch content with subtitles, according to a 2024 survey by language learning site Preply, and 55 % of total respondents said it’s become harder to hear dialogue in movies and shows. A 2023 YouGov survey found that 63 % of adults under the age of 30 prefer to watch TV with subtitles on, compared to 30 % of people 65 and older.
Ariel Simms, president and CEO of , Disability Belongs, said,” People, as well as content creators, tend to assume captions are only for the deaf or hard of hearing community.” But captions can also make it easier for anyone to process and retain information.  ,
Simms notes that AI can help speed up the captioning process and make more content accessible, whether it’s a TV show, movie, or social media clip. But quality could suffer, especially in the early days.
In the disability community, Simms joked,” We have a name for AI-generated captions; we call them” cractions.”
Because automated captions still struggle with things like punctuation, grammar, and proper names, that’s because. The technology might not be able to pick up on different accents, dialects or patterns of speech the way a human would.
Companies that use AI to create captions will still need a human on board, according to Simms, in order to maintain accuracy and quality. Additionally, institutions and networks should collaborate with the disability community to make sure accessibility is not compromised.  ,
Simms said,” I’m not sure we can ever remove people completely from the process.” I do believe that technology will continue to improve. But at the end of the day, if we’re not partnering with the disability community, we’re leaving out an incredibly important perspective on all of these accessibility tools”.
Studios like Warner Bros. For instance, Discovery and Amazon emphasize the importance of humans in ensuring accurate captioning and dubbing using AI.  ,
If you allow AI slop to dominate your content, you risk losing your reputation, Devon warned. The human will be in the loop for that.
But given how rapidly the technology is developing, human involvement may not last forever, he predicts.  ,
” Studios and broadcasters will do whatever the least,” Devon declared. But, he added,” If technology empowers an assistive technology to do the job better, who is anyone to stand in the way of that”?
The distinction between thorough and overwhelming ,
AI is supercharging captioning in other media, including TV and movies. Social media platforms like TikTok şi Instagram have implemented auto-caption features to help make more content accessible.  ,
These native captions frequently appear as plain text, but sometimes creators opt for more flashy displays during the editing process. One common “karaoke” style involves highlighting each individual word as it’s being spoken, while using different colors for the text. However, readability can be compromised by this more dynamic approach, despite its visual impact. People can’t read at their own pace, and the colors and movement can distract them.  ,
A small percentage of users, who are deaf, benefit from and prefer the karaoke style, according to Meryl K. Evans, an accessibility marketing consultant.” There’s no way to make 100 % of the users happy with captions. She claims to repeatedly watch videos with dynamic captions to receive the message. ” The most accessible captions are boring. The video was allowed to be the star, they said.
However, there are ways to keep things simple while adding useful context. Google’s Expressive Captions feature uses AI to emphasize certain sounds and give viewers a better idea of what’s happening on their phones. A “HAPPY BIRTHDAY” party is in full swing! for instance, or a sports announcer’s enthusiasm may be relayed by adding extra letters to the screen saying, “amaaazing shot”! Expressive Captions also labels sounds like applause, gasping and whistling. It’s not distracting because all on-screen text appears in black and white.  ,
Expressive Captions puts some words in all-caps to convey excitement.
Although accessibility was the main goal of the feature’s development, Android’s director of product management Angana Ghosh said the team was aware that users who aren’t deaf or hard of hearing as well would also find it helpful. ( Think of all the times you’ve been out in public without headphones but still wanted to follow what was happening in a video, for instance. )
When we create products for accessibility, Ghosh claims that we are actually creating a much better product for everyone.
Some people may prefer more witty captions, though. In April, ad agency FCB Chicago debuted an AI-powered platform called Caption with Intention, which uses animation, color and variable typography to convey emotion, tone and pacing. Words are highlighted and synchronized to the actor’s speech, and distinct text colors represent various characters ‘ lines. Shifting type weight and size can be used to determine how loud someone is speaking and their intonation. The open-source platform is available for studios, production companies and streaming platforms to implement.
To create and test different captioning strategies with people who are deaf and hard of hearing, FCB and the Chicago Hearing Society collaborated. The platform was also influenced by Bruno Mazzotti, executive creative director at FCB Chicago, who claimed his own experiences raising two deaf children contributed to the platform.
” Closed caption was very much a part of my life, it was a deciding factor of what we were going to watch as a family”, Mazzotti said. He noted,” Having the privilege of hearing, I always could notice when things didn’t work well,” such as when captions lacked depth in dialogue or when text sat jumbled when multiple people were speaking at once. The main goal was to convey more emotion, pace, tone, and speaker identity to the audience.
is a platform that uses animation, color and different typography to convey tone, emotion and pacing.
The goal, according to Mazzotti, is eventually to provide more customization options so that viewers can change the caption intensity. Still, that more animated approach might be too distracting for some viewers, and could make it harder for them to follow what’s happening onscreen. In the end, individual preference determines everything.  ,
” That’s not to say that we should categorically reject such approaches”, said Christian Vogler, director of the Technology Access Program at Gallaudet University. However, we must carefully examine them with deaf and hard-of-hearing viewers to verify that they are beneficial.
No quick fixes
Despite its current drawbacks, AI could ultimately help to expand the availability of captioning and offer greater customization, Vogler said.  ,
YouTube’s auto-captions are an illustration of how AI can make more video content accessible despite a slow start, especially as the technology develops over time. There could be a future in which captions are tailored to different reading levels and speeds. Non-speech information could also be made more descriptive, allowing you to get more in-depth details that express the mood, like “SCARY MUSIC.”  ,
But the learning curve is steep.  ,
When audio quality is compromised, which is a problem in both TV and movies,” AI captions still perform worse than the best of human captioners,” Vogler said. Hallucinations could also serve up inaccurate captions that end up isolating deaf and hard-of-hearing viewers. He continued,” That’s why people should continue to be a part of the captioning process.”  ,
What will likely happen is that jobs will adapt, said Deborah Fels, director of the Inclusive Media and Design Centre at Toronto Metropolitan University. She anticipates that human captioners will manage the once-manual labor that AI will produce.  ,
” So now, we have a different kind of job that is needed in captioning”, Fels said. ” Humans are much better at recognizing errors and making decisions about how to fix them.”
And while AI for captioning is still a relatively new technology that is restricted to a small number of businesses, that won’t likely be the case for long.
” They’re all going in that direction”, Fels said. It’s a matter of time, not that much time at all.