Same-Language Mode — Real-Time Transcription at 10x Lower Cost
March 2026
Live Translate Live is built for real-time translation between two languages — but the same engine solves a different problem almost by accident. Set your source and target language to the same value and the app switches to same-language transcription mode: a single full-screen scrolling marquee that displays live speech-to-text captions with no translation step at all. Credits burn 10x slower, the display gets bigger, and the tool quietly turns into one of the cheapest live captioning setups you can run in a browser.
Most people discover it the same way. They sign up to translate something — a visit from relatives, a medical appointment, a stream — and then realize the more persistent problem in their life is that someone in the room can't hear the TV. Or the meeting. Or the speaker at the front of the room. Same-language mode was the answer hiding in the settings the whole time.
The Moment It Clicks
A user wrote in recently with a simple description: "My mother has hearing loss. We watch British mysteries together and the accents plus the music mean she misses half of every episode. The built-in TV captions are fine for shows with captions — but a lot of what we watch is streamed from sources that don't caption reliably, or the captions are delayed, or they're in the wrong language." She propped her phone up next to the TV, set both languages to English, and let the marquee scroll live captions below the picture. Her mother can follow along again. That's the use case in one paragraph.
It's the same pattern over and over: someone needs captions for speech that isn't being captioned by anything else. A grandfather listening to a grandchild read a school essay. A manager running a staff meeting where one employee has an auditory processing disorder. A pastor whose congregation includes several hard-of-hearing members. A conference talk where the speaker's accent is hard to parse. Nothing about this is a translation problem — but it is a real-time speech-to-text problem, and that's exactly what the translation pipeline is doing before it translates anything.
How to Activate Same-Language Mode
There is nothing extra to install or configure. Open Live Translate Live, and set both the "Your Language" and "Their Language" dropdowns to the same language — for example, both set to English. That's it. The app automatically detects the match and switches to transcription-only mode.
When same-language mode is active, the marquee changes from a split two-row display to a single full-height panel. Only your speech-to-text transcription scrolls across the screen — no translation row, no split layout, just one clean stream of captions at roughly double the vertical size. On a TV or projector, that difference is the difference between "readable" and "readable from the couch."
Credit Math — Why It's 10x Cheaper
Translation mode runs two full speech-recognition pipelines (one for each speaker) plus a Google Cloud Translation call for every phrase. Same-language mode drops the translation call entirely and collapses to a single speech-recognition pass. Our billing reflects that: translation mode charges 1 credit per second, same-language mode charges 1 credit per 10 seconds. Ten times longer on the same wallet.
Put in dollars, the same credit packs cover dramatically different runtimes depending on which mode you're in:
| Mode | Credits / second | $1 (900 credits) | $3 (~11,000 credits) | Best for |
|---|---|---|---|---|
| Translation | 1 / sec | ~15 minutes | ~3 hours | Two speakers, different languages |
| Same-language | 1 / 10 sec | ~2.5 hours | ~30 hours | Captions, accessibility, transcription |
Credits are billed only while audio is actively being transcribed — silences and paused sessions don't burn credits. Actual runtime varies depending on how continuously people are speaking. See pricing for the full credit-pack list.
A dollar gets you a whole afternoon of captioned TV. Three dollars gets you most of a working week of Zoom captions. The pricing is still pay-as-you-go and credits don't expire, so the same pack that funds a year of occasional captioning sessions also still covers the occasional translation conversation when you need it.
Use Cases — Who's Actually Using It
The use-case list started small and keeps widening. A few patterns have emerged clearly enough to describe in detail:
Accessibility at Home
This is the largest single group. A family member with hearing loss, a household that wants captions for streaming services that don't caption well, or a visually reinforced version of a conversation in a noisy kitchen where the dishwasher is running. The setup is almost always the same: a phone or tablet on a stand next to (or below) the TV, aimed at the TV speakers, with both languages set to the same value and the marquee filling the screen. Credits last long enough that you can forget it's running.
A variation: captioning foreign-language content back into its original language. Watching a French film with the original French audio? Same-language mode set to French gives you live French captions even when the streaming service only offers English subtitles. It's not a replacement for professionally authored subtitles, but for unscripted or rare content it often beats the alternative of nothing.
Zoom, Google Meet, and Other Meetings
Google Meet has captions. Zoom has captions. Teams has captions. They're all pretty good. What they don't do is show up on a second screen where you can read along while taking handwritten notes, or where a colleague with hearing loss can look at them without fighting the call's active-speaker framing. Same-language mode running on a tablet beside your laptop solves that — and it also picks up audio from external speakers the platform caption system often misses (someone on speakerphone across the conference room, a remote participant whose audio is being played through a Bluetooth puck).
Running captions during a meeting is also a surprisingly effective note-taking aid. When someone says "the QBR is set for Q3 and the SLA on that is 99.95" and the captions show you the words as you heard them, it's much easier to pull the right acronyms into your own notes. Conversation history in your account saves the full transcript for later reference.
Live Events and Houses of Worship
Same-language mode on a large display — a projector, a wall-mounted TV, a dedicated monitor at the back of the room — turns the app into a lightweight live captioning system. A church running a Sunday service can have the sermon scroll in real time for hard-of-hearing congregants. A conference can put captions on a side screen. A board meeting can caption the chair's remarks for distant listeners. The pricing makes this feasible for small organizations that couldn't afford a dedicated CART service — $3 covers a month of weekly two-hour services with room to spare.
Streaming on OBS
Same-language mode works cleanly as an OBS browser source. See the OBS + smart TV setup guide for the full walkthrough — the short version is that the standalone marquee URL can be added as a browser source just like any other overlay, and with both languages set to the same value you get a clean captioning ribbon for an English (or Spanish, or Japanese) stream. The single-row layout leaves more vertical space for your game feed or webcam, and the 10x credit rate means a multi-hour stream costs pennies.
Accent and Terminology Reinforcement
Same-language mode is genuinely useful for situations where the speech is in a language you know but still takes effort to follow — a conference talk with a strong accent, a technical lecture packed with unfamiliar jargon, a medical or legal conversation where the exact words matter. Seeing the written form alongside the spoken form catches things the ear would skip over: proper nouns, drug names, statute numbers, acronyms. English-speaking attendees at an English-language conference often find captions more useful than translated captions would be, because the bottleneck isn't language, it's density.
Device Setup — What Works Where
Same-language mode is device-agnostic, but certain setups work better than others depending on the use case:
- Phone on a stand, near a TV — cheapest accessibility setup. A $10 desk stand holds a phone upright next to the TV so both people can see the captions without holding anything. Works with the phone mic picking up TV audio; works better if you pair the phone to the TV's Bluetooth or use a wired aux line.
- Tablet on a stand, beside a laptop — best for meetings. A 10-inch tablet running the marquee gives readable caption text without eating laptop screen real estate. Use the tablet's own mic or the laptop's mic; either works.
- Laptop propped up, second monitor — good for home offices. Open the standalone marquee URL on a second monitor so it runs independently of whatever you're doing on the main screen.
- Smart TV via Chromecast or built-in browser — best for live events and houses of worship. Chromecast the standalone marquee URL from a phone, or open it directly in the TV's browser if it has one. The full-screen single-panel layout is designed for exactly this setup.
- OBS browser source — for streamers. Add the standalone marquee URL as a browser source and position it as a lower-third or full-width ribbon. Credit consumption is trivial for typical stream lengths.
Limitations — What to Expect
It's worth being upfront about where live speech-to-text falls short, because same-language mode puts the captions directly in front of you rather than hiding them in a translation layer.
- Proper nouns, numbers, and acronyms are the hardest class of words. Street names, unfamiliar last names, long numeric strings, medical terms, product SKUs — these are where errors cluster. Common speech is very accurate; specialist vocabulary is hit or miss.
- No speaker diarization. The marquee doesn't distinguish between speakers. Everything everyone says flows into one stream. For one-speaker scenarios (a lecture, a sermon, a TV show with one dominant voice) this is fine; for a roundtable with five people talking it produces a readable but un-attributed transcript.
- Punctuation is inferred, not perfect. The speech engine places periods and commas based on pauses and prosody. Long flowing sentences sometimes land as one run-on; short clipped phrases sometimes land as fragments. It's readable but not publication-ready.
- Latency is sub-second but not zero. Expect captions to appear roughly half a second to a second behind the spoken word, with faster settlement for short clear phrases and slower settlement for long rambling ones. This is true of every live captioning system — human or machine — and generally fine for comprehension.
- Background noise matters. Clean audio in = clean captions out. A phone placed across the room from a mumbling TV will produce worse captions than a phone held six inches from a clear speaker. If accuracy matters, shorten the distance between mic and source.
- No offline mode. The speech engine runs in the cloud, so you need an internet connection. This is a real limitation for some use cases and is worth knowing upfront.
How It Compares to Dedicated Captioning Apps
Same-language mode isn't the only real-time captioning option. A fair comparison:
- YouTube auto-captions — free, excellent, but only work for content already on YouTube. No good for live TV, Zoom, in-person conversations, or anything that isn't a YouTube video.
- Otter.ai — purpose-built transcription tool with strong speaker diarization and searchable transcripts. Better for professional meeting notes; more expensive, subscription-based, and less suited to ambient "put captions on a screen" use cases.
- Google Meet / Zoom / Teams captions — free and good, but locked inside their respective platforms. Can't caption a podcast playing through speakers, a TV show, an in-person conversation, or a stream.
- Live Transcribe (Android) / Live Captions (iOS, macOS, Windows) — excellent on-device options that don't require internet. Better for privacy-sensitive use cases; no shared display mode and no multi-device flexibility.
Same-language mode on Live Translate Live is best when the thing you need is a shareable, readable captioning ribbon on a screen that anyone in the room can see — not a private on-device caption layer and not a platform-specific meeting feature. It's deliberately a complement to those tools, not a replacement. If you already have Live Captions running fine on your laptop for yourself, you don't need this. If you need captions on the wall for a room of people, this is the simplest way to get them.
FAQ
Can I use it with multiple speakers?
Yes, but with a caveat. Multiple people can speak into the same microphone and the marquee will transcribe all of it into a single stream. It does not label who said what — there's no speaker diarization. For one dominant speaker with occasional interjections (a sermon, a meeting chair, a teacher) this works well. For a freewheeling four-way conversation it produces a readable transcript that mashes everyone together without attribution.
Does it work offline?
No. The speech-recognition pipeline runs in the cloud, so you need an internet connection. If offline captioning is a hard requirement, the built-in OS captioning tools (Live Captions on iOS / macOS / Windows, Live Transcribe on Android) run on-device and are excellent for personal use.
How accurate is the transcription?
For clear audio of common speech, accuracy is high — comparable to YouTube auto-captions and Google Meet captions, which use similar underlying engines. The weak spots are proper nouns, numbers, acronyms, and heavily accented or overlapping speech. Accuracy scales with audio quality: a phone near the speaker will caption better than a phone across the room.
Can I save the transcript?
Yes. Every session is saved to your conversation history, which you can review and export later. See the features page for the current export options. For power users, the history includes timestamps, so you can use it as both a live captioning tool and an after-the-fact transcription tool.
Full-Screen Single-Panel Display
In translation mode, the marquee splits into two rows: one for each language. In same-language mode, the entire screen is used for a single row of scrolling text. This means larger text and better readability — especially useful when the display is across a room or projected on a wall.
All of the existing marquee features still work: adjustable scroll speed, font scaling, push-to-talk mode, standalone marquee URLs, OBS browser source compatibility, and the vis-à-vis flipped-display mode if you want captions readable from both sides of a table. The only difference from translation mode is the layout — one row instead of two — and the credit rate.
Works With Everything You Already Use
Same-language mode is not a separate feature — it is the same app in a different configuration. Everything you have set up continues to work:
- Standalone marquee — Open the marquee in a separate browser tab or window, or share the URL with a smart TV
- OBS browser source — Your existing OBS setup works unchanged; the marquee automatically switches to single-panel mode
- Push-to-talk — Hold the button to transcribe, release to stop
- Conversation history — All transcriptions are saved to your history for later review
Switch back to translation mode at any time by selecting a different target language. The marquee will restore the two-row layout and resume translating at the standard rate.
Try It Now
Sign in to Live Translate Live, set both languages to the same value, and start speaking. You will see a single full-screen marquee scrolling your words in real time — at a fraction of the cost. Whether you need live captions for accessibility, a presentation display, streaming overlay, or just a way to see your own speech on screen, same-language mode has you covered.
Try for $1 — no subscription. Credits don't expire, so one pack covers months of occasional captioning. For a full walkthrough of the broadcast setup, see the marquee on OBS and smart TV guide.