New: High-level announcements are live
    Open
    how ai language apps actually work

    How AI Language Apps Actually Work: A Technical Breakdown

    Every AI language app claims to "listen to you speak." Almost none of them actually do. Here's what's really happening when you talk to Duolingo, Speak, Praktika, and the rest — and why the architecture matters more than the marketing.

    If you've used any AI-powered language app in the last three years, you've probably noticed they all say roughly the same thing: "Practice speaking with AI." "Your AI conversation partner." "Real conversations powered by artificial intelligence." It sounds great. The reality is more complicated. Because the way these apps process your voice determines everything — how accurate the feedback is, how natural the conversation feels, whether your accent gets heard or flattened, and whether you're actually learning to speak or just learning to pass a transcript analysis. This article pulls back the curtain. No jargon without explanation, no vague hand-waving. Here's how AI language apps actually work, from the moment you open your mouth to the moment the app responds.

    The Three-Hop Pipeline: How 95% of Apps Work

    Almost every AI language app on the market — Duolingo, Speak, Praktika, ELSA, TalkPal, Langua, Talkio — uses the same fundamental architecture. It's called the STT-LLM-TTS pipeline, and it works in three steps:

    Step 1: Speech-to-Text (STT)

    When you speak into the app, your voice doesn't go to an AI that understands speech. It goes to a speech-to-text model — a system whose only job is to convert your audio into a text transcript.

    The most common STT model is OpenAI's Whisper, though some apps use Google's Speech-to-Text API or Amazon Transcribe. These models are impressive technology. They can transcribe clear English with over 95% accuracy. But they were built for transcription, not for language learning, and the distinction matters enormously.

    Here's what happens during transcription:

    • Your audio waveform gets segmented into small chunks (usually 30-second windows)
    • Each chunk gets converted into a mel spectrogram — a visual representation of the frequency content over time
    • A neural network (typically a Transformer encoder-decoder) maps that spectrogram to the most likely sequence of text tokens
    • The output is a string of text. Just text. The audio itself is discarded.

    This is the first point of information loss. Everything that made your speech yours — your pitch, your rhythm, your hesitation patterns, the way you aspirated that consonant, whether you whispered or spoke normally — gets reduced to the same flat transcript that would result from someone typing those words.

    A native Korean speaker saying "감사합니다" and a beginner mangling every vowel can produce the same transcript. The STT model's job is to figure out what you meant to say, not how you said it. That's a feature for transcription. It's a catastrophic flaw for language learning.

    Step 2: Large Language Model (LLM)

    The text transcript — not your voice, not your audio, just the text — gets sent to a large language model. Usually GPT-4, Claude, or Gemini. The LLM sees something like:

    User said: "Je voudrais un café, s'il vous plaît"
    Context: User is practicing French at intermediate level
    Task: Respond naturally and provide feedback
    

    The LLM generates a text response. It might say something like "Très bien ! Un café, tout de suite. Vous prenez du sucre ?" along with some feedback about your grammar or vocabulary.

    This is the second point of information loss, but it's also where something genuinely useful happens. LLMs are extraordinarily good at language. They can generate natural-sounding responses, adapt difficulty, roleplay scenarios, and provide grammatical corrections. This is the part of the pipeline that actually works well.

    The problem isn't what the LLM does with the text. The problem is what the LLM never saw: your actual voice.

    Step 3: Text-to-Speech (TTS)

    The LLM's text response gets fed to a text-to-speech engine, which converts it back into audio that plays through your phone's speaker.

    TTS technology has gotten remarkably good. Services like ElevenLabs, Play.ht, and Azure Neural TTS can generate speech that sounds nearly human. Langua even uses cloned native speaker voices, which is arguably the best TTS implementation in the language learning space.

    But there's a fundamental limitation: TTS is performing. It's not conversing. The audio output has no awareness of the audio input. The "conversation" you're having is actually: your voice → text → different text → synthesized voice. It's two translation steps with a text exchange in the middle.

    The Latency Tax

    Each hop takes time:

    • STT: 200-500ms to transcribe your speech
    • LLM: 300-1000ms to generate a response
    • TTS: 200-500ms to synthesize audio

    Total: 700ms to 2 seconds. On a bad connection or with a longer utterance, it can hit 3 seconds.

    Research on conversational turn-taking shows that natural human conversation operates on a 200-400ms response window. When responses take longer than about 700ms, the conversational rhythm breaks. Your brain stops treating it as a conversation and starts treating it as a request-response interaction — like talking to Siri. You become more self-conscious, more deliberate, less natural. The very thing these apps are supposed to practice — natural speaking — gets undermined by the architecture.


    • Your audio waveform gets segmented into small chunks (usually 30-second windows)
    • Each chunk gets converted into a mel spectrogram — a visual representation of the frequency content over time
    • A neural network (typically a Transformer encoder-decoder) maps that spectrogram to the most likely sequence of text tokens
    • The output is a string of text. Just text. The audio itself is discarded.
    • STT: 200-500ms to transcribe your speech
    • LLM: 300-1000ms to generate a response
    • TTS: 200-500ms to synthesize audio

    What Gets Lost in Translation

    The three-hop pipeline doesn't just add latency. It fundamentally changes what the AI can perceive about your speech. Here's a non-exhaustive list of what gets discarded when your voice becomes text:

    Pronunciation Nuance

    STT models are trained to be robust to pronunciation variation. That's good for transcription (you want Siri to understand you regardless of your accent) but terrible for language learning (you want the system to notice your accent and help you improve it).

    When you say "gracias" with an English "r" instead of a Spanish rolled "r," Whisper transcribes it as "gracias" either way. When you mix up the three Korean stops ㄱ, ㄲ, and ㅋ — which differ by aspiration and tenseness, not voicing — the STT model often maps all three to the same text token. When you nail the rising-falling tone of Mandarin's third tone versus butchering it into a flat low tone, the transcript says "mǎ" regardless.

    Some apps add a pronunciation scoring layer on top — typically a forced alignment model that compares your audio to a reference pronunciation. ELSA does this well for English. But forced alignment only works for languages with extensive training data, and it evaluates phonemes in isolation rather than in conversational flow. It can tell you that you mispronounced the "ü" in "grün." It can't tell you that your overall intonation pattern sounds English rather than German.

    Tonal Information

    This is where the three-hop pipeline most dramatically fails. In tonal languages — Mandarin (4 tones), Vietnamese (6 tones), Cantonese (6-9 tones depending on analysis), Thai (5 tones), Yoruba (3 tones) — the same sequence of consonants and vowels can mean completely different things depending on pitch.

    The Mandarin syllable "ma" means "mother" (first tone, high flat), "hemp" (second tone, rising), "horse" (third tone, dipping), or "scold" (fourth tone, falling). In actual speech, tone sandhi rules change these tones based on what comes before and after them. The third tone before another third tone becomes a second tone. These rules are automatic for native speakers and nearly impossible for learners to master through text-based study.

    STT models handle tonal languages by using context to disambiguate. If you're talking about riding, the model figures you probably said "horse" not "mother." This is, again, great for transcription. For language learning, it means the one thing you most need feedback on — your tones — is the thing the system actively ignores.

    Whispered Speech

    Whispered speech has a fundamentally different acoustic profile than normal speech. The vocal cords don't vibrate, so there's no fundamental frequency (F0). Formant frequencies shift. The entire spectral envelope changes.

    STT models are not trained on whispered speech. Their training data is overwhelmingly normal-volume conversational speech and read-aloud recordings. When you whisper into an app that uses STT, accuracy drops dramatically — often below 50% for anything other than English. The app either can't understand you or produces a garbled transcript that the LLM then responds to nonsensically.

    This matters because a huge number of language learners want to practice quietly. In shared apartments, late at night, in open-plan offices, on public transit. The three-hop pipeline makes this impossible.

    Conversational Dynamics

    When you pause mid-sentence to search for a word, a human conversation partner picks up on it. They might wait, offer a hint, or complete the sentence. STT models treat your pause as the end of your utterance and start transcribing.

    When you self-correct — "Yo quiero... no, yo quisiera un café" — a human tutor recognizes that you caught your own mistake and switched to the subjunctive. The STT model transcribes the whole thing as a single utterance, and the LLM may or may not notice the self-correction in the text.

    When your voice rises with uncertainty, a native speaker hears it as a question or a sign that you need help. The STT model doesn't encode prosody in its output.

    These micro-signals are the fabric of actual conversation. Stripping them out means the AI is responding to what you said, not how you said it.


    The Alternative: Speech-to-Speech

    There's a different approach. Instead of converting speech to text, processing text, and converting back to speech, you can process audio natively. Audio in, audio out, no text intermediary.

    This is what Yapr does with Gemini's multimodal audio capabilities. The pipeline looks like:

    1. You speak
    2. Your raw audio goes directly to a multimodal model that processes audio natively
    3. The model generates an audio response directly

    One hop instead of three. No transcription step. No TTS synthesis. The model that understands what you said is the same model that responds, and it operates on the audio signal itself rather than on a lossy text approximation.

    What Changes

    Latency drops to sub-second. One model inference instead of three sequential steps. Yapr's response time is consistently under one second, which puts it inside the natural conversational turn-taking window. Conversations feel like conversations, not like dictating into a machine.

    Pronunciation feedback gets real. Because the model processes your actual audio, it can distinguish between sounds that produce the same transcript. It can hear the difference between your aspirated and unaspirated Korean consonants. It can detect whether your Mandarin third tone is actually dipping or just low. It can tell that your Spanish "rr" is an English approximant, not a trill. The feedback is on your actual pronunciation, not on what Whisper decided you probably meant.

    Whispered speech works. A multimodal audio model trained on diverse audio data — including whispered, quiet, and non-standard speech — can process whispered input without collapsing. This is the fundamental architectural reason why Yapr's whisper mode exists and no STT-based app offers it. It's not a feature they chose not to build; it's a feature their architecture can't support.

    Tonal languages are actually teachable. When the model hears your audio directly, it can process pitch contours, tone patterns, and tone sandhi in context. It doesn't need to disambiguate through text — it can evaluate whether you produced the right tone, not just whether the right word appeared in a transcript.

    Conversation feels natural. The model can hear your hesitation, your self-corrections, your prosodic cues. It can respond to how you're speaking, not just what the transcript says you spoke. It can wait when you're searching for a word rather than cutting you off when you pause.


    Why Didn't Everyone Build It This Way?

    If native audio processing is better for language learning, why does almost every app use the three-hop pipeline?

    Timing. The STT-LLM-TTS pipeline has been buildable since GPT-3.5 (early 2023). Reliable multimodal audio processing only became viable with Gemini's multimodal audio capabilities in late 2024. Speak ($162M raised), Praktika ($38M raised), and ELSA all committed to their architectures before the alternative existed. Rebuilding a shipping product's core pipeline is a multi-quarter engineering project with significant risk.

    Infrastructure. The three-hop pipeline uses commodity APIs. You can buy STT from Google, LLM from OpenAI, TTS from ElevenLabs, wire them together, and ship an MVP in a few weeks. Building on native multimodal audio requires deeper integration with fewer providers and less off-the-shelf tooling.

    Inertia. When your entire product — feedback system, curriculum, analytics, pronunciation scoring — is built on text transcripts, switching to native audio means rethinking everything downstream. It's not just swapping one API for another. The data structures change. The feedback mechanisms change. The way you measure proficiency changes.

    None of this means the three-hop pipeline is bad technology. It's mature, well-understood, and works well enough for many use cases. If you're building a basic vocabulary drill or a simple chatbot, STT-LLM-TTS is fine.

    But for serious speaking practice — especially for heritage speakers, tonal language learners, introverts who need whisper mode, or anyone who wants the conversation to feel real — the architecture matters. And right now, the architecture most apps use is optimized for building quickly, not for learning effectively.


    How to Test Your App's Architecture

    You don't need to be an engineer to figure out which pipeline your language app uses. Three simple tests:

    The Whisper Test

    Whisper into the app. If it can't understand you or accuracy drops dramatically, it's using STT. Native audio processing handles whispered speech.

    The Tone Test (for tonal languages)

    Say the wrong tone deliberately. Say "mā" (mother, first tone) when you should say "mǎ" (horse, third tone). If the app doesn't catch it and responds as if you said the right thing, the STT model disambiguated via context and the pronunciation error was invisible.

    The Pause Test

    Pause for 3-4 seconds in the middle of a sentence, then continue. If the app treats your first half as a complete utterance and responds before you finish, it's using STT with an endpoint detection model. Native audio processing can distinguish between a hesitation pause and the end of an utterance.


    What This Means for You

    If you're shopping for a language app, ask one question: does this app hear me, or does it read a transcript of me?

    The marketing will always say "AI-powered conversation." The architecture determines whether that conversation is real or simulated. A real conversation means the AI processes your actual voice — your accent, your rhythm, your tone, your hesitations. A simulated conversation means the AI reads a text approximation of what you said and generates a text response that gets spoken back to you.

    Both can be useful. But they're not the same thing, and for speaking practice specifically, the difference is the whole game.

    Yapr supports 47 languages with native speech-to-speech processing. No text middleman, sub-second response times, and it actually hears you — including when you whisper. If you want to test the difference yourself, start at yapr.ca.


    Frequently Asked Questions

    What is STT-LLM-TTS?

    STT-LLM-TTS stands for Speech-to-Text, Large Language Model, Text-to-Speech. It's the three-step pipeline used by most AI language apps. Your voice is transcribed to text, the text is processed by an AI, and the AI's text response is converted back to speech. Each step introduces latency and information loss.

    Which language apps use speech-to-speech AI?

    As of 2026, Yapr is the primary consumer language app using native speech-to-speech processing powered by Gemini's multimodal audio. Most other apps — including Speak, Praktika, ELSA, Duolingo, TalkPal, and Talkio — use the STT-LLM-TTS pipeline.

    Why can't Duolingo understand my accent?

    Duolingo uses speech-to-text models that are designed to be *robust* to accent variation, meaning they try to transcribe your speech correctly regardless of your accent. This is the opposite of what language learners need — you want the system to *notice* your accent so it can help you improve. Native audio processing evaluates your actual pronunciation rather than just checking if the transcript matches.

    Does speech-to-speech AI work for tonal languages?

    Yes — and it's significantly better than STT-based approaches for tonal languages like Mandarin, Vietnamese, Cantonese, and Thai. STT models often disambiguate tones through context (guessing which word you meant), which hides pronunciation errors. Speech-to-speech processing can evaluate your actual pitch contours and tone production directly.

    What is whisper mode in language apps?

    Whisper mode allows you to practice speaking at a whisper volume. STT-based apps generally can't handle this because their transcription models weren't trained on whispered speech. Yapr's native audio processing can interpret whispered speech, making it possible to practice in shared spaces, late at night, or anywhere you can't speak at full volume.

    Start Speaking Today

    Try Yapr free — real conversations, 47 languages, zero judgment.