Realtime translation
Yap United
A real-time translation app built for shared-device conversations, with Gemini Live audio streaming, ElevenLabs voices, and location-based community chat.
Yap United
Shared-device conversations, live mode, and multilingual voice output that holds up in motion.
What shipped
- Built both turn-based translation and hands-free live mode for real conversations on a shared device.
- Supported 15 languages end-to-end across transcription, translation, and voice output.
- Added per-user voice assignment, location-based community chat, and social moderation controls around the core translation flow.
Core stack
Translation apps often break the rhythm of an actual conversation. Yap United needed to handle live voice, turn-taking, and multilingual community behavior without making people fight the interface.
I built a dual-mode speech system: a turn-based flow for shared-device conversations and a Gemini Live pipeline for hands-free mode. Audio is recorded with Expo, translated with Gemini, voiced with ElevenLabs, and routed to the correct earbud side with reconnect and backoff handling when the live session drops.
Yap United supports 15 languages end-to-end, lets each user keep a distinct voice identity, and extends beyond translation with community zones, moderation controls, and non-Latin script handling that keeps the conversation usable under real conditions.
Need this level of product depth for your own build?
The same mix of product direction, interface work, and backend systems can be applied to your next app.