Skip to main content

Overview

Voice Mode enables real-time voice conversations with Apollo Brain. It uses OpenAI’s GPT-4o Realtime API for voice processing, with Apollo Brain’s tools available as function calls — so you can ask questions, create tasks, and manage your workspace entirely by voice.

How It Works

Modes

Hold a button to speak, release to send. Best for noisy environments.

Audio Processing

  • Echo Gate — prevents the speaker output from being picked up by the microphone, avoiding feedback loops
  • Mute Control — quickly mute your microphone
  • Real-time indicators — visual feedback showing when the system is listening, processing, or speaking

Example Conversations

You: “What are my tasks for today?” Brain: “You have 5 tasks due today. Two are high priority: the API endpoint migration and the login bug fix. Would you like me to go through them?”
You: “Create a task to review the deployment pipeline, assign it to me, high priority, due end of week” Brain: “I’d like to create a task: ‘Review deployment pipeline’, assigned to you, high priority, due Friday. Should I go ahead?”
You: “Show me the AWS costs for this month” Brain: “Your total AWS costs for February are 4,230.ThebiggestcontributorsareECSat4,230. The biggest contributors are ECS at 2,100 and RDS at $890. Would you like a detailed breakdown?”
Voice Mode requires a stable internet connection for WebRTC streaming. Audio quality depends on your microphone hardware.