BoltAI Documentation
HomepagePricingChangelogDownload
  • Overview
  • Features
  • License
  • Getting Started
    • Setup
    • Your First Chat
    • AI Command
    • AI Inline
    • Use another AI Service
  • Chat UI
    • Overview
    • Basic Chat
    • Document Analysis
    • Advanced Voice Mode (beta)
    • Image Generation
    • Chat Configuration
    • AI Assistant
    • AI Plugins
    • App Appearance
    • Folder & Sidebar
    • Keyboard Shortcuts
    • Import from ChatGPT
    • Import from Claude.ai
    • Import / Export
    • Database Maintenance
    • Locations
    • Feature Flags
    • Community Icon
  • AI Command
    • Overview
    • Customize an AI Command
    • Alternative Profile
    • AI Command Behaviors
    • Instant Command
    • Bulk Editing Commands
    • FAQs
  • AI Inline
    • Overview
    • Inline Assistant
    • Inline Prompt
    • Inline Whisper
    • Advanced Configurations
  • Plugin
    • Overview
    • MCP Servers
    • Google Search
    • Web Browsing
    • Memory
    • Perplexity Search
    • Kagi Search
    • Brave Search
    • You.com Search
    • AppleScript
    • Shell Access
    • FFmpeg
    • DALL·E
    • Replicate
    • Whisper
    • Whisper (via Groq)
    • WolframAlpha
    • Gemini Code Execution
  • BoltAI on Setapp
    • Setapp Limitation
    • AI Inline on Setapp
    • Troubleshooting
  • BoltAI Mobile
    • Getting Started
    • MCP Servers (mobile)
  • Guides
    • How to create an OpenAI API Key
    • How to setup Web Search Plugin for BoltAI
    • How to set up BoltAI without an OpenAI API Key
    • How to generate Azure OpenAI API key
    • How to use Azure OpenAI API key in BoltAI
    • How to create an OpenRouter API key
    • How to set up a custom OpenAI-compatible Server in BoltAI
    • How to use Mistral AI on macOS with BoltAI
    • How to use Perplexity AI on mac with BoltAI
    • How to use Anthropic Claude on macOS with BoltAI
    • How to use Replicate AI on macOS with BoltAI
    • How to use Jina DeepSearch with BoltAI
    • How to migrate data to another Mac
    • How to back up your database
    • Cloud Sync Workaround
  • Troubleshooting
    • How to fix "This license key has reached the activation limit"
    • How to fix "You exceeded your current quota, please check your plan and billing details"
    • How to fix Accessbility permission
    • How to completely uninstall BoltAI
    • Can't select text in conversation prompt
    • API keys not persisted?
    • Download Previous Versions
  • Company
    • Run by a human
Powered by GitBook
On this page
  • Prerequisites
  • How do I start a voice conversation?
  • How to change the AVM configurations?
  • How to use AVM with Azure OpenAI Service?
  • FAQs

Was this helpful?

  1. Chat UI

Advanced Voice Mode (beta)

Your guide to voice chats in BoltA

PreviousDocument AnalysisNextImage Generation

Last updated 6 months ago

Was this helpful?

Voice conversations allow you to have a spoken conversation with GPT-4o using the You can ask question or have discussions through voice input and receive spoken response from GPT-4o.

BoltAI supports both OpenAI and Azure OpenAI Service.

Prerequisites

Advanced Voice Mode (AVM) relies on OpenAI's Realtime API and requires a valid OpenAI API key, or a valid Azure OpenAI deployment.

To setup AVM with Azure OpenAI Service, follow the guide below.

How do I start a voice conversation?

To start a voice conversation, select the Voice icon on the bottom right of the chat window.

You will be taken to a screen with an animated orb in the center.

  1. Click Start

  2. Grant BoltAI the microphone permission

  3. Start your conversation

How to change the AVM configurations?

Tweak AVM System Prompt:

BoltAI automatically uses your conversation's System Prompt for the voice conversation. With this, you can reuse your already-defined AI Assistant.

Change AVM assistant's voice:

To change the assistant's voice, click the gear button on top right of the AVM screen. You can choose one the the following voices: Shimmer (default), Alloy or Echo.

Don't forget to reconnect after updated.

Show realtime cost:

Realtime API can be really expensive when using for a long time. BoltAI automatically calculates the cost and updated the conversation. In the AVM settings dialog, you can choose to show the estimated costs in realtime.

How to use AVM with Azure OpenAI Service?

In BoltAI, go to Settings > Advanced > Advanced Voice Mode, check "Use Azure OpenAI Service" and fill the form.

Make sure the API endpoint and key are correct. BoltAI won't verify the configurations.

Known Issue: Usage data seems to be incorrect for Azure OpenAI Service.

FAQs

For how long can I have voice chats?

The Realtime API currently sets a 15 minute limit for session time for WebSocket connections. After this limit, the server will disconnect. In this case, the time means the wallclock time of session connection, not the length of input or output audio.

Can I use plugins (function calling)?

Not yet.

Which model is used for AVM?

BoltAI currently uses gpt-4o-realtime-preview-2024-10-01. Note that the LLM model of the current chat configuration has no effect in AVM.

Can I get a transcription of the voice conversation?

Yes. After the conversation finished, you can find the full transcription of the conversation in the current active chat.

I cannot connect to the Realtime API?

Other caveats:

Due to macOS privacy policy with embeded scripts, the OS might ask for the microphone permission every time.

First, make sure you have your deployment ready. Follow from Azure to deploy yours. Once deployed, you should have your API Endpoint and API Key.

Please check your internet connection and make sure your API account has enough credits. If the issue persist, please file a bug request at

this official guide
https://boltai.com/ideas
OpenAI's Realtime API.
Start a voice conversation