Hume AI
Hume AI builds empathic voice and emotional AI models. The platform exposes four published APIs: Voices, Octave Text-to-Speech, Empathic Voice Interface (EVI / speech-to-speech), and Expression Measurement (multimodal emotion analysis). REST + WebSocket interfaces are documented with public OpenAPI and AsyncAPI specifications at https://dev.hume.ai/.
APIs
Hume Voices API
REST API for managing custom voices used by Hume's TTS and EVI products. Supports voice creation, listing, and deletion. Bearer token auth.
Hume Octave Text-to-Speech API
REST API for synthesizing expressive speech using Octave. Supports streamed JSON/file and standard JSON/file responses, plus voice conversion endpoints.
Hume Empathic Voice Interface (EVI) API
Real-time emotionally intelligent speech-to-speech voice interface delivered over WebSockets. Includes configuration, prompt, and tool management plus chat history, chat group o...
Hume Expression Measurement API
Multimodal emotion analysis API for face, voice, language, and burst expression. Offers batch inference jobs over images/audio/video/text, and streaming inference.