Hume AI introduces Voice Control, empowering developers to personalise AI voices without coding, enhancing voice customisation and emotional intelligence.
Hume AI, an emerging player in the realm of emotionally intelligent voice interfaces, has recently debuted a groundbreaking feature known as Voice Control. This experimental tool enables developers and users to create personalised AI voices by adjusting various vocal characteristics without necessitating any coding, prompt engineering, or expertise in sound design.
Voice Control builds upon the innovations established by Hume’s previous product, the Empathic Voice Interface 2 (EVI 2). This earlier model made strides in enhancing voice naturalness, emotional responsiveness, and overall customisation. According to Alan Cowen, co-founder of Hume and a former member of the Google DeepMind team, “the release of Voice Control addresses a key pain point in the AI industry: the reliance on preset voices, which often fail to meet the specific needs of brands or applications.” Cowen further highlighted that both EVI 2 and Voice Control sidestep the ethical pitfalls associated with voice cloning by providing tools for the development of distinct, expressive voices.
Developers utilising Voice Control can modify voices along ten different dimensions, which include attributes such as gender expression (Masculine/Feminine), confidence levels, assertiveness, enthusiasm, and more. This level of granularity allows for a highly tailored voice experience that can be finely tuned through real-time adjustments using virtual sliders.
Currently accessible via Hume’s virtual playground with a complimentary user sign-up, Voice Control offers a user-friendly interface that represents a shift from traditional text prompts to a more intuitive sliding scale for modulating voice attributes. This approach captures the nuanced ways humans perceive vocal qualities while maintaining the complexity of emotional expression.
The launch coincides with significant advancements made in EVI 2, which improved latency by 40%, reduced operational costs by 30%, and broadened the available voice modulation capabilities. The earlier model’s features included in-conversation prompts and multilingual support, which has been leveraged in applications ranging from customer service to virtual tutoring.
Voice Control’s implementation promises to enhance interaction with voice-based systems by permitting developers to select a foundational voice, customise its traits, and instantly preview these modifications. This ensures consistent replication and stability, pivotal for real-time applications like chatbots or digital assistants.
In an increasingly competitive market that includes formidable competitors such as OpenAI and ElevenLabs, Hume AI’s emphasis on emotional intelligence and customisation sets it apart. These attributes foster differentiation in a landscape often dominated by pre-set voices. Hume’s ongoing developments aim to expand Voice Control further with additional adjustable dimensions and a greater variety of base voices, driving innovation in voice AI technology.
As Hume AI enhances its offerings with tools prioritising customisation and emotional sophistication, it solidifies its status as a notable leader in voice AI innovation. The availability of Voice Control marks a significant advancement in the evolution of AI-driven voice solutions, providing developers with a powerful resource to meet diverse user needs in various business applications.
Source: Noah Wire Services
- https://www.aibase.com/news/13637 – This article explains the launch of Hume AI’s ‘Voice Control’ feature, its integration with Empathic Voice Interface 2 (EVI2), and the ability to adjust various voice characteristics without coding or sound design expertise.
- https://www.aibase.com/news/13637 – It details how Voice Control addresses the issue of preset voices in the AI industry and highlights the ethical advantages over voice cloning.
- https://dev.hume.ai/docs/empathic-voice-interface-evi/overview – This documentation provides an overview of Hume’s Empathic Voice Interface (EVI), including its capabilities, such as real-time interactions, emotional intelligence, and multilingual support.
- https://dev.hume.ai/docs/empathic-voice-interface-evi/overview – It explains the technical aspects of EVI, including WebSocket and REST APIs, and the integration methods for developers.
- https://www.hume.ai/blog/introducing-voice-control – This blog post introduces the Voice Control feature, its interpretability-based method for AI voice customization, and the ability to adjust 10 voice dimensions.
- https://www.hume.ai/blog/introducing-voice-control – It discusses the user-friendly interface, real-time adjustments, and the preservation of voice characteristics when modifying specific parameters.
- https://www.hume.ai/blog/introducing-voice-control – The post also outlines the future plans for expanding Voice Control, including additional dimensions and base voices.
- https://www.hume.ai – This page provides an overview of Hume AI’s products, including EVI 2, and its capabilities in voice modulation, emotional intelligence, and real-time interactions.
- https://www.hume.ai – It highlights the improvements in EVI 2, such as reduced latency and operational costs, and the broadened voice modulation capabilities.
- https://www.aibase.com/news/13637 – This article mentions the competitive landscape and how Hume AI’s focus on emotional intelligence and customization sets it apart from competitors like OpenAI and ElevenLabs.
- https://www.hume.ai/blog/introducing-voice-control – The blog post emphasizes the significance of Voice Control in the evolution of AI-driven voice solutions and its potential to meet diverse user needs in various business applications.












