AI voices are an important factor in making customers feel satisfied with your automated voice services. In this article, we take an in-depth look at choosing AI voices for your voicebot your customers will be happy to interact with.
How your voicebot speaks and sounds is one of the first things your customers will notice when they call. Most customers will expect the voicebot to have a “human-like” voice. A robotic-sounding voice will have an impact on call success. If the voice sounds like a robocall, there is a higher chance that customers will hang up.
There are a variety of ways in which voicebot speech can be configured and each configuration has slightly different outcomes in terms of the final customer experience. Below, we compare some common voicebot solutions and show how they handle both general and personalized conversations.
Voice synthesis for voice AI
Voice synthesis uses existing voice samples to generate a synthetic voice for an AI virtual assistant. Using the samples, the voice synthesizer is able to extract pronunciation to enunciate a wider range of words than are covered by the original recordings. This technology makes it easy to create a virtual assistant speaking voice that can handle a wide range of variables in conversation.
Personalizing customer conversations and voice synthesis
If you want your virtual agents to greet customers with a personalized message, e.g. to use the customer’s name stored in your CRM records, you need to work with variables that can generate dynamic content. The variable in this example would be the customer’s name.
For a voicebot to accurately say someone’s name, it needs either:
- A large library of voice recordings of names recorded by the voice actor for the voicebot
- A synthesized voice that can approximate the pronunciation of the variable content, based on existing recordings and some smart machine learning
Pre-recorded samples sound more natural but the cost to produce them is relatively high. In some cases, there will be so much variable, dynamic content that it would be difficult to record all of it affordably up front.
Synthesized voices offer a lot more vocabulary coverage at a much lower cost but sometimes the voice synthesis can sound unnatural to human listeners. It’s an area of technology that is developing quickly, however, and the quality of synthetic voices is improving rapidly.
If personalized conversations are a planned part of your customer experience with your virtual agents, you will need to consider the most appropriate strategy for voice synthesis for your business.
Industry examples of synthetic voices for AI
Here are three examples of synthetic voices for AI-powered voicebots. The first is from Amazon (AWS), the second is from Google, and the last example is from a third-party vendor, Resemble.ai.