
The next generation of artificial intelligence-powered chatbots may be so good at answering your questions that you won't be able to tell them apart from their real-life counterparts.
Five9, the firm behind a new answering machine, believes that its new technology, which use artificial intelligence to break down phrases into sounds and tones, would save businesses money on labor expenditures.
The use of a human voice to teach the AI is critical to ensuring a positive caller experience.
What is the mechanism behind it?
Five9 auditioned performers in London for their newest voice and chose Joseph Vaughn to record a series of scripts for the firm.
The AI computer was then able to reproduce not only phrases but also different emotions by breaking down the audio into sounds and tones rather than words.
The program has been trained to recognize word combinations and tones in the caller's speech, as well as their emotional condition.
"We are capturing all of the audio data and all of the combination of frequencies and vibrations that are inherent to a voice, that as a human we would recognise is a voice, but the machine just is guessing sounds," Rhyan Johnson, the engineer from Wellsaid labs who is involved in the project, stated.
"Eventually, the sounds and patterns combine to form something we recognize as a human voice. We might strive for perfection, but because the human voice is imperfect, we'll settle for human naturalness "Added he.
Five9 claims that its AI agents have handled more than 82 million calls for healthcare providers such as Covid Clinic, big merchants such as Pizza Hut, insurance firms, banks, small enterprises, and state and municipal governments.
Next year, their revolutionary Virtual Voiceover technology will be accessible.