Here’s a way you can achieve 100% accuracy without touchtone technology by allowing the caller to speak to your IVR. It’s a mix of speech recognition and agent-assisted IVR technology.
If you’ve ever used an interactive voice response (IVR) system – and unless you’ve been living on a desert island, you have – you’ll know that there are some truly bad systems out there. The earliest voice-driven IVRs that used speech recognition often required users to shout into the telephone, repeating themselves until they nearly spontaneously combusted in frustration when the system returned the message, “I didn’t understand your response” for the sixth time.
Luckily, technology has come a long way.
Once upon a time, the number of response combinations that the systems could understand was very limited, which is why it had to restrict your responses (“say ‘one’ for the customer service department” instead of simply asking you to describe what you’re looking for in natural language). Those days are nearly behind us, thanks to newer solutions which offer a conversational natural language solution that allows people to speak to computers as if they were live agents.
Speech Technology Group (www.speechtechnologygroup.com) offers an agent-assisted IVR system, that works in conjunction with the powerful Microsoft speech engine. The combination of the two offers the highest accuracy and extraordinary value.
These types of systems leverage a combination of automated speech recognition (ASR) and “human-assisted understanding”, which improve accuracy and natural-language understanding by supplementing speech recognition when it can’t perform. In traditional speech-recognition applications, all requests get routed directly to an ASR engine. When the engine can’t recognize something, it keeps re-prompting the caller, or eventually gives up and transfers the call to a live agent. This limitation causes poor application design and performance – and frustrates callers. This approach helps you to overcome this application-design limitation.
This is an example of software and human intelligence working together. Rather than relying entirely on software to handle calls, these type of systems automatically hand speech that the ASR can’t cope with over to human agents, who select an appropriate response.
Who would have thought that humans interacting vocally with computers could be a source of anything but a nervous breakdown?