Only a few years ago, you thought that guy walking down the street apparently talking to himself was off his meds. Now, you’re rocking your Bluetooth headset every day without even thinking about it (even if you still annoy some of us on the train).
But that’s talking with other people, for pete’s sake–are you still phoning with your phone in 2015? Today, you can ask it to do almost anything just by speaking “OK Google” or “Hey Siri”: conduct a search, make a restaurant reservation, send a text, or do almost anything you used to have to type into a search box or tap into an app.
You probably already knew you could try doing all that, but here’s what you may not know: Most of it doesn’t suck anymore. If you haven’t tried Google Voice Search, Apple’s Siri, Microsoft’s Cortana, or even Amazon.com’s Echo “smart” speaker recently, you may be surprised how much better they’re working than even six months ago. Not only do they seem to understand words better, even in noisy situations, they also appear to produce more accurate results in many cases.
All that’s thanks to big improvements in machine learning, in particular a branch of artificial intelligence called deep learning, that’s been applied to speech recognition in the last couple of years. “Recent AI breakthroughs have cracked the code on voice, which is approaching 90% as good as human accuracy,” says Tim Tuttle, CEO of Expect Labs, which began offering its MindMeld cloud-based service last year to help any device or app create voice interfaces.
It’s great for us smartphone owners, but the stakes couldn’t be higher for companies. …