Will Apple’s Siri be like Facetime; widely praised and less-widely used? Or will it be like the App Store or even touch screens that ushered in new ways of interacting with handsets? Folks in the speech recognition and virtual assistant market are hoping it’s the latter. Executives at Vlingo and Nuance both got on the phone today to discuss how they’re reacting to Apple’s moves and what it means for the industry.
The little guy
Vlingo, which introduced a virtual assistant product called Vlingo Virtual Assistant this year, responded to Apple’s integration of its Siri acquisition on Tuesday by making its iOS application free. Prior to this, some tasks, such as checking the weather or asking Vlingo to find nearby restaurants were free, but others such as dictating emails or texts were not. The Siri product does all this and more. Given that 4 million of Vlingo’s 10 million users are on the iOS platform, it stood to lose significantly if it continued to charge what Apple now natively offers for free.
CEO Dave Grannan told me that the entrance of Apple isn’t a bad thing, however. “I think it’s a significant market maker for virtual assistants,” he said. “When Apple does something, it’s generally a tipping point and signals that something is going mainstream.” That’s good for his business. For example, when Google announced voice features for Android, Vlingo added users, Grannan said. However, as impressed as he was by the demonstrations, he also sounded a note of caution.
He wondered how far Apple went with the natural language interface, and if users might end up being too conversational and getting frustrated if the product doesn’t work. He explained that a virtual assistant has to offer two things: one easy and one really hard. The easier technology is basic speech recognition, and the harder bit is adding context to the words spoken and figuring out what the user wants to do. That requires a semantic engine and artificial intelligence that’s continually getting better but is still not perfect.
“If you don’t guide the users between some guard rails for natural language processing, it can cause problems,” Grannan says. “For those reasons we have tended to shy away from extreme natural language conditioning for our users, but that’s a balance [Apple will] have to strike in guiding their users.”
The big fish
Meanwhile, Nuance, a much bigger speech recognition and virtual assistant provider, is also excited about Apple’s potential to influence the adoption of voice as a broadly used user interface on mobile and other consumer devices. Mike Thompson, SVP & GM, mobile division at Nuance, says the company currently has over 100 million requests for mobile speech transactions on its service, and believes Apple’s move shows how speech is a “mainstream interface for mobile phones and mainstream consumer devices.”
Nuance counts Apple as a customer of its software and has worked with Siri in the past, but won’t disclose the details of the relationship it has today with Apple and Siri. But Thompson says, “When Apple puts its stamp of approval and design on something the rest of the world follows, and we expect a lot of innovation coming in deep natural language understanding where the phone takes action almost magically.”
This magic comes at cost in terms of back-end processing and connection to the network. The Siri service will require a connection to the web to not only understand the speech, but also to figure out what steps to take once it knows what was said. Given that most tasks such as asking Siri to tell your wife you are running late, or helping you find the best vegetarian restaurant in Seattle, require broadband, this isn’t really a hardship. Although it does mean if you’re lonely and without a data connection, even Siri won’t talk to you.