On Tuesday, Apple will change the way humans interact with electronic devices. All over again.
Perhaps the biggest announcement at Apple’s iPhone event on Tuesday will be Assistant, Apple’s evolution of the Siri Personal Assistant Software. Siri, you’ll remember, is the company Apple picked up for a rumored $200 million in April of last year for, in Steve Jobs’ words, its “Artificial Intelligence”, not search or speech recognition.
During Siri’s brief two months on its own, it described itself as a ‘VPA’:
Virtual Personal Assistants (VPAs) represent the next generation interaction paradigm for the Internet. In today’s paradigm, we follow links on search results. With a VPA, we interact by having a conversation. We tell the assistant what we want to do, and it applies multiple services and information sources to help accomplish our task. Like a real assistant, a VPA is personal; it uses information about an individual’s preferences and interaction history to help solve specific tasks, and it gets better with experience.
Apple has long wanted to bring an Artificial Intelligence-based Personal Assistant to the masses. In the late 80’s, Apple made the Knowledge Navigator series of videos (example below) to showcase this ambition.
In the video, the professor mentions that someone wrote an article 5 years ago trashing Jill’s research (watch from 1:25 min onwards, at 1:50 min he mentions more details) – The computer says the doctor’s name and says his article in 2006 – which means the professor is in 2011. Ha! Thanks PBHK!
The world has come a long way since then, but as you’ll see on Tuesday, Apple had remarkable foresight way back in 1987.
We had the chance to speak to Siri’s co-founder and board member, Norman Winarsky…
First, Some Background:
The device input methods we’ve used in the past have first been the keyboard, then the mouse and more recently, there is the touch interface. All of these methods, while not invented by Apple, were “mainstreamed” by Steve Jobs’ company over the last thirty-five years.
But humans didn’t evolve to communicate with keyboards or mice or even a touch screen. We’ve contorted our bodies to deal with our computer tools (Hi RSI!) but really we’re hard-wired for talking and listening – functions we, as a species, have been doing for tens of thousands of years.
Unfortunately, we haven’t yet invented a computer that can understand what we say, and more importantly, use that information to go find answers and relay that information back to us. That would require not only recognition of the language but the Artificial Intelligence to understand it, use it, and return something of value.
Well, that’s not entirely true. In 2003, the US Government began the most ambitious Artificial Intelligence program in its history called the “Cognitive Assistant that Learns and Organizes” or CALO program. The name was inspired by the Latin word “calonis”, which means “soldier’s servant”. Funded by DARPA as part of its Personal Assistant that Learns project, the program ran for five years and brought together more than 300 researchers from 25 of the top university and commercial research institutions, with the goal of “building a new generation of cognitive assistants that can reason, learn from experience, be told what to do, explain what they are doing, reflect on their experience, and respond robustly to surprise.”
The program was coordinated through SRI International in Menlo Park, CA. As the program ended in 2007, SRI took the knowledge gained by the CALO and some of its key players and formed Siri. SRI’s Norman Winarsky, the man uniquely positioned at the crossroads of the CALO project and the company spun off out of it talked to us about the implications of Apple mainstreaming ‘Assistant’.
9to5Mac: What was your role in putting together Siri?
Norm: As CALO was coming to an end, we realized that there were incredible commercial opportunities to build a smart personal assistant from what we learned over the five years of the CALO project. My job was getting funding (VC’s were Morgenthaler and Menlo Ventures) and assembling the team headed by Dag Kittlaus a former Motorola Executive. With him came Semantic Web genius Tom Gruber and Chief Architect of CALO Adam Cheyer. At the time of Apple purchase, the team was at 19 and growing. All three co-founders still work at Apple with much of the rest of the original team. I obviously stayed at SRI after the purchase.
9to5Mac: Can you tell us a little bit about getting picked up by Apple? What was the process? How did they evaluate the company? Are any of the financials available?
Norm: I am bound by non-disclosure on all of the information from the sale that is not public including the [rumored $200 million] sale price. What is notable is that Apple closed its purchase of Siri just two months after we went public with our app. You can probably draw your own conclusions from that.
9to5Mac: How important is Nuance speech recognition to the Siri technology?
Norm: It is a lot less important than you’d probably think. When we first built Siri, we use Vlingo for speech recognition and as such, at the time of purchase the speech recognition component is modular. Theoretically, if a better speech recognition comes along (or Apple buys one), they could likely replace Nuance without too much trouble. That being said, Nuance has far and away the most IP in speech synthesis technologies in the industry. We should know, SRI launched Nuance as one of our incubated companies in 1995 and it IPO’d in 2000.
9to5Mac: What kind of power does the Siri AI take? Could it have caused the delay of the next iPhone?
Norm: I’m not familiar with Apple’s roadmap and any delays but I can say that AI takes a lot of computing power. The Siri software needs to cache data, needs to access a big dataset at wide bandwidth and needs a big processor to crunch all of the numbers. When we originally released Siri for the iPhone 3GS, we had to perform all kinds of optimizations and shortcuts to get it to work efficiently. All I can say is that it will likely run much better on a faster phone.
9to5Mac: Is this Siri ‘Assistant’ a big deal?
Norm: Let me first say I have no knowledge of what Apple plans to do with the Siri purchase. I read the rumors just like everyone else and it appears that Apple is getting ready to reveal what it has done with Siri over the past year and a half (we were actually expecting it at WWDC). Make no mistake: Apple’s ‘mainstreaming’ Artificial Intelligence in the form of a Virtual Personal Assistant is a groundbreaking event. I’d go so far as to say it is a World-Changing event. Right now a few people dabble in partial AI enabled apps like Google Voice Actions, Vlingo or Nuance Go. Siri was many iterations ahead of these technologies, or at least it was two years ago. This is REAL AI with REAL market use. If the rumors are true, Apple will enable millions upon millions of people to interact with machines with natural language. The PAL will get things done and this is only the tip of the iceberg. We’re talking another technology revolution. A new computing paradigm shift.
It reminds me of another SRI Project: Doug Engelbart, Inventor of Mouse augmented human ability back in the ’60s. Just as Steve Jobs took that technology and ran with it, we believe that Apple will use Siri to start another revolution.
9to5Mac: Thanks for your time Norm. This reminds us of the Steve Jobs computer is a bicycle for the mind quote:
Some other interesting data on the founders:
The three founders are all still at Apple, though they work on other projects. We found Dag Kittlaus’ comments that he’s now making “the next big thing into a really big thing” interesting (below).
Adam Cheyer demonstrated (PDF) the CALO Express application in 2007, just before starting Siri. The application ran on Windows CE because it was aimed at government use.
Perhaps Tom Gruber at Semantic Web in 2008 just before Siri went public was most interesting:
Come back to 9to5mac.com Tuesday at 10am Pacific for all of the announcements.
- iOS 5 Assistant based on Siri technology, may tap into contacts, calendars, e-mail (9to5mac.com)
- iOS 5′s Siri-like system navigation is called ‘Assistant’, uses device-info to handle actions (9to5mac.com)
- Taking a page from Google, Apple’s Siri-based voice Assistant activates when you put the iPhone to your ear (9to5mac.com)