Apple's New Data Center Focus of Nuance Voice Recognition Partnership?


Late last week, we reported on claims that Apple is in discussions with voice recognition firm Nuance Communications regarding some sort of partnership presumably linked to rumors that Apple is integrating significant voice capabilities into iOS 5.

TechCrunch now follows up to report that the deal seems to revolve around Apple utilizing Nuance's technology in its new North Carolina data center to drive centrally-hosted voice services. The partnership is said to likely be introduced at Apple's Worldwide Developers Conference (WWDC) early next month.

In digging into the information about the relationship between the two companies, we had heard that Apple might actually already be using Nuance technology in their new (but yet to be officially opened) massive data center in North Carolina. Since then, we've gotten multiple independent confirmations that this is indeed the case. And yes, this is said to be the keystone of a partnership that Apple is likely to announce with Nuance at WWDC next month.

More specifically, we're hearing that Apple is running Nuance software - and possibly some of their hardware - in this new data center. Why? A few reasons. First, Apple will be able to process this voice information for iOS users faster. Second, it will prevent this data from going through third-party servers. And third, by running it on their own stack, Apple can build on top of the technology, and improve upon it as they see fit.

As was reported previously, Nuance is a leader in voice recognition technology and holds a number of key patents in the field, making Apple's interest in a partnership a natural fit given its acquisition of Siri last year and rumors of the company trying to incorporate aspects of voice recognition and artificial intelligence into its systems.

Top Rated Comments

(View all)

115 months ago
iOS5's killer feature: Real time conversational voice recognition

voice recognition is so useless and dumb, can u imagine all those ppl talking to their gadget on the tram


Short sighted people are always amusing to point at when everybody's adopted the new technology they said was useless and dumb just months prior to it becoming a hit.

Voice recognition hasn't taken off for a couple of reasons:

1 - It's not quite accurate enough so you have to speak at an unnatural speed and tone.

2 - Current technology is mostly one way so you still have to view the screen to communicate properly because you can't yet trust that the system will be correct within an acceptable ratio.

Apple's acquisition of Siri solves both of these.

Imagine having a conversation with your iPhone without thinking in pre-determined commands and being confident that it understands you and will act appropriately.

Instead of:

- Click and hold a button
- Wait for prompt
- "Play Song 'Beat it' "
- Verify that it got it right

- Begin next command...

You'll be able to do this:

"iPhone, play 'Beat it', make an appointment for lunch with Wendy on Tuesday at noon. Oh, also I need to get groceries today. I need potatoes, pasta, steaks and apples and send this to my wife too."

iPhone replies: Playing 'Beat It' by Michael Jackson. I've created an iCal appointment for 12pm this Tuesday for 'Lunch with Wendy'. I've set up a Task named "Groceries" and listed "potatoes, pasta, steaks and apples" and have shared it with Debbie.

This kind of real time interpretation requires a lot of power and constant refinement not possible in a battery powered mobile device. Instead, a small file containing a mono audio snipit can be sent to Apple's servers, interpreted and commands sent back within a few seconds, making this possible.

Look to iOS5 as being the beginning of a true personal assistant that you can speak to but that is also able to speak with you in a natural conversation.

Apple was indeed years ahead of the competition with the introduction of the iPhone and I doubt they've been sitting on their hands waiting for Android to catch up.

They've been working on the next killer feature that will once again push the boundaries and propel Apple way ahead of the pack. This type of conversational speech recognition definitely has the ability to do this.
Rating: 13 Votes
115 months ago
Hello, Computer?

Rating: 9 Votes
115 months ago
If this helps in Phone Sex, I'm in bitch.
Rating: 7 Votes
115 months ago
Based on all the rumors, what DOESN'T the NC data center do?
Rating: 7 Votes
115 months ago
Kirk: "Mr. Spock, full sensor scan of object".

Spock: "Computer, structural analysis, and make up of object to port of Enterprise".

Computer: "Computing".

Computer: "Object appears to contain early 21st century remains of Humanoids."
"There appears to be a digital recording encoded on silicone." "Mossberg, our friend, is no longer writing good things about us."

Spock: Looks at Kirk, and raises 1 eyebrow. :apple:
Rating: 6 Votes
115 months ago

Yeah, I couldn't even imagine people talking to their PHONE in public.... :rolleyes:


:D

Hopefully my phone will only respond to my voice, or if I walk past ipedro on the street my phone might start playing Michael Jackson's Beat It.
Rating: 5 Votes
115 months ago

Uh, are YOU going to give me 2TB of cloud storage? I'm at 1TB, now, haven't got all the video and photos on disk, yet.

Ok, I'll have to stop using the current vr, then. Thanks for the tip. I'll also tell my coworker to delete Dragon from his phone, obviously it couldn't actually be working for him. :rolleyes:


Dragon works over the network. The speech recognition engine is not running on the iPhone. It is just sending audio to the server for processing.

The Voice Control built into iOS is similar to what is running on the Mac. It is not very sophisticated and can't handle dictation or the kinds of things being discussed here.

Nuance is big software - in terms of disk space required, CPU utilization, and memory requirements. Ask anyone who has integrated it with telephony systems (I have). You wouldn't want this running on the iPhone. It would have to be running at all times to be responsive. It would use a lot if not all of the iPhone's resources and it would kill the battery.

Speech recognition is not some web 2.0 mashup that some aspiring hacker can whip up in a few lines of PHP code. It is many decades and billions of dollars of research. It is one of the hardest problems in computer science and Nuance is practically the only player left in the game.
Rating: 5 Votes
115 months ago
How long before TechCrunch reports that there are rumors that Osama bin Laden was hiding in Apple's NC data center?

(too soon?) :rolleyes:
Rating: 5 Votes
115 months ago
Iphone, tell me what the next iphone will have!
Rating: 5 Votes
115 months ago
100% right

Short sighted people are always amusing to point at when everybody's adopted the new technology they said was useless and dumb just months prior to it becoming a hit.

Voice recognition hasn't taken off for a couple of reasons:

1 - It's not quite accurate enough so you have to speak at an unnatural speed and tone.

2 - Current technology is mostly one way so you still have to view the screen to communicate properly because you can't yet trust that the system will be correct within an acceptable ratio.

Apple's acquisition of Siri solves both of these.

Imagine having a conversation with your iPhone without thinking in pre-determined commands and being confident that it understands you and will act appropriately.

Instead of:

- Click and hold a button
- Wait for prompt
- "Play Song 'Beat it' "
- Verify that it got it right

- Begin next command...

You'll be able to do this:

"iPhone, play 'Beat it', make an appointment for lunch with Wendy on Tuesday at noon. Oh, also I need to get groceries today. I need potatoes, pasta, steaks and apples and send this to my wife too."

iPhone replies: Playing 'Beat It' by Michael Jackson. I've created an iCal appointment for 12pm this Tuesday for 'Lunch with Wendy'. I've set up a Task named "Groceries" and listed "potatoes, pasta, steaks and apples" and have shared it with Debbie.

This kind of real time interpretation requires a lot of power and constant refinement not possible in a battery powered mobile device. Instead, a small file containing a mono audio snipit can be sent to Apple's servers, interpreted and commands sent back within a few seconds, making this possible.

Look to iOS5 as being the beginning of a true personal assistant that you can speak to but that is also able to speak with you in a natural conversation.

Apple was indeed years ahead of the competition with the introduction of the iPhone and I doubt they've been sitting on their hands waiting for Android to catch up.

They've been working on the next killer feature that will once again push the boundaries and propel Apple way ahead of the pack. This type of conversational speech recognition definitely has the ability to do this.



100% right
Rating: 4 Votes

[ Read All Comments ]