I watched the future happen right before my eyes on Christmas morning.
My 17-year-old nephew Walton got an iPad for Christmas. I had shown him Google Voice Search previously, so he downloaded the app immediately. For the rest of the day, whenever some question arose, he asked Google on the iPad by speaking out loud in regular language. In nearly every case, he got the answer he was looking for, and pretty fast, too.
The generation older than me got information from computers by programming queries on cards or tape, submitting their "job" to the mainframe priesthood, then coming back later for the answer.
My generation learned as adults to craft Boolean search queries for search engines, which gave us not so much answers as a very long list of guesses. We had to sift through search results and cherry-pick which one of the many returned links might satisfy our curiosity.
Walton's generation is the first for which getting answers from computers won't involve "submitting" anything or slogging through possible answers. They'll simply have a conversation with their phone. They won't type anything. They'll talk. And the phones will simply interact with them or take dictation.
More than that, phones are growing human-like personalities, including the ability to understand and speak natural language--and even empathize.
The technology is interesting enough, but even more interesting is how this will affect the human mind.
When mobile phones first became ubiquitous 15 years ago or so, they were objects or machines psychologically separate and distinct from us. They were "tools" that we "used."
As smartphones became popular, got ever smarter and became ever more central to our everyday lives, phones stopped being psychologically separate from us. They became part of us -- serving as prosthetic memory. They also gave us a sixth sense, superhuman abilities, such as the ability to communicate with people far away.
In fact, that's where phones fit into our lives today -- they're part of us. We feel naked and incomplete without them. When we don't have phones in our pockets, we still hear them ring and feel them vibrate, like an amputee still feels his missing leg.
But the evolution of phones will separate them from us again. But this time, instead of being separate "tools" that we use, they'll be "people" that we know.
Apple, for example, is making huge strides with its Siri personal assistant. Siri is slow and limited, but it represents a great first step toward an interface that lets you talk in natural language, and get answers in a way that resembles personality. Siri jokes around, mixes up various ways to phrase responses and generally simulates human interaction with the user, to some limited degree.
Google is doing amazing work in the development of its Google Now feature in Android Jelly Bean and higher. It is backed by a foundational project called the Google Knowledge Graph. Google Now learns about you, and takes the initiative to suggest and inform. You interact with Google Now by talking, much as you might with Siri.
These two examples will be viewed in hindsight as first steps toward the ubiquitous human-like smartphone, which understands, talks back, learns, grows and shoots the breeze with you like a friend.
Of course, these same capabilities, which use compute power far away in remote data centers, will be available on tablets, laptops and desktop computers. But since voice is the main interface, we’ll usually interact via phone.
Interestingly, the quality that will make our minds buy into the illusion of human-like personality is something that Apple and Google haven't been able to simulate yet: emotion.