John Markoff

In the end, the humans on Jeopardy! surrendered meekly. Facing certain defeat at the hands of a room-size IBM computer last week, Ken Jennings, famous for winning 74 games in a row on the TV quiz show, acknowledged the obvious. ?I, for one, welcome our new computer overlords,? he wrote on his video screen, borrowing a line from a Simpsons episode.

From now on, if the answer is ?the computer champion on Jeopardy!, the question will be, ?What is Watson?? For IBM, the showdown was not merely a well-publicised stunt and a $1 million prize, but proof that the company has taken a big step toward a world in which intelligent machines will understand and respond to humans, and perhaps inevitably, replace some of them.

Watson, specifically, is a ?question answering machine? of a type that artificial intelligence researchers have struggled with for decades?a computer akin to the one on Star Trek that can understand questions posed in natural language and answer them.

Watson showed itself to be imperfect, but researchers at IBM and other companies are already developing uses for Watson?s technologies that could have significant impact on the way doctors practice and consumers buy products.

?Cast your mind back 20 years and who would have thought this was possible?? said Edward Feigenbaum, a Stanford University computer scientist and a pioneer in the field.

In its Jeopardy! project, IBM researchers were tackling a game that requires not only encyclopedic recall, but the ability to untangle convoluted and often opaque statements, a modicum of luck, and quick, strategic button pressing.

The contest, which was taped in January at the company?s T J Watson Research Laboratory before an audience of IBM executives and company clients, played out in three televised episodes. At the end of the first day, Watson was in a tie with Brad Rutter, another ace human player, at $5,000 each, with Jennings trailing with $2,000. But on the second day, Watson went on a tear. By night?s end, Watson had a commanding lead with a total of $35,734, compared with Rutter?s $10,400 and Jennings? $4,800.

But victory was not cemented until late in the third match, when Watson was in Nonfiction. ?Same category for $1,200? it said in a manufactured tenor, and lucked into a Daily Double. Jennings grimaced. Even later in the match, however, had Jennings won another key Daily Double it might have come down to Final Jeopardy, IBM researchers acknowledged. The final tally was $77,147 to Jennings? $24,000 and Rutter?s $21,600.

More than anything, the contest was a vindication for the academic field of computer science, which began with great promise in the 1960s with the vision of creating a thinking machine and which became the laughingstock of Silicon Valley in the 1980s, when a series of heavily funded start-up companies went bankrupt.

Despite its intellectual prowess, Watson was by no means omniscient. During Final Jeopardy, the category was US Cities and the clue was: ?Its largest airport is named for a WW II hero; its second largest for a WW II battle.? Watson drew guffaws from many in the audience when it responded ?What is Toronto??????

The string of question marks indicated that the system had very low confidence in its response, IBM researchers said, but because it was Final Jeopardy, it was forced to give a response. The machine did not suffer much damage. It had wagered just $947 on its result.

?We failed to deeply understand what was going on there,? said David Ferrucci, an IBM researcher who led the development of Watson. ?The reality is that there?s lots of data where the title is US cities and the answers are countries, European cities, people, mayors. Even though it says US cities, we had little confidence that that?s the distinguishing feature.?

The researchers also acknowledged that the machine had benefited from the ?buzzer factor.?

Both Jennings and Rutter are accomplished at anticipating the light that signals it is possible to ?buzz in,? and can sometimes get in with virtually zero lag time. The danger is to buzz too early, in which case the contestant is penalised and ?locked out? for roughly a quarter of a second.

Watson, on the other hand, does not anticipate the light, but has a weighted scheme that allows it, when it is highly confident, to hit the buzzer in as little as 10 milliseconds, making it very hard for humans to beat. When it was less confident, it took longer to buzz in. In the second round, Watson beat the others to the buzzer in 24 out of 30 Double Jeopardy questions.

?It sort of wants to get beaten when it doesn?t have high confidence,? Dr Ferrucci said. ?It doesn?t want to look stupid.? Both human players said that Watson?s button pushing skill was not necessarily an unfair advantage. ?I beat Watson a couple of times,? Rutter said.

When Watson did buzz in, it made the most of it. Showing the ability to parse language, it responded to, ?A recent best seller by Muriel Barbery is called This of the Hedgehog,? with ?What is Elegance??

It showed its facility with medical diagnosis. With the answer: ?You just need a nap. You don?t have this sleep disorder that can make sufferers nod off while standing up,? Watson replied, ?What is narcolepsy??

The coup de gr?ce came with the answer, ?William Wilkenson?s ?An Account of the Principalities of Wallachia and Moldavia? inspired this author?s most famous novel.? Jennings wrote, correctly, Bram Stoker, but realised that he could not catch up with Watson?s winnings and wrote out his surrender.

Both players took the contest and its outcome philosophically. ?I had a great time and I would do it again in a heartbeat,? said Jennings. ?It?s not about the results; this is about being part of the future.?

Read Next