Final Jeopardy

Home > Other > Final Jeopardy > Page 19
Final Jeopardy Page 19

by Stephen Baker


  Medicine was one of the most promising areas but also among the toughest to crack. The natural job for Watson would be as a diagnostic aid, taking down the symptoms in cases like Ferrucci’s and producing lists of possible conditions, along with recommended treatments. Already, many doctors facing puzzling symptoms were consulting software tools known as medical decision trees, which guided them toward the most likely diagnoses and recommended treatments. Some were available as applications on smart phones. A medical Watson, though, would plunge into a much deeper pool of data, much of it unstructured. Conceivably, it would come up with hidden linkages. But even that job, according to Robert Wachter, the chief of hospital medicine at the University of California, San Francisco, was bound to raise serious questions. “Doctors like the idea of having information available,” he said. “Where things get more psychologically fraught is when a damned machine tells them what to do.” What’s more, once analysis is automated, he said, the recommendation algorithm is likely to include business analysis. In other words, the medical Watsons might come back not with the statistically most effective treatment but the most cost-effective one. Even if this didn’t happen, many would remain suspicious. And what if Watson had sky-high confidence in a certain diagnosis—say, 97 percent? Would doctors get in trouble if they turned a deaf ear to it? Would they face lawsuits if they ignored the advice and it later turned out the machine was right?

  Then, of course, there was the possibility of disastrous mistakes resulting from a computer’s suggestions. Even if a bionic assistant scrupulously labeled all of its findings as hypotheses, some of them—just like Watson’s answers in Jeopardy—were bound to be nutty, generating ridicule and distrust. Others, perhaps more dangerous, would be wrong while appearing plausible. If a treatment recommended by a machine killed a patient, confidence in bionic assistants could plummet.

  The other issue, sure to come up in many industries, boils down to a struggle for power, and even survival, in the workplace. “As every profession embraces systems that take humans out of it,” Wachter said, “the profession gets commoditized.” He noted the example of commercial aviation, where pilots who were once considered stars have ended up spending much of the time in flight simply backing up the machines that are actually flying the planes. The result? “Pilots’ pensions have been cut and they’re paid less, because they’re largely interchangeable,” he said. “Doctors don’t want to see that happening to them.”

  For IBM, this very scenario promises growth. With more than $4 billion in annual revenue, the health care practice within IBM Global Services has the size of a Fortune 500 company. It runs large data centers for hospitals and insurance companies. It also helps them analyze the data, looking for patterns of symptoms, treatments, and diseases—as well as ways to cut costs. This is part of a trend toward statistical analysis in the industry and the rapid growth of so-called evidence-based medicine. But one of the most valuable streams of data—the doctor’s notes—rarely makes it into the picture, said Joseph Jasinski, who heads research for IBM’s health care division. This is where the doctor writes down what he or she sees and thinks. Sometimes it is stored in a computer, but only, Jasinski said, “as a blob of text.” In other words, it’s unstructured data, Watson’s forte. “There’s a strong belief in the community that if you could study clinical notes, you could analyze patient similarities,” he said. Neurologists’ notes—going back to Ferrucci’s case—could have pointed to common symptoms between patients with the suicide disease and others with knots in a muscle just below their shoulder blade. This analysis could expand, comparing symptoms and treatments, and later study the outcomes. What works? What falls flat? Which procedures appear to waste money?

  Despite the growth of evidence-based medicine, many of these studies are hard to carry out, especially in the splintered $2.3 trillion American system. The doctor prescribes the treatment and the insurance company pays for it, but all too often neither of them gets the most vital feedback: how it worked. The feedback loop, in the language of statisticians, rarely closes. The most promising sites for this type of analysis, Jasinski said, are self-contained hospital networks, which keep voluminous records on patients and do extensive follow-up. They would include the Veterans’ Administration, the Mayo Clinic, and Kaiser Permanente in the United States. Many countries with national health care systems also have promising data. Denmark, where IBM has been running the data since 2006, could provide a national laboratory. There, a medical Watson could diagnose diseases, suggest treatments that have proven successful, and steer doctors away from those that have led to problems. Such analyses could save lives, Jasinski said. ”We kill a hundred thousand people a year from preventable medical errors.”

  In fact, the potential for next-generation computers in medicine stretches much further. Within a decade, it should cost less than $100 to have an individual’s entire genome sequenced. Some people will volunteer to have this done. (Already, companies like 23andMe, a Silicon Valley startup, charge people $429 for a basic decoding.) Others, perhaps, will find themselves pressed, or even compelled, by governments or insurers, to submit their saliva samples. In either case, computers will be studying, correlating, and answering questions about growing collections of this biological information.

  At the same time, we’re surrounding ourselves with sensors that provide streams of data about our activities. Coronary patients wear blood pressure monitors. Athletes in endurance sports cover themselves with electronics that produce torrents of personal data, reading everything from calorie burn to galvanic skin response, which is associated with stress. Meanwhile, companies are rushing into the market for personal monitoring. Zeo, a Rhode Island company, sells a bedside device that provides a detailed readout every morning of a person’s sleeping patterns, including rapid-eye movement, deep sleep, and even trips to the bathroom. Intel is outfitting the homes of elderly test subjects with sensors to measure practically every activity possible, from their nocturnal trips to the bathroom to the words they type on their computers. And each person who carries a cell phone unwittingly provides detailed information on his or her daily movements and routines—behavioral data that could prove priceless to medical researchers. Even if some of this data is shielded by privacy rules and withheld from the medical industry, much of it will be available. Machines like Watson will be awash in new and rising rivers of data.

  But in the autumn of 2010, as Watson prepared for its culminating Jeopardy match, it had yet to land its first hospital job, and its medical abilities remained largely speculative. “We have to be cautious here,” Jasinski said. Though full of potential, Watson was still untested.

  It may seem frivolous for the IBM team to have worked as hard as it did to cut down Watson’s response time from nearly two hours to three seconds. All of that engineering, and those thousands of processors were harnessed, just to be able to beat humans to a buzzer in a quiz show. Yet as Watson casts about for work, speed will be a crucial factor. Often it takes a company a day or two to make sense of the data it collects. It can seem remarkable, because the data provides a view of sales or operations that was unthinkable even a decade ago. But still, the delay means that today doesn’t come into focus until tomorrow or next week. The goal for many businesses now is to process and respond to data in real time—in the crucial seconds that a quick investment could net $10 million or the right treatment could save a patient’s life. Chris Bailey, director of the Advanced Computing Lab at SAS, a major producer of analytics software, says the focus is on speed. “Our goal is to make the systems run a thousand or a million times faster,” he said. “That enables us to look at a million times more input.” With this speed, companies increasingly will be able to carry out research, and even run simulations, while the customer is paying for a prescription or withdrawing funds.

  Computers with speed and natural language are poised to transform business processes, perhaps entire industries. Compared to what’s ahead, even today’s state of the art looks slugg
ish. Consider this snapshot of the data economy, circa 2011: A man walks into a pharmacy to renew his blood pressure medication. He picks up some toiletries while he’s there. He hands the cashier his customer loyalty card, which lowers his bill by a dollar or two, and then pays with his Visa card. This shopping data goes straight to Catalina Marketing in St. Petersburg, Florida, which follows the purchases of 190 million shoppers in America. Catalina scans the long list of items that this consumer has bought in the last three years and compares his patterns with those of millions of others. While he’s standing at the register, it calculates the items most likely to interest him. Bundled with the receipt the cashier hands him, he finds several coupons—maybe one for oatmeal, another for a toothpaste in a new upside-down dispenser. If and when he uses them, Catalina learns more about him and targets him with ever-greater precision.

  That might sound like a highly sophisticated process. But take a look at how Catalina operates, and you’ll see it involves a painfully slow roundtrip, from words to numbers and then back again. “Let’s say Kraft Foods has a new mac and cheese with pimentos,” said Eric Williams, Catalina’s chief technology officer. The goal is to come up with a target group of potential macaroni eaters, perhaps a million or two, and develop the campaign most likely to appeal to them. The marketers cannot summon this intelligence from their computers. They hand the instructions—the idea—to Catalina’s team of seventy statisticians. For perhaps a week, these experts hunt for macaroni prospects in the data. Eventually they produce lists, clusters, and correlations within their target market. But their statistical report is not even close to a marketing campaign. For that, the marketers must translate the statisticians’ results back into words and ideas. “Trying to interpret what these people find into common language is quite a feat,” Williams said. Eventually, a campaign takes shape. Catalina concocts about six hundred to eight hundred of them a year. They’re effective, often doubling or tripling customer response rates. But each campaign, on average, gobbles up a month of a statistician’s work.

  Williams’s fantasy is to have a new type of computer in his office. Instead of delivering Kraft’s order to his statisticians, he would simply explain the goals, in English, to the machine. It would pile through mountains of data in a matter of seconds and come back with details about potential macaroni buyers. The language-savvy machine wouldn’t limit its search to traditional data, the neatly organized numbers featuring purchases, dates, and product codes. It might read Twitter or scan social networks to see what people are writing about their appetites and dinner plans. After this analysis, Williams’s dream machine could return with a list of ten recent marketing campaigns that have proven the most effective with the target group. “If I don’t have to go to statisticians and wait while they run the data, that would be huge,” Williams said. Instead of eight hundred campaigns, Catalina might be able to handle eighty thousand, or even a million—and offer them at a fraction of today’s cost. “You’re talking about turning marketing on its head.”

  His dream machine, of course, sounds like a version of Watson. Its great potential, in marketing and elsewhere, comes from its ability to automate analysis—to take people, with their time-consuming lunch breaks and vacations, their disagreements and discussions, and drive them right out of the business. The crucial advantage is that Watson—and machines like it—eliminate the detour into the world of numbers. They understand and analyze words. Machines like this—speedy language whizzes—will open many doors for business. The question for IBM is what Watson’s place will be in this drama, assuming it has one.

  In the near term, Watson’s job prospects are likely to be in call centers. Enhanced with voice recognition software and trained in specific products and services, the computer could respond to phone calls and answer questions. But more challenging jobs, such as bionic marketing consulting, are further off. For each industry, researchers working with consulting teams will have to outfit Watson with an entirely new set of data and run through batteries of tests and training sets. They’ll have to fine-tune the machine’s judgment—the degree of confidence it generates for each response—and adapt hardware to the job. Will customers want access to mini-Watsons on site or perhaps gain access to a bigger one through a distant data center, a so-called cloud-based service? At this point, no one can say. Jurij Paraszczak, director of Industry Solutions and Emerging Geographies at IBM Research, sees versions of Watson eventually fitting into a number of industries. But such work is hardly around the corner. “Watson’s such a baby,” he said.

  The history of innovation is littered with technologies that failed because of bad timing or rotten luck. If that $16,000 Xerox computer, with the e-mail and the mouse, had hit the market a decade later, in 1991, cheaper components would have lowered the cost by a factor of five or ten and a more informed public might have appreciated its features.

  Technology breakthroughs can also consign even the most brilliant and ambitious projects to museum pieces or junk. In 1825, the first load of cargo floated from Buffalo to Albany on the new Erie Canal. This major engineering work, the most ambitious to date in the Americas, connected the farms and nascent industries of the Great Lakes to the Hudson River, and on to the Atlantic Ocean. It positioned New York State as a vital thoroughfare for commerce, and New York City as the nation’s premier port. The news could hardly have been worse for the business and government leaders in New York’s neighbor to the south, Pennsylvania, and its international port, Philadelphia. They’d been outmaneuvered. The only way to haul cargo across Pennsylvania was by Conestoga wagon, which often took up to three weeks. So that very year, the Pennsylvanians laid plans to build their own waterway connecting Philadelphia to the Great Lakes. They budgeted $10 million, an immense sum at the time and $3 million more than the cost of the Erie Canal.

  The Pennsylvania Canal faced one imposing obstacle: the Alleghenies. These so-called mountains were smaller and rounder than the Rockies or Alps, but they posed a vertical challenge for canal designers. Somehow the boats would have to cross from one side to the other. So with engineering far more ambitious than anything the New Yorkers had attempted, the Pennsylvanians constructed a complex series of iron-railed ramps and tunnels. This horse-powered system would hoist the boats back and forth over the hills and, in a few places, through them. By 1834, boats crossing the state left the water for a thirty-six-mile cross-country trek, or portage, which lifted them up fourteen thousand feet and back down again. For the criss-crossing on ramps, the boats had to be coupled and uncoupled thirty-three times. For those of us not accustomed to Conestoga wagons, this would seem excruciatingly slow, even after stationary steam engines replaced the horses. The novelist Charles Dickens made the crossing in 1842. He set off on a boat and found himself, in short order, peering down a mountain cliff. “Occasionally the rails are laid upon the extreme verge of the giddy precipice and looking down from the carriage window,” he wrote, “the traveler gazes sheer down without a stone or scrap of fence between into the mountain depths below.” Dickens wouldn’t soon forget Pennsylvania’s $10 million innovation.

  This canal project used current technology and heroic engineering to carry out work on a tight deadline. Would newer technology supplant it? It was always possible. After all, the Industrial Revolution was raging in Britain and making inroads into America. Things were changing quickly, which was precisely why Pennsylvania could not afford to wait. As it turned out, within a decade steam-powered trains rendered the canal obsolete. Yet the project laid out the vision and preliminary engineering for the next generation of transport. The train route over the Alleghenies, which followed much of the same path, was considered so vital to national security that Union soldiers kept guard over it during the Civil War. And in 1942, as part of the failed Operation Pastorius, Nazi agents targeted it for sabotage.

  Will Watson be the platform for the next stage of computing or, like the Pennsylvania Canal, a bold and startling curiosity to be picked over, its pieces going into cheap
er and more efficient technologies down the road? The answer may depend in part on IBM’s rivals.

  Google is the natural competitor, but it comes at the problem from an entirely different angle. While IBM builds a machine to grapple with one question at a time, Google is serving much of networked humanity. Electricity is a major expense, and even a small increase of wattage per query could put a dent in its profits. This means that even as Google increasingly looks to respond to queries with concrete answers, it can devote to each one only one billionth the processing power of Watson, or less. Peter Norvig, the research director, said that Google’s big investments in natural language and machine translation would lead the company toward more sophisticated question-answering for its mass market. As its search engine improves its language skills, he said, it will be able to carry out smarter hunts and make better sense of its users’ queries. The danger for IBM isn’t head-to-head competition from Google and other search engines. But as searching comes closer to providing answers to queries in English, a number of tech startups and consultants will be able to jury-rig competing question-answering machines, much the way James Fan built his Basement Baseline at the dawn of the Jeopardy project.

  As Google evolves, Norvig said, it will start to replicate some of Watson’s headier maneuvers, combining data from different sources. “If someone wants per capita income in a certain country, or in a list of countries, we might bring two tables together,” he said. For that, though, the company might require more detailed queries. “It might get to the point where we ask users to elaborate, and to write entire sentences,” he said. In effect, the computer will be demanding something closer to a Jeopardy clue—albeit with fewer puns and riddles.

  This would represent a turnaround. For more than a decade, the world’s Web surfers have learned to hone their queries. In effect, they’ve used their human smarts to reverse-engineer Google’s algorithms—and to understand how a search engine “thinks.” Each word summons a universe of connections. Looking at each one like a circle in a Venn diagram, the goal is to organize three or four words—3.5 is the global average—whose circles have the smallest possible overlap. For many, this analysis has become almost reflexive. Yet as the computer gets smarter, these sophisticated users stand to get poorer results than those who type long sentences, even paragraphs, and treat the computer as if it were human.

 

‹ Prev