Authors: Stephen Baker
Despite the growth of evidence-based medicine, many of these studies are hard to carry out, especially in the splintered $2.3 trillion American system. The doctor prescribes the treatment and the insurance company pays for it, but all too often neither of them gets the most vital feedback: how it worked. The feedback loop, in the language of statisticians, rarely closes. The most promising sites for this type of analysis, Jasinski said, are self-contained hospital networks, which keep voluminous records on patients and do extensive follow-up. They would include the Veterans' Administration, the Mayo Clinic, and Kaiser Permanente in the United States. Many countries with national health care systems also have promising data. Denmark, where IBM has been running the data since 2006, could provide a national laboratory. There, a medical Watson could diagnose diseases, suggest treatments that have proven successful, and steer doctors away from those that have led to problems. Such analyses could save lives, Jasinski said. ”We kill a hundred thousand people a year from preventable medical errors.”
In fact, the potential for next-generation computers in medicine stretches much further. Within a decade, it should cost less than $100 to have an individual's entire genome sequenced. Some people will volunteer to have this done. (Already, companies like 23andMe, a Silicon Valley startup, charge people $429 for a basic decoding.) Others, perhaps, will find themselves pressed, or even compelled, by governments or insurers, to submit their saliva samples. In either case, computers will be studying, correlating, and answering questions about growing collections of this biological information.
At the same time, we're surrounding ourselves with sensors that provide streams of data about our activities. Coronary patients wear blood pressure monitors. Athletes in endurance sports cover themselves with electronics that produce torrents of personal data, reading everything from calorie burn to galvanic skin response, which is associated with stress. Meanwhile, companies are rushing into the market for personal monitoring. Zeo, a Rhode Island company, sells a bedside device that provides a detailed readout every morning of a person's sleeping patterns, including rapid-eye movement, deep sleep, and even trips to the bathroom. Intel is outfitting the homes of elderly test subjects with sensors to measure practically every activity possible, from their nocturnal trips to the bathroom to the words they type on their computers. And each person who carries a cell phone unwittingly provides detailed information on his or her daily movements and routinesâbehavioral data that could prove priceless to medical researchers. Even if some of this data is shielded by privacy rules and withheld from the medical industry, much of it will be available. Machines like Watson will be awash in new and rising rivers of data.
But in the autumn of 2010, as Watson prepared for its culminating
Jeopardy
match, it had yet to land its first hospital job, and its medical abilities remained largely speculative. “We have to be cautious here,” Jasinski said. Though full of potential, Watson was still untested.
It may seem frivolous for the IBM team to have worked as hard as it did to cut down Watson's response time from nearly two hours to three seconds. All of that engineering, and those thousands of processors were harnessed, just to be able to beat humans to a buzzer in a quiz show. Yet as Watson casts about for work, speed will be a crucial factor. Often it takes a company a day or two to make sense of the data it collects. It can seem remarkable, because the data provides a view of sales or operations that was unthinkable even a decade ago. But still, the delay means that today doesn't come into focus until tomorrow or next week. The goal for many businesses now is to process and respond to data in real timeâin the crucial seconds that a quick investment could net $10 million or the right treatment could save a patient's life. Chris Bailey, the chief information officer at SAS, a major producer of analytics software, says the focus is on speed. “Our goal is to make the systems run a thousand or a million times faster,” he said. “That enables us to look at a million times more input.” With this speed, companies increasingly will be able to carry out research, and even run simulations, while the customer is paying for a prescription or withdrawing funds.
Computers with speed and natural language are poised to transform business processes, perhaps entire industries. Compared to what's ahead, even today's state of the art looks sluggish. Consider this snapshot of the data economy, circa 2011: A man walks into a pharmacy to renew his blood pressure medication. He picks up some toiletries while he's there. He hands the cashier his customer loyalty card, which lowers his bill by a dollar or two, and then pays with his Visa card. This shopping data goes straight to Catalina Marketing in St. Petersburg, Florida, which follows the purchases of 190 million shoppers in America. Catalina scans the long list of items that this consumer has bought in the last three years and compares his patterns with those of millions of others. While he's standing at the register, it calculates the items most likely to interest him. Bundled with the receipt the cashier hands him, he finds several couponsâmaybe one for oatmeal, another for a toothpaste in a new upside-down dispenser. If and when he uses them, Catalina learns more about him and targets him with ever-greater precision.
That might sound like a highly sophisticated process. But take a look at how Catalina operates, and you'll see it involves a painfully slow roundtrip, from words to numbers and then back again. “Let's say Kraft Foods has a new mac and cheese with pimentos,” said Eric Williams, Catalina's chief technology officer. The goal is to come up with a target group of potential macaroni eaters, perhaps a million or two, and develop the campaign most likely to appeal to them. The marketers cannot summon this intelligence from their computers. They hand the instructionsâthe ideaâto Catalina's team of seventy statisticians. For perhaps a week, these experts hunt for macaroni prospects in the data. Eventually they produce lists, clusters, and correlations within their target market. But their statistical report is not even close to a marketing campaign. For that, the marketers must translate the statisticians' results back into words and ideas. “Trying to interpret what these people find into common language is quite a feat,” Williams said. Eventually, a campaign takes shape. Catalina concocts about six hundred to eight hundred of them a year. They're effective, often doubling or tripling customer response rates. But each campaign, on average, gobbles up a month of a statistician's work.
Williams's fantasy is to have a new type of computer in his office. Instead of delivering Kraft's order to his statisticians, he would simply explain the goals, in English, to the machine. It would pile through mountains of data in a matter of seconds and come back with details about potential macaroni buyers. The language-savvy machine wouldn't limit its search to traditional data, the neatly organized numbers featuring purchases, dates, and product codes. It might read Twitter or scan social networks to see what people are writing about their appetites and dinner plans. After this analysis, Williams's dream machine could return with a list of ten recent marketing campaigns that have proven the most effective with the target group. “If I don't have to go to statisticians and wait while they run the data, that would be huge,” Williams said. Instead of eight hundred campaigns, Catalina might be able to handle eighty thousand, or even a millionâand offer them at a fraction of today's cost. “You're talking about turning marketing on its head.”
His dream machine, of course, sounds like a version of Watson. Its great potential, in marketing and elsewhere, comes from its ability to automate analysisâto take people, with their time-consuming lunch breaks and vacations, their disagreements and discussions, and drive them right out of the business. The crucial advantage is that Watsonâand machines like itâeliminate the detour into the world of numbers. They understand and analyze words. Machines like thisâspeedy language whizzesâwill open many doors for business. The question for IBM is what Watson's place will be in this drama, assuming it has one.
In the near term, Watson's job prospects are likely to be in call centers. Enhanced with voice recognition software and trained in specific products and services, the computer could respond to phone calls and answer questions. But more challenging jobs, such as bionic marketing consulting, are further off. For each industry, researchers working with consulting teams will have to outfit Watson with an entirely new set of data and run through batteries of tests and training sets. They'll have to fine-tune the machine's judgmentâthe degree of confidence it generates for each responseâand adapt hardware to the job. Will customers want access to mini-Watsons on site or perhaps gain access to a bigger one through a distant data center, a so-called cloud-based service? At this point, no one can say. Jurij Paraszczak, director of Industry Solutions and Emerging Geographies at IBM Research, sees versions of Watson eventually fitting into a number of industries. But such work is hardly around the corner. “Watson's such a baby,” he said.
The history of innovation is littered with technologies that failed because of bad timing or rotten luck. If that $16,000 Xerox computer, with the e-mail and the mouse, had hit the market a decade later, in 1991, cheaper components would have lowered the cost by a factor of five or ten and a more informed public might have appreciated its features.
Technology breakthroughs can also consign even the most brilliant and ambitious projects to museum pieces or junk. In 1825, the first load of cargo floated from Buffalo to Albany on the new Erie Canal. This major engineering work, the most ambitious to date in the Americas, connected the farms and nascent industries of the Great Lakes to the Hudson River, and on to the Atlantic Ocean. It positioned New York State as a vital thoroughfare for commerce, and New York City as the nation's premier port. The news could hardly have been worse for the business and government leaders in New York's neighbor to the south, Pennsylvania, and its international port, Philadelphia. They'd been outmaneuvered. The only way to haul cargo across Pennsylvania was by Conestoga wagon, which often took up to three weeks. So that very year, the Pennsylvanians laid plans to build their own waterway connecting Philadelphia to the Great Lakes. They budgeted $10 million, an immense sum at the time and $3 million more than the cost of the Erie Canal.
The Pennsylvania Canal faced one imposing obstacle: the Alleghenies. These so-called mountains were smaller and rounder than the Rockies or Alps, but they posed a vertical challenge for canal designers. Somehow the boats would have to cross from one side to the other. So with engineering far more ambitious than anything the New Yorkers had attempted, the Pennsylvanians constructed a complex series of iron-railed ramps and tunnels. This horse-powered system would hoist the boats back and forth over the hills and, in a few places, through them. By 1834, boats crossing the state left the water for a thirty-six-mile cross-country trek, or portage, which lifted them up fourteen thousand feet and back down again. For the criss-crossing on ramps, the boats had to be coupled and uncoupled thirty-three times. For those of us not accustomed to Conestoga wagons, this would seem excruciatingly slow, even after stationary steam engines replaced the horses. The novelist Charles Dickens made the crossing in 1842. He set off on a boat and found himself, in short order, peering down a mountain cliff. “Occasionally the rails are laid upon the extreme verge of the giddy precipice and looking down from the carriage window,” he wrote, “the traveler gazes sheer down without a stone or scrap of fence between into the mountain depths below.” Dickens wouldn't soon forget Pennsylvania's $10 million innovation.
This canal project used current technology and heroic engineering to carry out work on a tight deadline. Would newer technology supplant it? It was always possible. After all, the Industrial Revolution was raging in Britain and making inroads into America. Things were changing quickly, which was precisely why Pennsylvania could not afford to wait. As it turned out, within a decade steam-powered trains rendered the canal obsolete. Yet the project laid out the vision and preliminary engineering for the next generation of transport. The train route over the Alleghenies, which followed much of the same path, was considered so vital to national security that Union soldiers kept guard over it during the Civil War. And in 1942, as part of the failed Operation Pastorius, Nazi agents targeted it for sabotage.
Will Watson be the platform for the next stage of computing or, like the Pennsylvania Canal, a bold and startling curiosity to be picked over, its pieces going into cheaper and more efficient technologies down the road? The answer may depend in part on IBM's rivals.
Google is the natural competitor, but it comes at the problem from an entirely different angle. While IBM builds a machine to grapple with one question at a time, Google is serving much of networked humanity. Electricity is a major expense, and even a small increase of wattage per query could put a dent in its profits. This means that even as Google increasingly looks to respond to queries with concrete answers, it can devote to each one only one billionth the processing power of Watson, or less. Peter Norvig, the research director, said that Google's big investments in natural language and machine translation would lead the company toward more sophisticated question-answering for its mass market. As its search engine improves its language skills, he said, it will be able to carry out smarter hunts and make better sense of its users' queries. The danger for IBM isn't head-to-head competition from Google and other search engines. But as searching comes closer to providing answers to queries in English, a number of tech startups and consultants will be able to jury-rig competing question-answering machines, much the way James Fan built his Basement Baseline at the dawn of the
Jeopardy
project.