In other words, today’s machines are sophisticated enough to glean hidden insights from complex imagery—perspective that would otherwise evade even the most experienced human. With the right training, machines are able to show human doctors things they cannot see. “For example, what’s the pattern of a certain infection caused by a certain bacteria?” Kanevsky says. Such a pattern might be so subtle that it can’t be identified by a human, yet it may be discernible to a machine that’s drawing from a large enough dataset.
Kanevsky gives another example, recalling a recent patient who was injured in a propane-tank explosion. “Burns are tricky,” he says, and for a number of reasons. One of the first things he and his colleagues had to do in response was to estimate the extent of the person’s injuries. One of the traditional ways to do that in plastic surgery is to assume that 1 percent of a person’s body is equivalent to the surface area of the palm of that person’s hand. It isn’t precise or even necessarily accurate—“We use a very crude estimation,” Kanevsky says—but it’s important for doctors to guess. Knowing how much of a person’s body is burned is a predictor for mortality, and helps doctors make critical decisions about the best course of treatment; including assessing how much IV fluid is necessary in the initial days after the injury and determining what kind of surgical response is appropriate. Algorithms can already determine how deep a burn is—and accurately predict how long it will take for a burn to heal.
Already, “previously unimaginable” applications for machine learning are “within grasp” for individual patient care, according to a February essay about machine learning in the Journal of the American Medical Association. There are at least dozens of startups specializing in artificial intelligence with a focus on health-care. Yet there is still much work to be done before clinicians can turn over their clinics to machines. (In other words, docs, don’t worry about being replaced by bots just yet.) It takes time to teach a machine. Supervised learning requires giving a computer feedback, confirming when it’s right, tagging and cataloguing images, and “training the algorithm so it can say, ‘This is a wound that looks like it will heal,’” as Kanevsky puts it. With enough attention and pristine data, a machine’s way of seeing can quickly supersede human ability. For example, the deep-learning startup Enlitic boasts an algorithm that’s 50 percent more accurate than human radiologists at detecting lung cancer, according to the company.
In other settings, machines have proved to be at least as good as human doctors at crucial tasks. “In many applications, the performance of the machine learning-based systems is comparable to that of experienced radiologists,” wrote the authors of a 2012 paper published in the journal of Medical Image Analysis. (And that was four years ago, practically an eternity by AI standards.)