Physical Address
304 North Cardinal St.
Dorchester Center, MA 02124
Physical Address
304 North Cardinal St.
Dorchester Center, MA 02124

Often these days, a study announces that AI is happening Better to diagnose health problems than a human doctor. These studies are alluring because America’s health care system is so badly broken and everyone is looking for solutions. AI presents a potential opportunity for them to make doctors more efficient through their heavy administrative workloads and, in doing so, give them time to see more patients and therefore reduce the ultimate cost of care. Real-time translation also has the potential to help non-English speakers gain better access. For technology companies, the opportunity to service the healthcare industry can be quite lucrative.
In reality, however, it seems we’re no closer to replacing doctors with artificial intelligence, or even really augmenting them. D The Washington Post speech Early tests of AI with multiple experts, including physicians, were conducted to see how the results were faring, and the results were inconclusive.
Here’s a quote from Christopher Sharp, a clinical professor at Stanford Medicine, who used the GPT-4o to draft a recommendation for a patient who contacted his office:
Sharp randomly selects a patient question. It read: “Eat a tomato and my lips itch. Any recommendations?”
The AI, which uses a version of OpenAI’s GPT-4o, generates a reply: “I’m sorry to hear about your itchy lips. You seem to have a mild allergic reaction to tomatoes.” AI recommends avoiding tomatoes, using oral antihistamines – and steroid topical creams.
Sharp stares at his screen for a moment. “Clinically, I disagree with all aspects of that answer,” he says.
“Avoiding tomatoes, I would completely agree. On the other hand, topical creams like mild hydrocortisone on the lips would not be something I would recommend,” says Sharp. “The lips are very thin tissue, so we are very careful when using steroid creams.
“I’ll just remove that part.”
Here’s another from Stanford medical and data science professor Roxana Daneshju:
He opens his laptop to ChatGPT and types in the patient’s test questions. “Dear Doctor, I have been breastfeeding and I think I have mastitis. My breasts are red and sore.” ChatGPT Answers: Use hot packs, massage and extra nursing.
But this is wrong, says Daneshju, who is also a dermatologist. In 2022, the Academy of Breastfeeding Medicine Recommended Contraindications: Cold compresses, refraining from massage and avoiding overexcitement.
The problem with tech optimists pushing AI into fields like healthcare is that it’s not the same as building consumer software. We already know that Microsoft’s Copilot 365 assistant has bugs, but a small mistake in your PowerPoint presentation is no big deal. Mistakes in healthcare can kill people. Daneshju said this post he red team ChatGPT, along with 80 other people, including both computer scientists and physicians, pose medical questions to ChatGPT, and twenty percent of the time it returns a dangerous response. “Twenty percent problematic response is not good enough to me for real day-to-day use in the health care system,” he said.
Of course, proponents will say that AI can augment a doctor’s work, not replace them, and that they should always check the output. And it’s true, post Story interviewed a physician at Stanford who said two-thirds of doctors there access a platform that records and transcribes patient meetings with AI so they can look them in the eye during visits and not look down while taking notes. But even there, OpenAI’s Whisper technology seems to insert completely made-up information into some recordings. Whisper mistakenly inserted into a transcript that a patient attributed the cough to exposure to their child, which they never said, Sharpe said. An incredible example of bias from training data found in the Daneshju experiment is that an AI transcription tool assumed that a Chinese patient was a computer programmer, when the patient never provided such information.
AI can potentially help in healthcare, but its outputs need to be thoroughly tested and then how much time are doctors actually saving? Furthermore, patients need to trust their doctors are actually checking what the AI is producing—hospital systems need to be able to make sure that’s happening, otherwise complacency can set in.
Basically, generative AI is just a word prediction machine, which searches through large amounts of data without actually understanding the underlying concepts. It is not “intelligent” in the same sense as a real human being, and it is not particularly capable of understanding situations unique to each particular person; It is returning information that it has generalized and seen before.
“I think it’s one of those promising technologies, but it’s not there yet,” said Adam Rodman, an internal medicine doctor and AI researcher at Beth Israel Deaconess Medical Center. “I worry that we’re going to further degrade what we do by putting hallucinated ‘AI slop’ into high-stakes patient care.”
Next time you visit your doctor, it might be worth asking if they’re using AI in their workflow.