3 things The Pitt episode 8:00 AM gets right about AI in medicine, and 1 big thing it gets wrong

Share This Post

People keep telling me to watch The Pitt, and now that the show is tackling generative AI in healthcare, I must. As Mashable’s tech editor, I field a lot of questions about generative AI, and I also spend a lot of time talking to people who are extremely enthusiastic about artificial intelligence, and people who are extremely hostile to AI.

So, what does the latest episode of The Pitt Season 2 get right about AI in medicine, and what does it get wrong? 

How AI factors into The Pitt episode “8:00 AM.”

Gerran Howell, Amielynn Abellera, Noah Wyle, Sepideh Moafi, and Supriya Ganesh in "The Pitt."


Credit: Warrick Page / HBO Max

AI wasn’t the entire focus of Season 2, episode two, “8:00 AM.” We also get to see a nun with gonorrhea in her eye, an unhoused man with a colony of maggots inside his moldy arm cast, and a clear view of an erection, which, of course, has lasted more than four hours.

However, one of the plotlines focuses on newcomer Dr. Baran Al-Hashimi’s (Sepideh Moafi) attempt to modernize the ER and integrate AI into patient care. She introduces an AI app that automatically listens to patient visits and summarizes the pertinent details in their charts.

When an excited student says, “Oh my god, do you know how much time this will save?”, Dr. Al-Hashimi has an answer: With the new AI app, the ER doctors will spend 80 percent less time charting. Later in the episode, the good doctor tells Dr. Robby (Noah Wyle) that it will also allow physicians to spend 20 percent more time with patients.

So far, so good, but the app immediately makes a mistake. It documents the wrong medication in the patient’s chart, substituting a similar-sounding medication instead.

This doesn’t diminish Dr. Al-Hashimi’s enthusiasm at all. “Generative AI is 98 percent accurate at present,” she says. “You must always carefully proofread and correct minor errors. It’s excellent but not perfect.”

AI transcription is really good (and it still makes mistakes).

Dr. Al-Hashimi states that generative AI is 98 percent accurate. But is this really true?

Fact-checking Dr. Al-Hashimi’s claim is actually tricky, because it’s not entirely clear what she means. If she’s referring purely to AI transcription, she’s closer to the truth. That’s a task that generative AI excels at.

Last year, a group of researchers conducted a systematic review of 29 studies that measured AI transcription accuracy specifically in healthcare settings. (You can check out the entire review, published in the BMC Medical Informatics and Decision Making journal.)

Some of those studies did find accuracy rates of 98 percent or higher, and some found significant time savings for doctors. However, those studies involved controlled, quiet environments. In multi-speaker environments with a lot of crosstalk and medical jargon, like you’d find in a crowded emergency room, accuracy rates were much lower, sometimes as low as 50 percent.

Still, rapid advancements in large-language models are improving AI transcriptions all the time. So, we could be generous and say Dr. Al-Hashimi’s claim is close to the truth, for the latest LLM models, in certain settings.

Generative AI is definitely not 98 percent accurate.

Let’s take the latest version of ChatGPT as an example. When the GPT-5.2 model was released a few months ago, OpenAI published documentation on the model’s tendency to hallucinate and provide false information.

According to OpenAI, its GPT-5.2 Thinking model has an average hallucination rate of 10.9 percent. That’s really high, especially considering that OpenAI wants ChatGPT to help you with medical questions. (The company recently launched ChatGPT Health, a “dedicated experience in ChatGPT designed for health and wellness.”) 

Now, when GPT-5.2 Thinking is given access to the internet, its hallucination rate drops to 5.8 percent. But would you trust a doctor who’s wrong 5.8 percent of the time — and only when they can use the internet to check their work? And would you want your AI healthcare app to be connected to the internet at all?

Generative AI may one day be 98 percent accurate. But we’re not there yet.

Generative AI can’t replace doctors.

Noah Wyle and Supriya Ganesh in "The Pitt."


Credit: Warrick Page / HBO Max

Dr. Robby and other characters have a conversation about “gut” instincts in this episode. That’s something that generative AI can’t replicate, and one reason that many people don’t want to see AI tools replacing human workers entirely, whether in the arts or medicine. 

The episode also spends a lot of time showing examples of empathy. The best doctors don’t just have an encyclopedic knowledge of their specialty. They don’t just have good gut instincts. They’re true healers, who know that holding a patient’s hand at the right moment can be just as important as the right diagnosis.

To be fair, most AI healthcare tools I’ve seen aren’t trying to replace doctors. Instead, they want to give healthcare workers more diagnostic tools and save them time. As Dr. Al-Hashimi says repeatedly in the episode, she wants to give doctors more time at their patients’ bedside.

AI can help save doctors time.

Radiology, or medical imaging, is one of the most promising applications for generative AI in healthcare settings. When radiologists at Northwestern University implemented a custom generative AI tool to help analyze X-rays and CT scans, their productivity increased by 40 percent — and without compromising accuracy.

I think even a lot of AI skeptics would agree that’s a positive result, for patients and doctors.

The Pitt does seem to be setting up Dr. Al-Hashimi as something of a villain — or, at least, a foil for Dr. Robby — but good medicine and generative AI aren’t necessarily in conflict. As with any tool, it can be extraordinarily helpful — or extraordinarily dangerous.

The Pitt Season 2 is now streaming on HBO Max, with new episodes being released every Thursday at 9 p.m. ET.


Disclosure: Ziff Davis, Mashable’s parent company, in April 2025 filed a lawsuit against OpenAI, alleging it infringed Ziff Davis copyrights in training and operating its AI systems.

Subscribe The Newsletter

Get updates and learn from the best

More To Explore

Mashable

Verizon outage cause revealed

Verizon has confirmed that the extreme service outage was not caused by a cybersecurity breach but was some kind of software problem. Details are still

Do You Want To Stay Connected?

drop a line and keep in touch