Content
summary Summary

Researchers put Google's latest video AI, Veo-3, to the test with real surgical footage, revealing a disconnect between the model's visual output and its understanding of medical procedures.

Ad

Veo-3 was prompted to predict how a surgery would unfold over the next eight seconds, using just a single image as input. To measure its performance, an international team created the SurgVeo benchmark, using 50 real videos from abdominal and brain surgeries.

Four-stage Surgical Plausibility Pyramid: Visual plausibility, instrument operation, tissue feedback, surgical intention
The Surgical Plausibility Pyramid used in the SurgVeo benchmark breaks down evaluation into four levels: visual appearance, instrument handling, tissue response, and surgical intent. | Image: Chen et al.

To rate Veo-3, four experienced surgeons watched the AI-generated clips and scored them on four criteria: visual appearance, instrument use, tissue feedback, and whether the actions made medical sense.

SurgVeo Pipeline: Data preparation, Veo-3 video generation, and expert scoring in four plausibility dimensions
The SurgVeo pipeline covers every step, from preparing real surgical data to generating Veo-3 videos and having experts rate them across four medical plausibility categories. | Image: Chen et al.

Strong visuals, poor surgical logic

Veo-3 produced videos that looked authentic at first glance - some surgeons called the quality "shockingly clear." But on closer inspection, the content fell apart. In abdominal surgery tests, the model scored 3.72 out of 5 for visual plausibility after one second. But as soon as medical accuracy was required, its performance dropped.

Ad
Ad
Six comparison frames: real vs. generated with errors in image quality, instrument guidance, operation, target selection, tissue response, and intention
Side-by-side comparisons show real and AI-generated surgical frames, highlighting typical mistakes like visual artifacts, incorrect instrument use, and medically implausible actions. | Image: Chen et al.

For abdominal procedures, instrument handling earned just 1.78 points, tissue response only 1.64, and surgical logic was lowest at 1.61. The AI could create convincing images, but it couldn't reproduce what actually happens in an operating room.

Brain surgery reveals even bigger gaps

The challenge was even greater for brain surgery footage. From the first second, Veo-3 struggled with the fine precision required in neurosurgery. For brain operations, instrument handling dropped to 2.77 points (compared to 3.36 for abdominal) and surgical logic fell as low as 1.13 after eight seconds.

Error distribution of generated surgical videos: Proportions of intention, instrument, feedback, surgical, and image quality errors per track
Error analysis shows that most problems in AI-generated surgical videos come from faulty reasoning and instrument handling, while issues with image quality are rare. | Image: Chen et al.

The team also broke down the types of errors. Over 93 percent were related to medical logic: the AI invented tools, imagined impossible tissue responses, or performed actions that made no clinical sense. Only a small fraction of errors (6.2 percent for abdominal and 2.8 percent for brain surgery) were tied to image quality.

Researchers tried giving Veo-3 more context, such as the type of surgery or the exact phase of the procedure. The results showed no meaningful or consistent improvement. According to the team, the real problem isn't the information provided, but the model's inability to process and understand it.

Visual medical understanding is still out of reach

The SurgVeo study shows how far current video AI is from real medical understanding. While future systems could one day help train doctors, assist with surgical planning, or even guide procedures, today's models are nowhere near that level. They produce videos that look real, but lack the knowledge to make safe or meaningful decisions.

Recommendation

The researchers plan to release the SurgVeo benchmark on GitHub, inviting other teams to test and improve their models.

The study also highlights the risks of using synthetic AI-generated videos for medical training. Unlike Nvidia's approach, where AI videos help train robots for general tasks, in healthcare, these kinds of AI hallucinations could be dangerous. If a system like Veo-3 generates videos that look plausible but show medically incorrect procedures, it could teach robots or trainees the wrong techniques.

The results also make it clear that the concept of video models as "world models" is still far off. Current systems can imitate how things look and move, but they don't have a reliable grasp of physical or anatomical logic. As a result, their videos might seem convincing at a glance, but they can't capture the real logic or cause-and-effect behind surgery.

 

Ad
Ad
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.

Meanwhile, text-based AI is already showing real gains in medicine. In one study, Microsoft's "MAI Diagnostic Orchestrator" delivered diagnostic accuracy four times higher than experienced general practitioners in complex cases, although the study notes some methodological limitations.

Support our independent, free-access reporting. Any contribution helps and secures our future. Support now:
Bank transfer
Summary
  • Google's video model Veo 3 creates realistic-looking videos of surgical procedures, but it fails to demonstrate accurate medical knowledge or correct simulation of operations.
  • More than 93 percent of the mistakes made by the AI involved medical logic, such as inventing instruments, showing impossible tissue reactions, and performing actions that do not make sense.
  • The study concludes that video models like Veo 3 are currently not suitable for simulating medical expertise or being used as a training tool in medical contexts.
Sources
Jonathan writes for THE DECODER about how AI tools can improve both work and creative projects.
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.