A recent study, detailed in a paper titled "Write on Paper, Wrong in Practice: Why LLMs Still Struggle with Writing Clinical Notes" by Rohan Paul, reveals that large language models (LLMs) face significant challenges in generating clinical notes due to a fundamental mismatch with existing healthcare workflows. The research, which included a six-week pilot involving ten therapists and thirty interviews across twenty staff, highlights that automated note-taking tools often fail to integrate effectively into the nuanced realities of clinical practice.
The core task, transforming therapists' scratch notes into structured SOAP (Subjective, Objective, Assessment, and Plan) notes, proved more complex than anticipated. While seemingly straightforward, the pilot found that "scratch notes ranged from full sentences to cryptic cues or none," complicating automated processing. This variability, combined with practical constraints like limited time for documentation during school visits, underscored the disconnect between the ideal input for LLMs and real-world clinical data.
A significant finding was that time inefficiencies stemmed not from the writing itself, but from "duplicate forms in the electronic record and multiple artifacts." Clinicians also expressed a strong preference for autonomy and maintaining personal templates, often bypassing the tools entirely when "edits felt heavier than starting fresh." This led many to "rewrit[e] inputs for the model, turning it into a formatter that mixed sections and erased time savings.
Even advanced models like Llama 3 8B, fine-tuned with Low Rank Adaptation on deidentified SOAP pairs, and Microsoft Copilot, struggled to meet user expectations. According to the study, these models "still mismatched inputs and expectations." The paper attributes these failures to a "weak match" across the "Fit between Individual, Task, and Technology" framework, suggesting that current LLM solutions are not sufficiently adaptable to diverse clinical environments.
The conclusive takeaway from the research emphasizes that for automated clinical note-taking to succeed, "tools must flex to workflows and policies or automated notes stay fragile." This highlights the critical need for future AI development in healthcare to prioritize deep integration with existing human processes and adapt to the dynamic, often unstructured, nature of clinical documentation rather than imposing rigid new systems.