What’s the best way to use AI in Your General Practice? Balancing Innovation with “Human In The Loop” Human-AI teaming.

As doctors, we’re fond of telling patients not to eat junk food. Fast food is tempting because it’s fast and easy, but it’s really unhealthy. Doctors are increasingly facing similar decisions and temptations to “eat the AI fast food” in their clinical practice. AI has already and will continue to rapidly transform healthcare; General Practitioners (GP’s) are increasingly turning to AI-driven tools to save time, enhance documentation, and support clinical decision-making. However, the integration of AI into your patient care, especially in high-stakes environments, demands a thoughtful, ethical approach that prioritises patient safety and upholds fundamental rights. “Human in The Loop” is an important guideline, but what does this really mean in best practice? What should you, and should you NOT do?

The EU AI Act and Human In The Loop

Let’s start with laws and legal guidelines, the EU AI Act (implemented August 2024), sets clear legal standards for the development and deployment of AI systems within the EU and for EU citizens[1][2]. For high-risk AI applications, like AI-driven medical devices and clinical decision support systems, the Act mandates that these technologies must be designed so that natural persons (i.e., human professionals) can effectively oversee their functioning throughout their use[3][4][5]. This “human in the loop” requirement is not just a technicality; it is a critical safeguard to prevent or minimize risks to health, safety, and fundamental rights. But what does that really mean?

Under Article 14 of the AI Act, high-risk AI systems must:

  • Be designed and developed with appropriate human-machine interface tools to enable effective human oversight[3][4][5].
  • Allow human overseers to monitor, interpret, and, if necessary, override AI outputs, including the ability to stop the system or reverse its decisions[4][5].
  • Ensure that oversight measures are commensurate with the risks, autonomy, and context of use, meaning that more complex or impactful decisions require more robust human oversight mechanisms[3][4].

The Act also emphasises the importance of training and authorising human overseers, ensuring they have the competence and authority to intervene when necessary[5]. While the Act provides a strong foundation for ethical AI use, it leaves some ambiguity regarding the specifics of human oversight responsibilities, highlighting the ongoing need for clear, practical guidance in clinical settings[5]. Our aim is to define some easy-to-follow best-practice guidelines.

The Risks of AI-Generated Doctor-Patient Interaction Summaries

One area of particular concern is the use of AI to automate the creation of doctor-patient interaction summaries. While these tools promise to reduce administrative burden and improve documentation efficiency, they introduce significant risks that are not always immediately apparent.

Hallucinations and Factual Errors

AI models, particularly large language models (LLMs), are prone to “hallucinations” i.e. generating factually incorrect statements or fabricating details that were never discussed during a consultation[6][7][8]. Recent studies have shown that AI-generated medical summaries frequently contain inaccuracies, including incorrect patient information, misrepresented symptoms, and erroneous medication instructions[6][8]. In one real-world example, an AI transcription tool introduced exaggerated details into a patient’s medical record, which were only discovered upon careful review by the patient herself[7].

Errors of Omission

Perhaps even more concerning is the risk of errors of omission. AI systems may fail to detect or record important statements made during a consultation, leading to incomplete or misleading medical records[9]. Because these omissions are not obvious—especially in long, detail-driven summaries—they are especially difficult for doctors to detect during post-call reviews. This can have serious clinical consequences, as vital information may be overlooked, potentially affecting diagnosis, treatment, and patient safety[9]. Because this technology is so new

Clinical and Legal Implications

The variability and unpredictability of AI-generated summaries mean that even with identical inputs, the output can differ in content, emphasis, and organisation[9]. This variability can subtly influence clinical decision-making, nudging doctors toward different diagnostic or treatment paths based on what the AI chooses to include or exclude. Moreover, the legal implications are significant: if AI-generated inaccuracies or omissions go unnoticed, they may only come to light when medical decisions are challenged or when records are used as evidence in legal proceedings[7].

The Way Forward: Ethical AI in General Practice

To use AI ethically in general practice, it is essential to:

  • Maintain robust human oversight as required by laws the EU AI Act, ensuring that all AI-generated outputs are reviewed and validated by qualified professionals[3][4][5].
  • Do not rely on AI to summarise long complex interactions especially after the interaction is completed AI generated mistakes like hallucinations and omissions are more likely and more difficult to for doctors to detect when reading AI-generated notes and summaries of longer consultations, especially if there is a time-lag between interaction and summary.[9][10]
  • Educate staff and patients about the limitations and risks of AI tools, fostering a culture of vigilance and critical review[7][9].
  • Preserve original records whenever possible, so that discrepancies can be investigated and corrected[7].

Conclusion - Real-time Human-AI teaming.

AI has the potential to revolutionise general practice, but its use must be guided by strong ethical principles and regulatory frameworks. The persistent challenges of AI hallucination and omissions in AI-generated summaries underscore the need for real-time human doctor in the loop confirmation when creating medical notes and patient charts. Doctors are at significant professional risk when using AI to summarise and recommend care plans especially after long consultations. 

  1. https://health.ec.europa.eu/ehealth-digital-health-and-care/artificial-intelligence-healthcare_en 
  2. https://www.goodwinlaw.com/en/insights/publications/2024/11/insights-lifesciences-dpc-how-the-eu-ai-act-could-affect-medtech 
  3. https://artificialintelligenceact.eu/article/14/     
  4. https://www.mishcon.com/eu-uk-ai-navigator/article-14      
  5. https://www.euaiact.com/key-issue/4      
  6. https://www.clinicaltrialsarena.com/news/hallucinations-in-ai-generated-medical-summaries-remain-a-grave-concern/  
  7. https://judgeschlegel.substack.com/p/medical-records-meet-ai-a-looming     
  8. https://www.aiforeducation.io/blog/ai-hallucination-concerns  
  9. https://jamanetwork.com/journals/jama/article-abstract/2814609 Goodman KE, Yi PH, Morgan DJ. AI-Generated Clinical Summaries Require More Than Accuracy. JAMA. 2024;331(8):637–638. doi:10.1001/jama.2024.0555   
  10. https://apnews.com/article/ai-artificial-intelligence-health-business-90020cdf5fa16c79ca2e5b6c4c9bbb14

Subscribe to our Blog

Keep up to do with F365 updates and content!

Join Function 365 Today!

LET'S GO!
Function 365 Landscape Logo - Practice Management Software
user linkedin facebook pinterest youtube rss twitter instagram facebook-blank rss-blank linkedin-blank pinterest youtube twitter instagram