Diagnostic errors are among the most serious problems in everyday medical practice. AI systems – especially large language models (LLMs) like ChatGPT-4, Gemini, or Claude 3 – offer new ways to efficiently support medical diagnoses. Yet these systems also entail considerable risks – for example, they can “hallucinate” and generate false information.