top of page


AI technologies and machine learning are already widely used and researched in medicine, yet still seem (both in practice and theory) to resemble an opaque black box rather than a trustworthy control mechanism. MedAIcine addresses key challenges and tensions regarding the responsible design and use of AI in medical imaging.

AI systems are already widely used and researched in medicine. The possible areas of use for artificial intelligence (AI) appear to be almost unlimited: For example, the AI-supported processing and rapid evaluation of large amounts of data enables robot-assisted surgery, the use of chatbots and screening apps as diagnostic aids, or the continuous monitoring of chronic diseases with the help of medical wearables, such as fitness trackers, which measure, record and interpret the patient's vital signs. In this context, AI acts as a kind of "sparring partner" for physicians in their clinical decision-making in the context of human-machine interaction (Helmholtz 2022). Thanks to specially developed algorithms and computer programs that possess deep-learning technology, AI thus has the potential to effectively improve medical care in terms of individual prevention, screening, diagnostics, prognosis and therapy.

Although the use of AI in medicine may sound promising at first, ethical considerations point to certain risks of the current use and design of AI in medicine: Lack of transparency, explicability and fairness, but also insufficient protection of patients' privacy - or their sensitive health data - are just a few examples of the specific challenges in dealing with AI in medicine. For example, what data set is the AI-assisted diagnosis based on? Are the training data representative of the individuals being treated (implicit bias)? Has the General Data Protection Regulation (GDPR) been complied with when collecting the data? But it is not only technological aspects that take on an important role in the assessment of responsible AI. Genuinely philosophical questions, such as those about the good life, good coexistence, or freedom of action, must also be considered in the context of AI research and critically reflected upon in light of the now digital environment.


Bernhard Bauer

Professor für Softwaremethodik für verteilte Systeme an der Universität Augsburg

Kerstin Schlögl-Flierl

Inhaberin des Lehrstuhls für Moraltheologie an der Universität Augsburg

Paula Ziethmann

Wissenschaftliche Mitarbeiterin an Universität Augsburg

Benjamin Rathgeber

Professor für Natur- und Technikphilosophie mit Schwerpunkt KI an der HFPH

Michael Reder

Professor für Praktische Philosophie an der HFPH mit dem Schwerpunkt Völkerverständigung

Sophie Jörg

Wissenschaftliche Mitarbeiterin an der Hochschule für Philosophie München

Alena Buyx

Professorin für Medizinethik und Gesundheitstechnologien an der TUM

Ruth Müller

Associate Professor of Science & Technology Policy an der Technischen Universität München

Svenja Breuer

Wissenschaftliche Mitarbeiterin an der Technischen Universität München

bottom of page