Skip to main content Skip to main navigation

IASEAI‘2026 Conference: Interdisciplinary Collaboration for Safe and Trustworthy AI

| Foundations of Systems AI | Neuro-mechanistic Modeling | Saarbrücken | Darmstadt

Large language models (LLMs) such as ChatGPT, Claude and Llama have made AI part of our daily lives. But even indirectly, this technology is already influencing our lives. It is making important decisions behind the scenes. However, AI is not infallible. At the second annual conference of the International Association for Safe & Ethical AI (IASEAI’2026), researchers from DFKI and CERTAIN (Centre for European Research in Artificial Intelligence) presented their latest research on safe AI and organised a workshop aimed at strengthening future interdisciplinary collaboration in this important field. The conference took place in Paris from 24 to 26 February.

© Kevin Baum

Already today, algorithms determine who receives a loan, what news we see and how the police organise their work. Given the growing real-world impact of AI systems, it is becoming increasingly important to focus on safety and ethical guidelines. As part of the IASEAI’2026 Conference, the IASEAI addressed this issue by bringing together renowned international experts from academia, politics, civil society and industry to promote research and exchange, discuss policy frameworks and ultimately advance the development of safe and ethical AI systems. Speakers included Yoshua Bengio, a pioneer in AI research and winner of the Turing Award, and Geoffrey Hinton, a winner of the Nobel Prize. The conference was organised by Mark Nitzberg of UC Berkeley, among others.

Researchers from DFKI and CERTAIN presented GRACE (Governor for Reason-Aligned Containment) at the conference: a system architecture designed to ensure that autonomous AI agents act not only effectively but also appropriately. GRACE separates moral evaluation from the actual decision-making process. A so-called “moral module” uses logic-based rules to determine which actions are permitted under which circumstances. It learns these rules from the feedback of a “moral instructor,” an external moral authority (typically a human). The “decision module” envelops an AI agent and enables it to choose the best practical steps within those boundaries. A “guardian” monitors compliance with the guidelines. Authors Felix Jahn (DFKI, Saarland University), Yannic Muskalla (DFKI, Saarland University), Lisa Dargasz (DFKI), Dr. Patrick Schramowski (DFKI, CERTAIN), and Dr. Kevin Baum (DFKI, Saarland Informatics Campus, CERTAIN) using an AI therapy assistant.

Another key factor in ensuring AI safety is the integration of human oversight. To explore this central safety aspect from an interdisciplinary perspective and discuss it in the context of the EU AI Regulation, Dr Kevin Baum and Dr Patrick Schramowski, together with Prof Dr Markus Langer of the University of Freiburg and Dr Johann Laux of Oxford University, organised the workshop 'Human Oversight of AI Variants, Technologies, and Conditions'. The central question was how human oversight differs from other forms of human involvement, and how effective, appropriate and meaningful oversight can be achieved for current systems and future AI agents.

A look into the workshop at UNESCO house© Kevin Baum