Schedule
Wrap-Up
Friday November 15, 2024 16:00 - 16:15 CET Hybrid Conference
Panel 2: Future of Mobility - AI Quality & Certification
Corina Pascu, Dr.-Ing. Arwed Schmidt, Adam Smith, Dr. Arndt von Twickel
Friday November 15, 2024 15:15 - 16:00 CET Hybrid Conference
Coffee Break
Friday November 15, 2024 14:45 - 15:15 CET Hybrid Conference
Session 3: New Challenges - LLMs & AI-based assistance systems
Florian Strohm, Dr. Bernhard Nessler, Prof. Dr. Andreas Zell
Friday November 15, 2024 13:45 - 14:45 CET Hybrid Conference
Panel 1: Regulatory Readiness - EU AI Act Alignment
Dr. Till Klein, Georg Stettinger, Prof. Dr. Beatrix Weber, Patrick Gilroy
Friday November 15, 2024 13:00 - 13:45 CET Hybrid Conference
Lunch Break
Friday November 15, 2024 12:00 - 13:00 CET Hybrid Conference
Session 2: Use Case Focus - Automotive/Mobility Sector Testing
Lukas Birkemeyer, Dr. Emil Gracic, Kathrin Grosse, Dr. Padma Iyenghar, Fabian Langer, Gregor Pawelke
Friday November 15, 2024 10:40 - 12:00 CET Hybrid Conference
Fabian Langer - AIMobilityAudit – AI Evaluation in Automotive: The AIMobilityAudit project, led by the German Federal Office for Information Security (BSI) in partnership with ZF and TÜV Informationstechnik GmbH (TÜVIT), targets the development of processes for the security-focused evaluation of AI systems in vehicles, particularly for automated driving functions respectively assistant systems. Building on the initial AIMobilityAuditPrep project, which defined 50 evaluation requirements and created a basic framework for testing, AIMobilityAudit aims to fathom practical security assessments, e.g., on dedicated test tracks, and adjustments to the requirements thereupon. The project addresses the unique challenges of AI in vehicles, including the risk of deliberate adversarial attacks for the security and safety of (part-)autonomous vehicles. Through collaboration with industry, academia, and government, AIMobilityAudit shall establish a comprehensive foundation for AI security audits in autonomous vehicles, ensuring that AI systems are both resilient and trustworthy. The project’s outcomes will be groundwork for a technical guideline for the derivation of AI-specific evaluation requirements and their subsequent analysis, both in the digital domain as in real-world testing, supporting safer integration of AI into mobility. The presentation will give the AIMobilityAudit’s key challenges and the most valuable insights of the project.
Lukas Birkemeyer - Scenario Generation for Testing Automated Driving Systems: Established testing strategies in the automotive domain are not sufficient for testing Automated Driving Systems (ADS) due to the extremely large number of possible scenarios the ADS might face. Instead, the SOTIF-standard (ISO 21448) establishes scenario-based testing as state-of-the-art, explicitly using scenarios as test cases in the testing process to evaluate the system under test's behavior while interacting with its environment. Instead of focusing on requirements, scenario-based considers the system under test a black box enabling testing of AI components. However, the SOTIF standard does not specify how to generate scenarios, which hinders its practical application. In this talk, we elaborate on the state-of-the-art regarding scenario generation and discuss whether generated scenarios are SOTIF-compliant. Subsequently, we propose a combinatorial scenario generation approach to address the unsolved challenges of modeling an overall scenario space and covering it with a selection of scenarios. We analyze, compare, and discuss sampling strategies for generating SOTIF-compliant scenario suites. We apply mutation testing to assess the scenario suite's ability to detect potential failures in a rule-based system. Finally, we elaborate on the potential and challenges of transferring our findings to testing black-box systems.
Kathrin Grosse - AI Security Testing for Autonomous Vehicles (AVs): AI security has received a lot of attention in the recent years, but securing complex systems like autonomous vehicles remains a critical challenge. We will approach this topic first from the perspective of AI security and corresponding reported incidents. However, real-world incidents are hard to collect. We thus take a second route and revisit the literature of demonstrated attacks on AVs. One aspect that is picked for discussion in the context of AV security testing is how realistic attacks are. Even under this condition, there is a large variety of attacks to be considered-pointing us towards the need for good test and system designs.
Padma Iyenghar/Emil Gracic/Gregor Pawelke - How to achieve ASIL A and ASIL B conformance of machine learning products?:This work introduces a novel design and testing concept for ML products with the goal of achieving the lower levels of safety in automotive - ASIL A and ASIL B. It provides a systematic procedure to close the gaps between ISO 26262 and ISO PAS 8800, which manifest in absence of a clear guidance for design and testing of ML products (ISO 26262) as well as in absence of the guidance for ASIL A or ASIL B conformity of them (ISO PAS 8800).The procedure is based on four main pillars: 1. Defining proper lifecycle phases for ML development, 2. Introducing the most relevant desired (safety) properties and suitable test methods for each of the phases, 3. Evaluated rigor regarding the capability of each method for systematic testing of the desired properties, and finally a derivation of ASIL A or ASIL B related recommendations for each method
Coffee Break
Friday November 15, 2024 10:20 - 10:40 CET Hybrid Conference
Session 1: Technical State of Play - Auditing Trustworthy AI
Matthias König, Rainer Plaga, Dr. Christoph Poetsch, Prof. Dr. Wojciech Samek, Dr. Arndt von Twickel
Friday November 15, 2024 09:20 - 10:20 CET Hybrid Conference
Christoph Poetsch/Matthias König - Systematic AI Assessment: To systematically assess AI systems, we need a structured approach that provides both an overarching general framework to horizontally structure the field of AI assessment and specific test methods to actually test different AI systems. Additionally, it’s essential to vertically bridge these two aspects, ensuring that they work together cohesively. This talk introduces an AI Assessment Matrix designed to address these challenges, which in addition serves as a key component of the AI Assessment Framework currently being developed by the TÜV AI.Lab. The first part of the presentation will introduce the systematic core ideas of this Matrix as an organisational structure for AI assessment practice. In the second part, the vertical concretisation will be exemplified with regard to the testing dimension of robustness.
Welcome
Friday November 15, 2024 09:00 - 09:20 CET Hybrid Conference