VeriLearn 2023 : Verifying Learning AI Systems Workshop @ ECAI'23
Call For Papers
We would like to announce the call for papers for the Verifying Learning AI Systems Workshop @ ECAI'23
While there is no uniformly agreed-upon definition of what constitutes safe or trustworthy AI, it is clear that such systems should exhibit certain properties. For example, systems should be robust to minor perturbations to their inputs and there should be some transparency about how a system arrives at a prediction or decision. More importantly, it is becoming increasingly common for deployed AI models to have to conform to requirements (e.g., legal) and/or exhibit specific properties (e.g., fairness). That is, it is necessary to verify that a model complies with these requirements. In the software engineering community, verification has been long studied with the goal of assuring that software fully satisfies the expected requirements. Therefore, a key open question in the quest for safe AI is how verification and machine learning can be combined to provide strong guarantees about software that learns and that adapts itself on the basis of past experience? Finally, what are the boundaries of what can be verified, and how can and should system design be enhanced by other mechanisms (e.g., statistics on benchmarks, procedural safeguards, accountability) to produce the desired properties?
The goal of the Verifying Learning AI Systems (VeriLearn) workshop is to bring together researchers interested in these questions. The workshop will be held in conjunction with the 26th European Conference on Artificial Intelligence, which will take place in Krakow Poland.
**Topics of Interest**
This workshop solicits papers on the following non-exhaustive list of topics:
• Representations and languages that facilitate reasoning and verification.
• Applications and extensions of software verification techniques in the context of machine learning.
• Verifying safety in dynamic systems or models.
• Reasoning about learned models to assess, e.g., their adherence to requirements.
• Learning models that are safe by design.
• Assessing the robustness of AI systems.
• Ways to evaluate aspects of AI systems that are relevant from a trust and safety perspective.
• Out of distribution detection and learning with abstention.
• Certification methodologies for AI systems.
• Concepts, approaches, and methods for identifying and dealing with the limits of verifiability.
• Approaches and case studies where verification is important for addressing ethical, privacy and societal concerns about AI.
• Case studies showing illustrative applications where verification is used to tackle issues related to safety and trustworthiness.
**Submission Instructions and Dates**
We solicit two types of papers:
- Long papers can be a maximum of 6 pages of content and an unlimited number of references in the ECAI 2023 formatting style and should report on novel, unpublished work that might not be quite mature enough for a conference or journal submission.
- Extended abstracts can be 2 pages in ECAI formatting style and summarize recent publications fitting the workshops.
Submissions should be anonymous. Papers are to be submitted in pdf format at https://cmt3.research.microsoft.com/VeriLearn2023
Paper submission deadline: 20/06/2023 @ 23:59pm CET
Jesse Davis (firstname(dot)firstname.lastname@example.org)
• Jesse Davis, KU Leuven,
• Bettina Berendt, TU Berlin, director of the Weizenbaum Institute for the Networked Society, KU Leuven
• Hendrik Blockeel, KU Leuven
• Luc De Raedt, KU Leuven
• Benoit Frenay, University of Namur
• Fredrik Heintz, Linköping University
• Jean-Francois Raskin, Université Libre de Bruxelles