EU Horizon 2020
Horizon 2020
HomeNewsCase StudiesPeopleKey Prior PublicationsPublications


June 2021 - Research Opportunities: We are seeking exceptional researchers interested in formulating theories, models and algorithms for probabilistic verification and synthesis to enable robust AI.

If this sounds like you, and you hold a PhD in computer science, mathematics or related discipline and post-qualification research experience, we'd love to hear from you. Please contact Marta ( or Anita ( to learn more about opportunities on the FUN2MODEL project.

May 2021 - Paper accepted at UAI 2021! Here we propose a framework to provide safety certification for given control policies, and synthesize control policies that improve the certification bounds.
May 2021 - Event: Marta was awarded the prestigious 2019 BCS Lovelace Medal for her research in probabilistic and quantitative verification and so led the 2020 Lovelace Lecture “Probabilistic model checking for the data-rich world.” A recording can be found here.

The BCS Lovelace Lecture was sponsored by The Ada Lovelace Institute

April 2021 - Papers: three papers accepted at IJCAI 2021!

Here we build on abduction-based explanations for machine learning and develop a method for computing local explanations for neural network models in natural language processing (NLP).

Here we demonstrate provable guarantees on the robustness of decision rules, paving the way towards provably causally robust decision-making systems.

Here we introduce the first method for verifying the time-unbounded safety of neural networks controlling dynamical systems.

April 2021 - Welcome. We warmly welcome DPhil Student Emanuele La Malfa, to the FUN2MODEL project. Emanuele's work will focus on robustness and explainability for natural language processing (NLP).
March 2021 - Software Release: PRISM 4.7 is now available, including support for POMDPs, improved accuracy reporting and more.
March 2021 - Welcome. We warmly welcome researcher, Rui Yan, to the FUN2MODEL project. Rui's work will focus on probabilistic verification and synthesis, including Bayesian and psychological games.
December 2020 - News: Marta becomes a Fellow of the European Laboratory for Learning and Intelligent Systems (ELLIS). ELLIS Fellows advance science, provide strategic advice and leadership, and act as ambassadors of ELLIS.
November 2020 - Event: Professor David Parker, co-PI on FUN2MODEL, gives a keynote talk at iFM 2020 titled "Verification with Stochastic Games: Advances and Challenges".
October 2020 - Event: Marta is a panelist at the Royal Society's Briefing for Making Europe a Leader in AI: in conversation with Venki Ramakrishnan, Antoine Petit and Martin Stratmann.
October 2020 - Welcome. We are delighted that graduate student, Elias Benussi, has joined the FUN2MODEL project. Elias will focus on fairness in AI.
October 2020 - Paper accepted at FORMATS 2020! Here we propose MOSAIC, an algorithm for measuring the safety of deep reinforcement learning controllers in stochastic settings.
October 2020 - Paper accepted at EMNLP 2020! Here we focus on robustness of text classification against word substitutions.
September 2020 - Event: Marta gives a webinar on 'Safety and robustness for deep learning with provable guarantees' in the ICE-TCS Reykjavik University series.
September 2020 - Event: Marta gives a keynote speech at ASE 2020 on 'Safety and robustness for deep learning with provable guarantees'.
September 2020 - Event: Marta gives a plenary talk at DNA26 on 'Probabilistic verification and synthesis for reliable molecular circuit designs'.
September 2020 - Event: Marta gives a keynote speech at KR 2020 on 'Probabilistic model checking for strategic equilibria-based decision making'.
August 2020 - Welcome. We are delighted to have researcher, Gabriel Santos, join the FUN2MODEL project. Gabriel will focus on strategic reasoning and game-theoretic techniques in AI.
August 2020 - Paper accepted at QEST 2020! Here we propose multi-coalitional verification techniques for concurrent stochastic games.
July 2020 - Event: Marta gives a Webinar on "When to trust a self-driving car" at The National Academy of Sciences, India (NASI) - Delhi Chapter attended by over 440+ participants. Please see here to watch a recording.
June 2020 - Event: Marta has been invited to sit on the Global Partnership on Artificial Intelligence (GPAI) Working Group on Responsible AI, nominated by the European Commission.

GPAI is an international and multi-stakeholder initiative to guide the responsible development and use of artificial intelligence consistent with human rights, fundamental freedoms, and shared democratic values, as reflected in the OECD Recommendation on AI.

Please see here for more details.

May 2020 - Paper accepted at UAI 2020! Here we show how to compute worst-case adversarial guarantees for Bayesian Neural Networks (BNNs).
May 2020 - Paper by Clare Lyle accepted at ICML 2020! Here we consider the problem of learning abstractions that generalize in block MDPs, families of environments with a shared latent state space, and dynamics structure over that latent space, but varying observations.
April 2020 - Welcome. We are delighted to welcome researcher, Andrea Patane, to the project. Andrea brings with him expertise in safety verification of Bayesian models and the role played by uncertainty in adversarial prediction settings. He will focus on data-driven perception modelling and cognitive reasoning, for integration within agent-based models.
April 2020 - Paper accepted at CAV 2020!

Here we present a major new release of the PRISM-games model checker, featuring multiple significant advances in its support for verification and strategy synthesis of stochastic games.

Download the tool and case studies here.

March 2020 - Event: Marta becomes a member of the Royal Society’s Working Group on Digital Technology and the Planet which aims to help the UK play a leading role in data-enabled innovation and the adoption of digital technologies to tackle climate change.
February 2020 - Paper by Min Wu accepted at CVPR 2020! Here we consider the robustness of deep neural networks on videos.
January 2020 - Paper accepted at AISTATS 2020! Here we show how to compute worst-case adversarial guarantees for classification with Gaussian processes.
January 2020 - Event: Marta gives a keynote speech at ERTS 2020 titled 'Safety verification for deep neural networks with provable guarantees'.
January 2020 - Event: Marta is an invited speaker at VMCAI 2020.
January 2020 - Software Release: PRISM-games 3.0 is now available, providing concurrent stochastic games, equilibria, real-time models and many new examples. More information here.
October 2019 - Welcome. We warmly welcome graduate student, Emanuele La Malfa, who joins the project as an Associate Member and will formally join in April 2021. Emanuele is working on robustness and explainability of AI, with specific focus on natural language processing (NLP) models.
October 2019 - Welcome. We are delighted that graduate student, Benjie Wang, has joined the FUN2MODEL project. Benjie is particularly interested in causal modelling as a means to enhance the robustness and explainability of deep learning.
October 2019 - Welcome. We are excited to appoint postdoctoral researcher, Luca Laurenti, to the project. Luca will work on on developing probabilistic verification and synthesis methods for deep learning, with a particular focus on Baysian neural networks.
1st October 2019 - The Fun Begins! FUN2MODEL kicks off and work begins as we aim to make major advances in the quest towards provably robust and beneficial AI.
March 2019 - Announcement: We are delighted to announce that Professor Marta Kwiatkowska has been awarded a highly competitive European Research Council Advanced Investigators Grant for a new five-year project FUN2MODEL.

Our objectives are to develop novel probabilistic verification and synthesis techniques to guarantee safety, robustness and fairness for complex decisions based on machine learning, formulate a comprehensive, compositional game-based modelling framework for reasoning about systems of autonomous agents and their interactions, and evaluate the techniques on a variety of case studies.

This is the second ERC Advanced Grant awarded to Marta Kwiatkowska: from 2010 until 2016 she held the grant VERIWARE. Please see the press release here and here for more details.