RESPONSIBLE AI & CORONA

RESPONSIBLE AI & CORONA

A research project to evaluate and promote responsible use of AI in the fight against Corona

WHY this project?

WHY THIS PROJECT?

  • AI can contribute to tackling the Corona-crisis

    AI can play an important role during this crisis. AI applications can contribute to the understanding of the spread of the virus, to the search for a vaccine and to treatment of COVID-19. AI can also guide policy measures and gain insight in their long-term impact on society and the economy.

  • If it doesn’t work, it can do harm

    In times of crisis, we tend to turn more quickly to ‘invasive’ technology and to ‘turn a blind eye’ when it comes to fundamental rights, ethical values and even effectiveness. The motto is often: no harm, no foul. Unfortunately, this is different with AI. If it does not help, it may very well harm.

  • Desparate times, desparate measures?

    We see false contradictions such as: we have to choose between ethics or health, between fundamental rights or the opening up of the economy. And it is often thought that an invasive technology that is harmful, will be dismantled after the crisis. In reality this is often not the case and we risk setting unwanted precedents for the future.

  • This crisis needs Responsible AI

    Despite the urgency of this crisis, it is important that AI is developed and deployed in a responsible manner. Robustness, effectiveness, transparency and explainability, but also fundamental rights, inclusion and ethics are adamant to ensure that AI actually helps in tackling the Corona-crisis without causing harm to society along the way.

What are the objectives?

Systematic assessment of the opportunities and challenges of AI to tackle the Corona-crisis and devise AI strategies for the current as well as future epidemics.

Highlight responsible, useful, reliable and safe AI & Corona applications, while at the same time draw attention to unsafe, irresponsible or undesirable AI.

Increase knowledge and awareness among policymakers, stakeholders and society about the importance of responsible AI, especially in times of crisis.

What does the project involve?

What does the project involve?

1

The Observatory will contain monitors to provide insights into AI applications that are being deployed / launched in connection with the Corona crisis.

2

The Quickscans will entail a quick first assessment of a number of carefully selected AI-applications deployed in the Corona crisis. We will look at elements such as ethical impact, legal impact, societal impact, etc.

3

“Why we need responsible AI in times of Corona and how to achieve it.” – Project Inception Paper by Catelijne Muller, Virginia Dignum and Noah Schoppl.

4

Interviews and information will be provided on a continuous basis throughout the project to raise awareness and understanding of AI during Corona and the need to ensure its responsible development and use.

5

Several Expert Sessions will be organized for selected groups to raise knowledge and awareness of AI during Corona, including its opportunities and challenges.

Catelijne Muller, Project Lead

Expertise: AI & Society, Law and Policy

President of ALLAI. Master of Laws; EU Policy expert; Member of EU High Level Expert Group on AI; AI Rapporteur of the European Economic and Social Committee; Expert advisor on AI to the Council of Europe.

Virginia Dignum, Scientific Lead

Expertise: Social and Ethical AI

Co-founder and board member of ALLAI; Professor of Ethical and Social Artificial Intelligence; Member of the High Level Expert Group on AI. Member of the Global Partnership on AI. Book: “Responsible AI”.

Multi-disciplinary expertise

Prof. Dagmar Monett

Expertise: Artificial Intelligence

Full Professor of Computer Science HWR Berlin; Co-founder AGI Sentinel Initiative, AGISI.org; AI expert at Ms.AI, “Artificial Intelligence for and with Women.”

Prof. Filippo Santoni de Sio

Expertise: Ethics of Technology

Associate Professor in Ethics of Technology TU Delft; adjunct professor in Ethics of Transportation Politecnico di Milano; Rapporteur of the EU Expert Group on ethical issues of driverless mobility.

Bart Geerts, MD

Expertise: AI & Healthcare

Anaesthetist; CEO and founder of Healthplus AI, a company that builds, validates and scales machine-learning to transform health data into valuable insights for patients and doctors.

Andreas Theodorou, PhD

Expertise: Responsible Artificial Intelligence

Postdoc. Researcher and Research Engineer in Responsible AI Umeå University; CEO and Co-Founder of VeRAI AB, a company that develops methods for responsible AI.

Prof. Frank Dignum

Expertise: Social Simulations & AI

Wallenberg chair in AI at Umeå University in Sweden; affil. Utrecht University; honorary principal research fellow of the University of Melbourne; EurAI fellow.

Gabrielle Speijer, MD

Expertise: AI, Healthcare & Data

Radiation-Oncologist at the Haga hospital; Founder of health innovation company CatalyzIT. She’s board member ICT&health international/NL, HIMSS Global Future50 community.

Nathalie Smuha, LLM.

Expertise: AI & EU Regulation and Law

Researcher at the KU Leuven Faculty of Law and the Leuven.AI Institute; Coordinator of the EU High Level Expert Group on AI; Admitted to the New York Bar.

Prof. Jan van Gemert

Expertise: Image Recognition

Lead of Computer Vision lab at TUDelft. His research focuses on visual inductive priors for deep learning for automatic image and video understanding.

Prof. Catholijn Jonker

Expertise: Explainable AI

Professor of Interactive Intelligence at TUDelft. Board member of the European Artificial Intelligence Association.

Prof. Frederik Zuiderveen Borgesius

Expertise: Privacy and Data Protection

Professor ICT and Law Radboud University Nijmegen where he is affiliated with iHub, a multidisciplinary research hub on Security, Privacy, and Data Governance.

Assistant Prof. Roel Dobbe

Expertise: Safety, Sustainability and Democratization of AI

Assistant Professor at TUDelft. Research: security, sustainability and democratization of algorithmic systems and complex digital infrastructures. Former Postdoc fellow AI ​​Now Institute in New York.

Christofer Talvitie

Team member: Observatory Research

AI Governance research intern at ALLAI. Christofer will research and assess AI-applications on the Observatory.

Maria Avramidou

Team member: Observatory Research

Researcher at ALLAI. Maria will research and assess AI-applications on the Observatory.

Mónica Fernández

Team member: Observatory Research

Intern at ALLAI. Mónica will research and pre-assess AI-applications on the Observatory.

What our Experts say

“Especially in times of crisis, it is essential to acknowledge that – even if AI has a great potential to do good – its development and use needs to be lawful, ethical and robust to actually materialize that potential.”

“AI can support decision making of governments, but only if its contributions can be explained and adapted to a new reality.”

“Emergencies like the Corona pandemic put some of our basic and political values under pressure: freedom, responsibility, privacy. In these times it is even more important to keep reflecting on the meaning of these concepts and how to embed them in AI-based technologies.”

“AI can and will exploit all patterns and biases in data; what I admire about this project is that it helps to identify unwanted AI bias.”

“The corona crisis motivates the rapid development of AI applications that increase the power of the government and technology companies. The ALLAI Project Responsible AI & Corona is therefore of great importance to keep one’s finger on the pulse and protect our fundamental rights, democracy and values. ”

“Responsible AI needs responsible humans, both in pandemic times and beyond. I am certain the ALLAI project ‘Responsible AI & Corona’ will set the tone in this respect.”

“Doctors are the trusted link: they determine the quality of care and consequently the tools, including artificial intelligence. Leading sustainable IT orchestration from the clinical perspective as an expert in ‘Responsible AI & Corona’ in open collaboration and cross-domain is crucial now for society as a whole.”

       This project is supported by: