This document is intended to serve as a starting point for reflection, guidance and discussion
for the ICAI labs in Amsterdam. It is a living document in which senior artificial intelligence (AI)
researchers leading an ICAI Lab in Amsterdam express what they consider the core values
underlying their work with an aim to facilitate constructive and responsible AI research in The
Netherlands and beyond. We welcome feedback to this document and will update it
In accordance with the Advisory report on Ethical and Legal Aspects of Informatics Research, by
the Royal Netherlands Academy of Arts and Sciences, our researchers have a core responsibility
to ensure that the research they conduct is ethically sound. Like all science, AI is not practiced
in a vacuum, but in interaction with society. We therefore want to be explicit about our core
values and the guidelines we follow to uphold these values. They apply to our researchers, the
contents of our research as well as to the choice of our partners.
Considering that the development of this field of research is characterised by a fast pace and
high uncertainty, we continuously improve our core values and guidelines to ensure they reflect
the expectations of today’s society. ICAI would like to nurture existing and invites new
dialogues with different sectors about the ethical challenges presented by AI research, and
thereby facilitate ethically sound and responsible AI research.
- We strive for a positive social impact and economic value of our research (“AI for social good“), as expressed in the Sustainable Development Goals of the United Nations, and
as endorsed by the Dutch universities.
- We subscribe to the Ethics Guidelines for Trustworthy AI as formulated by the High-
Level Expert Group on AI of the European Commission.
- We are conscious of the transformative competencies of AI in shaping society, creating
new values and reconciling tensions and dilemmas. We contribute to social awareness
about how the world is changing and will change under the influence of AI by initiating
and conducting the social debate about the role of AI in society.
- We defend academic freedom and independence as formulated in the Dutch code of
conduct, and we publish our scientific results, including publications, datasets, and
experimental code, under open access, with the exception of trade secrets and IP
protection, and sensitive data, on which we make transparent agreements.
- We are transparent about where we stand academically, where we are going and how.
We are open about dilemmas we may face, how we will deal with them when they arise
and what choices we make.
- We regard the continuous training of scientific talent as one of our core objectives.
- We engage in dialogues with and build bridges to both emerging and established
scientists in other countries and cultures, contributing to national and international
communities for knowledge sharing, collaboration and innovation.
- We strengthen our local ecosystem for AI research and its applications, in collaboration
with universities, industry, government, civil society and social institutions.
- We promote an academic culture that recognizes and rewards academics for learning by
doing and for bridging the gaps between a) fundamental and applied research, b)
scientific fields, c) research and education, and d) science and society through AI.
Marieke van Erp (Cultural AI Lab)
Stratis Gavves (POP AART, QUVA)
Theo Gevers (ATLAS Lab)
Sennay Ghebreab (Civic AI Lab)
Paul Groth (AIRLab Amsterdam, Discovery Lab)
Laura Hollink (Cultural AI Lab)
Frank van Harmelen (Discovery Lab)
Joris Mooij (Mercury Machine Learning Lab)
Jacco van Ossenbruggen (Civic AI Lab, Cultural AI Lab)
Maarten de Rijke (AIRLab Amsterdam)
Clarisa Sánchez (AI for Oncology)
Cees Snoek (AIM Lab, ATLAS Lab, QUVA)
Marcel Worring (AIM Lab)