License: Creative Commons Attribution 3.0 Unported license (CC BY 3.0)
When quoting this document, please refer to the following
DOI: 10.4230/DagRep.9.4.59
URN: urn:nbn:de:0030-drops-113046
URL: http://dagstuhl.sunsite.rwth-aachen.de/volltexte/2019/11304/
Go back to Dagstuhl Reports


Fisher, Michael ; List, Christian ; Slavkovik, Marija ; Weiss, Astrid
Weitere Beteiligte (Hrsg. etc.): Michael Fisher and Christian List and Marija Slavkovik and Astrid Weiss

Ethics and Trust: Principles, Verification and Validation (Dagstuhl Seminar 19171)

pdf-format:
dagrep_v009_i004_p059_19171.pdf (8 MB)


Abstract

This report documents the programme of, and outcomes from, the Dagstuhl Seminar 19171 on "Ethics and Trust: Principles, Verification and Validation". We consider the issues of ethics and trust as crucial to the future acceptance and use of autonomous systems. The development of new classes of autonomous systems, such as medical robots, "driver-less" cars, and assistive care robots has opened up questions on how we can integrate truly autonomous systems into our society. Once a system is truly autonomous, i.e. learning from interactions, moving and manipulating the world we are living in, and making decisions by itself, we must be certain that it will act in a safe and ethical way, i.e. that it will be able to distinguish 'right' from `wrong' and make the decisions we would expect of it. In order for society to accept these new machines, we must also trust them, i.e. we must believe that they are reliable and that they are trying to assist us, especially when engaged in close human-robot interaction. The seminar focused on questions of how does trust with autonomous machines evolve, how to build a `practical' ethical and trustworthy system, and what are the societal implications. Key issues included: Change of trust and trust repair, AI systems as decision makers, complex system of norms and algorithmic bias, and potential discrepancies between expectations and capabilities of autonomous machines.
This workshop was a follow-up to the 2016 Dagstuhl Seminar 16222 on Engineering Moral Agents: From Human Morality to Artificial Morality. When organizing this workshop we aimed to bring together communities of researchers from moral philosophy and from artificial intelligence and extend it with researchers from (social) robotics and human-robot interaction research.

BibTeX - Entry

@Article{fisher_et_al:DR:2019:11304,
  author =	{Michael Fisher and Christian List and Marija Slavkovik and Astrid Weiss},
  title =	{{Ethics and Trust: Principles, Verification and Validation (Dagstuhl Seminar 19171)}},
  pages =	{59--86},
  journal =	{Dagstuhl Reports},
  ISSN =	{2192-5283},
  year =	{2019},
  volume =	{9},
  number =	{4},
  editor =	{Michael Fisher and Christian List and Marija Slavkovik and Astrid Weiss},
  publisher =	{Schloss Dagstuhl--Leibniz-Zentrum fuer Informatik},
  address =	{Dagstuhl, Germany},
  URL =		{http://drops.dagstuhl.de/opus/volltexte/2019/11304},
  URN =		{urn:nbn:de:0030-drops-113046},
  doi =		{10.4230/DagRep.9.4.59},
  annote =	{Keywords: Verification, Artificial Morality, Social Robotics, Machine Ethics, Autonomous Systems, Explain-able AI, Safety, Trust, Mathematical Philosophy,}
}

Keywords: Verification, Artificial Morality, Social Robotics, Machine Ethics, Autonomous Systems, Explain-able AI, Safety, Trust, Mathematical Philosophy,
Freie Schlagwörter (englisch): Robot Ethics, Human-Robot Interaction
Collection: Dagstuhl Reports, Volume 9, Issue 4
Issue Date: 2019
Date of publication: 30.09.2019


DROPS-Home | Fulltext Search | Imprint | Privacy Published by LZI