A collection of reports, blogs, papers and other written material about artificial intelligence and robot ethics.
Policy Documents & Institutions – ethical, legal and socio-economic issues of robotics and artificial intelligence
Index compiled by Vincent C. Müller (email@example.com)
A comprehensive list of policy documents published between 2006 and 2019.
Ethics guidelines for trustworthy AI
European Commission, April 2019
This report highlights that trustworthy AI should respect all applicable laws and regulations, as well as a series of specific requirements (human agency oversight, robustness and safety, privacy and data governance, transparency, diversity and non-discrimination, fairness, society and environmental wellbeing and accountability) that are set out in the report”
Report, Ethics guidelines for trustworthy AI, European Commission Single Digital Market Expert Group on AI, April 2019, 41 pges
Perspectives on Issues in AI Governance
Google, Undated – probably late 2018
This report highlights “five areas where government, in collaboration with wider civil society and AI practitioners, has a crucial role to play in clarifying expectations about AI’s application on a context-specific basis. These include explainability standards, approaches to appraising fairness, safety considerations, requirements for human-AI collaboration, and general liability frameworks.”
Report, Perspectives on Issues in AI Governance, Google, Late 2018, 30 pges
Statement to the United Nations on behalf of the LAWS open letter signatories
A statement was read on the floor of the United Nations during the August, 2018 Convention on Certain Conventional Weapons (CCW) meeting, in which delegates discussed a possible ban on lethal autonomous weapons (LAWS).
The signatories are nearly 4,000 AI and robotics researchers and scientists from around the world who have called on the United Nations to move forward to negotiations to consider a legally binding instrument on lethal autonomous weapons.
On September 12th 2018, the European Parliament passed (by an 82% majority) a resolution calling for an international ban on lethal autonomous weapons systems (LAWS).
Article, Statement to the United Nations on behalf of the LAWS open letter signatories, August 2018, 1 page
AI Governance: A Research Agenda
“This report outlines an agenda for this research, dividing the field into three research clusters. The first cluster, the technical landscape, seeks to understand the technical inputs, possibilities, and constraints for AI. The second cluster, AI politics, focuses on the political dynamics between firms, governments, publics, researchers, and other actors. The final research cluster of AI ideal governance envisions what structures and dynamics we would ideally create to govern the transition to advanced artificial intelligence.”
Report, AI Governance: A Research Agenda, Governance of AI Program Future of Humanity Institute University of Oxford, August 2018, 53 pages
An Overview of National AI Strategies
Tim Dutton, July 2018
Article summarising the key policies and goals of over 25 country’s national strategis in AI. (Includes: Australia, Canada, China, Denmark, EU Commission, Finland, France, Germany, India, Italy, Japan, Kenya, Malaysia, Mexico, New, Zealand, Nordic-Baltic Region, Poland, Russia, Singapore, South Korea, Sweden, Taiwan,, Tunisia, UAE,United Kingdom, United States)
Article, An Overview of National AI Strategies, Tim Dutton, Politics and AI, July 2018, 25 minute read
Report on Consultation on the Centre for Data Ethics and Innovation
UK Government, June 2018
The government launched a consultation outlining proposals for the Centre for Data Ethics and Innovation in June 2018. The consultation ran for 12 weeks and closed on 5 September 2018. They received 104 responses and feedback through a series of roundtable discussions.
All suggestions and comments were reviewed and assessed by the government as part of its response published in November 2018. The response focuses primarily on clarifying the Centre’s existing functions and strengthening its reporting and recommendation functions.
RSA – The Ethics of Ceding more Power to Machines
Brhmie Balaram, May 2018
The RSA’s report, launched 31st May 2018, argues that the public needs to be engaged early and more deeply in the use of AI if it is to be ethical. There is a real risk that if people feel like decisions about how technology is used are increasingly beyond their control, they may resist innovation, even if this means they could lose out on benefits.
Visit this article to link to the full RSA report ‘Artificial Intelligence: Real Public Engagement’
IEEE Ethically Aligned Design: A Vision For Prioritizing Wellbeing With Artificial Intelligence And Autonomous Systems
For Public Discussion – By 7th May 2018
Version 2 of this report is available by registering at:
Public comment on version 1 of this document was invited by March 2017 to encourages technologists to prioritize ethical considerations in the creation of autonomous and intelligent technologies. The document was created by committees of The IEEE Global Initiative for Ethical Considerations in Artificial Intelligence and Autonomous Systems, comprised of over one hundred global thought leaders and experts in artificial intelligence, ethics, and related issues.
Version 2 presents the following principles/recommendations:
Candidate Recommendation 1 – Human Rights
To best honor human rights, society must assure the safety and security of A/IS so that they are designed and operated in a way that benefits humans:
1. Governance frameworks, including standards and regulatory bodies, should be established to oversee processes assuring that the use of A/IS does not infringe upon human rights, freedoms, dignity, and privacy, and of traceability to contribute to the building of public trust in A/IS.
2. A way to translate existing and forthcoming legal obligations into informed policy and technical considerations is needed. Such a method should allow for differing cultural norms as well as legal and regulatory frameworks.
3. For the foreseeable future, A/IS should not be granted rights and privileges equal to human rights: A/IS should always be subordinate to human judgment and control.
Candidate Recommendation 2 – Prioritizing Wellbeing
A/IS should prioritize human well-being as an outcome in all system designs, using the best available, and widely accepted, well-being metrics as their reference point.
Candidate Recommendation 3 – Accountability
To best address issues of responsibility and accountability:
1. Legislatures/courts should clarify issues of responsibility, culpability, liability, and accountability for A/IS where possible during development and deployment (so that manufacturers and users understand their rights and obligations).
2. Designers and developers of A/IS should remain aware of, and take into account when relevant, the diversity of existing cultural norms among the groups of users of these A/IS.
3. Multi-stakeholder ecosystems should be developed to help create norms (which can mature to best practices and laws) where they do not exist because A/IS-oriented technology and their impacts are too new (including representatives of civil society, law enforcement, insurers, manufacturers, engineers, lawyers, etc.).
4. Systems for registration and record-keeping should be created so that it is always possible to find out who is legally responsible for a particular A/IS. Manufacturers/operators/owners of A/IS should register key, high-level parameters, including:
• Intended use
• Training data/training environment (if applicable)
• Sensors/real world data sources
• Process graphs
• Model features (at various levels)
• User interfaces
• Optimization goal/loss function/reward function
Standard Reference for Version 2
The IEEE Global Initiative on Ethics of Autonomous and Intelligent Systems. Ethically
Aligned Design: A Vision for Prioritizing Human Well-being with Autonomous and Intelligent Systems, Version 2. IEEE, 2017. http://standards. ieee.org/develop/indconn/ec/autonomous_ systems.html.
Report, The IEEE Global Initiative on Ethics of Autonomous and Intelligent Systems. Ethically
Aligned Design: A Vision for Prioritizing Human Well-being with Autonomous and Intelligent Systems, Version 2. IEEE, 2017, December 2017, 136 pages
HOUSE OF LORDS Select Committee on Artificial Intelligence
Report of Session 2017–19, AI in the UK: ready, willing and able?
Concludes that the UK can be a world leader in the development of artificial intelligence. Useful historical documentation of the UK governments previous funding in AI. Includes many references for the setting up of the Centre for Data Ethics and Innovation.
Report, Report of Session 2017–19, AI in the UK: ready, willing and able?, HOUSE OF LORDS Select Committee on Artificial Intelligence, April 2018, about 180 pages
The Malicious Use of Artificial Intelligence: Forecasting, Prevention, and Mitigation
Future of Humanity Institute, University of Oxford, Centre for the Study of Existential Risk, University of Cambridge, Center for a New American Security, Electronic Frontier Foundation, OpenAI (and more)
“Artificial intelligence (AI) and machine learning (ML) are altering the landscape of security risks for citizens, organizations, and states. Malicious use of AI could threaten digital security (e.g. through criminals training machines to hack or socially engineer victims at human or superhuman levels of performance), physical security (e.g. non-state actors weaponizing consumer drones), and political security (e.g. through privacy-eliminating surveillance, profiling, and repression, or through automated and targeted disinformation campaigns).”
Report, The Malicious Use of Artificial Intelligence: Forecasting, Prevention, and Mitigation, Contributions from several organisations, February 2018, 99 pages
Growing the artificial intelligence industry in the UK
This independent review, carried out for the UK government (Business Secretary and Culture Secretary) by Professor Dame Wendy Hall (Professor of Computer Science at the University of Southampton, UK) and Jérôme Pesenti (CEO BenevolentTech), reports on how the Artificial Intelligence industry can be grown in the UK.
The recommendations cover:
- Improving data access
- Improving the supply of skills
- Maximising UK AI research
- Supporting the uptake of AI
Notably, a search of the recommendations for the words ‘ethics’, ‘morality’, ‘caution’, ‘safety’, ‘principle’ and ‘harm’ all return zero results. Transparency and accountability are mentioned in relation to supporting the uptake of AI (recommendation 14 in the main report). The report refers the reader to The Royal Society and the British Academy review on the needs of a 21st century data governance system (see below).
In the main report (page 14) it states:
‘Trust, ethics, governance and algorithmic accountability: Resolving ethical and societal questions is beyond the scope and the expertise of this industry-focused review, and could not in any case be resolved in our short time-frame.
However, building public confidence and trust will be vital to successful development of UK AI. Therefore this Review stresses the importance of industry and experts working together to secure and deserve public trust, address public perceptions, gain public confidence, and model how to deliver and demonstrate fair treatment. Fairness will be part of gaining economic benefits, and addressing ethical issues effectively to support wider use of AI could be a source of economic advantage for the UK.’
Page 66 of the main report:
‘As noted above, AI can also create new situations with new implications for fairness, transparency and accountability. AI could also change the nature of many areas of work.
AI in the UK will need to build trust and confidence in AI-enabled complex systems. There is already collective activity to work towards guidelines in ethics for automation, but we can expect this field to grow and change. A publicly visible expert group drawn from industry and academia, which engages with these issues would help to build that trust and confidence.’
Page 68 has a section on ‘Explainability of AI-enabled uses of data’ which concludes with the recommendation:
‘￼Recommendation 14: The Information Commissioner’s Office and the Alan Turing Institute should develop a framework for explaining processes, services and decisions delivered by AI, to improve transparency and accountability.’
‘Further on, it is possible that new applications of AI may hold solutions on transparency and explainability, using dedicated AIs to track and explain AI-driven decisions.’
Report,Growing the artificial intelligence industry in the UK, Professor Dame Wendy Hall, October 2017
The Royal Society and the British Academy review on the needs of a 21st century data governance system.
‘The amount of data generated from the world around us has reached levels that were previously unimaginable. Meanwhile, uses of data-enabled technologies promise benefits, from improving healthcare and treatment discovery, to better managing critical infrastructure such as transport and energy.
These new applications can make a great contribution to human flourishing but to realise these benefits, societies must navigate significant choices and dilemmas: they must consider who reaps the most benefit from capturing, analysing and acting on different types of data, and who bears the most risk.’
Article and Report, The Royal Society, Data management and use: Governance in the 21st century – a British Academy and Royal Society project, October 2017
Report of COMEST on Robot Ethics
World Commission on the Ethics of Scientific Knowledge and Technology (COMEST)
United Nations Educational, Scientific and Cultural Organisation, September 2017
This 62 page report concludes with recommendations in the following area:
VI.1. A technology-based ethical framework
VI.2. Relevant ethical principles and values
VI.2.1. Human Dignity
VI.2.2. Value of Autonomy
VI.2.3. Value of Privacy
VI.2.4. ‘Do not harm’ Principle
VI.2.5. Principle of Responsibility
VI.2.6. Value of Beneficence
VI.2.7. Value of Justice
VI.3. COMEST specific recommendations on robotics ethics
VI.3.1. Recommendation on the Development of the Codes of Ethics for Robotics and Roboticists
VI.3.2. Recommendation on Value Sensitive Design
VI.3.3. Recommendation on Experimentation
VI.3.4. Recommendation on Public Discussion
VI.3.5.Recommendation on Retraining and Retooling of the Workforce
VI.3.6. Recommendations related to Transportation and Autonomous Vehicles
VI.3.7. Recommendations on Armed Military Robotic Systems (‘Armed Drones’)
VI.3.8. Recommendations on Autonomous Weapons
VI.3.9. Recommendations on Surveillance and Policing
VI.3.10. Recommendation relating to Private and Commercial Use of Drones
VI.3.11. Recommendation on Gender Equality
VI.3.12. Recommendation on Environmental Impact Assessment
VI.3.13. Recommendations on Internet of Things
Report, Report of COMEST on Robot Ethics, COMEST, UN, September 2017
Prioritizing Human Well-being in the Age of Artificial Intelligence
On 11 April 2017, IEEE hosted a dinner debate at the European Parliament in Brussels called, Civil Law Rules on Robotics: Prioritizing Human Well-being in the Age of Artificial Intelligence. The event featured experts from The IEEE Global Initiative for Ethical Considerations in Artificial Intelligence and Autonomous Systems (“The IEEE Global Initiative”) and was hosted by Member of European Parliament (MEP) Mady Delvaux (The Progressive Alliance of Socialists and Democrats, Luxembourg), who served as Rapporteur on the Parliament’s Civil Law Rules on Robotics report. Among other recommendations, the report proposed a system of registration for advanced robots managed by a potential EU Agency for Robotics and Artificial Intelligence (AI).
The report also suggested autonomous robots be granted the status of electronic personhood under a liability framework regarding the actions of these devices and their users. This idea has been largely misconstrued as a form of robot rights although the way “personhood” is described in the report is similar to the legal notion of corporate personhood. The confusion and heightened interest surrounding this issue paved the way for an in depth discussion on how to ascribe and measure value for technology and the well-being of the people who use it.
Report, Prioritizing Human Well-being in the Age of Artificial Intelligence, IEEE, April 2017
ASILOMAR AI Principles
A set of principles signed by 1,273 AI/Robotics researchers and 2541 others at the BAI 2017 conference, the Future of Life Institute’s second conference on the future of artificial intelligence in January 2017
‘Artificial intelligence has already provided beneficial tools that are used every day by people around the world. Its continued development, guided by the following principles, will offer amazing opportunities to help and empower people in the decades and centuries ahead.’
Conference Report, ASILOMAR AI Principles, Future of Life Institute, January 2017
BS 8611:2016 – Robot Ethics Standard
Robots and robotic devices. Guide to the ethical design and application of robots and robotic systems
BS 8611 gives guidelines for the identification of potential ethical harm arising from the growing number of robots and autonomous systems being used in everyday life. The standard also provides additional guidelines to eliminate or reduce the risks associated with these ethical hazards to an acceptable level. The standard covers safe design, protective measures and information for the design and application of robots.
This standard for robot and robotics device designers and managers, and the general public. It BS was written by scientists, academics, ethicists, philosophers and users to provide guidance on specifically ethical hazards associated with robots and robotic systems and how to put protective measures in place. It recognizes that these potential ethical hazards have a broader implication than physical hazards, so it is important that different ethical harms and remedial considerations are considered. The new standard builds on existing safety requirements for different types of robots, covering industrial, personal care and medical.
Hardcopy and PDF document, BS 8611:2016 – Robots and robotic devices, British Standards Institute, April 2016
Available from British Standards Institute Shop
EPSRC Principles of Robotics – Regulating robots in the real world
Engineering and Physical Sciences Research Council (EPSRC) principles
Evidence submitted to the House of Lords AI Committee 6th September 2017
Doteveryone – Written evidence (AIC0148)
Evidence around inequality for APPG AI
by Laura James, Doteveryone, October 2017
A presentation given to the All Party Parliamentary Committee on AI, on 16th October 2017
The Quest for Roboethics – A Survey
Last Update April 2019
“The aim of this paper is give a brief account of subjects, projects, groups and authors dealing with ethical aspects of robots. I first start with recent research on roboethics in two EU projects namely ETHICBOTS (2005-2008) and ETICA (2009-2011). I report on the activities of Roboethics.org and particularly of the Technical Committee (TC) on Roboethics of the IEEE and list some ethical issues and principles currently discussed. I also report briefly on the Machine Ethics Consortium.”
Article, The Quest for Roboethics – A Survey, Rafael Capurro, April 2019
How AI can be a force for good
Mariarosaria Taddeo1,2,3, Luciano Floridi1,2
24 Aug 2018
“Artificial intelligence (AI) is not just a new technology that requires regulation. It is a powerful force that is reshaping daily practices, personal and professional interactions, and environments. For the well-being of humanity it is crucial that this power is used as a force of good. Ethics plays a key role in this process by ensuring that regulations of AI harness its potential while mitigating its risks.”
Journal Article : How AI can be a force for good, Mariarosaria Taddeo1,2,3, Luciano Floridi1,2, Science, Vol. 361, Issue 6404, pp. 751-752 DOI: 10.1126/science.aat5991
What jobs will still be around in 20 years?
‘Jobs won’t entirely disappear; many will simply be redefined. But people will likely lack new skillsets required for new roles and be out of work anyway’
Newspaper Article, What jobs will still be around in 20 years? Read this to prepare your future, The Guardian, June 2017
The Doomesday Invention
Article, The Doomesday Invention, Raffi Khatchadourian – Review of Nick Bostrom’s warning about AI, The New Yorker, November 2015
Tech Ethics in Practice
Article, Tech Ethics in Practice, doteveryone, Laura James, March 2018
Microsoft’s AI Blog
Regulator Against the Machine
Artificial intelligence (AI) is part of today’s zeitgeist: whether it be parallels with ‘I, Robot’ dystopias or predictions about its impact on society. But for all the potential, the development of machines that learn as they go remains slow in health care.
Blog post, Regulator Against the Machine, British Journal of Health Computing, November 2017
What do Robots Believe? – Ways of Knowing
How do we know what we know? This article considers: (1) the ways we come to believe what we think we know (2) the many issues with the validation of our beliefs (3) the implications for building artificial intelligence and robots based on the human operating system.
Blog Post, Human Operating System 4 – Ways of Knowing, Rod Rivers, September 2017
Cennydd Bowles (Author), September 2018
“Based on Cennydd’s years of research and consulting, Future Ethics transforms modern ethical theory into practical advice for designers, product managers, and software engineers alike. Cennydd uses the three lenses of modern ethics to focus on some of the tech industry’s biggest challenges: unintended consequences and algorithmic bias, the troubling power of persuasive technology, and the dystopias of surveillance, autonomous war, and a post-work future.”
Book, Future Ethics, ISBN-10: 1999601912 ISBN-13: 978-1999601911, Paperback, 242 pages, September 2018
Robot Ethics 2.0: From Autonomous Cars to Artificial Intelligence 1st Edition
Patrick Lin (Editor), Keith Abney (Editor), Ryan Jenkins (Editor), October 2017
Expanding discussions on robot ethics ‘means listening to new voices; robot ethics is no longer the concern of a handful of scholars. Experts from different academic disciplines and geographical areas are now playing vital roles in shaping ethical, legal, and policy discussions worldwide. So, for a more complete study, the editors of this volume look beyond the usual suspects for the latest thinking. Many of the views as represented in this cutting-edge volume are provocative–but also what we need to push forward in unfamiliar territory.’
Book, Robot Ethics 2.0: From Autonomous Cars to Artificial Intelligence 1st Edition, ISBN-13: 978-0190652951 ISBN-10: 0190652950, 440 pages, October 2017
Robot Ethics: The Ethical and Social Implications of Robotics
(Intelligent Robotics and Autonomous Agents series)
Patrick Lin (Editor), Keith Abney (Editor), George A. Bekey (Editor)
‘Starting with an overview of the issues and relevant ethical theories, the topics flow naturally from the possibility of programming robot ethics to the ethical use of military robots in war to legal and policy questions, including liability and privacy concerns. The contributors then turn to human-robot emotional relationships, examining the ethical implications of robots as sexual partners, caregivers, and servants. Finally, they explore the possibility that robots, whether biological-computational hybrids or pure machines, should be given rights or moral consideration.’
Book, Robot Ethics: The Ethical and Social Implications of Robotics, January 2014
Moral Machines: Teaching Robots Right from Wrong
1st Edition by Wendell Wallach (Author), Colin Allen (Author)
‘Computers are already approving financial transactions, controlling electrical supplies, and driving trains. Soon, service robots will be taking care of the elderly in their homes, and military robots will have their own targeting and firing protocols. Colin Allen and Wendell Wallach argue that as robots take on more and more responsibility, they must be programmed with moral decision-making abilities, for our own safety. Taking a fast paced tour through the latest thinking about philosophical ethics and artificial intelligence, the authors argue that even if full moral agency for machines is a long way off, it is already necessary to start building a kind of functional morality, in which artificial moral agents have some basic ethical sensitivity. But the standard ethical theories don’t seem adequate, and more socially engaged and engaging robots will be needed. As the authors show, the quest to build machines that are capable of telling right from wrong has begun.
Moral Machines is the first book to examine the challenge of building artificial moral agents, probing deeply into the nature of human decision making and ethics.’
Book, Moral Machines: Teaching Robots Right from Wrong, ISBN-13: 9780195374049, first published January 2009
Meet the robot that helps you learn from afar
Mike Scialom, Cambridge Independent
PUBLISHED: 13:33 11 March 2018
The AV1 robot avatar – An innovative concept means an interactive education for students who can’t make it to lectures
AI: ‘It’s not just a playground’
Mike Scialom, Cambridge Independent
PUBLISHED: 11:04 16 November 2017
Is Prof Stephen Hawking right to say AI is a danger to humanity? Dr Stella Pachidi says machine intelligence needs closer scrutiny
Data is the fuel for AI
Newsletter Post, Data is the fuel for AI, so let’s ensure we get the ethics right, Birgitte Andersen, City AM Newsletter, December 2017
BBC News: MEPs vote on robots’ legal status – and if a kill switch is required
MEPs have called for the adoption of comprehensive rules for how humans will interact with artificial intelligence and robots.
The report makes it clear that it believes the world is on the cusp of a “new industrial” robot revolution.
It looks at whether to give robots legal status as “electronic persons”.
Designers should make sure any robots have a kill switch, which would allow functions to be shut down if necessary, the report recommends.
Meanwhile users should be able to use robots “without risk or fear of physical or psychological harm”, it states.
BBC News, MEPs vote on robots’ legal status – and if a kill switch isrequired, By Jane Wakefield, Technology reporter, January 2017
BBC News, Sex robots: Experts debate the rise of the love droids
Would you have sex with a robot? Would you marry one? Would a robot have the right to say no to such a union?
These were just a few of the questions being asked at the second Love and Sex with Robots conference hastily rearranged at Goldsmiths University in London after the government in Malaysia – the original location – banned it.
BBC News, Sex robots: Experts debate the rise of the love droids, By Jane Wakefield, Technology reporter, December 2016