A collection of reports, blogs, papers and other written material about robot ethics.
Consultation on the Centre for Data Ethics and Innovation – Consultation Outcome
The creation of the Centre for Data Ethics and Innovation was announced by the Chancellor in the 2017 Autumn Budget. The consultation outlined proposals for the Centre and sought views on the way in which the centre would operate and its priority areas of work. The consultation closed in September 2018 and the government reported and responed in November 2018.
Report, UK Government Response to Consultation on the Centre for Data Ethics and Innovation, Department of Digital, Culture, Media & Sport, November 2018, 1 page with links to consultation response
Statement to the United Nations on behalf of the LAWS open letter signatories
A statement was read on the floor of the United Nations during the August, 2018 Convention on Certain Conventional Weapons (CCW) meeting, in which delegates discussed a possible ban on lethal autonomous weapons (LAWS).
The signatories are nearly 4,000 AI and robotics researchers and scientists from around the world who have called on the United Nations to move forward to negotiations to consider a legally binding instrument on lethal autonomous weapons.
On September 12th 2018, the European Parliament passed (by an 82% majority) a resolution calling for an international ban on lethal autonomous weapons systems (LAWS).
Article, Statement to the United Nations on behalf of the LAWS open letter signatories, August 2018, 1 page
RSA – The Ethics of Ceding more Power to MachinesBrhmie Balaram
The RSA’s report, launched 31st May 2018, argues that the public needs to be engaged early and more deeply in the use of AI if it is to be ethical. There is a real risk that if people feel like decisions about how technology is used are increasingly beyond their control, they may resist innovation, even if this means they could lose out on benefits.
Visit this article to link to the full RSA report ‘Artificial Intelligence: Real Public Engagement’
IEEE Ethically Aligned Design: A Vision For Prioritizing Wellbeing With Artificial Intelligence And Autonomous Systems
For Public Discussion – By 7th May 2018
Version 2 of this report is available by registering at:
Public comment on version 1 of this document was invited by March 2017 to encourages technologists to prioritize ethical considerations in the creation of autonomous and intelligent technologies. The document was created by committees of The IEEE Global Initiative for Ethical Considerations in Artificial Intelligence and Autonomous Systems, comprised of over one hundred global thought leaders and experts in artificial intelligence, ethics, and related issues.
Version 2 presents the following principles/recommendations:
Candidate Recommendation 1 – Human Rights
To best honor human rights, society must assure the safety and security of A/IS so that they are designed and operated in a way that benefits humans:
1. Governance frameworks, including standards and regulatory bodies, should be established to oversee processes assuring that the use of A/IS does not infringe upon human rights, freedoms, dignity, and privacy, and of traceability to contribute to the building of public trust in A/IS.
2. A way to translate existing and forthcoming legal obligations into informed policy and technical considerations is needed. Such a method should allow for differing cultural norms as well as legal and regulatory frameworks.
3. For the foreseeable future, A/IS should not be granted rights and privileges equal to human rights: A/IS should always be subordinate to human judgment and control.
Candidate Recommendation 2 – Prioritizing Wellbeing
A/IS should prioritize human well-being as an outcome in all system designs, using the best available, and widely accepted, well-being metrics as their reference point.
Candidate Recommendation 3 – Accountability
To best address issues of responsibility and accountability:
1. Legislatures/courts should clarify issues of responsibility, culpability, liability, and accountability for A/IS where possible during development and deployment (so that manufacturers and users understand their rights and obligations).
2. Designers and developers of A/IS should remain aware of, and take into account when relevant, the diversity of existing cultural norms among the groups of users of these A/IS.
3. Multi-stakeholder ecosystems should be developed to help create norms (which can mature to best practices and laws) where they do not exist because A/IS-oriented technology and their impacts are too new (including representatives of civil society, law enforcement, insurers, manufacturers, engineers, lawyers, etc.).
4. Systems for registration and record-keeping should be created so that it is always possible to find out who is legally responsible for a particular A/IS. Manufacturers/operators/owners of A/IS should register key, high-level parameters, including:
• Intended use
• Training data/training environment (if applicable)
• Sensors/real world data sources
• Process graphs
• Model features (at various levels)
• User interfaces
• Optimization goal/loss function/reward function
Standard Reference for Version 2
The IEEE Global Initiative on Ethics of Autonomous and Intelligent Systems. Ethically
Aligned Design: A Vision for Prioritizing Human Well-being with Autonomous and Intelligent Systems, Version 2. IEEE, 2017. http://standards. ieee.org/develop/indconn/ec/autonomous_ systems.html.
Report, The IEEE Global Initiative on Ethics of Autonomous and Intelligent Systems. Ethically
Aligned Design: A Vision for Prioritizing Human Well-being with Autonomous and Intelligent Systems, Version 2. IEEE, 2017, December 2017, 136 pages
Growing the artificial intelligence industry in the UK
This independent review, carried out for the UK government (Business Secretary and Culture Secretary) by Professor Dame Wendy Hall (Professor of Computer Science at the University of Southampton, UK) and Jérôme Pesenti (CEO BenevolentTech), reports on how the Artificial Intelligence industry can be grown in the UK.
The recommendations cover:
- Improving data access
- Improving the supply of skills
- Maximising UK AI research
- Supporting the uptake of AI
Notably, a search of the recommendations for the words ‘ethics’, ‘morality’, ‘caution’, ‘safety’, ‘principle’ and ‘harm’ all return zero results. Transparency and accountability are mentioned in relation to supporting the uptake of AI (recommendation 14 in the main report). The report refers the reader to The Royal Society and the British Academy review on the needs of a 21st century data governance system (see below).
In the main report (page 14) it states:
‘Trust, ethics, governance and algorithmic accountability: Resolving ethical and societal questions is beyond the scope and the expertise of this industry-focused review, and could not in any case be resolved in our short time-frame.
However, building public confidence and trust will be vital to successful development of UK AI. Therefore this Review stresses the importance of industry and experts working together to secure and deserve public trust, address public perceptions, gain public confidence, and model how to deliver and demonstrate fair treatment. Fairness will be part of gaining economic benefits, and addressing ethical issues effectively to support wider use of AI could be a source of economic advantage for the UK.’
Page 66 of the main report:
‘As noted above, AI can also create new situations with new implications for fairness, transparency and accountability. AI could also change the nature of many areas of work.
AI in the UK will need to build trust and confidence in AI-enabled complex systems. There is already collective activity to work towards guidelines in ethics for automation, but we can expect this field to grow and change. A publicly visible expert group drawn from industry and academia, which engages with these issues would help to build that trust and confidence.’
Page 68 has a section on ‘Explainability of AI-enabled uses of data’ which concludes with the recommendation:
‘￼Recommendation 14: The Information Commissioner’s Office and the Alan Turing Institute should develop a framework for explaining processes, services and decisions delivered by AI, to improve transparency and accountability.’
‘Further on, it is possible that new applications of AI may hold solutions on transparency and explainability, using dedicated AIs to track and explain AI-driven decisions.’
Report,Growing the artificial intelligence industry in the UK, Professor Dame Wendy Hall, October 2017
Report of COMEST on Robot Ethics
World Commission on the Ethics of Scientific Knowledge and Technology (COMEST)
United Nations Educational, Scientific and Cultural Organisation
This 62 page report concludes with recommendations in the following area:
VI.1. A technology-based ethical framework
VI.2. Relevant ethical principles and values
VI.2.1. Human Dignity
VI.2.2. Value of Autonomy
VI.2.3. Value of Privacy
VI.2.4. ‘Do not harm’ Principle
VI.2.5. Principle of Responsibility
VI.2.6. Value of Beneficence
VI.2.7. Value of Justice
VI.3. COMEST specific recommendations on robotics ethics
VI.3.1. Recommendation on the Development of the Codes of Ethics for Robotics and Roboticists
VI.3.2. Recommendation on Value Sensitive Design
VI.3.3. Recommendation on Experimentation
VI.3.4. Recommendation on Public Discussion
VI.3.5.Recommendation on Retraining and Retooling of the Workforce
VI.3.6. Recommendations related to Transportation and Autonomous Vehicles
VI.3.7. Recommendations on Armed Military Robotic Systems (‘Armed Drones’)
VI.3.8. Recommendations on Autonomous Weapons
VI.3.9. Recommendations on Surveillance and Policing
VI.3.10. Recommendation relating to Private and Commercial Use of Drones
VI.3.11. Recommendation on Gender Equality
VI.3.12. Recommendation on Environmental Impact Assessment
VI.3.13. Recommendations on Internet of Things
Report, Report of COMEST on Robot Ethics, COMEST, UN, September 2017
The Royal Society and the British Academy review on the needs of a 21st century data governance system.
‘The amount of data generated from the world around us has reached levels that were previously unimaginable. Meanwhile, uses of data-enabled technologies promise benefits, from improving healthcare and treatment discovery, to better managing critical infrastructure such as transport and energy.
These new applications can make a great contribution to human flourishing but to realise these benefits, societies must navigate significant choices and dilemmas: they must consider who reaps the most benefit from capturing, analysing and acting on different types of data, and who bears the most risk.’
Article and Report, The Royal Society, Data management and use: Governance in the 21st century – a British Academy and Royal Society project, October 2017
Prioritizing Human Well-being in the Age of Artificial Intelligence
On 11 April 2017, IEEE hosted a dinner debate at the European Parliament in Brussels called, Civil Law Rules on Robotics: Prioritizing Human Well-being in the Age of Artificial Intelligence. The event featured experts from The IEEE Global Initiative for Ethical Considerations in Artificial Intelligence and Autonomous Systems (“The IEEE Global Initiative”) and was hosted by Member of European Parliament (MEP) Mady Delvaux (The Progressive Alliance of Socialists and Democrats, Luxembourg), who served as Rapporteur on the Parliament’s Civil Law Rules on Robotics report. Among other recommendations, the report proposed a system of registration for advanced robots managed by a potential EU Agency for Robotics and Artificial Intelligence (AI).
The report also suggested autonomous robots be granted the status of electronic personhood under a liability framework regarding the actions of these devices and their users. This idea has been largely misconstrued as a form of robot rights although the way “personhood” is described in the report is similar to the legal notion of corporate personhood. The confusion and heightened interest surrounding this issue paved the way for an in depth discussion on how to ascribe and measure value for technology and the well-being of the people who use it.
Report, Prioritizing Human Well-being in the Age of Artificial Intelligence, IEEE, April 2017
ASILOMAR AI Principles
A set of principles signed by 1,273 AI/Robotics researchers and 2541 others at the BAI 2017 conference, the Future of Life Institute’s second conference on the future of artificial intelligence in January 2017
‘Artificial intelligence has already provided beneficial tools that are used every day by people around the world. Its continued development, guided by the following principles, will offer amazing opportunities to help and empower people in the decades and centuries ahead.’
Conference Report, ASILOMAR AI Principles, Future of Life Institute, January 2017
BS 8611:2016 – Robot Ethics Standard
Robots and robotic devices. Guide to the ethical design and application of robots and robotic systems
BS 8611 gives guidelines for the identification of potential ethical harm arising from the growing number of robots and autonomous systems being used in everyday life. The standard also provides additional guidelines to eliminate or reduce the risks associated with these ethical hazards to an acceptable level. The standard covers safe design, protective measures and information for the design and application of robots.
This standard for robot and robotics device designers and managers, and the general public. It BS was written by scientists, academics, ethicists, philosophers and users to provide guidance on specifically ethical hazards associated with robots and robotic systems and how to put protective measures in place. It recognizes that these potential ethical hazards have a broader implication than physical hazards, so it is important that different ethical harms and remedial considerations are considered. The new standard builds on existing safety requirements for different types of robots, covering industrial, personal care and medical.
Hardcopy and PDF document, BS 8611:2016 – Robots and robotic devices, British Standards Institute, April 2016
Available from British Standards Institute Shop
EPSRC Principles of Robotics – Regulating robots in the real world
Engineering and Physical Sciences Research Council (EPSRC) principles
Evidence submitted to the House of Lords AI Committee 6th September 2017
Doteveryone – Written evidence (AIC0148)
Evidence around inequality for APPG AI
by Laura James, Doteveryone
A presentation given to the All Party Parliamentary Committee on AI, on 16th October 2017
What jobs will still be around in 20 years?
‘Jobs won’t entirely disappear; many will simply be redefined. But people will likely lack new skillsets required for new roles and be out of work anyway’
Newspaper Article, What jobs will still be around in 20 years? Read this to prepare your future, The Guardian, June 2017
The Doomesday Invention
Article, The Doomesday Invention, Raffi Khatchadourian – Review of Nick Bostrom’s warning about AI, The New Yorker, November 2015
Tech Ethics in Practice
Article, Tech Ethics in Practice, doteveryone, Laura James, March 2018
Regulator Against the Machine
Artificial intelligence (AI) is part of today’s zeitgeist: whether it be parallels with ‘I, Robot’ dystopias or predictions about its impact on society. But for all the potential, the development of machines that learn as they go remains slow in health care.
Blog post, Regulator Against the Machine, British Journal of Health Computing, November 2017
What do Robots Believe? – Ways of Knowing
How do we know what we know? This article considers: (1) the ways we come to believe what we think we know (2) the many issues with the validation of our beliefs (3) the implications for building artificial intelligence and robots based on the human operating system.
Blog Post, Human Operating System 4 – Ways of Knowing, Rod Rivers, September 2017
Robot Ethics 2.0: From Autonomous Cars to Artificial Intelligence 1st Edition
Patrick Lin (Editor), Keith Abney (Editor), Ryan Jenkins (Editor)
Expanding discussions on robot ethics ‘means listening to new voices; robot ethics is no longer the concern of a handful of scholars. Experts from different academic disciplines and geographical areas are now playing vital roles in shaping ethical, legal, and policy discussions worldwide. So, for a more complete study, the editors of this volume look beyond the usual suspects for the latest thinking. Many of the views as represented in this cutting-edge volume are provocative–but also what we need to push forward in unfamiliar territory.’
Book, Robot Ethics 2.0: From Autonomous Cars to Artificial Intelligence 1st Edition, ISBN-13: 978-0190652951 ISBN-10: 0190652950, 440 pages, October 2017
Robot Ethics: The Ethical and Social Implications of Robotics
(Intelligent Robotics and Autonomous Agents series)
Patrick Lin (Editor), Keith Abney (Editor), George A. Bekey (Editor)
‘Starting with an overview of the issues and relevant ethical theories, the topics flow naturally from the possibility of programming robot ethics to the ethical use of military robots in war to legal and policy questions, including liability and privacy concerns. The contributors then turn to human-robot emotional relationships, examining the ethical implications of robots as sexual partners, caregivers, and servants. Finally, they explore the possibility that robots, whether biological-computational hybrids or pure machines, should be given rights or moral consideration.’
Book, Robot Ethics: The Ethical and Social Implications of Robotics, January 2014
Moral Machines: Teaching Robots Right from Wrong
1st Edition by Wendell Wallach (Author), Colin Allen (Author)
‘Computers are already approving financial transactions, controlling electrical supplies, and driving trains. Soon, service robots will be taking care of the elderly in their homes, and military robots will have their own targeting and firing protocols. Colin Allen and Wendell Wallach argue that as robots take on more and more responsibility, they must be programmed with moral decision-making abilities, for our own safety. Taking a fast paced tour through the latest thinking about philosophical ethics and artificial intelligence, the authors argue that even if full moral agency for machines is a long way off, it is already necessary to start building a kind of functional morality, in which artificial moral agents have some basic ethical sensitivity. But the standard ethical theories don’t seem adequate, and more socially engaged and engaging robots will be needed. As the authors show, the quest to build machines that are capable of telling right from wrong has begun.
Moral Machines is the first book to examine the challenge of building artificial moral agents, probing deeply into the nature of human decision making and ethics.’
Book, Moral Machines: Teaching Robots Right from Wrong, ISBN-13: 9780195374049, first published January 2009
Meet the robot that helps you learn from afar
Mike Scialom, Cambridge Independent
PUBLISHED: 13:33 11 March 2018
The AV1 robot avatar – An innovative concept means an interactive education for students who can’t make it to lectures
AI: ‘It’s not just a playground’
Mike Scialom, Cambridge Independent
PUBLISHED: 11:04 16 November 2017
Is Prof Stephen Hawking right to say AI is a danger to humanity? Dr Stella Pachidi says machine intelligence needs closer scrutiny
Data is the fuel for AI
Newsletter Post, Data is the fuel for AI, so let’s ensure we get the ethics right, Birgitte Andersen, City AM Newsletter, December 2017
BBC News: MEPs vote on robots’ legal status – and if a kill switch is required
MEPs have called for the adoption of comprehensive rules for how humans will interact with artificial intelligence and robots.
The report makes it clear that it believes the world is on the cusp of a “new industrial” robot revolution.
It looks at whether to give robots legal status as “electronic persons”.
Designers should make sure any robots have a kill switch, which would allow functions to be shut down if necessary, the report recommends.
Meanwhile users should be able to use robots “without risk or fear of physical or psychological harm”, it states.
BBC News, MEPs vote on robots’ legal status – and if a kill switch isrequired, By Jane Wakefield, Technology reporter, January 2017
BBC News, Sex robots: Experts debate the rise of the love droids
Would you have sex with a robot? Would you marry one? Would a robot have the right to say no to such a union?
These were just a few of the questions being asked at the second Love and Sex with Robots conference hastily rearranged at Goldsmiths University in London after the government in Malaysia – the original location – banned it.
BBC News, Sex robots: Experts debate the rise of the love droids, By Jane Wakefield, Technology reporter, December 2016