Abstract
Emerging technologies like robotics for war and peace stress our moral norms and generate much public interest and controversy. We use this interest to attract participants to an innovative on-line survey platform, designed for experimenting with public engagement in the ethics of technology. In particular, the N-Reasons platform addresses several issues in democratic ethics: the cost of public participation, the methodological issue of feasible reflective ethical equilibrium (how can individuals in a large group, take into account the ethical views of all others?), and the reliability of public participation processes. We sketch the motivation and design of the N-Reasons platform, stressing the need for a practical (fast, low-cost) instrument that makes equilibrium feasible. We focus on the Robot Ethics Survey that featured a set of nine ethical challenges raised by robotics for war and peace. Over 400 people in five disjoint groups participated in this on-line survey experiment. We analyze the results, both quantitatively and qualitatively, the participants’ decisions taken and the reasons supporting these decisions. Both decisions and reasons strongly distinguished lethal military robotics from peace-related robotics. Methodologically, both decisions and reasons over five distinct groups were remarkably consistent.
Similar content being viewed by others
Notes
We invite you to take the survey or to use it in your own teaching. We can provide ways to identify groups of users’ responses for debriefing.
N-Reasons operationalizes a rather minimal concept of a reason: a decision paired by the prompt “because” to a text field. Further conditions could be imposed (and were in early prototypes). On reflection, we decided it would be better if conditions were imposed by the participants rather than by the designers. This furthers goal 2 of our research: to remove expert content from the design, as it burdens our premises and weakens our conclusions.
Participants in the first four groups saw reasons ranked by votes; the fifth group used our platform’s newest display ranking algorithm, which blends popularity and recency to mitigate the primacy effect favoring early contributions (see Danielson 2011a).
In this survey, we did not feed back (a chart of) the developing decision, although we plan to experiment again with this form of feedback, used in our pre-N-Reasons experimental designs. One can see what this would look like at http://yourviews.ubc.ca/en/Robot_Ethics_Results, which also includes full-page displays of each question.
We could, alternatively, look at the sample, but this survey was not designed for sample-based representativeness. Note also that the need to have all participants see the same questions over the 2-year run of the survey kept us from updating the scenarios, some of which are getting a bit out of date.
We include question 7 about the trolley problem, as it is relevant to the issue of survey reliability. Note that question 7 attracted the most neutral decisions, often around protest reasons like “I am boycotting trolley problems.”
The last group (class 4) used a revised platform that allowed split votes, allowing fractional vote totals.
References
Ahmad, R. (2009). “The risks of cognitive enhancers: a normative theory”. Paper presented at the The 4th International Conference on Applied Ethics, Sapporo.
Ahmad, R., Bornik, Z., Danielson, P., Dowlatabadi, H., Levy, E., Longstaf, H., et al. (2005). Innovations in web-based public consultation: Is public opinion on genomics influenced by social feedback? Paper presented at the First International Conference on e-Social Science, National Centre for e-Social Science, University of Manchester
Ahmad, R., Bailey, J., Bornik, Z., Danielson, P., Dowlatabadi, H., Levy, E., et al. (2006). A web-based instrument to model social norms: NERD design and results. Integrated Assessment, 6(2), 9–36.
Ahmad, R., Bailey, J., & Danielson, P. (2010). Analysis of an innovative survey platform: comparison of the public’s responses to human health and salmon genomics surveys. Public Understanding of Science, 19(2), 155–165.
Arkin, R. C. (2010). Governing lethal behavior in autonomous robots. Boca Raton, Fl: Chapman and Hall/CRC.
Danielson, P. (2006). From artificial morality to NERD: models, experiments, & robust reflective equilibrium. Paper presented at the Artificial Life 10: Achievements and Future Challenges for Artificial Life, Bloomington, Indiana.
Danielson, P. (2009a). Metaphors and models for data mining ethics. In E. Eyob (Ed.), Social implications of data mining and information privacy: Interdisciplinary frameworks and solutions (pp. 33–47). Hershey: IGI Global.
Danielson, P. A. (2009b). Review of Wendell Wallach and Colin Allen, moral machines: Teaching robots right from wrong. Notre Dame Philosophical Reviews, 2009.03.01.
Danielson, P. A. (2009c). Can robots have a conscience? Review of Wendell Wallach and Colin Allen, moral machines: teaching robots right from wrong. Nature, 457(29), 540.
Danielson, P. (2010a). A collaborative platform for experiments in ethics and technology. In I. van der Poel & D. E. Goldberg (Eds.), Philosophy and engineering: an emerging agenda (pp. 239–252). Berlin: Springer.
Danielson, P. (2010b). Designing a machine to learn about the ethics of robotics: the N-reasons platform. Ethics and Information Technology, 12(3), 251–261.
Danielson, P., Ahmad, R., Bornik, Z., Dowlatabadi, H., & Levy, E. (2007). Deep, cheap, and improvable: Dynamic democratic norms & the ethics of biotechnology. In F. Adams (Ed.), Ethics and the life sciences. Charlottesville: Philosophy Documentation Center.
Danielson, P., Mesoudi, A., & Stanev, R. (2008). NERD and norms: framework and experiments. Philosophy of Science, 75(5), 830–842.
Danielson, P. A., Longstaff, H., Ahmad, R., Van der Loos, H. F. M., Mitchell, I. M., Oishi, M. M. K., et al. (2010). Case study: An assistive technology ethics survey. In M. M. K. Oishi, I. M. Mitchell, & H. F. M. Van der Loos (Eds.), Design and use of assistive technology: Social, technical, ethical, and economic challenges (pp. 75–93). New York: Springer.
Danielson, P. A. (2011a). N-Reasons: computer mediated ethical decision support for public participation. In: E. Einsiedel & K. O’Doherty (eds.), Publics & Emerging Technologies: Cultures, Contexts, and Challenge. Vancouver: UBC Press.
Danielson, P. A. (2011b). Prototyping N-Reasons: A computer mediated ethics machine. In M. Anderson & E. Anderson (Eds.), Machine ethics (p. 9). New York: Cambridge University Press.
Fishkin, J. S. (1997). The voice of the people: public opinion and democracy. New Haven: Yale University Press.
Fishkin, J. S. (2006). Strategies of public consultation. Integrated Assessment, 6(2), 57–72.
Ormandy, E. H., Schuppli, C. A., Weary, D. M. (2009). Regulation increases public acceptance of animal-based science. Paper presented at 7th World Congress on Alternatives & Animal Use in the Life Sciences, Rome, Italy.
Schuppli, C. A., & Weary, D. M. (2007). Multiple uses of pigs: an interactive survey to assess people’s attitudes towards animal use and genetic modification. Paper presented at the Moving Mountains, 46th Annual Symposium of The Association for Laboratory Animal Science, Calgary, Alberta.
Solove, D. J. (2008). The future of reputation: Gossip, rumor, and privacy on the internet. New Haven: Yale University Press.
Acknowledgments
This research was funded by Genome Canada through the offices of Genome British Columbia. Thanks to the NERD research team for contribution to the survey design and implementation and to the participants for their enthusiastic involvement.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Danielson, P. Engaging the Public in the Ethics of Robots for War and Peace. Philos. Technol. 24, 239–249 (2011). https://doi.org/10.1007/s13347-011-0025-8
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s13347-011-0025-8