Agre, P. (1997). Toward a critical technical practice: lessons learned in trying to reform AI. In Bowker, G., Star, S., Gasser, L., & Turner, W. (Eds.) Social science, technical systems and cooperative work: beyond the great divide, psychology press. pp. 131–157.
Agüera y Arcas, B., Todorov, A., & Mitchell, M. (2018). Do algorithms reveal sexual orientation or just expose our stereotypes? Medium https://link.medium.com/GO7FJgFgM1.
Ali, S.M. (2016). A brief introduction to decolonial computing. XRDS: Crossroads The ACM Magazine for Students, 22(4), 16–21.
AMA. (2018). Augmented intelligence in health care H-480.940. American Medical Association PolicyFinder.
Amodei, D., Olah, C., Steinhardt, J., Christiano, P., Schulman, J., & Mané, D. (2016). Concrete problems in AI safety. arXiv:160606565.
Angwin, J., Larson, J., Mattu, S., & Kirchner, L. (2016a). Machine bias. Propublica, May 23:2016.
Angwin, J., Larson, J., Mattu, S., & Kirchner, L. (2016). Machine bias. There’s software used across the country to predict future criminals. and it’s biased against blacks. ProPublica. https://www.propublica.org/article/machine-bias-risk-assessments-in-criminal-sentencing.
Arora, P. (2019). Decolonizing privacy studies. Television & New Media, 20(4), 366–378.
Ashcroft, B. (2006). The post-colonial studies reader. Taylor & Francis, Tiffin, H.
Asilomar Meeting. (2017). Asilomar AI principles. https://futureoflife.org/ai-principles/.
Awori, K., Bidwell, NJ., Hussan, TS., Gill, S., & Lindtner, S. (2016). Decolonising technology design. In Proceedings of the first African conference on human computer interaction, pp. 226–228.
Balaram, B., Greenham, T., & Leonard, J. (2018). Artificial intelligence: real public engagement. London: RSA. https://www.thersaorg/discover/publications-and-articles/reports/artificial-intelligence-realpublic-engagement.
Barocas, S., & Selbst, A.D. (2016). Big data’s disparate impact. California Law Review, 104, 671.
Belmont Commission. (1978). The Belmont report: ethical principles and guidelines for the protection of human subjects of research, vol 1. United States National Commission for the Protection of Human Subjects of Biomedical and Behavioral Research.
Benjamin, R. (2019). Race after technology: abolitionist tools for the new Jim Code. New York: John Wiley & Sons.
Bhambra, G.K., Nisancioglu, K., & Gebrial, D. (2018). Decolonizing the university. London: Pluto Press.
Bhattacharyya, G., & et al. (2018). Rethinking racial capitalism: questions of reproduction and survival. Maryland: Rowman & Littlefield International.
Boden, M.A. (2018). Artificial intelligence: a very short introduction. London: Oxford University Press.
Bonneuil, C. (2000). Development as experiment: science and state building in late colonial and postcolonial Africa, 1930-1970. Osiris, 15, 258–281.
Brandt, A.M. (1978). Racism and research: the case of the Tuskegee syphilis study, (pp. 21–29). New York: Hastings Center Report.
Bruhn, M., & Gallego, F.A. (2012). Good, bad, and ugly colonial activities: do they matter for economic development?. Review of Economics and Statistics, 94(2), 433–461.
Bueter, A. (2015). The irreducibility of value-freedom to theory assessment. Studies in History and Philosophy of Science Part A, 49, 18–26.
Bullington, J., & Lane, E. (2018). How a tech firm brought data and worry to New Orleans crime fighting. The New Orleans Times-Picayune. https://www.nola.com/news/crime_police/article_33b8bf05-722f-5163-9a0c-774aa69b6645.html.
Buolamwini, J., & Gebru, T. (2018). Gender shades: intersectional accuracy disparities in commercial gender classification. In Conference on fairness, accountability and transparency, pp 77–91.
Butorac, S.K. (2018). Hannah Arendt, James Baldwin, and the politics of love. Political Research Quarterly, 71(3), 710–721.
Canetti, R., Cohen, A., Dikkala, N., Ramnarayan, G., Scheffler, S., & Smith, A. (2019). From soft classifiers to hard decisions: how fair can we be?. In Proceedings of the conference on fairness, accountability, and transparency (pp. 309–318). New York: ACM.
Capurro, R. (2018). Intercultural information ethics. In Localizing the Internet (pp. 19–38). Munich: Wilhelm Fink Verlag.
Champion, T. (2005). Metropole and margin: the dependency theory and the political economy of the Solomon Islands, 1880–1980. In Centre and periphery (pp. 43–60). Evanston: Routledge.
Chen, I.Y., Szolovits, P., & Ghassemi, M. (2019). Can AI help reduce disparities in general medical and mental health care?. AMA journal of ethics, 21(2), 167–179.
Chiappa, S., & Isaac, W.S. (2019). A Causal Bayesian Networks Viewpoint on Fairness. In Kosta, E., Pierson, J, Slamanig, D., Fischer-Hübner, S., & Krenn, S. (Eds.) Privacy and identity management. Fairness, accountability, and transparency in the age of big data. Privacy and identity 2018. IFIP advances in information and communication technology, (Vol. 547 pp. 3–20). Cham.
Coates, J.F., & at al. (1985). Foresight in federal government policy making. Futures Research Quarterly, 1(2), 29–53.
Costanza-Chock, S. (2018). Design justice, AI, and escape from the matrix of domination. Journal of Design and Science.
Couldry, N., & Mejias, U.A. (2019a). The costs of connection: how data is colonizing human life and appropriating it for capitalism. Stanford: Stanford University Press.
Couldry, N., & Mejias U.A. (2019b). Data colonialism: rethinking big data’s relation to the contemporary subject. Television & New Media, 20(4), 336–349.
Dastin, J. (2018). Amazon scraps secret AI recruiting tool that showed bias against women. San Francisco, CA: Reuters Retrieved on October 9:2018.
Derrida, J. (1993). Politics of friendship. American Imago, 50 (3), 353–391.
D’Ignazio, C., & Klein, L.F. (2020). Data feminism. Cambridge: MIT Press.
DiSalvo, C. (2012). Adversarial design (design thinking, design theory). Cambridge: MIT Press.
Douglas, H. (2007). Rejecting the ideal of value-free science. In Kincaid, H., Dupré, J., & Wylie, A. (Eds.) Value-free science: ideals and illusions? chap 6 (pp. 120–141). Oxford: Oxford university press.
Dourish, P., & Mainwaring, S.D. (2012). Ubicomp’s colonial impulse. In Proceedings of the 2012 ACM conference on ubiquitous computing (pp. 133–142). New York: ACM.
Dourish, P., Finlay, J., Sengers, P., & Wright, P. (2004). Reflective HCI: towards a critical technical practice. In Conference on human factors in computing systems: CHI’04 extended abstracts on human factors in computing systems, (Vol. 29 pp. 1727–1728).
Directorate of Science, Technology, and Innovation in the Office of the President, Sierra Leone. (2019). Sierra Leone National Innovation & Digital Strategy 2019 - 2029. Digitization for all: Identity, Economy, and Governance.
Dutton, T. (2018). An overview of national AI strategies. Medium. https://link.medium.com/jqPZBjs7j2.
Dwork, C., Hardt, M., Pitassi, T., Reingold, O., & Zemel R. (2012). Fairness through awareness. In Proceedings of the 3rd innovations in theoretical computer science conference, pp 214–226.
Edgerton, D. (2007a). Creole technologies and global histories: rethinking how things travel in space and time. Journal of History of Science and Technology, 1(1), 75–112.
Edgerton, D.E. (2007b). The contradictions of techno-nationalism and techno-globalism: a historical perspective. New Global Studies 1(1).
El Khayat, R., & Khatibi, A. (2010). Open correspondence: an epistolary dialogue. UNO press translated by Babana-Hampton S, Orlando VK, Vogl M.
Elliott, K.C., & McKaughan, D.J. (2014). Nonepistemic values and the multiple goals of science. Philosophy of Science, 81(1), 1–21.
Escobar, A. (2011). Sustainability: design for the pluriverse. Development, 54(2), 137–140.
Ess, C. (2006). Ethical pluralism and global information ethics. Ethics and Information Technology, 8(4), 215–226.
Eubanks, V. (2018). Automating inequality: how high-tech tools profile, police, and punish the poor. New York: St. Martin’s Press.
Fanon, F. (1961). The wretched of the earth. United States: Grove/Atlantic, Inc.
Fanon, F. (1967). Black skin, white masks. New York: Grove press.
Feenberg, A. (2017). Critical theory of technology and STS. Thesis Eleven, 138(1), 3–12.
Floridi, L. (2019). Translating principles into practices of digital ethics: five risks of being unethical. Philosophy & Technology, 32(2), 185–193.
Floridi, L., Cowls, J., Beltrametti, M., Chatila, R., Chazerand, P., Dignum, V., Luetge, C., Madelin, R., Pagallo, U., Rossi, F., & et al. (2018). AI4People—an ethical framework for a good AI society: opportunities, risks, principles, and recommendations. Minds and Machines, 28(4), 689–707.
Forsythe, D. (2001). Studying those who study us: an anthropologist in the world of artificial intelligence. Stanford: Stanford University Press.
Forsythe, D.E. (1993). Engineering knowledge: the construction of knowledge in artificial intelligence. Social studies of science, 23(3), 445–477.
Freudenthal, E. (2019). Ebola’s lost blood: row over samples flown out of africa as’ big pharma’set to cash in. The Telegraph.
Friedman, B., Kahn, P.H., Borning, A., & Huldtgren, A. (2013). Value sensitive design and information systems. In Early engagement and new technologies: opening up the laboratory (pp. 55–95). Berlin: Springer.
Gabriel, I. (2020). Artificial intelligence, values and alignment. arXiv:200109768.
Gallopin, G. (1992). Science, technology and the ecological future of Latin America. World Development, 20(10), 1391–1400.
Gandhi, L. (2006). Affective communities: anticolonial thought, fin-de-siècle radicalism, and the politics of friendship. Durham: Duke University Press.
Garcia, E. (2019). The militarization of artificial intelligence: a wake-up call for the global south. Available at SSRN 3452323.
Gavighan, C., Knott, A., Maclaurin, J., Zerilli, J., & Liddicoat, J. (2019). Government use of artificial intelligence in New Zealand. New Zealand: Law Society.
Gebru, T., Morgenstern, J., Vecchione, B., Vaughan, J.W., Wallach, H., Daumé III, H., & Crawford, K. (2018). Datasheets for datasets. arXiv:180309010.
Gerrish, S. (2018). How smart machines think. Cambridge: MIT Press.
Gershgorn, D. (2019). Africa is building an AI industry that doesn’t look like silicon valley. Medium OneZero. https://bit.ly/2SBnQFm.
Gomberg-Muñoz, R.M. (2018). Review essay: law and migrant labor in the 20th century: Ghost workers and global capitalism. PoLAR: Political and Legal Anthropology Review.
Gopal, P. (2019). Insurgent empire: anticolonial resistance and british dissent. London: Verso Books.
Goyanes, R. (2018). Data for black lives is an organization using analytics as a tool for social change. Garage Magazine. https://garage.vice.com/en_us/article/kzn4jn/data-for-black-lives-is-an-organization-using-analytics-as-a-tool-for-social-change.
Gray, M.L., & Suri, S. (2019). Ghost work: how to stop silicon valley from building a new global underclass. Eamon Dolan Books.
Green, B. (2019). “Good” isn’t good enough. In NeurIPS workshop on AI for social good.
Greene, D., Hoffmann, A.L., & Stark, L. (2019). Better, nicer, clearer, fairer: a critical assessment of the movement for ethical artificial intelligence and machine learning. In Proceedings of the 52nd hawaii international conference on system sciences.
Hacking, I. (2015). Biopower and the avalanche of printed numbers. Biopower: Foucault and beyond, 65–80.
Hanna, A., Denton, E., Smart, A., & Smith-Loud, J. (2019). Towards a critical race methodology in algorithmic fairness. In Proceedings of the 2020 conference on fairness, accountability, and transparency.
Hao, K. (2019). An AI startup has found a new source of cheap labor for training algorithms: prisoners. Cambridge: MIT Tech Review . https://www.technologyreview.com/f/613246/an-ai-startup-has-found-a-new-source-of-cheap-labor-for-training-algorithms/.
Harding, S. (2011). The postcolonial science and technology studies reader. Durham: Duke University Press.
Harvey, D. (2004). The ‘new’ imperialism: accumulation by dispossession. Socialist Register, 40, 63–87.
Hogarth, I. (2018). AI nationalisms. https://www.ianhogarth.com/blog/2018/6/13/ai-nationalism.
IEEE Global Initiative. (2016). Ethically aligned design. IEEE Standards v1.
Irani, L., Vertesi, J., Dourish, P., Philip, K., & Grinter, R.E. (2010). Postcolonial computing: a lens on design and development. In Proceedings of the SIGCHI conference on human factors in computing systems (pp. 1311–1320). New York: ACM.
Isaac, W.S. (2017). Hope, hype, and fear: the promise and potential pitfalls of artificial intelligence in criminal justice. Ohio State Journal of Criminal Law, 15, 543.
ITU. (2019). United nations activities on artificial intelligence (AI). Geneva: International Telecommunication Union. https://www.itu.int/dms_pub/itu-s/opb/gen/S-GEN-UNACT-2019-1-PDF-E.pdf.
James, C.L.R. (1993). Beyond a boundary. Durham: Duke University Press.
Jansen, J. (2019). Decolonisation in universities: the politics of knowledge. Wits University Press.
Jasanoff, S., & Hurlbut, J.B. (2018). A global observatory for gene editing. Nature, 555(7697), 435–437.
Jobin, A., Ienca, M., & Vayena, E. (2019). The global landscape of AI ethics guidelines. Nature Machine Intelligence, 1(9), 389–399.
Johnston, K. (2019). A comparison of two smart cities: Singapore and Atlanta. Journal of Comparative Urban Law and Policy, 3, 191.
Jung, C., Kearns, M., Neel, S., Roth, A., Stapleton, L., & Wu, Z.S. (2019). Eliciting and enforcing subjective individual fairness. arXiv:190510660.
Kanth, D.R. (2019). India boycotts ‘Osaka Track’ at G20 summit. Live Mint. https://www.livemint.com/news/world/india-boycotts-osaka-track-at-g20-summit-1561897592466.html.
Katell, M., Young, M., Dailey, D., Herman, B., Guetler, V., Tam, A., Binz, C., Raz, D., & Krafft, P. (2020). Toward situated interventions for algorithmic equity: lessons from the field. In Proceedings of the 2020 conference on fairness, accountability, and transparency, pp 45–55.
Keyes, O. (2018). The misgendering machines: trans/HCI implications of automatic gender recognition. Proceedings of the ACM on Human-Computer Interaction, 2(CSCW), 88.
Kiros, T. (1992). Moral philosophy and development: the human condition in Africa, vol 61. Ohio Univ Ctr for International Studies.
Latonero, M. (2019). Stop surveillance humanitarianism. New York Times. https://www.nytimes.com/2019/07/11/opinion/data-humanitarian-aid.html.
Laudan, L. (1968). Theories of scientific method from Plato to mach: a bibliographical review. History of science, 7(1), 1–63.
Law, J., & et al. (1987). Technology and heterogeneous engineering: the case of Portuguese expansion. The social construction of technological systems:, New directions in the sociology and history of technology, 1, 1–134.
LeCun, Y., Bengio, Y., & Hinton, G. (2015). Deep learning. Nature, 521(7553), 436–444.
Lee, K.F. (2017). The real threat of artificial intelligence. The New York Times 24.
Lee, K.F. (2018). AI superpowers: China, Silicon Valley, and the New World Order. Houghton Mifflin Harcourt.
Lum, K., & Isaac, W. (2016). To predict and serve?. Significance, 13(5), 14–19.
Madras, D., Creager, E., Pitassi, T., & Zemel, R. (2019). Fairness through causal awareness: learning causal latent-variable models for biased data. In Proceedings of the conference on fairness, accountability, and transparency (pp. 349–358). New York: ACM.
Maldonado-Torres, N. (2007). On the coloniality of being: contributions to the development of a concept. Cultural studies, 21(2-3), 240–270.
Marda, V., & Narayan, S. (2020). Data in New Delhi’s predictive policing system. In Proceedings of the 2020 conference on fairness, accountability, and transparency, pp. 317–324.
McClintock, A. (1992). The angel of progress: pitfalls of the term “post-colonialism”. Social text (31/32), 84–98.
McDowell, C., & Chinchilla, M.Y. (2016). 30 partnering with communities and institutions. Civic Media: Technology, Design, Practice , 461.
Mezzadra, S., & Neilson, B. (2017). On the multiple frontiers of extraction: excavating contemporary capitalism. Cultural Studies, 31(2-3), 185–204.
Mignolo, W.D. (2007). Introduction: coloniality of power and de-colonial thinking. Cultural studies, 21(2-3), 155–167.
Mignolo, W.D. (2012). Local histories/global designs: coloniality, subaltern knowledges, and border thinking. Princeton: Princeton University Press.
Mikesell, L., Bromley, E., & Khodyakov, D. (2013). Ethical community-engaged research: a literature review. American Journal of Public Health, 103 (12), e7–e14.
Milan, S. (2013). Social movements and their technologies: wiring social change. Berlin: Springer.
Milan, S., & Treré, E. (2019). Big data from the south (s): beyond data universalism. Television & New Media, 20(4), 319–335.
Milan, S., & Van der Velden, L. (2016). The alternative epistemologies of data activism. Digital Culture & Society, 2(2), 57–74.
Mitchell, M., Wu, S., Zaldivar, A., Barnes, P., Vasserman, L., Hutchinson, B., Spitzer, E., Raji, I.D., & Gebru, T. (2019). Model cards for model reporting. In Proceedings of the conference on fairness, accountability, and transparency (pp. 220–229). New York: ACM.
Mitchell, S., Potash, E., Barocas, S., D’Amour, A., & Lum, K. (2018). Prediction-based decisions and fairness: a catalogue of choices, assumptions, and definitions. arXiv:181107867.
Mitchell, T.M. (2006). The discipline of machine learning. Tech. Rep CMU-ML-06-108, Machine Learning Department, Carnegie Mellon University.
Nabi, R., & Shpitser, I. (2018). Fair inference on outcomes. In Thirty-second AAAI conference on artificial intelligence.
Nandy, A. (1989). Intimate enemy: loss and recovery of self under colonialism. Oxford: Oxford University Press Oxford.
Ndlovu-Gatsheni, S.J. (2015). Decoloniality as the future of Africa. History Compass, 13(10), 485–496.
Nelson, A. (2002). Unequal treatment: confronting racial and ethnic disparities in health care. Journal of the National Medical Association, 94(8), 666.
Nissenbaum, H. (2001). How computer systems embody values. Computer, 34(3), 120–119.
Noble, S.U. (2018). Algorithms of oppression: how search engines reinforce racism. New York: NYU Press.
Nuremberg Military Tribunals. (1947). Permissible medical experiments. In Trials of war criminals before the nuremberg military tribunals under control council law No. 10, vol 2, U.S. government printing office, pp 181–182.
Nyabola, N. (2018). Digital democracy, analogue politics: how the Internet era is transforming politics in kenya. Zed Books Ltd.
Nyawa, J.M. (2019). The big brother is watching: Huduma Namba a threat to our rights and freedoms. Available at SSRN 3389268.
Obermeyer, Z., Powers, B., Vogeli, C., & Mullainathan, S. (2019). Dissecting racial bias in an algorithm used to manage the health of populations. Science, 366(6464), 447–453.
OECD. (2019). OECD principles on artificial intelligence. https://www.oecd.org/going-digital/ai/principles/.
O’Neil, C. (2016). Weapons of math destruction: how big data increases inequality and threatens democracy. Broadway Books.
Pappas, G.F. (2017). The limitations and dangers of decolonial philosophies: lessons from Zapatista Luis Villoro. Radical Philosophy Review.
Pathways for Prosperity. (2019). Digital diplomacy: technology governance for developing countries. Pathways for Prosperity Commission on Technology and Inclusive Development . https://pathwayscommission.bsg.ox.ac.uk/sites/default/files/2019-10/Digital-Diplomacy.pdf.
Philip, K., Irani, L., & Dourish, P. (2012). Postcolonial computing: a tactical survey. Science, Technology, & Human Values, 37(1), 3–29.
Van de Poel, I., & Kroes, P. (2014). Can technology embody values?. In The moral status of technical artefacts (pp. 103–124). Berlin: Springer.
Pollard, J., Mcewan, C., & Hughes, A. (2011). Postcolonial economies. London/New York: Zed Books.
Quijano, A. (2000). Coloniality of power and Eurocentrism in Latin Americas. International Sociology, 15(2), 215–232.
Quijano, A. (2007). Coloniality and modernity/rationality. Cultural studies, 21(2-3), 168–178.
Raji, I.D., & Dobbe, R. (2020). Concrete problems in AI safety, revisited. In ICLR workshop on ML in the real world.
Ramose, M.B. (1999). African philosophy through Ubuntu. Mond Books.
Ricaurte, P. (2019). Data epistemologies, the coloniality of power, and resistance. Television & New Media, 20(4), 350–365.
Richardson, R., Schultz, J., & Crawford, K. (2019). Dirty data, bad predictions: how civil rights violations impact police data, predictive policing systems and justice. New York University Law Review Online, Forthcoming.
Rock, D., & Grant, H. (2016). Why diverse teams are smarter. Harvard Business Review, 4(4), 2–5.
Ronconi, L. (2015). Enforcement and the effective regulation of labor. Tech. rep., IDB Working Paper Series.
Røpke, I. (2001). New technology in everyday life–social processes and environmental impact. Ecological economics, 38(3), 403–422.
Russell S, & Norvig P. (2016). Artificial intelligence: a modern approach. New Jersey: Prentice Hall.
Said, E.W. (1993). Culture and imperialism. Vintage.
Sap, M., Card, D., Gabriel, S., Choi, Y., & Smith, N.A. (2019). The risk of racial bias in hate speech detection. In Proceedings of the 57th annual meeting of the association for computational linguistics, pp 1668–1678.
Schroeder, D., Cook Lucas, J., Hirsch, F., Fenet, S., & Muthuswamy, V. (2018). Ethics dumping case studies from north-south research collaborations. Cham: Springer International Publishing.
Selbst, A.D., Boyd, D., Friedler, S.A., Venkatasubramanian, S., & Vertesi, J. (2019). Fairness and abstraction in sociotechnical systems. In Proceedings of the conference on fairness, accountability, and transparency, (pp. 59–68). New York: ACM.
Sengers, P., Boehner, K., David, S., & Kaye, J. (2005). In Proceedings of the 4th decennial conference on critical computing: between sense and sensibility, pp 49–58: Reflective design.
Sengers, P., McCarthy, J., & Dourish, P. (2006). Reflective HCI: Articulating an agenda for critical practice. In CHI’06 extended abstracts on Human factors in computing systems (pp. 1683–1686). New York: ACM.
Senior, E. (2018). The Caribbean and the medical imagination, 1764-1834: slavery, disease and colonial modernity Vol. 119. Cambridge: Cambridge University Press.
Shore, N. (2006). Re-conceptualizing the Belmont report: a community-based participatory research perspective. Journal of Community Practice, 14 (4), 5–26.
Siddiqui, A.U., & Singh, H.K. (2015). “Aadhar” management system. IITM Journal of Management and IT, 6(1), 40–43.
Sokoloff, N.J., & Pincus, F.L. (2008). Introduction: race, class, gender, and capitalism. Race, Gender & Class, 4–8.
Stark, L. (2019). Facial recognition is the plutonium of AI. XRDS: Crossroads. The ACM Magazine for Students, 25(3), 50–55.
Steiner, C.B. (1994). Technologies of resistance: structural alteration of trade cloth in four societies. Zeitschrift fü,r Ethnologie, pp. 75–94.
Stoler, A.L. (2008). Epistemic politics: ontologies of colonial common sense. In The philosophical forum, (Vol. 39 pp. 349–361). New Jersey: Wiley Online Library.
Taylor, E. (2016). Groups and oppression. Hypatia, 31(3), 520–536.
Thatcher, J., O’Sullivan, D., & Mahmoudi, D. (2016). Data colonialism through accumulation by dispossession: new metaphors for daily data. Environment and Planning D: Society and Space, 34(6), 990–1006.
Thrush, C. (2008). American curiosity: cultures of natural history in the colonial British Atlantic world. Environmental History, 13(3), 573.
Tilley, H. (2014). Conclusion: experimentation in colonial East Africa and beyond. The International Journal of African Historical Studies, 47(3), 495–505.
Tomašev, N, Cornebise, J, Hutter, F, Mohamed, S, Picciariello, A, Connelly, B, Belgrave, DCM, Ezer, D, van der Haert, FC, Mugisha, F, Abila, G, Arai, H, Almiraat, H, Proskurnia, J, Snyder, K, Otake-Matsuura, M, Othman, M, Glasmachers, T, de Wever, W, Teh, YW, Khan, ME, De Winne, R, Schaul, T, & Clopath, C. (2020). AI for social good: unlocking the opportunity for positive impact. Nature Communications, 11(1), 1–6. Nature Publishing Group.
Toyama, K. (2015). Geek heresy: rescuing social change from the cult of technology. New York: PublicAffairs.
UK National Health Service. (2019). Code of conduct for data-driven health and care technology.
UNCTAD. (2013). Information economy report 2013: The cloud economy and developing countries. Geneva: United Nations Conference on Trade and Development.
Vaithianathan, R., Maloney, T., Putnam-Hornstein, E., & Jiang, N. (2013). Children in the public benefit system at risk of maltreatment: identification via predictive modeling. American journal of preventive medicine, 45(3), 354–359.
Vinuesa, R., Azizpour, H., Leite, I., Balaam, M., Dignum, V., Domisch, S., Felländer, A, Langhans, S.D., Tegmark, M., & Nerini, F.F. (2020). The role of artificial intelligence in achieving the sustainable development goals. Nature Communications, 11(1), 1–10.
Vitak, J., Shilton, K., & Ashktorab, Z. (2016). Beyond the Belmont principles: ethical challenges, practices, and beliefs in the online data research community. In Proceedings of the 19th ACM conference on computer-supported cooperative work & social computing (pp. 941–953). New York: ACM.
Wa Thiong’o, N. (1992). Decolonising the mind: the politics of language in African literature. East African Publishers.
Wallerstein, I. (1987). World-systems analysis.
Washington, H.A. (2006). Medical apartheid: the dark history of medical experimentation on Black Americans from colonial times to the present. Doubleday Books.
Whittlestone, J., Nyrup, R., Alexandrova, A., Dihal, K., & Cave, S. (2019). Ethical and societal implications of algorithms, data, and artificial intelligence: a roadmap for research. London: Nuffield Foundation.
Winner, L. (1980). Do artifacts have politics? Daedalus, 121–136.
WMA General Assembly. (1964). Ethical principles for medical research involving human subjects. World Medical Association Declaration of Helsinki. https://www.ncbi.nlm.nih.gov/pmc/articles/PMC2566407/pdf/11357217.pdf.
Wong, P.H. (2012). Dao, harmony and personhood: towards a confucian ethics of technology. Philosophy & technology, 25(1), 67–86.
Young, M., Magassa, L., & Friedman, B. (2019). Toward inclusive tech policy design: a method for underrepresented voices to strengthen tech policy documents. Ethics and Information Technology, 21(2), 89–103.
Yuan, L. (2018). How cheap labor drives China’s A.I. ambitions. New York: New York Times. https://www.nytimes.com/2018/11/25/business/china-artificial-intelligence-labeling.html.
Zembylas, M. (2017). Love as ethico-political practice: inventing reparative pedagogies of aimance in “disjointed” times. Journal of curriculum and pedagogy, 14(1), 23–38.
Zuboff, S. (2019). The age of surveillance capitalism: the fight for a human future at the new frontier of power. London: Profile Books.