Abstract
As of 2021, more than 30 countries have released national artificial intelligence (AI) policy strategies. These documents articulate plans and expectations regarding how AI will impact policy sectors, including education, and typically discuss the social and ethical implications of AI. This article engages in thematic analysis of 24 such national AI policy strategies, reviewing the role of education in global AI policy discourse. It finds that the use of AI in education (AIED) is largely absent from policy conversations, while the instrumental value of education in supporting an AI-ready workforce and training more AI experts is overwhelmingly prioritized. Further, the ethical implications of AIED receive scant attention despite the prominence of AI ethics discussion generally in these documents. This suggests that AIED and its broader policy and ethical implications—good or bad—have failed to reach mainstream awareness and the agendas of key decision-makers, a concern given that effective policy and careful consideration of ethics are inextricably linked, as this article argues. In light of these findings, the article applies a framework of five AI ethics principles to consider ways in which policymakers can better incorporate AIED’s implications. Finally, the article offers recommendations for AIED scholars on strategies for engagement with the policymaking process, and for performing ethics and policy-oriented AIED research to that end, in order to shape policy deliberations on behalf of the public good.
Similar content being viewed by others
Availability of data and material
Not applicable
Code availability
Not applicable
Notes
Australia, Canada, Singapore, Denmark, Denmark, Taiwan, France, EU, UK, and South Korea have committed nearly 8 billion with the US contributing or planning to contribute at least 4 billion and China at least 14 billion. These are moving targets and low-end estimates, especially as private investment constitutes an even greater sum.
I do not include documents produced by intergovernmental bodies, such as the United Nations or European Union. While these documents are similar in nature, they are less tied to direct national institutions and strategies, and are therefore less analogous to the other documents. For further details and rationales regarding data collection and inclusion/exclusion criteria, please see Appendix 1.
Further details about codebook development and iteration are available in Appendix 1.
For example, Malta’s (2019) document initially notes that AI for healthcare may be amongst the highest impact projects and “most pressing domestic challenges” worthy of prioritization, but it does not proceed to include any substantive discussion or a subsection on healthcare. In comparison to the document’s discussion of other topics, and in comparison to other countries’ AI policy documents that discuss healthcare in more depth, this relatively more narrow treatment of the topic led to coding it as yellow. Similarly, Russia’s (2019) brief mention of using AI to “[improve] the quality of education services” does not provide enough detail to be clear about the role of AIED as a potential tool for teaching and learning, and so is considered too ambiguous to code as either green or red.
Any errors are the sole fault of the author.
I nevertheless captured these mentions in the research memo and discuss them at the end of this section.
Note that other policy sectors, like healthcare, also have dedicated agencies and other policy documents, but nevertheless receive more attention than education in AI policy strategies.
See The World Bank Country and Lending Groups classification for classifications by national income.
There are additional explanations to consider as well, though this study does not provide clear evidence to establish these. First, policymakers may simply never have been informed about AIED. AIED has typically been contained within an expert academic domain primarily accessible to computer scientists (Schiff 2021). Relatedly, while some AIED applications are beginning to hit mainstream classrooms, relatively few people, adults or children, have experience with them personally. In contrast, general education and its role in the labor market is something that nearly all members of society experience and can relate to, and a traditional focus of policymakers. Barring basic awareness, policymakers may not realize the transformative potential of AIED. If this is accurate, a clear prescription is to significantly increase efforts to inform policymakers about AIED and its ethical implications. This begs the question of whether the AIED community is prepared to do so, something which this article addresses in its Recommendations for AIED Researchers section.
The discussions of Spain, Mexico, Kenya, and India demonstrate how such a link between ethics and policy for AIED might be established even though most countries have not yet identified these connections explicitly.
Note that a similar approach has also been adopted by The Institute for Ethical AI in Education (2020), which, through a series of workshops and reports, has explored AIED policy by using a different AI ethics framework, the EU’s seven Ethics Guidelines for Trustworthy AI (European Commission 2019). This provides further support to the idea of approaching AI governance through an ethical lens.
References
Acemoglu, D., & Restrepo, P. (2019). The wrong kind of AI? Artificial intelligence and the future of labor demand. Working Paper 25682. National Bureau of Economic Research. https://doi.org/10.3386/w25682.
Adadi, A., & Berrada, M. (2018). Peeking inside the black-box: A survey on explainable artificial intelligence (XAI). IEEE Access, 6, 52138–52160. https://doi.org/10.1109/ACCESS.2018.2870052.
Ali, W. (2020). Online and remote learning in higher education institutes: A necessity in light of COVID-19 pandemic. Higher Education Studies, 10(3), 16–25.
Allen, G. C. (2019). Understanding China’s AI strategy. Center for a New American Security (blog). February 6, 2019. https://www.cnas.org/publications/reports/understanding-chinas-ai-strategy.
Alper, J. and U.S. National Academies of Sciences, Engineering, and Medicine, eds. 2016. Developing a national STEM workforce strategy: A workshop summary. Washington, DC: The National Academies Press.
Amiel, T., & Reeves, T. C. (2008). Design-based research and educational technology: Rethinking technology and the research agenda. Journal of Educational Technology and Society, 11(4), 29–40.
Article 19. 2021. “Emotional Entanglement: China’s Emotion Recognition Market and Its Implications for Human Rights.” A19/DIG/2021/001. London, UK: Article 19. https://www.article19.org/wp-content/uploads/2021/01/ER-Tech-China-Report.pdf.
Autor, D. H. (2015). Why are there still so many jobs? the history and future of workplace automation. Journal of Economic Perspectives, 29(3), 3–30. https://doi.org/10.1257/jep.29.3.3
Bagloee, S. A., Tavana, M., Asadi, M., & Oliver, T. (2016). Autonomous vehicles: Challenges, opportunities, and future implications for transportation policies. Journal of Modern Transportation, 24(4), 284–303. https://doi.org/10.1007/s40534-016-0117-3
Beauchamp, T. L., & Childress, J. F. (1979). Principles of biomedical ethics. Oxford University Press.
Bellamy, R. K., Dey, K., Hind, M., Hoffman, S. C., Houde, S., Kannan, K., Lohia, P. et al. (2018). AI fairness 360: An extensible toolkit for detecting, understanding, and mitigating unwanted algorithmic bias. [Cs], October. http://arxiv.org/abs/1810.01943.
Blockchain & Artificial Intelligence Taskforce, Kenya. 2019. “Kenya: Emerging Digital Technologies for Kenya: Exploration & Analysis.” Nairobi, Kenya: Blockchain & Artificial Intelligence taskforce. https://www.ict.go.ke/blockchain/blockchain.pdf.
Boeglin, J. (2015). The costs of self-driving cars: reconciling freedom and privacy with tort liability in autonomous vehicle regulation. Yale JL & Tech., 17, 171.
Boin, A. ’t Hart, P., & McConnell, A. (2009). Crisis exploitation: Political and policy impacts of framing contests. Journal of European Public Policy, 16, 81-106. https://doi.org/10.1080/13501760802453221
Borenstein, J., & Arkin, R. C. (2017). Nudging for good: Robots and the ethical appropriateness of nurturing empathy and charitable behavior. AI & SOCIETY, 32(4), 499–507.
Bozeman, B. (2002). Public-value failure: When efficient markets may not do. Public Administration Review, 62(2), 145–161. https://doi.org/10.1111/0033-3352.00165
Bravo-Biosca, A. (2019). Experimental Innovation Policy. Working Paper 26273. National Bureau of Economic Research. https://doi.org/10.3386/w26273.
British Embassy in Mexico, Oxford Insights, and C Minds. 2018. “Mexico: Towards an AI Strategy in Mexico: Harnessing the AI Revolution.” Mexico City, Mexico: British Embassy in Mexico, Oxford Insights, and C Minds. http://go.wizeline.com/rs/571-SRN-279/images/Towards-an-AI-strategy-in-Mexico.pdf.
Bucchi, M. (2013). Style in science communication. Public Understanding of Science, 22(8), 904–915. https://doi.org/10.1177/0963662513498202
Buchanan, R. (2020). Through growth to achievement: Examining Edtech as a solution to australia’s declining educational achievement. Policy Futures in Education, 18, 1026–1043. https://doi.org/10.1177/1478210320910293
Cairney, P., Oliver, K., & Wellstead, A. (2016). To bridge the divide between evidence and policy: reduce ambiguity as much as uncertainty. Public Administration Review, 76(3), 399–402. https://doi.org/10.1111/puar.12555
Caliskan, A., Bryson, J. J., & Narayanan, A. (2017). Semantics derived automatically from language corpora contain human-like biases. Science, 356(6334), 183–186. https://doi.org/10.1126/science.aal4230
Castelvecchi, D. (2016). Can we open the black box of AI? Nature News, 538(7623), 20. https://doi.org/10.1038/538020a
Castleberry, A., & Nolen, A. (2018). Thematic analysis of qualitative research data: Is it as easy as it sounds? Currents in Pharmacy Teaching and Learning, 10(6), 807–815. https://doi.org/10.1016/j.cptl.2018.03.019
Cath, C., Wachter, S., Mittelstadt, B., Taddeo, M., & Floridi, L. (2018). Artificial Intelligence and the ‘Good Society’: The US, EU, and UK approach. Science and Engineering Ethics, 24(2), 505–528. https://doi.org/10.1007/s11948-017-9901-7
Chae, Y. (2020). U.S. AI regulation guide: Legislative overview and practical considerations. The Journal of Robotics, Artificial Intelligence and Law, 3(1), 17–40.
Char, D. S., Shah, N. H., & Magnus, D. (2018). Implementing machine learning in health care—addressing ethical challenges. The New England Journal of Medicine, 378(11), 981–983. https://doi.org/10.1056/NEJMp1714229
Chatila, R., & Havens, J. C. (2019). The IEEE global initiative on ethics of autonomous and intelligent systems. In M. I. A. Ferreira, J. S. Sequeira, G. S. Virk, M. O. Tokhi, & E. E. Kadar (Eds.), Robotics and well-being (pp. 11–16). Cham: Springer.
Chouldechova, A. (2017). Fair prediction with disparate impact: a study of bias in recidivism prediction instruments. [Cs, Stat], February. http://arxiv.org/abs/1703.00056.
Cooper, A. F., & Flemes, D. (2013). Foreign policy strategies of emerging powers in a multipolar world: An introductory review. Third World Quarterly, 34(6), 943–962. https://doi.org/10.1080/01436597.2013.802501
Cuban, L. (1986). Teachers and machines: The classroom use of technology since 1920. Teachers College Press.
Daly, A., Hagendorff, T., Li, H., Mann, M., Marda, V., Wagner, B., Wang, W., & Witteborn, S. (2019). Artificial Intelligence, Governance and Ethics: Global Perspectives. SSRN Scholarly Paper ID 3414805. Rochester, NY: Social Science Research Network. https://papers.ssrn.com/abstract=3414805.
Dawson, D., E. Schleiger, J. Horton, J. McLaughlin, C. Robinson, G. Quezada, J. Scowcroft, & S. Hajkowicz. 2019. “Australia: Artificial Intelligence: Australia’s Ethics Framework.” Canberra, Australia: Data 61, CSIRO.
Dennis, M., Masthoff, J., & Mellish, C. (2016). Adapting progress feedback and emotional support to learner personality. International Journal of Artificial Intelligence in Education, 26(3), 877–931. https://doi.org/10.1007/s40593-015-0059-7
Department of Public Administration, Italy. 2018. “Italy: Artificial Intelligence at the Service of Citizens.” Rome, Italy: Task Force on Artificial Intelligence of the Agency for Digital Italy and Department for Public Administration, Italy. https://ia.italia.it/assets/whitepaper.pdf.
Dickard, N. (2003). The sustainability challenge: Taking EdTech to the next level. For full text: https://eric.ed.gov/?id=ED477720.
Dignum, V. (2018). Ethics in artificial intelligence: Introduction to the special issue. Ethics and Information Technology, 20(1): 1–3 https://doi.org/10.1007/s10676-018-9450-z
Downey, G. L., & Lucena, J. C. (1997). Engineering selves: Hiring in to a contested field of education. In N. Sakellariou & R. Milleron (Eds.), Ethics, politics, and whistleblowing in engineering (pp. 25–43). CRC Press.
Dutton, T. (2018). An overview of national AI strategies. Politics + AI (blog). June 28, 2018. https://medium.com/politics-ai/an-overview-of-national-ai-strategies-2a70ec6edfd.
Dutton, T., Barron, B., & Boskovic, G. (2018). Building an AI world: Report on national and regional AI strategies. Toronto, Canada: CIFAR. https://www.cifar.ca/docs/default-source/ai-society/buildinganaiworld_eng.pdf.
Ely, D. P. (1999). Conditions that facilitate the implementation of educational technology innovations. Educational Technology, 39(6), 23–27.
European Commission. (2019). Ethical guidelines for trustworthy AI. Brussels, Belgium: European Commission, High-Level Expert Group on Artificial Intelligence (AI HLEG). https://ec.europa.eu/newsroom/dae/document.cfm?doc_id=58477.
Federal Ministry of Education and Research, Germany. (2018). “Germany: National Strategy for Artificial Intelligence: AI Made in Germany.” Bonn, Germany: Federal Ministry of Education and Research, the Federal Ministry for Economic Affairs and Energy, and Federal Ministry of Labour and Social Affairs, Germany. https://www.ki-strategie-deutschland.de/home.html?file=files/downloads/Nationale_KI-Strategie_engl.pdf.
Fishman, B. J., Penuel, W. R., Allen, A. R., Cheng, B. R., & Sabelli, N. (2013). Design-based implementation research: an emerging model for transforming the relationship of research and practice. p. 21.
Fjeld, J., Achten, N., Hilligoss, H., Nagy, A., & Srikumar, M. (2020). Principled artificial intelligence: Mapping consensus in ethical and rights-based approaches to principles for AI. SSRN Scholarly Paper ID 3518482. Rochester, NY: Berkman Klein Center for Internet & Society. https://papers.ssrn.com/abstract=3518482.
Floridi, L., & Cowls, J. (2019). A unified framework of five principles for AI in society. Harvard Data Science Review. https://doi.org/10.1162/99608f92.8cd550d1
Frey, C. B., & Osborne, M. A. (2017). The future of employment: how susceptible are jobs to computerisation? Technological Forecasting and Social Change, 114(January), 254–280. https://doi.org/10.1016/j.techfore.2016.08.019
Friedman, B., Hendry, D. G. & Borning, A. (2017). A survey of value sensitive design methods. Foundations and Trends® in Human–Computer Interaction, 11(2): 63–125. https://doi.org/10.1561/1100000015.
Future of Life Institute. (2020). AI Policy—United States. Future of Life Institute. 2020. https://futureoflife.org/ai-policy-united-states/.
Gebru, T., Morgenstern, J., Vecchione, B., Wortman Vaughan, J., Wallach, H., Daumé III, H., & Crawford, K. (2020). Datasheets for Datasets. [Cs], January. http://arxiv.org/abs/1803.09010.
Gibert, M., Mondin, C., & Chicoisne, G. (2018). Montréal declaration of responsible AI: 2018 Overview of international recommendations for AI Ethics. University of Montréal. https://www.montrealdeclaration-responsibleai.com/reports-of-montreal-declaration.
Giuliani, D., & Ajadi, S. (2019). 618 Active Tech Hubs: The Backbone of Africa’s Tech Ecosystem. GSMA (blog). July 10, 2019. https://www.gsma.com/mobilefordevelopment/blog/618-active-tech-hubs-the-backbone-of-africas-tech-ecosystem/.
Graesser, A. C., D’Mello, S. K., & Strain, A. C. (2014). Emotions in advanced learning technologies. In L. Linnenbrink-Garcia & R. Pekrun (Eds.), International handbook of emotions in education (pp. 473–93). New York: Routledge.
Guidotti, R., Monreale, A., Ruggieri, S., Turini, F., Giannotti, F., & Pedreschi, D. (2018). A survey of methods for explaining black box models. ACM Computing Surveys, 51(5), 1–42. https://doi.org/10.1145/3236009
Guidaudon, V. (2000). European intergration and migration policy. Journal of Common Market Studies, 38(2), 251–71. https://doi.org/10.1111/1468-5965.00219
Hagendorff, T. (2020). The ethics of AI ethics: An evaluation of guidelines. Minds and Machines. https://doi.org/10.1007/s11023-020-09517-8
Hannafin, R. D., & Savenye, W. C. (1993). Technology in the classroom: The teacher’s new role and resistance to it. Educational Technology, 33(6), 26–31.
Hao, K. (2019). China has started a grand experiment in AI education. it could reshape how the world learns. MIT Technology Review, August 2, 2019. https://www.technologyreview.com/2019/08/02/131198/china-squirrel-has-started-a-grand-experiment-in-ai-education-it-could-reshape-how-the/.
Hao, K. (2020). The UK exam debacle reminds us that algorithms can’t fix broken systems. MIT Technology Review, August 20, 2020. https://www.technologyreview.com/2020/08/20/1007502/uk-exam-algorithm-cant-fix-broken-system/.
Harding, T., & Whitehead, D. (2013). Analysing data in qualitative research. In Z. Schneider & D. Whitehead (Eds.), Nursing and midwifery research: Methods and appraisal for evidence-based practice (4th ed., pp. 127–42). Sydney: Mosby Elsevier.
Harley, J. M., Lajoie, S. P., Frasson, C., & Hall, N. C. (2017). Developing emotion-aware, advanced learning technologies: a taxonomy of approaches and features. International Journal of Artificial Intelligence in Education, 27(2), 268–297. https://doi.org/10.1007/s40593-016-0126-8
Heyneman, S. P. (2004). International education quality. Economics of Education Review, 23(4), 441–452. https://doi.org/10.1016/j.econedurev.2003.10.002
Holmes, W., Bektik, D., Whitelock, D., & Woolf, B. P. (2018). Ethics in AIED: Who Cares?. In Artificial Intelligence in Education: 19th International Conference, AIED 2018, London, UK, June 27–30, 2018, Proceedings, Part II, Rosé, C. P., Martínez-Maldonado, R., Hoppe, H. U., Luckin, R., Mavrikis, M., Porayska-Pomsta, K., McLaren, B. and Du Boulay, B (ed) Vol. 10948. Lecture Notes in Computer Science. Cham: Springer International Publishing. https://doi.org/10.1007/978-3-319-93846-2.
Holmes, W., Porayska-Pomsta, K., Holstein, K., Sutherland, E., Baker, T., Shum, S. B., Santos, O. C., Rodrigo, M. T., Cukurova, M., Bittencourt, I. I., & Koedinger, K. R. (2021). Ethics of AI in education: Towards a community-wide framework. International Journal of Artificial Intelligence in Education. https://doi.org/10.1007/s40593-021-00239-1.
House of Lords, Select Committee on Artificial Intelligence. (2018). United Kingdom: AI in the UK: Ready, Willing and Able?. London, UK: House of Lords, Select Committee on Artificial Intelligence. https://publications.parliament.uk/pa/ld201719/ldselect/ldai/100/100.pdf.
IEEE. (2020). 7010–2020 - IEEE Recommended Practice for Assessing the Impact of Autonomous and Intelligent Systems on Human Well-Being. S.l.: IEEE. https://ieeexplore.ieee.org/servlet/opac?punumber=9084217.
Jobin, A., Ienca, M., & Vayena, E. (2019). The global landscape of ai ethics guidelines. Nature Machine Intelligence, 1(9), 389–399. https://doi.org/10.1038/s42256-019-0088-2
Jonsen, K., & Jehn, K. A. (2009). Using triangulation to validate themes in qualitative studies. Qualitative Research in Organizations and Management: An International Journal, 4(2), 123–150. https://doi.org/10.1108/17465640910978391
Joshi, Meghna, T. J., & Rangaswamy, N. (2018). Scaling Classroom IT Skill Tutoring: A Case Study from India. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems, 630:1–630:12. CHI ’18. New York, NY, USA: ACM. https://doi.org/10.1145/3173574.3174204.
Kaplan, S. (2008). Framing contests: Strategy making under uncertainty. Organization Science, 19(5), 729–752. https://doi.org/10.1287/orsc.1070.0340
Katyal, S. K. (2019). Private accountability in the age of artificial intelligence. UCLA Law Review, 66(1), 54–141.
Kingdon, J. W. (1995). Agendas, alternatives, and public policies (2nd ed.). HarperCollins College Publishers.
Korean Ministry of Science, ICT and Future Planning. (2016). South Korea: Mid-to long-term master plan in preparation for the intelligent information society: Managing the fourth industrial revolution. South Korea: Korean Ministry of Science, ICT and Future Planning (MSIP). http://english.msit.go.kr/cms/english/pl/policies2/__icsFiles/afieldfile/2017/07/20/Master%20Plan%20for%20the%20intelligent%20information%20society.pdf; https://medium.com/syncedreview/south-korea-aims-high-on-ai-pumps-2-billion-into-r-d-de8e5c0c8ac5.
Langheinrich, M. (2001). Privacy by design principles of privacy-aware ubiquitous systems. In G. D. Abowd, B. Brumitt, & S. Shafer (Eds.), Ubicomp 2001: Ubiquitous computing (pp. 273–91). Berlin, Heidelberg: Springer.
Lee, J.-W. (2001). Education for technology readiness: Prospects for developing countries. Journal of Human Development, 2(1), 115–151. https://doi.org/10.1080/14649880120050
Lencucha, R., Kothari, A., & Hamel, N. (2010). Extending collaborations for knowledge translation: Lessons from the community-based participatory research literature. Evidence and Policy: A Journal of Research, Debate and Practice, 6(1), 61–75. https://doi.org/10.1332/174426410X483006
Lincoln, Y. S., & Guba, E. G. (1985). Naturalistic Inquiry (Vol. 75). Sage.
Lincoln, Y. S., & Guba, E. G. (1986). but is it rigorous? Trustworthiness and authenticity in naturalistic evaluation. New Directions for Program Evaluation, 1986(30), 73–84. https://doi.org/10.1002/ev.1427
Lord, S. M., Mejia, J. A., Hoople, G., Chen, D., Dalrymple, O., Reddy, E., Przestrzelski, B., & Choi-Fitzpatrick, A. (2018). Creative Curricula for Changemaking Engineers. In 2018 World Engineering Education Forum—Global Engineering Deans Council (WEEF-GEDC), 1–5. https://doi.org/10.1109/WEEF-GEDC.2018.8629612.
Lowther, D. L., Inan, F. A., Daniel Strahl, J., & Ross, S. M. (2008). Does technology integration ‘work’ when key barriers are removed? Educational Media International, 45(3), 195–213. https://doi.org/10.1080/09523980802284317
Mahood, Q., Van Eerd, D., & Irvin, E. (2014). Searching for grey literature for systematic reviews: Challenges and benefits. Research Synthesis Methods, 5(3), 221–234. https://doi.org/10.1002/jrsm.1106
Malin, J., & Brown, C. (Eds.). (2019). The role of knowledge brokers in education: Connecting the dots between research and practice. Routledge.
Manheim, K. M., & Kaplan, L. (2018). Artificial intelligence: Risks to privacy and democracy. SSRN Scholarly Paper ID 3273016. Rochester, NY: Social Science Research Network. https://papers.ssrn.com/abstract=3273016.
Marcinkowski, F., Kieslich, K., Starke, C., & Lünich, M. (2020). Implications of AI (Un-)Fairness in Higher Education Admissions: The Effects of Perceived AI (Un-)Fairness on Exit, Voice and Organizational Reputation. In Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency, 122–30. Barcelona Spain: ACM. https://doi.org/10.1145/3351095.3372867.
McKinsey Global Institute. (2018). Notes From the AI Frontier: Modeling the Impact of AI on the World Economy. McKinsey Global Institute. https://www.mckinsey.com/~/media/McKinsey/Featured%20Insights/Artificial%20Intelligence/Notes%20from%20the%20frontier%20Modeling%20the%20impact%20of%20AI%20on%20the%20world%20economy/MGI-Notes-from-the-AI-frontier-Modeling-the-impact-of-AI-on-the-world-economy-September-2018.ashx.
Mehrabi, N., Morstatter, F., Saxena, F., Lerman, K., & Galstyan, A. (2019). A survey on bias and fairness in machine learning. [Cs], September. http://arxiv.org/abs/1908.09635.
Miles, M. B., Huberman, A. M., & Saldaña, J. (2014). Qualitative data analysis: A methods sourcebook (3rd ed.). SAGE Publications Inc.
Ministry of Economic Affairs and Employment, Finland. (2017). Finland: Finland’s Age of Artificial Intelligence Turning Finland into a Leading Country in the Application of Artificial Intelligence. 47/2017. Helsinki, Finland: Ministry of Economic Affairs and Employment. http://julkaisut.valtioneuvosto.fi/bitstream/handle/10024/160391/TEMrap_47_2017_verkkojulkaisu.pdf.
Ministry of Finance and Ministry of Industry, Business and Financial Affairs, Danish Government. (2019). Denmark: National Strategy for Artificial Intelligence. Copenhagen, Denmark: Ministry of Finance and Ministry of Industry, Business and Financial Affairs, Danish Government. https://investindk.com/-/media/invest-in-denmark/files/danish_national_strategy_for_ai2019.ashx?la=en.
Ministry of Science, Innovation and Universities, Spain. (2019). Spanish RDI Strategy in Artificial Intelligence. Madrid, Spain: Ministry of Science, Innovation and Universities. http://www.ciencia.gob.es/stfls/MICINN/Ciencia/Ficheros/Estrategia_Inteligencia_Artificial_EN.PDF.
Ministry of the Economy and Innovation. (2019). Lithuanian Artificial Intelligence Strategy: A Vision of the Future. Vilnius, Lithuania: Ministry of the Economy and Innovation. http://kurklt.lt/wp-content/uploads/2018/09/StrategyIndesignpdf.pdf.
Mitchell, M., Wu, S., Zaldivar, A., Barnes, P., Vasserman, L., Hutchinson, B., Spitzer, E., Raji, I. D. & Gebru, T. (2019). Model cards for model reporting. In Proceedings of the Conference on Fairness, Accountability, and Transparency - FAT* ’19, 220–29. https://doi.org/10.1145/3287560.3287596.
Mittelstadt, B. D., Allo, P., Taddeo, M., Wachter, S., & Floridi, L. (2016). The ethics of algorithms: Mapping the debate. Big Data and Society, 3(2), 205395171667967. https://doi.org/10.1177/2053951716679679
Moher, D., Liberati, A., Tetzlaff, J., Altman, D. G. & The PRISMA Group. (2009). Preferred reporting items for systematic reviews and meta-analyses: The PRISMA statement. PLOS Medicine, 6 (7): e1000097.https://doi.org/10.1371/journal.pmed.1000097
Morganstein, D., & Wasserstein, R. (2014). ASA Statement on value-added models. Statistics and Public Policy, 1(1), 108–110. https://doi.org/10.1080/2330443X.2014.956906
Morley, J., Floridi, L., Kinsey, L., & Elhalal, A. (2019). From what to how: An initial review of publicly available AI ethics tools, methods and research to translate principles into practices. Science and Engineering Ethics. https://doi.org/10.1007/s11948-019-00165-5
NITI Aayog, India. (2018). India: National Strategy for Artificial Intelligence #AIFORALL. Delhi, India: NITI Aayog. http://niti.gov.in/writereaddata/files/document_publication/NationalStrategy-for-AI-Discussion-Paper.pdf.
Noble, D. F. (1998). Digital diploma mills: The automation of higher education. Science as Culture, 7(3), 355–368.
Nye, B. D. (2015). Intelligent tutoring systems by and for the developing world: A review of trends and approaches for educational technology in a global context. International Journal of Artificial Intelligence in Education, 25(2), 177–203. https://doi.org/10.1007/s40593-014-0028-6
Oetzel, M. C., & Spiekermann, S. (2014). A systematic methodology for privacy impact assessments: A design science approach. European Journal of Information Systems, 23(2), 126–150. https://doi.org/10.1057/ejis.2013.18.
Office of the President, Russia. (2019). Russia: National Strategy for the Development of Artificial Intelligence Over the Period Extending up to the Year 2030. Moscow, Russia. http://kremlin.ru/events/president/news/60630.
Office of the Prime Minister, Malta. (2019). Malta: Towards An AI Strategy. Valletta, Malta: Parliamentary Secretariat for Financial Services, Digital Economy and Innovation, Office of the Prime Minister, Malta. http://malta.ai/wp-content/uploads/2019/04/Draft_Policy_document_-_online_version.pdf.
Oliver, K., Innvar, S., Lorenc, T., Woodman, J., & Thomas, J. (2014). A systematic review of barriers to and facilitators of the use of evidence by policymakers. BMC Health Services Research, 14(1), 2. https://doi.org/10.1186/1472-6963-14-2
Oswald, M., Grace, J., Urwin, S., & Barnes, G. C. (2018). Algorithmic risk assessment policing models: Lessons from the Durham hart model and ‘experimental’ proportionality. Information and Communications Technology Law, 27(2), 223–250. https://doi.org/10.1080/13600834.2018.1458455
Parkhurst, J. (2017). The politics of evidence :From evidence -based policy to the good governance of evidence. Taylor & Francis. https://library.oapen.org/handle/20.500.12657/31002.
Phoenix, J. H., Atkinson, L. G., & Baker, H. (2019). Creating and communicating social research for policymakers in government. Palgrave Communications, 5(1), 98. https://doi.org/10.1057/s41599-019-0310-1
Pinkwart, N. (2016). Another 25 years of AIED? Challenges and opportunities for intelligent educational technologies of the future. International Journal of Artificial Intelligence in Education, 26(2), 771–783. https://doi.org/10.1007/s40593-016-0099-7
Porter, M. E. (1996). Competitive advantage, agglomeration economies, and regional policy. International Regional Science Review, 19(1–2), 85–90. https://doi.org/10.1177/016001769601900208
Qatar Center for Artificial Intelligence. (2019). “Qatar: Blueprint: National Artificial Intelligence Strategy for Qatar.” Ar-Rayyan, Qatar: Qatar: Qatar Center for Artificial Intelligence (QCAI), Qatar Computing Research Institute (QCRI), Hamad Bin Khalifa University. https://qcai.qcri.org/wp-content/uploads/2019/02/National_AI_Strategy_for_Qatar-Blueprint_30Jan2019.pdf.
Radaelli, C. M. (2009). Measuring policy learning: Regulatory impact assessment in Europe. Journal of European Public Policy, 16(8), 1145–1164. https://doi.org/10.1080/13501760903332647
Raji, I. D., Smart, A., White, R. N., Mitchell, M., Gebru, T., Hutchinson, B., Smith-Loud, J., Theron, D., & Barnes, P. (2020). Closing the AI accountability gap: Defining an end-to-end framework for internal algorithmic auditing. In Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency, 33–44. FAT* ’20. Barcelona, Spain: Association for Computing Machinery. https://doi.org/10.1145/3351095.3372873.
Reisman, D., Schultz, J., Crawford, K., & Whittaker, M. (2018). Algorithmic impact assessments: A practical framework for public agency accountability. New York, NY: AI Now. https://ainowinstitute.org/aiareport2018.pdf.
Roberts, K., Dowell, A., & Nie, J.-B. (2019). Attempting rigour and replicability in thematic analysis of qualitative research data: A case study of codebook development. BMC Medical Research Methodology, 19(1), 66. https://doi.org/10.1186/s12874-019-0707-y
Roolaht, T. (2012). The characteristics of small country national innovation systems. In E. G. Carayannis, U. Varblane, & T. Roolaht (Eds.)In novation Systems in Small Catching-Up Economies: New Perspectives on Practice and Policy (pp. 21–37). Springer. https://doi.org/10.1007/978-1-4614-1548-0_2.
Rothstein, H. R., & Hopewell, S. (2009). Grey literature. The handbook of research synthesis and meta-analysis (pp. 103–25). Russell Sage Foundation.
Schendel, R., & McCowan, T. (2016). Expanding higher education systems in low- and middle-income countries: The challenges of equity and quality. Higher Education, 72(4), 407–411. https://doi.org/10.1007/s10734-016-0028-6
Schiff, D. (2021). Out of the laboratory and into the classroom: The future of artificial intelligence in education. AI & Society, 36(1), 331–348. https://doi.org/10.1007/s00146-020-01033-8
Schiff, D., Ayesh, A., Musikanski, L., & Havens, J. C. (2020a). IEEE 7010: A new standard for assessing the well-being implications of artificial intelligence. In 2020 IEEE International Conference on Systems, Man, and Cybernetics (SMC), 2746–53. https://doi.org/10.1109/SMC42975.2020.9283454.
Schiff, D., Biddle, J., Borenstein, J., & Laas, K. (2020b). What’s next for AI ethics, policy, and governance? A global overview. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society, 153–58. AIES ’20. New York, NY, USA: Association for Computing Machinery. https://doi.org/10.1145/3375627.3375804.
Schiff, D., Borenstein, J., Laas, K., & Biddle, J. (2021). AI ethics in the public, private, and NGO sectors: A review of a global document collection. IEEE Transactions on Technology and Society, 2, 31–42. https://doi.org/10.1109/TTS.2021.3052127
Schomberg, R. von. (2013). A vision of responsible research and innovation. In Responsible Innovation (pp. 51–74). Wiley-Blackwell. https://doi.org/10.1002/9781118551424.ch3.
Scott, J., Lubienski, C., DeBray, E., & Jabbar, H. (2014). The intermediary function in evidence production, promotion, and utilization: The case of educational incentives. In K. S. Finnigan & A. J. Daly (Eds.), Using research evidence in education (pp. 69–89). Cham: Springer.
Selbst, A. D., Boyd, D., Friedler, S. A., Venkatasubramanian, S., & Vertesi, J. (2019). Fairness and abstraction in sociotechnical systems. In Proceedings of the Conference on Fairness, Accountability, and Transparency, 59–68. FAT* ’19. New York, NY, USA: Association for Computing Machinery. https://doi.org/10.1145/3287560.3287598.
Sin, C. H. (2008). The role of intermediaries in getting evidence into policy and practice: some useful lessons from examining consultancy-client relationships. Evidence and Policy: A Journal of Research, Debate and Practice, 4(1), 85–103. https://doi.org/10.1332/174426408783477828
Smart Nation and Digital Government Office, Singapore. (2019). Singapore: National Artificial Intelligence Strategy: Advancing Our Smart Nation Journey. Singapore: Smart Nation and Digital Government Office. https://www.smartnation.sg/docs/default-source/default-document-library/national-ai-strategy.pdf.
Stahl, B. C., & Coeckelbergh, M. (2016). Ethics of healthcare robotics: towards responsible research and innovation. Robotics and Autonomous Systems, 86(December), 152–161. https://doi.org/10.1016/j.robot.2016.08.018
Stern, C., Jordan, Z., & McArthur, A. (2014). Developing the review question and inclusion criteria. AJN the American Journal of Nursing, 114(4), 53–56. https://doi.org/10.1097/01.NAJ.0000445689.67800.86
Terhart, E. (2013). Teacher resistance against school reform: Reflecting an inconvenient truth. School Leadership and Management, 33(5), 486–500. https://doi.org/10.1080/13632434.2013.793494
The Institute for Ethical AI in Education. (2020). Interim report: Towards a shared vision of ethical AI in education. The institute for ethical AI in education. https://www.buckingham.ac.uk/wp-content/uploads/2020/02/The-Institute-for-Ethical-AI-in-Educations-Interim-Report-Towards-a-Shared-Vision-of-Ethical-AI-in-Education.pdf.
Thomas, D. R. (2006). A general inductive approach for analyzing qualitative evaluation data. American Journal of Evaluation, 27(2), 237–246. https://doi.org/10.1177/1098214005283748
Lee, N. T., Resnick, P., & Barton, G. (2019). Algorithmic bias detection and mitigation: best practices and policies to reduce consumer harms. Brookings Institition. https://www.brookings.edu/research/algorithmic-bias-detection-and-mitigation-best-practices-and-policies-to-reduce-consumer-harms/.
Villani, C., Marc, S., Bonnet, Y., Charly, B., Anne-Charlotte Cornut, François Levin, & Rondepierre, B. (2018). “France: For a Meaningful Artificial Intelligence: Towards a French and European Strategy.” French Parliament. https://www.aiforhumanity.fr/pdfs/MissionVillani_Report_ENG-VF.pdf.
Warschauer, M., & Ames, M. (2010). Can One Laptop per Child save the world's poor?. Journal of International Affairs, 64(1): 33-51.
West, D. M. (2018). The future of work: Robots, AI, and automation. Brookings Institution Press.
White, M. D., & Marsh, E. E. (2006). Content analysis: A flexible methodology. Library Trends, 55(1), 22–45. https://doi.org/10.1353/lib.2006.0053
Whittaker, M., Crawford, K., Dobbe, R., Fried, G., Kaziunas, E., Mathur, V., West, S.M., Richardson, R., Schultz, J., & Schwartz, O. (2018). AI Now Report 2018. New York, NY: New York University. https://ainowinstitute.org/AI_Now_2018_Report.pdf.
Williamson, B., & Eynon, R. (2020). Historical threads, missing links, and future directions in AI in education. Learning, Media and Technology, 45(3), 223–235. https://doi.org/10.1080/17439884.2020.1798995
Winfield, A. (2019). Ethical standards in robotics and AI. Nature Electronics, 2(2), 46–48. https://doi.org/10.1038/s41928-019-0213-6
Women Leading in AI. (2019). Women leading in AI: 10 Principles of responsible AI. Women Leading in AI. https://womenleadinginai.org/wp-content/uploads/2019/02/WLiAI-Report-2019.pdf.
Woolf, B. P., Ivon, A., David, C., Winslow, B., & Kasia, M. (2010). Affective tutors: Automatic detection of and response to student emotion. In N. Roger, B. Jacqueline, & M. Riichiro (Eds.), Advances in intelligent tutoring systems (pp. 207–27). Berlin, Heidelberg: Springer.
Zarsky, T. Z. (2016). Incompatible: The GDPR in the age of big data. Seton Hall Law Review, 47, 995–1020.
Zeng, Y., Enmeng L., & Cunqing H. (2018). Linking artificial intelligence principles. [Cs], December. http://arxiv.org/abs/1812.04814.
Zhang, D., Saurabh, M., Brynjolfsson, E., Etchemendy, J., Ganguli, D., Grosz, B., Lyons, T., Manyika, J., Niebles, J. C., Sellitto, M., Shoham, Y., Clark, J., & Perrault, R. (2021). Artificial Intelligence Index Report 2021. AI Index Steering Committee, Human-Centered AI Institute, Stanford University. https://aiindex.stanford.edu/wp-content/uploads/2021/03/2021-AI-Index-Report_Master.pdf
Zhou, Y., & Danks, D. (2020). Different ‘intelligibility’ for different folks. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society, 194–99. AIES ’20. New York, NY, USA: Association for Computing Machinery. https://doi.org/10.1145/3375627.3375810.
Funding
The author declares no funding.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflicts of interest
The author declares no conflicts of interest or competing interests.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix 1: Methodology
Appendix 1: Methodology
This appendix provides more extensive details surrounding the data collection, screening, coding, and analysis strategy, as well as associated limitations, than is available in the main body of the article.
Data Collection
The data collection process differs from that typically employed in meta-analysis and review papers, such as recommended by the PRISMA Statement (Moher et al. 2009), largely because the documents evaluated here are so-called “gray literature” not found in traditional databases (Mahood et al. 2014; Rothstein and Hopewell 2009). Given the type of documents and substantive focus of the study, the data collection process relied on linkhubs, Google searches, and manual searching of certain documents and countries: e.g., “country name + AI policy strategy.” The search process is connected to ongoing research assessing AI ethics and policy (Schiff et al. 2021, 2020a, 2020b). The first author and colleagues have maintained a database of AI policy documents and additional AI ethics documents (available at https://dx.doi.org/10.21227/fcdb-pa48). The database was created in Fall 2018 and updated regularly until early 2020. Data collection focused on policy documents benefited especially from lists maintained by the Future of Life Institute (2020) and Tim Dutton (2018). These linkhubs contain updates about AI policy developments in dozens of countries, including how far along countries are in development of task forces, funding proposals, and formal AI policy strategy documents.
For each country noted in these linkhubs, I searched for and accessed documents and initiatives mentioned, and performed additional Google searches and manual searches for each country to ensure that the set of national AI policy strategies was as complete as possible. While it is possible that some countries were omitted, perhaps due to lack of language familiarity, the two key sources are invested in tracking national AI policy developments. All such candidate documents were thus captured in the database managed by the first author and colleagues. From this larger database, I extracted only documents produced by public sector organizations (e.g., countries, not corporations or non-governmental organizations). This resulted in 76 public sector AI documents that formed the candidate pool of national AI policy documents.
Screening Process
The screening process involved identifying criteria for types of documents, publication language, the population and phenomena of interest, and time period (Stern et al. 2014). The purpose of this study was to assess national AI policy strategies as they relate to education. Therefore, the study applied the following inclusion/exclusion criteria:
-
Documents needed to be complete and resemble a national AI policy strategy. Some such documents describe themselves as preliminary reports or blueprints, working towards more robust or formalized policy strategies. Nevertheless, many were sufficiently robust so as to be considered as policy strategies. On the other hand, countries that had only announced task forces, funding initiatives, created websites, or otherwise did not have a well-developed document analogous to that of other countries were not included. This ensures countries can be compared fairly and speak to sufficiently detailed items of policy, with a sizable average of 62 pages per document.
-
Documents needed to be in English, due to the author’s limited language proficiency. However, in a number of cases, governments had produced official English-language translations (e.g., Finland, Italy). While automated translation of non-English documents (e.g., Google Translate) may not be of sufficient quality, there was one unofficial but high-quality translation included in the final sample, of China’s AI policy strategy, performed by the Foundation for Law and International Affairs.
-
The study also excluded documents produced by inter-governmental organizations, such as the United Nations, the Organization for Economic Cooperation and Development, and the European Union. While these documents are no doubt important, they address a different scope, as they are relatively distant from national-level institutions, funding activities, and other policy activities, such as those involving education policy. This makes these documents less comparable to national-level AI policy strategies.
-
Finally, in a number of cases, countries produced multiple documents that were potentially relevant to AI policy. Only one document was selected per country. The chosen document was typically the most robust and the most recent, at times an evolution of a previous draft or more preliminary document. Further, some candidate documents were not representative of an overarching national AI policy strategy. For example, documents from Germany addressing autonomous vehicle policy and from Finland addressing work in the age of AI were excluded in preference of Germany’s National Strategy for AI and Finland’s Age of AI. This screening criteria helped to ensure that individual countries were not overrepresented, that information analyzed was not redundant, and that the most robust, high-quality, and comparable policy strategies were selected in each case.
Of the 76 candidate documents, eight did not resemble a complete national policy strategy document, one was not available in English, 13 were inter-governmental, and 30 were excluded in favor of more representative documents. Screening resulted in a final sample of 24 national AI policy strategies.
Codebook Development
After identifying the final sample, the analytical strategy began with the development of a preliminary set of topics, in the form of a codebook (Miles et al. 2014; Thomas 2006). These topics in the codebook were chosen based on the study’s conceptual scope and framework, the author’s subject matter knowledge, and previous exposure to AI policy strategies. The scope of interest was any discussion of education, construed as broadly as possible, such as youth and adult education, training and re-skilling, investing in educational systems, public education and awareness, the need to develop more AI talent (e.g., computer scientists, engineers), and social and ethical issues related to AI and education.
The initial codebook included 11 categories: Education as Priority Topic, K-12 Education, Post-Secondary Education, Adult Education and Training, General AI Literacy, Training AI Experts, Preparing Workforce, Intelligent Tutoring Systems, Pedagogical Agents and Learning Robots, Predictive Educational Tools, and AI for Healthcare. A best practice in qualitative research is to iterate and refine the codebook through testing on a small subset of the data (Roberts et al. 2019). Therefore, I randomly selected five documents—aiming for a meaningfully-sized and somewhat representative subset—and applied the thematic schema to them. This involved reading the documents to determine whether the coding schema could validly and straightforwardly reflect the way education was discussed in the documents, and to identify if the coding schema captured the full range of issues in the documents relevant to the article’s conceptual scope.
Based on this initial test, several categories were modified, removed, and collapsed as follows:
-
Education as Priority Topic and AI for Healthcare were retained, as they were easy to apply. Either topic might be explicitly noted as a priority topic in a document, for example, if a list of priority policy sectors was mentioned and education was among that list. Alternatively, education/healthcare were coded as priority topics if a significant subsection was dedicated to them, or if there was a similar amount of discussion relative to the length of the document as compared to other documents that did identify education (or healthcare) as an explicit priority.
-
K-12 Education, Post-Secondary Education, and Adult Education and Training were removed. These categories were originally designed to separate discussion of education by target age/population group. However, the test documents often did not identify the target age/group when discussing AI and education, making this distinction difficult to code accurately. Moreover, these population differences were deemed less relevant for the overall purpose of the article. For example, that documents emphasized the need to develop more AI researchers seemed more pressing to the document authors than whether this development happened in secondary or postsecondary educational institutions.
-
Training AI Experts and Preparing Workforce for AI were straightforward and were retained.
-
General AI Literacy was renamed to Public AI Literacy. The former was originally defined to emphasize development of general digital, STEM, and other skills in educational settings. The theme was relabeled and redefined to incorporate AI literacy in both educational (classroom) and ‘public’ settings, because both settings were discussed and justified to pertain to similar policy purposes.
-
The revised codebook collapsed Intelligent Tutoring Systems and Pedagogical Agents and Learning Robots into Teaching and Learning. Too few documents addressed these issues at the level of detail of individual AIED technologies or tools to allow for reliable identification, as the documents generally employed more abstracted terms and discussions.
-
The revised codebook also abstracted Predictive Educational Tools into Administrative Tools, as there were several examples of AIED tools mentioned that were better captured by the latter, broader terminology, such as the use of AI for inspection or assigning teachers to schools.
These adjustments resulted in a revised codebook with seven categories (a reasonable number for inductive studies) (Thomas 2006), described in the main body. The final coding categories were straightforward to apply to the data and captured relevant concepts within the study’s scope well.
An important note is that, despite an initial attempt to code for discussion of AIED ethics specifically given its importance to this study, discussion of these topics was too rare to justify having as a theme. Most discussion addressing ethics and education was focused on Education for AI purposes, such as training future machine learning experts to develop ethical design skills, rather than addressing ethical implications emanating from AIED. Nevertheless, I captured all mentions of ethics in the context of both Education for AI and AI for Education in my memos, and considered the presence and absence of these topics as part of the interpretive work.
Coding Approach
Next, I applied the codebook to the 24 documents in the sample (approximately 1491 pages total). Each document was read closely and assessed manually along the seven topics using a simple form of content analysis. This consisted of evaluating each document for the presence or absence of each theme (White and Marsh 2006), largely a binary exercise, though some documents were coded as borderline cases. In Table 1, a country is marked as green when a theme was reflected, red when absent, and yellow when the case was sufficiently ambiguous or borderline.
For example, Malta’s (2019) document initially notes that AI for healthcare may be amongst the highest impact projects and “most pressing domestic challenges” worthy of prioritization, but it does not proceed to include any substantive discussion or a subsection on healthcare. In comparison to the document’s discussion of other topics, and in comparison to other countries’ AI policy documents that discuss healthcare in more depth, this relatively more narrow treatment of the topic led to coding it as yellow. Similarly, Russia’s (2019) discussion of using AI to “[improve] the quality of education services” does not provide enough detail to be clear about the role of AIED as a potential tool for teaching and learning, and so is considered to be too ambiguous to code as either green or red.
Analysis Approach
Relevant quotes from the documents were captured in a research memo and organized under the seven categories (Thomas 2006) to support higher-order conceptual and thematic interpretation. Additional quotes of interest and minor categories were included here as well, such as any mentions of ethics related to education. From this, I synthesized insights from the frequency and character of these topics, applying a thematic analytic approach (Castleberry and Nolen 2018) to identify major findings. This interpretive exercise involves considering second-order meanings or explanations for the patterns identified in the data (Miles et al. 2014), including the finding that AIED’s ethical implications are neglected. I present results for each topic in the main article, along with interpretation of key findings within and across topics, to support a broader discussion of the role of education and ethics in AI policy in the subsequent sections.
Limitations
Because the documents were coded by a single researcher, it is not possible to, for example, assess inter-rater reliability. Further, the conceptualization of the study, codebook development, and interpretation were not subject to the perspectives of other researchers or experts outside of the peer review process. However, quantitative measures of reliability are only sometimes considered essential in qualitative research (Castleberry and Nolen 2018), and a single coder approach can be appropriate and, in cases, even preferrable (Harding and Whitehead 2013). Multiple researchers may not be necessary to provide sufficient consistency and credibility, as single researchers can provide a unitary and consistent perspective, albeit one dependent on that author’s subjective assessments. For example, research using semi-structured interviews with dozens of coding categories and many degrees of detail (e.g., scoring attributes from 1–10) benefit especially from the assessment of interrater reliability, particularly if the codes are challenging to conceptually separate or define. In this study, however, the number of topics is small, the level of detail simple, and the concepts are fairly easy to conceptually separate.
Moreover, in qualitative research, there are common criteria of research rigor used as alternatives to traditional quantitative criteria of validity and reliability. For example, one widely used set of criteria comes from Lincoln and Guba (1985), who propose credibility as an alternative to internal validity, transferability as an alternative to external validity, dependability as an alternative to reliability, and confirmability as an alternative to objectivity. To satisfy these criteria, the analysis employed several recommended strategies (Lincoln and Guba 1986). Within-method triangulation across multiple documents (Jonsen and Jehn 2009) and the use of direct quotes as descriptive evidence provide rich support to demonstrate claims, supporting their credibility, dependability, and transferability. Further, because the data are publicly available, as opposed to privately held interview data, for example, they are open to scrutiny and confirmation or disconfirmation. However, in part because of researchers’ individual positions and biases (Castleberry and Nolen 2018), it is possible that other researchers would identify different coding categories or identify different salient themes. As such, the single researcher approach is a limitation of this study, discussed in the study’s limitations section. Future research examining the role of education in AI policy would be welcome in assessing the extent to which the findings presented here are indeed credible, dependable, confirmable, and transferable.
Rights and permissions
About this article
Cite this article
Schiff, D. Education for AI, not AI for Education: The Role of Education and Ethics in National AI Policy Strategies. Int J Artif Intell Educ 32, 527–563 (2022). https://doi.org/10.1007/s40593-021-00270-2
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s40593-021-00270-2