Abstract
Smart cities do not exist without robotics and Artificial Intelligence (AI). As the case of the COVID-19 pandemic shows, they can assist in combating the novel coronavirus and its effects, and preventing its spread. However, their deployment necessitate the most secure, safe, and efficient use. The purpose of this article is to address the regulatory framework for AI and robotics in the context of developing resilient organisations in smart cities during the COVID-19 pandemic. The study provides regulatory insights necessary to re-examine the strategic management of technology creation, dissemination, and application in smart cities, in order to address the issues regarding the strategic management of innovation policies nationally, regionally, and worldwide. To meet these goals, the article analyses government materials, such as strategies, policies, legislation, reports, and literature. It also juxtaposes materials and case studies, with the help of expert knowledge. The authors emphasise the imminent need for coordinated strategies to regulate AI and robots designed for improving digital and smart public health services globally.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Avoid common mistakes on your manuscript.
1 Introduction
Smart cities do not exist without Artificial Intelligence (AI) and robotics. It is not only a matter of definition, because this correlation also has a practical aspect, as AI is becoming an increasingly important component of smart cities. This has been especially important during the COVID-19 pandemic, where AI and robots have been assisting in combating the novel coronavirus with its effects, and preventing its spread. However, as the experience of subsequent lockdowns and distancing measures demonstrates, the deployment of AI and robotics–as a matter of technological advancement and the level of technology available–must be tailored to the situation (pandemic), necessitating the most secure, safe, and efficient use of smart emerging technologies and AI.
With this in mind, the primary purpose of this article is to address the regulatory framework for AI and robotics in the context of developing resilient mechanisms in smart cities during the COVID-19 pandemic. The study provides regulatory insights to re-examine the strategic management of technology creation, dissemination, and application in smart cities, in order to address the issues regarding the strategic management of innovation policies nationally, regionally, and worldwide. This is particularly relevant to the deployment of AI solutions and the establishment of resilient organisations in smart cities to handle the regulatory challenges of the pandemic. To meet these goals, the article analyses a range of materials, such as strategies, policies, legislation, reports, and literature and also juxtaposes numerous case studies. This is done with the help of expert knowledge, enhanced by the activity of the Global Pandemic Network which brings together scholars from universities all over the world, in order to conduct research on legal, economic, and social issues related to pandemics (Benjamin et al. 2021).
The article seeks to expand the discussion on the governance of AI and robotics during the pandemic and provides an added value to the discussion on recent changes in the institutional environment. For these needs, the following research questions are discussed:
-
1.
what are the contemporary legal issues underlying AI and robotics regulatory choices?
-
2.
how have these issues translated into the use of AI and robotics in the fight against the COVID-19 pandemic?
-
3.
what are the determinants for creating a regulatory model of AI and robotics for the needs of a development of resilient organisations in smart cities?
In this light, the structure of this paper is based on three pillars that showcase a discussion of a general regulatory framework for AI and robots, covering such issues as privacy and data protection (Sect. 2), case studies of AI and robots used to combat the COVID-19 pandemic (Sect. 3), and resilience regulatory modes based on public disaster management (Sect. 4). The final section of this study emphasises the main issues and provides some follow-up remarks.
2 Development of policy approach for AI and robotics
Development of AI and robotics in smart cities requires an organised policy approach (see Tsuji 2018; Sokołowski 2022). This paradigm derives from the need to protect fundamental rights and freedoms, including the right to life, liberty and security of a person and privacy, or the right to freedom of opinion and expression (see Ufert 2020). In fact, the development of AI systems and robotics can profoundly undermine the value structure of some legal frameworks, especially in those where democratic constitutions guarantee a high level of protection of the rights and interests of an individual. Observing the pervasive capacity of such emerging technologies, legal analysis has developed along several lines. With reference to the introduction of systems designed to control people, conflicting with their freedom of movement and the protection of their personal sphere, regulators worldwide seek to strengthen the rules on privacy by default and design. Issues related to liability for harms caused by AI tools also require delineating the roles and individual responsibilities of the actors who use these tools, both in public and private activities. Moreover, algorithms can affect individuals’ freedom of choice, self-determination, and awareness, imposing the need for transparency rules to make users aware. The listed concerns are also sparking a debate on the distinction of roles between humans and machines that transcend law and touch the boundaries of ethics and philosophy, innovating the legal approach to be implemented in the future. This compels policymakers to take responsibility for providing the appropriate approach. What measures, however, should be implemented?
Depending on the socio-economic structure adopted, the degree of public approach inevitably varies, in case of technology policy as well (see Barge-Gil and Modrego-Rico 2008); however, two extreme points can be differentiated on the axis of this impact: total subordination and complete release, the total opposites (Sokołowski 2018a). These liberal and interventionist approaches, both rooted in the market- and state-based approaches (see Sokołowski 2016) are delimited by an intervening space filled by public law regulation (see Barnett 1986), a regulatory zone representing mixed economy (Sokołowski 2020a; see Harris 1990). The history of the development of such areas as telecommunications, energy, or aviation shows that the regulatory action is becoming the preferred approach (Sokołowski 2020a; Dempsey and Gesell 2013; Kolasa-Sokołowska 2022; Dempsey 1990). This tendency also applies to AI (see Scherer 2015; Clarke 2019), where the race to AI has also spurred a race towards AI regulation (Smuha 2021; see Chawla et al. 2022).
In contrast to the regulatory approach, one may find deregulation. It is the process of removing public components from a specific area, sector, or policy (Sokołowski and Heffron 2022). “For some, the problem will always be that the markets were not ‘free’ enough from government interference and a further reduction in regulation is needed” notices Thomas (2006, p 1975). This means that, in the most severe scenario, neither a public agency (regulator) nor the instruments of public regulation, such as control, commands, sanctions, etc., exist (Sokołowski 2018a, p 595). It can lead to a complete release in extreme cases, which, when paired with fraudulent behaviour and market manipulation (see Windolf 2004), can lead to a significant crisis (see Duane 2002; Sokołowski 2020a, pp 174–175).
A particularly appealing alternative to business is a policy that enables the avoidance of responsibility, e.g., by applying a soft regulatory approach – a “light touch regulation” (Fisk 2011, p 556) – which takes the risks of non-compliance by relying on the independent achievement of the set goals (Heffron et al. 2018, p 1193). While this approach may work well in the short term, for instance in a new technology development phase, or in the current COVID-19 pandemic (see Bachtiger et al. 2020), it necessitates a high level of trust in regulated entities – as the regulatory arrangement can be used by the regulated entity to relieve a burden of responsibility regarding duties – a regulatory capture (see Galloway 2020, pp 59–60). In the long run, it can result in misconducts, frequently discovered only in the final stages (Sokołowski and Heffron 2022). As a result, leaving the policy implementation in the hands of the regulated businesses (self-regulation) is not the best idea (see Lauri 2020), especially when no enforcement is available under the soft approach or the enforcement is of a very weak nature. Because the regulators’ and regulated parties’ interests differ in various ways (Bella et al. 2021), it is quite likely that this accommodative approach will fail (Sokołowski and Heffron 2022).
Nevertheless, there is a risk that during a pandemic, addressed under the state-of-emergency laws or some other extraordinary framework, overregulation characterised by the excessive creation of legislation and legal overweight will occur (Sokołowski 2018a, see 2020b, p 596). Cooperation can aid in preventing overregulation (see Sokołowski 2018a, p 596). In this improvement process, it will be useful not only to build a global data enforcement framework, based on the principles of privacy by design and privacy by default, but also to enforce basic principles for the upcoming regulatory assets. The said principles include transparency, interpretability, accountability, explicability, auditability, traceability, and neutrality or fairness (Bassan 2019; Kritikos 2020). The anticipated solution relates to the application of a global framework for the regulation of AI and robots, especially in order to define the conditions of their use, ensure the fairness of global competition, and conform their use to the protection of health and human rights according to the human centric approach (Lauri 2021b). Smart cities, with their smart infrastructures, are ideal testing grounds for implementing this approach (Cook et al. 2018; Obringer R, Nateghi R 2021). Table 1 summarises the benefits and drawbacks of the discussed policy approaches.
Let us refer this to issues such as privacy and data protection. When numerical statistics are collected and analysed with a help of AI or by AI itself, and then combined with other personal data, they become “personally identifiable information” (Shimpo 2020). This necessitates the development and implementation of appropriate approach, including legal, to protect the right to privacy (Florencio and Ramanathan 2001, p 105). This is particularly essential given that it is regarded as a universal human right (Blasi Casagran 2017, p 228) and considered basic in many jurisdictions around the world (see Parker 2010). Here, in setting the rules to deal with new problems, EU regulatory initiatives have been ahead of the curve in many respects, for instance in terms of tackling climate change (see Perez de las Heras 2013; Sokołowski 2018b). This also concerns the area of data protection (see Pajuste 2019), in which there has been a remarkable degree of legislative upgrading (in the EU), from the Data Protection Directive 95/46/EC (Data Protection Directive 1995) to the EU General Data Protection Regulation (GDPR) 2016/679 (GDPR 2016). The degree of transition and upgrading of the legislation is remarkable. In addition, the EU has been deliberating on an AI Regulation and is looking ahead to lead the debate on AI regulation in the future. Recently, the European Commission (2021a) proposed new rules regarding the promotion of regulation for trustworthy AI. The first-ever legal framework on AI, combined with the new Coordinated Plan of Member States, will ensure people’s and enterprises’ safety and basic rights, while also boosting AI use, investment, and innovation across the EU. Moreover, the proposed new EU rules on machinery (European Commission 2021b) will complement these efforts by adjusting the safety rules to boost users’ confidence in the next generation of products (European Commission 2021c).
Regulated regimes can help to mitigate the scope of these issues, but some concerns may still persist (see Dickinson et al. 2021). For instance, if a problem arises because of AI making an autonomous decision, such as taking the wrong action and causing damage, can we hold it liable under traditional product liability? Even if we can claim product liability for a robot equipped with AI, how should we view the fact that in the future AI might make decisions that we cannot predict? Even if we can sue for product liability for AI-enabled robots, how should we compensate for the damage caused by the AI programme itself running out of control or operating in an uncontrollable state? This is especially true when these robots are constantly connected to the network and used in the Internet of Things (IoT) applications daily; as a result, the problem will be present everywhere.
Because the IoT aims to influence the use of everyday objects via the Internet, we must consider legal issues when these objects are connected to robots, and the robots themselves are connected to the network and controlled by AI for use in our daily lives (autonomous robots). The issues surrounding autonomous robots equipped with AI, whether general-purpose or specialised, are expected to differ from those surrounding industrial robots. However, the primary legal issues surrounding robots continue to be those pertaining to industrial robots. Furthermore, some believe that existing robot safety standards will be adequate to deal with any new issues that arise. Let us now examine these issues through the lens of the COVID-19 pandemic’s use of robots and AI.
3 AI and robots in the fight with COVID-19: governance issues
In 2018, the World Health Assembly Resolution on Digital Health recognized the value of digital technologies to reach the Sustainable Development Goals (see WHO 2019; WHO 2021). In 2020, as the COVID-19 pandemic spread worldwide, several forms of technological applications started to be massively implemented. Among them Kritikos (2020) lists: AI, blockchain, open-source technologies, telehealth technologies, three-dimensional printing, gene-editing technologies, nanotechnology, synthetic biology, and, lastly, drones and robots. Without disparaging the initial elements of this list, the final few have a wide range of applications in fighting the novel coronavirus. For instance, according to the Robotics for Infectious Diseases consortium, more than 150 robots are being used to combat COVID-19 (Vargo et al. 2021).
Indeed, during the pandemic, digital technologies have been widely introduced in a number of areas of intervention that characterise smart cities (Pacheco Rocha et al. 2019; Lauri 2021b) and their intrinsic purpose of improving the quality of life in densely populated urban contexts (see Lytras et al. 2019b). This can be summarised as follows (Murphy et al. 2020, 2021): public safety (for compulsive quarantine enforcement, disinfection of public spaces, identification of infected individuals, public service announcements, and traffic flow monitoring); clinical care (for point-of-care disinfection, observational telepresence, delivery and inventory, interventional telepresence, patient and family socialization, and patient and visitor admissions); continuity of work and education (for sanitisation at work or school, for telepresence, for private health surveillance and security); laboratory and supply chain automation (for delivery activities, laboratory automation, management of infection materials); quality of life (delivery of food and other purchases, attending social events, interpersonal socialization); and non-hospital care (for delivery to quarantine, socialisation in quarantine, and public health surveillance). Table 2 summarises the technological applications used during the COVID-19 pandemic in relation to the public interest to be fulfilled, considering their type of use and the specific activities performed (Murphy et al. 2020, 2021).
Early detection and infection monitoring by AI and robots, in particular, have been crucial in the fight against COVID-19 (Pazzaglia et al. 2021); however, this activity has aroused concerns about the compatibility with legal compliance. Thus, it is no coincidence that most legal studies have been concerned with studying their utility for public health, considering issues of efficacy, equity, and privacy (Landau 2021). While the collection of real-time public health data has been advantageously implemented in both assisting the policymakers in the planning process and informing the public about the evolution of the pandemic’s spread (Budd et al. 2020), the governments’ access to various data, including a person’s geolocation, has raised numerous privacy issues (see Gerke et al. 2020; Hassandoust et al. 2021; Chan and Saqib 2021). These applications are not limited to a single country (see Chakraborty et al. 2020). For example, since October 2020, the EU ensures interoperability of COVID-19 contact tracing and warning apps in order to facilitate free movement as an integral part of the Single Market (European Commission 2020a).
Nevertheless, during the peak months of 2020 and 2021 globally, different models emerged for approaching the tracking activity. One is the Chinese model, while the other is the European model. Several different approaches can be found in the Asian context. Considering China, parallel with the outbreak of the contagion the Chinese authorities started to lay down strong measures to track the movement of people who had visited the Wuhan market. It was done through tools such as mobile phones, mobile payment applications, social real-time data on people’s location (Whitelaw et al. 2020) and facial recognition. This also allowed the authorities to forecast the transmission of the virus and orient border checks and surveillance strategies. For example, China used the AliPay HealthCode app for automatic communication and the enforcement of quarantine measures by limiting transactions permitted for high-risk users (Kupferschmidt and Cohen 2020).
While this system has enabled a drastic containment of the pandemic, the need to increase data protection is a topic of growing interest in Chinese law (Greenleaf 2020). The initial response began to emerge in October 2020, when parliamentarians began debating the Personal Information Protection Law (PIPL) to regulate the collection and use of personal data. In August 2021 the law was adopted, and will go into effect at the beginning of November 2021. PIPL is also intended to enhance the exchange of data with countries that have a higher level of protection and do not tolerate collection systems with no users’ consent or through non-transparent processes. In particular, Article 49 of the law stipulates that “[p]ersonal information handlers shall establish mechanisms to accept and handle applications from individuals to exercise their rights. Where they reject individuals’ requests to exercise their rights, they shall explain the reason”. The above mentioned “mechanisms” proceed in the direction of the European-derived legal meaning of privacy by design and privacy by default (Bifulco 2018). More specifically, such mechanisms would seem to recall those automatisms referred to by the GDPR (Article 25) that should allow data controllers (and consequently data processors) to carry out a processing operation by providing, from the outset (by design), the tools and correct settings to protect personal data, so that the framework of principles is respected by default. The PILP, along with the Data Security Law (implemented from since September 2021) mark two major regulations set to govern China’s smart cities in the coming years. These rules will affect the big tech companies who are the main actors in smart cities, as the new law will change the value of data and have a significant impact on their business and relations with institutions.
Moreover, countries such as South Korea have integrated AI and robotics into the government-coordinated containment and mitigation processes for early disease detection. These include surveillance, testing, contact tracing, and strict quarantine, even using geolocalisation and video surveillance measures (Zastrow 2020). South Korea, in other terms, has adopted a soft policy of voluntary containment, with widespread dissemination of information to citizens. The system is based on the central government’s existing smart city project, and is being developed in consultation between various ministries and the Centres for Disease Control and Prevention. The Korean system enjoyed a higher level of resilience when compared to others. And indeed, Korean law, amended after the 2015 MERS outbreak, provides a specific legal basis to allow authorities to access camera data, GPS tracking data from phones and cars, credit card transactions and other personal data for infectious disease control purposes. Access to this data by health professionals must still be authorised by law enforcement authorities, but the most recent changes (as of March 2020) allow also direct access by health authorities. The real time data and monitoring can support administration in management of smart services and control for better governance (see Kumar et al. 2020b). A public service enhanced by AI allows for community interaction that is tailored to the end users’ perceptions and abilities, and promotes individuals’ involvement in the community (Lytras et al. 2021). This is especially critical when dealing with the pandemic, which necessitates cohesion, collaboration, and coordination. Thus, the data that flows to the authorities not only supports the government’s (both central and local) activities in combating the spread of the coronavirus, but also keeps the population constantly informed – by the authorities – of these activities, and of the spread of the contagion. The secured data flow represents a tool for cooperation between the authorities and citizens, which can also help to maintain a balance between lockdown rules and normal life in times of high social tension.
Based on the experience of South Korea, some European states have begun to design a soft control system, that included: controlling quarantined persons through geolocation; tracking the routes of infected persons to identify those at risk; disseminating information to the public on the movements of infected persons to alert those at risk and invite them to undergo diagnostic tests. Such an approach has made it possible to comply with the requirement of “proportionality” between data protection and the interests of individuals (as highlighted in GDPR). As emphasised in the European Data Protection Supervisor’s guidelines (European Data Protection Supervisor 2020), compliance with the GDPR’s regulatory framework on privacy does not allow for strong pervasiveness of technological tools in the EU. For example, data collection in Norway through the Smittestopp app has been stopped due to its “disproportion to the task” (Budd et al. 2020). Therefore, it should come as no surprise that one of the envisaged solutions has been the implementation of soft apps, such as the Italian one called Immuni. The app is based on technical requirements aimed at balancing privacy and personal rights (De Falco and Maddalena 2020) with the detection action carried out with the support of an algorithm, thanks to the use of Bluetooth technology (European Commission 2020b). The soft approach is based on the following assumptions: first, the freedom of the user to download the app or not (without prejudice to those who evade); second, the transparency towards the subject regarding the use that is made of the users’ data; third, the determinacy and exclusivity of the data as far as statistical or scientific aims are concerned; fourth, data storage on a governmental server, for the duration of the pandemic; fifth, the reciprocity of anonymity, as citizens are limited to receiving a notice only in the event of interaction with an infected person; and finally, the selectivity, the minimisation of data and its pseudonymisation (Article 26 of the GDPR) according to the decentralised PEPPT (Pan-European Privacy Preserving Proximity Tracing initiative, 2020) model (Bonomi 2020).
As previously observed, AI and robotics have even been used to prevent rule violations in order to stop the spread of COVID-19, such as general lockdown and quarantine for people exposed to or infected with the virus. Even in legal systems with highly developed privacy-protection regulations, governments have used telephone traffic data obtained from internet providers (data retention) to tackle the pandemic (Oliver et al. 2020), given the need to repress behaviours capable of undermining public health. Only data pseudonymisation and anonymisation could alleviate privacy concerns in this scenario.
Prevention and monitoring activities are the prerequisite for stabilising the new situation, but a further step in building a resilient system is bolstering healthcare services. This is especially important in metropolises and urban complexes due to human congestion. As a result, contact tracking applications may aid in the development of smart cities, benefiting public transportation and related industries while also providing valuable insights for city management (see Schmidtke 2020, p 200). The definition of a clear framework in terms of privacy by design and by default constitutes the basis for introducing different AI and robotics systems useful for enhancing smart cities and improving citizens’ well-being. Table 3 summarises the various approaches used by different legal systems to track the activity of infected individuals, considering the technology involved, the main objective pursued by legislators, and highlighting advantages and disadvantages.
Yet another facet of healthcare and the use of AI in the COVID-19 pandemic is the utilisation of robots for assisting healthcare workers. For instance, disinfection of spaces in public buildings such as schools and hospitals, but also delivering food and medical supplies (see Bogue 2020). Using robots in medical activities has a positive impact on improving the smart city’s health resilience. The rapid use of IoT devices has facilitated the collection of health-related big data (see Lytras et al. 2019a). Many medical facilities have begun to fully digitise electronic health records for clinician testing orders, referrals, and patient scheduling in order to improve the efficacy and efficiency of both medical and administrative healthcare processes (see Flynn et al. 2020). Deep learning has been used to diagnose COVID-19 using X-ray pictures (Wang et al. 2020). AI can be used to track the spread of COVID-19 and predict a patient’s needs. Through computational biology and the use of data analytics, mathematical modelling and computational simulation have helped to study and research the pandemic (Kumar et al. 2020a).
Furthermore, many medical facilities have started to introduce robots in therapies. Such is the case of Loccioni, a company which used a robot to prepare a monoclonal to treat COVID-19 patients at the Hospital of Ancona (Italy). Robots can autonomously carry out the most complex operations in order to guarantee the correct composition of the therapy and intercept any possible errors during validation, transcription, preparation and delivery. Based on physicians’ reports, eligible patients are received in special rooms set up at the infectious diseases department, in a protected environment with negative pressure. The therapy requires utmost care and precision during the drug preparation procedure. The personalised preparation of injectable drugs represents a critical aspect for healthcare facilities as it involves numerous risks for the safety of patients and operators, as well as significant costs and possible organisational inefficiencies. The entire drug pathway, from prescription to administration, is controlled through sophisticated automated measurement systems that ensure high accuracy, complete traceability of operations and integrity of information. The prescription is digitised and the preparation phase takes place in a fully automated manner, in a dedicated and constantly monitored work environment. As a result, these therapies are confirmed by quality certificates, offer maximum safety in terms of sterility and accuracy of the injectable drugs prepared, allow safe management of clinical data and the production phase, and reduce clinical risk (Yaniv et al. 2017).
Indeed, since the spread of COVID-19, previously harmless tasks may pose serious health risks. In places too dangerous for them, humans are being replaced by robots, which are considered more reliable and cost-effective. However, any advantages in terms of health risk prevention are matched by the risk of job losses for all those whose tasks are going to be taken over by technologies (Ramirez 2021). “Retraining unemployed people was never easy, but it is more challenging now that technological disruption is spreading so rapidly, widely, and unpredictably” highlights Floridi (2017, p 3). This fact is linked to a broader reflection on the loss of humanity in certain activities and relationships, which, along with issues of privacy and security, is part of the debate that many legal systems are facing in preparing a regulatory framework for the use of AI (Bassan 2019).
The application of AI and robots, discussed here, is a tool for strengthening the resilience of the public health service on several fronts (Auby 2020). Moreover, it offers at least three methodological considerations helpful for understanding the coordinates on which to develop the regulatory framework. First, the public–private partnership created to develop the robots between the public (in this case, the hospital) and the private company combines the expertise of the national service system and the know-how of the private company. This promotes an increase in the organisational efficiency and ergonomics of the process (Valaguzza and Parisi 2020). It also makes it possible to move away from the dependence of the service on public resources, which are often insufficient, and to be able to rely on the economic investments and resources of the private sector. The adopted method of shared governance also allows the reengineering of processes through the sharing of best practices, in order to bring innovation to the public sector, considered extremely conservative both for the scarcity of resources available and still underdeveloped culture of innovation. Second, AI and robot introduction in ordinary medical activities prevents human errors and accurately controls the appropriateness of the medical prescription. This is a form of “preventive medicine actions” and can be useful to create “personalized services”, adaptable to the patient and highly efficient, as recommended by the European Communication on Digital Health Services (2018). Third, the implementation of robotics simplifies documentation management by making information more usable for the benefit of the patient. There is thus an advantage in terms of transparency of the service provided and of knowability, creating a more collaborative environment and reinforcing trust between treatment facilities and patients.
4 AI, robots, and smart cities: COVID-19 resilience regulatory model
A re-examination of the strategic management of technology creation, dissemination, and application in smart cities is required to build resilient organisations in smart cities during the COVID-19 pandemic with the help of robotics and AI (de Pablos et al. 2022). Disaster management provided by public authorities when dealing with the effects of hurricanes, earthquakes, or tsunamis constitutes a useful benchmark (Sokołowski 2020b). This refers to: the application of AI and robotics to enhance the public’s ability to respond to disasters, policymaking under unusual circumstances (see Schneider 1992), or remedies (ex-post disaster assistance or ex-ante regulation) to limit loss exposure (see Priest 1996, p 219) or alleviate disaster effects (Malawani et al. 2020). A pandemic, if treated as a calamity that may reoccur – like in the case of SARS-COV-3, SARS-COV-4, or any other infectious disease – makes preparedness the key element of regulatory approach (see WHO 2019), and a critical component of true smart cities, which are primarily targeted by the current pandemic’s negative effects (being large clusters of people). As a result, it is necessary to consider the future challenges now, while, at the same time, continuing to implement the measures aimed at combating the current pandemic. In this regard, considering a pandemic as a natural disaster that may reoccur demonstrates the validity of referring to a regime designed to counteract natural disasters (see Sokołowski 2020b; Dixit 2020; Tsuji 2021).
Furthermore, many parallels can be found in the current pandemic between activities related to those undertaken by public authorities. For example, following Hurricanes Katrina and Rita in 2005, federal regulatory agencies recognised that, due to extraordinary circumstances, flexibility in the application of rules and simplifying several applications were required (Sokołowski 2020b). This also concerns certain regulatory reliefs offered to professionals vital in disaster response or recovery, e.g. by adjusting licensing requirements, or freezing inspections (Sokołowski 2020b). AI and robotics offer a wide range of possibilities in this area. This is about simplifying procedures, making them more responsive as well as contactless, and conducting them online. Innovative chatbots can offer a straightforward support in administrative procedures, for example, when applying for licences or certificates (see van Noordt and Misuraca 2019), while different AI applications can perform inspections comparable to those carried out by humans (for instance, a drone – an unmanned aerial vehicle equipped with a camera conducts technical monitoring of a power line). However, this requires a regulatory environment that recognises the equivalence of such activities to those carried out in a traditional manner. Smart cities are the perfect environment for introducing such improvements.
Furthermore, as in disaster prevention, AI offers enormous modelling possibilities, providing expert forecasts on pandemic development. These models can be utilised on a voluntary basis; however, a legal approach should guide their development. This could be done, for example, by offering specifications for their use (including the scope of the analysed data), as well as listing institutions that should use them (for instance, by making it mandatory for health establishments). This is also a source of concern for city authorities, particularly those in metropolitan areas, as the health-care administration (naturally at the forefront) is not the only one working to combat the pandemic. Other institutions are also striving to ensure compliance with the law and standards, as well as transparency and clarity of rules regarding consumers and competition in extraordinary times (Sokołowski 2020b).
Moreover, fighting the pandemic demonstrates the importance of well-functioning coordination systems; coordinating policies can improve the effectiveness of crisis response (OECD 2020, p 2). This is especially true in those circumstances when the central government plays a larger role – worldwide examples of actions performed during the COVID-19 pandemic illustrate that this involvement frequently overshadows activities of other entities, e.g., local authorities (Sokołowski 2020b). In such situations, AI technologies can improve the coordination mechanism of a multi-actor administration system, making it more effective and responsive. The widespread adoption of AI should be a post-pandemic standard, transforming traditional administration into true e-administration (see Wierzbowski et al. 2021). The same applies to the transformation of traditional cities into smart ones (see Bobadilla et al. 2018).
This also concerns the structure of administration, both central and local. If – apart from coordination mechanisms – a specialised anti-pandemic authority is established (for example, an agency responsible for combating COVID-19), it can, in addition to all necessary expert knowledge obtained from the health administration, serve as a valuable benchmark of an e-administration scenario. In such conditions, AI applications can not only help with the creation of a structure solely responsible for countering COVID-19 (or future, similar events), but also accelerate the process of transformation to e-administration at different levels (van Noordt and Misuraca 2019). This, of course, also applies on a city level, as no real smart city can exist without e-administration (Lauri 2021a).
Finally, AI has a wide range of applications connected to the knowledge-based approach (see Fig. 1), which could result in adopting a system of rules, standards, authorisation, permissions, and guidance dedicated to COVID-19, based on best available practices as much as feasible (see Sokołowski 2020b). These should be accompanied by pandemic-specific monitoring, surveillance, and enforcement that is as safe (non-physical, online, etc.) as possible, free of unnecessary administrative hassle and with deadlines suspended or extended (Sokołowski 2020b). Of course, it must be scaled to the challenge – the recent COVID-19 variants, especially the quickly spreading omicron, make it far more difficult to adjust state logistics to the size of the problem (as is the case, for example, in South Korea). With such an approach, AI can help authorities, also in cities, become smarter and more resilient organisations, guided by pragmatic and responsive regulation. Nevertheless, this process requires some universal standards. The urgent need for coordinated, global, digital and smart public health strategies has been highlighted both by the WHO, in its global strategy on digital health 2020–2025 (WHO 2021) and by the EU, which called for a pan-European approach on the use of data for COVID-19 (European Commission 2020a, b) currently also being implemented through collaborative research projects (Tacconelli et al. 2022). Moreover, coordination serves as an ancillary element to bridge the digital divide by ensuring access to mobile communication and internet services, particularly in low- and middle-income countries, as well as for minorities and people with lower socioeconomic status. Indeed, an unequal access to technology can exacerbate inequalities between countries in terms of their preparedness to fight future pandemics, which can jeopardise the resilience of all the areas of the world. This also concerns cities, as the smart ones are at the forefront, while the “analogue” ones are lagging behind (Thomas et al. 2021).
5 Conclusion
The rapid advancement of science and technology, the increased use of information technology, and the development of network-related technologies have all resulted in significant improvements to our daily lives, opening several legal issues. This also applies to governance of robotics and AI that have also helped profoundly in the frontline fight against COVID-19 in the urban environments. These actions, however, have frequently caused or exacerbated legal issues related to the employment of AI and robotics.
As discussed in this study, looking at issues in an institutional system AI solutions can improve the coordination mechanism of a multi-actor administration system, making it more effective and responsive. For instance, private public partnerships combine the know-how of the private enterprise with the knowledge of the national service system, optimising the possibilities of delivering the best results in terms of preventing human errors, appropriateness of medical actions, and ease of documentation administration. However, as these new and emerging technologies have been introduced into societies and cities and their use has increased, certain problems have arisen – not only from illegal activities or misuse (that should be regulated by law), but also from the lack of rules governing the use of the said technologies. This applies not only to surveillance systems aimed at preventing the spread of a pandemic, which – as can be seen – may follow “soft” or “strong” approaches depending on regulatory frameworks; it refers also to the digitisation of services (as in the use of robotics for health services).
Given this scenario, what are the determinants for creating a regulatory model of AI and robotics for the needs of a development of resilient organisations in smart cities? As identified in the paper, among current legal issue there is a need for a responsive regulatory framework of a universal character (at least at a level of principles, as due to the diversity of legal systems, it is challenging to attain complete universality of solutions at the global level), which can simultaneously hold together the protection of privacy and the rights of individuals and the fulfilment of public interests. Indeed, as it turns out, with AI and modern technologies consistent with a regulatory system, it is possible to improve the resilience of health systems and work and social activities, which are essential prerequisites for contextualising smart cities in an institutional system. This calls for smart regulation, driven by knowledge-based approach, with disaster management as a benchmark and preparedness included in regulatory approach, bringing flexibility in the application of rules, simplifying procedures and offering regulatory reliefs, and moving procedures online, with AI-enhanced administrative procedures, inspections, and decisions, and last but not least AI-improved coordination mechanisms.
References
Auby J-B (2020) Administrative law facing digital challenges. Euro Review Digital Admin Law 1:7–16
Bachtiger P, Peters NS, Walsh SL (2020) Machine learning for COVID-19—asking the right questions. Lancet Digital Health 2(8):e391–e392
Barge-Gil A, Modrego-Rico A (2008) Are technology institutes a satisfactory tool for public intervention in the area of technology? A neoclassical and evolutionary evaluation. Eviron Plann C Gov Policy 26(4):808–823
Barnett RE (1986) Foreword: four senses of the public law-private law distinction. Har J Law Public Policy 9(2):267–276
Bassan F (2019) Potere dell’algoritmo e resistenza dei mercati in Italia: la sovranità perduta sui servizi [Algorithm power and market resistance in Italy lost sovereignty over services]. Rubbettino, Soveria Mannelli
Bella FA, Lauri C, Capra G (2021) The role of COVID-19 soft law measures in Italy: much ado about nothing? Euro J Risk Reg 12(1):93–110
Benjamin A, Iamiceli P, Kasimbazi E, Lin J, Robinson N, Scotti E, Lauri C, Tigre MA (2021) Editorial. Legal Policy Pandemics. J Global Pand Net 1:5–9
Bifulco R (2018) Intelligenza artificiale, internet e ordine spontaneo [Artificial intelligence, the internet and spontaneous order]. In: Pizzetti F (ed) Intelligenza artificiale, protezione dei dati personali e regolazione [Artificial intelligence, personal data protection and regulation]. Giappichelli, Turin, pp 383–400
Blasi Casagran C (2017) Global data protection in the field of law enforcement: an EU perspective. Routledge, Abingdon - New York, NY
Bobadilla LMO, Nieto MPR, Molano JIR (2018) From traditional cities to smart cities. In: Uden L, Hadzima B, Ting I-H (eds) Knowledge Management in Organizations: 13th International Conference, KMO 2018, Žilina, Slovakia, August 6–10, 2018, Proceedings. Springer, Cham, pp 661–673
Bogue R (2020) Robots in a contagious world. Ind Robot 47(5):673–642
Bonomi MS (2020) L’app Immuni: tra tutela della salute e protezione dei dati personali [The Immuni app: between health protection and personal data protection]. In: Federalismi.it. https://www.federalismi.it/nv14/articolo-documento.cfm?Artid=43707. Accessed 31 May 2021
Budd J, Miller BS, Manning EM et al (2020) Digital technologies in the public-health response to COVID-19. Nat Med 26(8):1183–1192
Chakraborty P, Maitra S, Nandi M, Talnikar S (2020) Contact Tracing in Post-Covid World. Indian Statistical Institute Series, Springer, Singapore
Chan EY, Saqib NU (2021) Privacy concerns can explain unwillingness to download and use contact tracing apps when COVID-19 concerns are high. Comput Hum Behav 119:106718
Chawla Y, Shimpo F, Sokołowski MM (2022) Artificial intelligence and information management in the energy transition of India: lessons from the global IT heart. Digit Policy Regul Gov 24(1):17–29
Clarke R (2019) Regulatory alternatives for AI. Comput Law Secur Rev 35(4):398–409
European Commission (2020a) Coronavirus: EU interoperability gateway. https://ec.europa.eu/commission/presscorner/detail/en/ip_20_1904. Accessed 31 May 2021
European Commission (2020b) Commission Recommendation of 8.4.2020b on a common Union toolbox for the use of technology and data to combat and exit from the COVID-19 crisis, in particular concerning mobile applications and the use of anonymised mobility data, C(2020b) 2296 final
European Commission (2021a) Proposal for a Regulation of the European Parliament and of the Council laying down harmonised rules on artificial intelligence (Artificial Intelligence Act) and amending certain Union legislative acts, COM (2021a) 206 final, 21.4.2021a
European Commission (2021b) Proposal for a Regulation of the European Parliament and of the Council on machinery products, COM (2021b) 202 final, 21.4.2021b
European Commission (2021c) Europe fit for the digital age: artificial intelligence. https://ec.europa.eu/commission/presscorner/detail/en/IP_21_1682. Accessed 6 Aug 2021c
Cook D, Duncan G, Sprint G, Fritz S (2018) Using smart city technology to make healthcare smarter. Proc IEEE 106(4):708–722
de Pablos O, Chui KT, Lytras MD (2022) Digital innovation for healthcare in COVID-19 pandemic: strategies and solutions. Academic Press, Cambridge, MA
Dempsey PS (1990) Airline deregulation and laissez-faire mythology: economic theory in turbulence. J Air Law Com 56(2):305
Dempsey PS, Gesell LE (2013) Public policy and the regulation of commercial aviation. Coast Aire Publications, Chandler, AZ
Dickinson H, Smith C, Carey N, Carey G (2021) Exploring governance tensions of disruptive technologies: the case of care robots in Australia and New Zealand. Policy Society 40(2):232–249
Data Protection Directive (1995) Directive 95/46/EC of the European Parliament and of the Council of 24 October 1995 on the protection of individuals with regard to the processing of personal data and on the free movement of such data, OJ L 281, 23.11.1995
Dixit P (2020) Synergising international public health law and international disaster law. Euro J Risk Reg. 22:1–11
Duane TP (2002) Regulation’s rationale: learning from the California energy crisis. Yale J Reg 19(2):471–540
De Falco D, Maddalena ML (2020) La politica del tracciamento dei contatti e dei test per Covid-19 alla luce delle ultime direttive OMS: nessun ostacolo giuridico impedisce di utilizzare il “modello coreano” anche in Italia [The policy of contact tracing and testing for COVID-19 in the light of the latest WHO guidelines: no legal obstacle prevents the use of the “Korean model” also in Italy]. In: Federalismi.it. https://www.federalismi.it/nv14/articolo-documento.cfm?Artid=41629. Accessed 31 May 2021
European Data Protection Supervisor (2020) Guidelines 04/2020 on the use of location data and contact tracing tools in the context of the COVID-19 outbreak, C(2020) 2296 final, 8.4.2020
Fisk D (2011) The 2008 financial collapse: lessons for engineering failure. Eng Fail Anal 18(2):550–556
Florencio PS, Ramanathan ED (2001) Secret code: the need for enhanced privacy protections in the United States and Canada to prevent employment discrimination based on genetic and health information. Osgoode Hall Law Journal 39:77–116
Floridi L (2017) Robots, jobs, taxes, and responsibilities. Philos Technol 30(1):1–4
Flynn EF, Kuhn E, Shaik M et al (2020) Drive-through COVID-19 testing during the 2020 pandemic: a safe, efficient, and scalable model for pediatric patients and healthcare workers. Academic Pediat 20:753–755
Galloway W (2020) Planning for disaster – lessons from the 2011 Tohoku Disaster. In: Black-Branch JL, Fleck D (eds) Nuclear non-proliferation in international law. Asser Press, The Hague, pp 53–72
Gerke S, Shachar C, Chai PR, Cohen IG (2020) Regulatory, safety, and privacy concerns of home monitoring technologies during COVID-19. Nat Med 26(8):1176–1182
Greenleaf G (2020) China issues a comprehensive draft data privacy law. Privacy Laws & Business International Report. 26:6–10
Harris L (1990) The mixed economy of a democratic South Africa. Invest Anal J 19(33):32–42
Hassandoust F, Akhlaghpour S, Johnston AC (2021) Individuals’ privacy concerns and adoption of contact tracing mobile applications in a pandemic: a situational privacy calculus perspective. J Am Med Inform Assoc 28:463–471
Heffron RJ, McCauley D, de Rubens GZ (2018) Balancing the energy trilemma through the Energy Justice Metric. Appl Energy 229:1191–1201
Kolasa-Sokołowska K (2022) Artificial intelligence and risk preparedness in the aviation industry. In: Bielicki D (ed) Regulating artificial intelligence in industry. Routledge, Abingdon-New York, NY, pp 114–126
Kritikos M (2020) Ten technologies to fight coronavirus. In: European Parliamentary Research Service. https://www.europarl.europa.eu/RegData/etudes/IDAN/2020/641543/EPRS_IDA(2020)641543_EN.pdf. Accessed 31 May 2021
Kumar A, Gupta PK, Srivastava A (2020a) A review of modern technologies for tackling COVID-19 pandemic. Diabetes Metab Syndr 14(4):569–573
Kumar H, Singh MK, Gupta MP, Madaan J (2020b) Moving towards smart cities: solutions that lead to the smart city transformation framework. Technol Forecast Soc Chang 153:119281
Kupferschmidt K, Cohen J (2020) Can China’s COVID-19 strategy work elsewhere? Science 367(6482):1061–1062
Landau SE (2021) People count: contact-tracing apps and public health. MIT Press, Cambridge, MA
Lauri C (2020) Indennizzi automatici e apatia razionale. Il public enforcement nelle sfide poste dalla complessità dei mercati (a proposito di Cons. St. 879/2020) [Automatic compensation and rational apathy. Public enforcement in the challenges posed by the complexity of markets (about Cons. State, Section VI, n. 879/2020)]. In: Carbone A, Zampetti E, Aperio Bella F (eds) Dialoghi di diritto amministrativo [Dialogues of Administrative Law]. Roma Tre Press, Rome, pp 155–188
Lauri C (2021a) Expert knowledge and smart city administration european review of digital administration law. Erdal. 2(1):57–76
Lauri C (2021b) Smart city. In: Bifulco R, Celotto A, Olivetti M (eds) Digesto delle discipline pubblicistiche [Digest of public law]. UTET, Turin, pp 377–396
Lytras MD, Chui KT, Visvizi A (2019a) Data analytics in smart healthcare: the recent developments and beyond. Appl Sci 9(14):2812
Lytras MD, Visvizi A, Sarirete A (2019b) Clustering smart city services: perceptions, expectations, responses. Sustainability 11:1669
Lytras MD, Visvizi A, Chopdar PK et al (2021) Information management in smart cities: turning end users’ views into multi-item scale development, validation, and policy-making recommendations. Int J Inf Manage 56:102146
Murphy RR, Gandudi VBM, Adams J (2020) Robots are playing many roles in the coronavirus crisis – and offering lessons for future disasters. The Conversation. https://theconversation.com/robots-are-playing-many-roles-in-the-coronavirus-crisis-and-offering-lessons-for-future-disasters-135527. Accessed 31 May 2021
Murphy RR, Gandudi VBM, Adams J, Clendenin A, Moats J (2021) Adoption of Robots for Disasters: Lessons from the Response to COVID-19. Foundations and Trends 9(2):130–200
Obringer R, Nateghi R (2021) What makes a city ‘smart’ in the Anthropocene? Sustainable Cities and Society, A critical review of smart cities under climate change. https://doi.org/10.1016/j.scs.2021.103278
Oliver N, Lepri B, Sterly H et al (2020) Mobile phone data for informing public health actions across the COVID-19 pandemic life cycle. Sci Adv 6(23):0764
Pacheco Rocha N, Dias A, Santinha G, Rodrigues M, Queirós A, Rodrigues C (2019) Smart cities and public health: a systematic review. Procedia Comp Sci 164:516–523. https://doi.org/10.1016/j.procs.2019.12.214
Pajuste T (2019) The protection of personal data in a digital society: the role of the GDPR. In: Susi M (ed) Human rights, digital society and the law. Routledge, Abingdon–New York, NY, pp 303–315
Parker T (2010) Are we protected? The adequacy of existing legal frameworks for protecting privacy in the biometric age. In: Kumar A, Zhang D (eds) Ethics and policy of biometrics: third international conference on ethics and policy of biometrics and international data sharing. Springer, Berlin, pp 40–46
Pazzaglia G, Mameli M, Rossi L, Paolanti M, Mancini A, Zingaretti P, Frontoni E (2021) People counting on low cost embedded hardware during the SARS-CoV-2 pandemic. In: Pattern recognition. ICPR international workshops and challenges: virtual event, January 10–15, 2021, Proceedings, part II. Springer International Publishing, pp 521–533
Perez de las Heras B, (2013) Beyond Kyoto: the EU’s contribution to a more sustainable world economy. Eur Law J 19(4):577–591
Ramirez JJ (2021) Against automation mythologies: business science fiction and the ruse of the robots. Routledge, Abingdon – New York, NY
General Data Protection Regulation (GDPR) Regulation (2016) Regulation (EU) 2016/679 of the European Parliament and of the Council of 27 April 2016 on the protection of natural persons with regard to the processing of personal data and on the free movement of such data, and repealing Directive 95/46/EC (General Data Protection Regulation), OJ L 119, 4.5.2016
Scherer MU (2015) Regulating artificial intelligence systems: risks, challenges, competencies, and strategies. Harvard J Technol 29(2):353–400
Schmidtke HR (2020) Location-aware systems or location-based services: a survey with applications to COVID-19 contact tracking. Journal Rel Intel Environm 6(4):191–214
Shimpo F (2020) The importance of ‘smooth’data usage and the protection of privacy in the age of AI, IoT and autonomous robots. Global Privacy Law Review 1:49–54
Smuha NA (2021) From a ‘race to AI’to a ‘race to AI regulation’: regulatory competition for artificial intelligence. Law Innov Technol 13(1):57–84
Sokołowski MM (2016) Regulation in the European electricity sector. Routledge, Abingdon – New York, NY
Sokołowski MM (2018a) Regulatory dilemma: between deregulation and overregulation. In: Jagielski J, Kijowski D, Grzywacz M (eds) Prawo administracyjne wobec współczesnych wyzwań. Księga jubileuszowa dedykowana Profesorowi Markowi Wierzbowskiemu [Administrative law facing contemporary challenges: Jubilee Anniversary Publication dedicated to Professor Marek Wierzbowski]. C.H. Beck, Warsaw, pp 591–598
Sokołowski MM (2018b) Burning out coal power plants with the Industrial Emissions Directive. J World Energy Law Bus 11(3):260–269
Sokołowski MM (2020a) Balancing energy regulation: a day-watchman approach. In: Grzeszczak R (ed) Economic freedom and market regulation: in search of proper balance. Nomos, Baden-Baden, pp 167–186
Sokołowski MM (2020b) Regulation in the COVID-19 pandemic and post-pandemic times: day-watchman tackling the novel coronavirus. Transf Govern People. 15(2):206–208. https://doi.org/10.1108/TG-07-2020-0142
Sokołowski MM (2022) Artificial intelligence and climate-energy policies of the EU and Japan. In: Bielicki D (ed) Regulating and Japan. Regulating artificial intelligence in industry. Routledge, Abingdon-New York, NY, pp 138–155
Sokołowski MM, Heffron RJ (2022) Defining and conceptualising energy policy failure: the when, where, why, and how. Energy Policy 161:112745
Tacconelli E, Gorska A, Carrara E, Davis RJ, Bonten M, Friedrich AW, Glasner C, Goossens H, Hasenauer J, Haro Abad JM, Penalvo JL, Sanchez-Niubo A, Sialm A, Scipione G, Soriano G, Yazdanpanah Y, Vorstenbosch E, Jaenisch T (2022) Challenges of data sharing in European Covid-19 projects: a learning opportunity for advancing pandemic preparedness and response. Lancet Reg Health Euro 2022:100467. https://doi.org/10.1016/j.lanepe.2022.100467
Thomas S (2006) The grin of the Cheshire cat. Energy Policy 34(15):1974–1983
Thomas JJ, Geropanta V, Karagianni A, Panchencko V, Vasant P (2021) Smart cities and machine learning in urban health. IGI Global, Hershey
Tsuji Y (2018) Review safety of medical device robots in Japan. Asia Pacific J Health Law Eth 12(1):1–27
Tsuji Y (2021) Japanese government actions against COVID-19 under the directives of constitutional and administrative law. Cardozo International Comparat Law 4(1):1–34
Ufert F (2020) AI regulation through the lens of fundamental rights: how well does the GDPR address the challenges posed by AI? European Papers. J Law Integrat. 2020:1087–1097
Valaguzza S, Parisi E (2020) Public private partnerships: governing common interests. Edward Elgar Publishing, Cheltenham
van Noordt C, Misuraca G (2019) New wine in old bottles: chatbots in government. In: Panagiotopoulos P, Edelmann N, Glassey O et al (eds) Electronic Participation. Springer, Cham, pp 49–59
Vargo D, Zhu L, Benwell B, Yan Z (2021) Digital technology use during COVID-19 pandemic: a rapid review. Human Behav Emerg Technol 3(1):13–24
Wang L, Lin ZQ, Wong A (2020) COVID-net: a tailored deep convolutional neural network design for detection of COVID-19 cases from chest X-ray images. Sci Rep 10(1):1–12
Whitelaw S, Mamas MA, Topol E, Van Spall HG (2020) Applications of digital technology in COVID-19 pandemic planning and response. Lancet Digital Health 2(8):435
WHO (2019) Pandemic influenza preparedness framework: annual progress report: 1 January-31 December 2018. World Health Organization
WHO (2021) Global Strategy on Digital Health 2020–2025. World Health Organization
Wierzbowski M, Galán-Vioque R, Gamero Casado E et al (2021) Challenges and prospects of e-governance in Poland and Spain. Elect Gover 17(1):1–26
Windolf P (2004) Corruption, fraud, and corporate governance: a report on Enron. In: Grandori A (ed) Corporate governance and firm organization: microfoundations and structural forms. Oxford University Press, Oxford, pp 159–190
Yaniv AW, Orsborn A, Bonkowski JJ et al (2017) Robotic iv medication compounding: recommendations from the international community of APOTECAchemo users. Am J Health Syst Pharm 74(1):e40–e46
Zastrow M (2020) South Korea is reporting intimate details of COVID-19 cases: has it helped? Nature. https://doi.org/10.1038/d41586-020-00740-y
Acknowledgements
This work was supported by JST Moonshot R&D Grant Number JPMJMS2215.
Author information
Authors and Affiliations
Contributions
MMS proposed the paper’s outline and contributed to the regulatory model put forward by this study. CL contributed to AI and robotics case studies used in the COVID-19 pandemic. FS contributed to the development of privacy and data protection issues. All authors contributed to data collection, analysis, interpretation, and writing.
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Lauri, C., Shimpo, F. & Sokołowski, M.M. Artificial intelligence and robotics on the frontlines of the pandemic response: the regulatory models for technology adoption and the development of resilient organisations in smart cities. J Ambient Intell Human Comput 14, 14753–14764 (2023). https://doi.org/10.1007/s12652-023-04556-2
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s12652-023-04556-2