Keywords

1 Introduction: The Harmful Ecosystem of Migration Management Technologies

We are Black and the border guards hate us. Their computers hate us too. Adissu, living without immigration status in Brussels, BelgiumFootnote 1

Tucked away on a quiet street minutes from a major train station in Brussels, a house is at first indistinguishable from its nondescript neighbours. However, inside this ‘squat’ lives a bustling community. Made up of climate justice organisers, self-described anarchists, and social justice advocates, L’Autre Caserne provides shelter, food, and support to undocumented people who find themselves living in the Belgian capital. With a massive cardboard clock that reads ‘Revolution Time’ and ‘No One is Illegal’ stickers everywhere, the bright three-story building has skylight windows and even a salsa dance room, contrasting with the ever-present threat of eviction and arrest. On a sunny Sunday afternoon, over 30 undocumented people gathered together to share snacks and stories of their migration journeys, while sheltering in place during the Covid-19 pandemic.Footnote 2 Most people present were from Eritrea or Ethiopia, as the previous group of Syrians vacated their rooms a few weeks back, moving on and trying their luck with the elusive passage to the UK – a Promised Land that seems simultaneously full of opportunity yet unattainable except for a lucky few. The UK remains difficult to reach during the coronavirus pandemic, where irregular passage on a boat across the English Channel or on the back of a lorry all come with the risk of increased surveillance (Meaker, 2020) and potential indefinite detention, if apprehended.

Since the coronavirus lockdown started, this community in central Brussels has been sheltering in place together for nearly three months but many said they were itching to try their luck elsewhere, no matter the cost.

‘I am tired’, says Negasi, a young man in his twenties from Ethiopia. ‘I am tired and I want to go to the UK’. Negasi has been in Belgium for two years without papers and in Nuremberg, Germany, for five years before that. But this is not Negasi’s first time in Belgium: he was deported to Germany once before, after being apprehended by the Belgian police for sleeping in a park when he was homeless. His fingerprints set off a notification in the EURODAC system (the (European Asylum Dactyloscopy Database which records and matches people’s fingerprints across the EU) since he had previously filed for asylum in Germany. He said the biometric collection process was invasive, but he did not know what do to: ‘How can I say no when the police handcuffed me, brought me to the station, and forced me to give them my fingerprints?’ He knows of a few friends who even went as far as to burn their fingerprints off to avoid detection but Negasi thinks ‘this don’t solve the problem’, as lack of identification usually will mean longer detention. As a young Black man, Negasi remembers his interactions with the Belgian police vividly during his 30 days in jail before being deported to Germany. ‘They hit and kicked me and said “if you’re not happy with this life, go back to your fucking country.” They really have no respect for us, not today and not ever’.

Negasi, like many others in the community has had a complicated journey to Belgium. When he left Ethiopia, he made his way to Sudan, Chad, and Libya, before taking a boat across the Mediterranean in 2014. His journey echoes a similar path that brought Amari to Brussels. In his forties and serving as the community’s Amharic interpreter, Amari doesn’t remember how long he has been in Belgium exactly; he entered Europe through Italy and tried to join his sister in Sweden but was returned and made his way to Belgium.

According to Amari, the general feeling is that ‘people think they have better chances if they go upwards’, into western and northern Europe. Many try to forget the horrors they experienced on their way. Amari was jailed for a year and a half in an underground Libyan bunker because he did not have enough money to pay the smugglers to take him across the Mediterranean (Mannocchi, 2019). He eventually escaped and joined a friend who was able to secure them both a passage to Italy. His face changed immediately when recalling his time in Libya and it was clear that the scars of that time remain: ‘It’s incredibly difficult to live in this world right now’. To cope, he is helping the ‘squat’ organisers make a 3D model of the detention facility from memory, both to document what happened to him and to perhaps help with his asylum claim, if he ever has a chance to file one. For now, Amari remains stuck in Belgium but has his eye on the UK or ‘whatever country takes me’. He was planning to try and catch a lorry to the UK sometime during the following week, aware of the risks but unwilling to remain in stasis any longer because of the coronavirus. Amari has experienced torture in a Libyan jail yet says ‘living without papers in a racist country is another torture. I only need papers to live. I want to study now. I want to have a reason to live’.

At multiple points in their journey that landed them in Brussels, this community interacted with surveillance technology and biometrics. Kaleb, a married man in his thirties who ‘misses [his] beloved wife Liya everyday’ reflected on how violent and impersonal the immigration and asylum system is. Twirling his wedding ring during the entire time we spoke, Kaleb reflected on feeling ‘like a piece of meat without a life, just fingerprints and eye scans’. As Kaleb spoke, the others in the circle nodded: ‘It is the human touch that keeps us warm – now all around us it’s cold as fuck’. Kaleb has not seen his wife in four years.

The group was disturbed by the coldness and dehumanisation of technology. As Eshe, a young woman who did not stay long at our meeting put it, ‘their computers are making decisions for us’. The use of drones in the Mediterranean and the English Channel prompted many people to shake their heads, with Amari making others laugh: ‘now we have flying shit instead of more help’. When discussing the use of tools like AI lie detectors at the border or drones patrolling the Mediterranean, the group discussion got heated: ‘We are Black and the police [border guards] hate us. Their computers hate us too’. The threat of police presence and increased surveillance is ever-present in an occupied building full of undocumented people – in fact, the police have been coming to the squat repeatedly over the last few weeks as the coronavirus lockdowns eased, once to investigate a death from an overdose of a young Moroccan man without papers and a few other times because the neighbours made a noise complaint.

Many in the community were not aware of what exact technologies they may have come in contact with over the course of the many months – and in some cases years – of travel. But Kaleb reflected on how he found it ‘weird’ that new tools were being used to control migration when instead ‘the world government should work on the main problem, like fighting and wars’ – problems that force people to seek a better life in the first place.Footnote 3

2 Real World Impacts of Technological Experiments in Pandemic Times

Adissu, Kaleb, and Eshe’s stories highlight the far-reaching individual impacts of migration surveillance technologies, exacerbated in the securitised and bio-surveilled world of Covid-19. Their reflections show just how little engagement there has been with the experiences of communities at the sharp edges of these unregulated technological developments. Each individual perspective is also firmly situated in broader ecosystems of migration management and border surveillance technologies, encouraging zones of experimentation with little oversight and accountability.

People on the move like Adissu are stuck in an ever-growing panopticon of technological experiments increasingly making their way into migration management. A whole host of actors and players operate in the development and deployment of migration control technologies, obscuring responsibility and liability, exacerbating racism and discrimination, and obfuscating meaningful mechanisms of redress. Communities made under-resourced and marginalised such as non-citizens, refugees, and people on the move often have access to less robust human rights protections and fewer resources with which to defend those rights and thus become technological testing grounds.

Now, as governments move toward increasing biosurveillance (Cliffe, 2020) to contain Covid-19’s spread, we are seeing a growth in tracking projects and automated surveillance technology, justified by the need to control the virus (Lewis & Mok, 2020). However, if previous use of technology is any indication, refugees and people crossing borders will be disproportionately targeted and negatively affected. Various proposed tools sound quite far-fetched, such as virus-targeting robots (The Current, 2020), cellphone tracking (Romm, 2020), and AI-based thermal cameras (Cox, 2020). However, given the robust migration management technology ecosystem that already exists, these Covid-specific tools can all easily be used against people crossing borders, with far-reaching human rights impacts. This use of technology to manage and control migration is also shielded from scrutiny because of its emergency nature. The basic protections that exist for more politically powerful groups that have access to mechanisms of redress and oversight are often not available to people crossing borders. The current global digital rights space also does not sufficiently engage with migration issues, at best only tokenising the involvement from both migrants and groups working with this community.

Ultimately, the primary purpose of the technologies used in migration management is to track, identify, and control those crossing borders. The issues around emerging technologies in the management of migration are not just about the inherent use of technology but rather about how it is used and by whom, with states and private actors setting the stage for what is possible and which priorities matter. The data-gathering inherent in the development of these technologies also includes the expansion of existing mass-scale databases underpinning these practices to sensitive data, especially biometrics. Such data and technology systems provide an enabling infrastructure for many automated decision-making projects with potentially harmful implications. The development and deployment of migration management is ultimately about decision-making by powerful actors on communities with few resources and mechanisms of redress.

The introduction of new technologies impacts both the processes and outcomes associated with decisions that would otherwise be made by administrative tribunals, immigration officers, border agents, legal analysts, and other officials responsible for the administration of immigration and refugee systems, border enforcement, and refugee response management. Border enforcement and immigration and refugee decision-making sit at an uncomfortable legal nexus: the impact on the rights and interests of individuals is often very significant, even where the degree of deference is high and the procedural safeguards are weak. There is also a serious lack of clarity surrounding how courts will interpret administrative law principles like natural justice, procedural fairness, and standard of review where an automated decision system is concerned or where an opaque use of technology operates. As such, it is concerning that we are already seeing a rise in Covid-specific technologies aimed at strengthening border security, aiding in data gathering, and automatic discretionary processes and decision-making with little oversight and accountability.

The political economy in which this technological development and deployment occurs also cannot be ignored. The unequal distribution of benefits from technological development privileges the private sector as the primary actor in charge of development, with states and governments wishing to control the flows of migrant populations benefiting from these technological experiments. Governments and large organisations are the primary agents who benefit from data collection (Okediji, 2018) and affected groups remain the subject, relegated to the margins. It is therefore not surprising that the regulatory and legal space around the use of these technologies remains murky and underdeveloped, full of discretionary decision-making, privatised development, and uncertain legal ramifications.

This chapter draws on data collected during a year-long study with European Digital Rights (EDRi) to interrogate the effects of migration management technologies on the lives and rights of people on the move and to foreground the lived experiences of these communities (see Molnar, 2020). This chapter reflects some of the interviews with over 40 refugees, asylum seekers, migrants without status, and people on the move conducted in Brussels, Belgium, and various locations in Greece over the summer and early fall of 2020.Footnote 4 The first section canvases some of the human rights ramifications of migration- and Covid-tech, followed by an analysis of some of the private sector drivers behind this turn to biosurveillance and migration management technologies, arguing that this pandemic emergency provides the perfect foil to fast-track problematic surveillance and automated digital technologies without taking into account their impacts on the rights and lives of people on the move.

3 Human Rights Ramifications of Migration- and Covid-Tech

Like a piece of meat without a life, just fingerprints and eye scans. Kaleb, living without immigration status in Brussels, Belgium

Technologies such as automated decision-making, biometrics, and unpiloted drones are increasingly controlling migration and affecting millions of people on the move. This allure of using technological interventions at and around the border highlights the very real impacts on people’s rights and lives, exacerbated by a lack of meaningful governance and oversight mechanisms of these technological experiments.

Even before the pandemic, surveillance and automated decision-making technologies had been increasingly used in securing border spaces, infringing on people’s life and liberty and their freedom of movement. For example, Frontex, the European Border and Coast Guard Agency, announced an expanded border strategy bolstered by a new regulation (Frontex, 2019) which relies on increased staff and new technology. An EU-funded project, ROBORDER (n.d.), explicitly ‘aims to create a fully functional autonomous border surveillance system with unmanned mobile robots including aerial, water surface, underwater and ground vehicles’. The EU borders are not the only site of drone technology. In the US, politicians have presented similar ‘smart-border’ technologies as a more ‘humane’ alternative to the Trump Administration’s calls for a physical wall. Most recently, this includes a partnership between the US Customs and Border Protection, Google Cloud AI, and Anduril Industries to create a new ‘virtual’ wall of surveillance towers and drones, a move that has been endorsed by the Biden Administration (Fang & Biddle, 2020; also Bernd, 2021). However, these technologies can have drastic results. For example, border control policies that use new surveillance technologies along the US-Mexico border have actually doubled migrant deaths (Boyce et al., 2019) and pushed migration routes towards more dangerous terrain through the Arizona desert, creating what anthropologist Jason De Leon calls a ‘land of open graves’ (De Leon & Wells, 2015). With similar surveillance technology increasingly used on the shores of Europe to facilitate interceptions and pushbacks of boats (Forensic Architecture, 2020; Heller & Jones, 2014; Keady-Tabbal & Mann, 2020), a similar increase of watery graves will likely occur (see Border Violence Monitoring Network, n.d.). Given that the International Organization for Migration (IOM) has reported that due to recent shipwrecks, over 20,000 people have died trying to cross the Mediterranean since 2014 (Black, 2020), we can only imagine how many more bodies will wash upon the shores of Europe as the situation worsens along the EU’s borders. The use of these technologies by border enforcement is only likely to increase in the ‘militarised technological regime’ (Csernatoni, 2018) of border spaces, without appropriate public consultation, accountability frameworks, and oversight mechanisms. This increased reliance on border securitisation and surveillance through new technologies, as clearly underscored by the EU’s New Migration Pact (European Commission, 2020) and its focus on border enforcement and deterrence, also works to send a clear message that human lives are expendable to protect Europe’s borders.

We should not underestimate the far-reaching impacts of new technologies on the lives and rights of people on the move. The right to life and the right liberty, the right to be free from discrimination, the right to privacy, and a host of other fundamental internationally protected rights are highly relevant to technological experimentation in migration and refugee contexts.Footnote 5 For example, aspects of training data which are mere coincidences in reality may be treated as relevant patterns by a machine-learning system, leading to arbitrary, incorrect, or discriminatory outcomes.Footnote 6 Given the problematic track record that automated technologies already have on race and gender, similar issues likely occur in migration surveillance and decision-making. Proxies for discrimination, such as country of origin, can be used to make problematic inferences leading to discriminatory outcomes.

Algorithms are vulnerable to the same decision-making concerns that plague human decision-makers: transparency, accountability, discrimination, bias, and error (Tufekci, 2015). The opaque nature of immigration and refugee decision-making creates an environment ripe for algorithmic discrimination. Decisions in this system – from whether a refugee’s life story is ‘truthful’ to whether a prospective immigrant’s marriage is ‘genuine’ – are highly discretionary, and often hinge on assessment of a person’s credibility (Satzewich, 2015). To the extent that these technologies will be used to assess ‘red flags’, ‘risk’ and ‘fraud’, they also raise definitional issues, as it remains unclear what the parameters of these markers will be.

These risks are not merely speculative. Biases at the border have far-reaching results if they are embedded in the emerging technologies being used experimentally in migration. For example, in airports in Hungary, Latvia, and Greece, a new pilot project by a consortium called iBorderCtrl introduced AI-powered lie detectors at border checkpoints (Picheta, 2018).Footnote 7 The project claimed that passengers’ faces would be monitored for signs of lying, and if the system became more ‘sceptical’ through a series of increasingly complicated questions, the person would be selected for further screening by a human officer. Canada and Romania have also experimented with a similar border-screening ‘emotion-recognition’ project called AVATAR (Daniels, 2018; Kendrick, 2019).Footnote 8

However, it is unclear how these systems will be able to handle cultural differences in communication or account for trauma and its effects on memory, such as when dealing with a traumatised refugee claimant unable to answer questions clearly.Footnote 9 Refugee claims and immigration applications are filled with nuance and complexity, qualities that may be lost on automated technologies, leading to serious breaches of internationally and domestically protected human rights in the form of bias, discrimination, privacy breaches, and due process and procedural fairness issues, among others. It remains unclear how the right to a fair and impartial decision-maker and the right to appeal a decision will be upheld during the use of automated decision-making systems.Footnote 10

The complexity of human migration is not easily reducible to an algorithm. Yet states are willing to experiment with these new unregulated technologies in the space of migration precisely because it is a discretionary space of opaque decision-making. The development and deployment of technologies also reinforce the North–South power asymmetries and concretise which locations are seen as innovation centres, while spaces like conflict zones and refugee camps become sites of experimentation under the guise of ‘humanitarianism’ and ‘empowerment of migrants’ through innovation.Footnote 11 Technological innovations exude the promises of increased fairness and efficiency. Moreover, much of migration management is also enacted by international organisations such as the UNHCR and IOM. As non-state actors operating under various legal and quasi-legal authorities and regulations globally, international organisations are ‘arenas for acting out power relationships’ (Evans & Wilson, 1992) without being beholden to the responsibilities that states have to protect human rights. States that operate through international organisations can also ‘launder’ their legal responsibility for acts or omissions that are attributed to the organisation (Benvenisti, 2018). With the proliferation of migration control technologies, international organisations are overly empowered to administer technology without having to abide by rights-protecting laws and principles, resulting in problems with compliance (See, for example, Raustiala & Slaughter, 2002).

The appetite for these advances also reveals the fissures of imbalanced power relations in society. Technological development does not occur in a vacuum but replicates existing power hierarchies and differentials. Technology is not inherently democratic, and issues of informed consent and right of refusal are particularly salient in humanitarian and forced migration contexts when, for example, refugees in Jordan have their irises scanned in order to receive their weekly rations under the justification of efficiency, while not being able to refuse biometric registration (Staton, 2016). Technologies of migration management also operate in an inherently global context. They reinforce institutions, cultures, policies, and laws, and exacerbate the gap between the public and the private sector, where the power to design and deploy innovation comes at the expense of oversight and accountability.

Unfortunately, an exceptional crisis moment like a global pandemic seems to be accelerating the development of technological interventions in and around the border without appropriate safeguards and oversights. This push to render people on the move as viable testing communities for further surveillance opens up spaces for greater justifications for border enforcement and securitisation, particularly in frontier zones with little oversight and accountability, rife for the private sector to foreground profit-making as the priority agenda.

4 Politics, Pandemics, and Privatisation of Migratory Spaces

Their computers are making decisions for us. Eshe, living without status in Brussels, Belgium

Pandemic responses are political (Khatin, 2020; also, Arundhati, 2020). Refugees have long been tied to tropes of bringing disease and illness (The World, 2019), underscored by growing xenophobia and racism (UNHCR, 2020). Not only are these formulations inaccurate; they also legitimise incursions on human rights (Herrera, 2019). If previous use of technology is any indication, refugees and people crossing borders will be disproportionately targeted and Covid-tracking technology can be easily co-opted to impact communities made marginalised (Lukacs, 2020). Making people on the move more trackable and detectable justifies the use of more technology and data collection in the name of public health and national security.

Most importantly, technological solutions do not address the root causes of displacement, forced migration, and economic inequality, all of which exacerbate the spread of global pandemics like Covid-19 (Molnar & Naranjo, 2020). Coupled with extraordinary state powers, the incursion of the private sector’s solutions presented to manage both migration and the pandemic leaves open the possibility of grave human rights abuses and far-reaching effects on civil liberties, particularly for communities on the margins.

5 Pandemics at the Frontier: Greek Refugee Camps in the Age of Surveillance

Certain places serve as the perfect testing grounds for new technologies, places where regulation is limited and where an ‘anything goes’ frontier attitude informs the development and deployment of surveillance at the expense of humanity. Greece is one of these places, a surveillance sandbox at the frontiers of Europe.

Moria, Europe’s largest refugee camp, burned to the ground on 9 September 2020. I visited Lesvos in the aftermath to document the building on a new containment centre and begin mapping out how this locale fits into broader narratives of technological experimentation. After the fire, thousands of people were sequestered on a barren stretch of road (Lavella, 2020) without food or water, tear-gassed (BBC News, 2020), and then herded into a new camp hastily built on the grounds of an old shooting range on a windswept peninsula (Solomon, 2020). This rocky outcropping is the newest site of containment on Europe’s borders, one housing over 9000 people displaced during a global pandemic, with no idea when or how they will be able to leave.

However, instead of opening the island camps and coming up with a meaningful plan for years of inaction, the EU’s new Migration Pact explicitly doubles down on containment and border security, opening the door to increasingly draconian tools of surveillance using new technologies – including the use of facial recognition – not just on adults but also on minors (European Commission, 2020). More and more, violent uses of technology work to push European borders farther afield (Howden et al., 2019), contributing to policies of border externalisation (Hernàndez, 2020), making Europe’s migration issues someone else’s problem. These policies have direct and dire consequences – drownings in the Mediterranean (Trilling, 2020), pushbacks to Libya (Kingsley, 2020) and Turkey, including using floating tents (Keady-Tabbal & Mann, 2020), and years-long detention in decrepit camps like Moria and other sites on islands like Samos, Chios, and Kos.

Frontier countries like Greece, ‘Europe’s Shield’ (Jamieson et al., 2020), act as testing grounds for new technologies and surveillance mechanisms. In October 2019, new legal rules were adopted as regards the deployment of drones by the Hellenic Police (Presidential Decree 98/2019), including the use of drones to monitor migration in border regions. Groups like Homo Digitalis in Greece claim that the new rules do not address the challenges arising from the applicable data protection legislation and have filed an open letter to the Ministry of Citizen Protection requesting more information about the deployment of drones by the Hellenic Police (Homo Digitalis, 2020). Drones, along with cameras and various other surveillance technologies are also being used and tested along the Evros land border with Turkey, in an increasing push to militarise migration management (Human Rights 360, 2020; also, Kalafatis, 2020; also Gatopoulos & Kantouris, 2021).

In September 2020, Frontex also announced that it was piloting a new aerostat maritime surveillance system (Frontex, 2020b), using Greece as a testing ground. The current pandemic conditions must also not be discounted, as they will likely expedite and exacerbate the turn to technological solutions at the border. We are already seeing the border industry pushing for increased adoption of ‘contactless biometrics’ for ‘regular’ travellers as a way of stopping contagion (PR Newswire, 2020). Frontex in particular has been clear in its messaging to position itself as an agency apt at both controlling migration as well as the spread of Covid-19. According to a May 2020 press release, ‘if we cannot control the external borders, we cannot control the spread of pandemics in Europe. Frontex plays a key role in ensuring effective protection of the external borders of the European Union not only against cross-border crime but also against health threats’ (Frontex, 2020a). The EU continues to make its priorities clear: containment, surveillance, and technosolutionism at the expense of human rights (Molnar, 2021).

The appetite for migration management technologies remains present in Greece, with the Hellenic Ministry of Migration and Asylum co-hosting and supporting the World Border Security Congress in the fall of 2021, a gathering of public and private actors eager to address ‘threats’ such as ‘ISIS threatens to send 500,000 migrants to Europe’ and ‘Migrants and refugees streaming into Europe from Africa, the Middle East, and South Asia’.Footnote 12 And as recently as December 2020, tech giant Palantir proudly announced a partnership with the Greek government, stating: ‘Our partnership with the Greek government was borne out of necessity once the pandemic began, and we look forward to broadening it for years to come’ (Business Wire, 2020). It is currently unclear what exactly this partnership entails but it bears to remember that Palantir is the same company that has been facilitating ICE detentions and deportations at the US-Mexico border (Frenkel, 2018) and has been implicated in a host of other human rights abuses (Hemmadi, 2019; O’Brien, 2020).

The use of technology is never neutral – it reinscribes the way that powerful actors make decisions that affect thousands of people. Along with Big Tech, big money is also involved in the management of borders, with private security companies making major inroads (Privacy International, n.d.) with lucrative contracts procured by governments for shiny new tech experiments presented as a way to strengthen border security. These technological experiments also play up the ‘us’ vs ‘them’ mentality at the centre of migration management policy. Instead of long-term viable redistribution of resources across the EU and timely processing of people’s asylum, turning to techno-solutionism (Molnar & Naranjo, 2020) and migration surveillance will only exacerbate deterrence mechanisms already so deeply embedded in the EU’s migration strategy.

Various people in our interviews on Lesvos were concerned about proposed screening and surveillance mechanisms to keep people contained, tracked, and managed and the normalisation of surveillance in the aftermath of the Moria fire. Yet no one was clear on exactly how the next few months would shape up, particularly regarding increasingly technological incursions, data collection, and surveillance. However, the planned rollout in 2021 of so-called Multi-Purpose Reception and Identification Centres (MPRICs) on the islands of Lesvos, Samos, Leros, Chios, and Kos all seem to indicate a turn to both containment and surveillance, creating what are essentially closed prison camps complemented by ‘camera surveillance with motion analysis algorithms monitoring the behaviour and movement of centre residents’ (Europe Must Act, 2021; Petridi, 2021). Clearly, migration management through technology is firmly on the agenda.

Given the far-reaching and high-risk nature of these technological experiments, the unique context of migration should be the central consideration when analysing which human rights should be taken into consideration when exploring new technologies, given the very real risks to life, liberty, and security, as well as heightened privacy considerations, particularly during an unprecedented global pandemic. Yet states can justify technological experiments in migration control precisely because migrants are not able to exercise the same rights as citizens and because they are seen as a useful tool through which to exercise powers of sovereignty in an increasingly destabilised world.

6 Big Tech and Big Profit in the Privatisation of Migration Management

Ultimately, the primary purpose of the technologies used in migration management is to track, identify, and control those crossing borders – and now doing so in the middle of a global health crisis. The issues around emerging technologies in the management of migration are not just about the inherent use of technology but rather about how it is used and by whom, with states and private actors setting the stage for what is possible and which priorities matter. The data-gathering inherent in the development of these technologies also includes the expansion of existing mass-scale databases that underpin these practices to sensitive data, especially biometrics. Such data and technology systems provide an enabling infrastructure for many automated decision-making projects with potentially harmful implications. The development and deployment of migration surveillance technologies is ultimately about decision-making by powerful actors on communities with few resources and mechanisms of redress, which becomes justified by the state of emergency such as a global pandemic.

Politics also cannot be discounted, as migration management is inherently a political exercise. Migration data is already being politicised to support greater interventions in defence of threatened national sovereignty (Scheel & Ustek-Spilda, 2018), reinforcing politics of exclusion. The state’s ultimate power to decide who may enter and under what conditions (ibid.) is bolstered by ongoing beliefs in technological impartiality. However, there is an inherent tension between the claimed prerogative of nation-states over sovereignty and the malleable nature of technology. In its fluidity, technology is inherently oppositional to borders, and by extension sovereignty. Indeed, oftentimes it impinges on the very definition of ‘humanness’ in the digital era (Zureik & Hindle, 2004). Ultimately, the primary purpose of the technologies used in migration management is to track, identify, and control those crossing borders. The unequal distribution of benefits that accrue in technological development work to create monopolies of knowledge and consolidate power and authority vested in the sovereign state. These monopolies exist because there is no unified global regulatory regime governing the use of new technologies, creating laboratories for high-risk experiments with profound impacts on people’s lives.

Even before the pandemic, the human body itself increasingly became the embodiment of progressively violent and discriminatory border regimes. For example, in March 2020, the US announced that it would begin collecting DNA samples from hundreds of thousands of migrants apprehended along the US-Mexico border (Hauslohner, 2020). This practice will dramatically expand a federal database of individual genetic information used by law enforcement.Footnote 13 Migrant rights groups in Mexico have also reported the use of digital technologies to spread misinformation that stigmatises migrants as criminals or carriers of diseases (Cadena, 2020), including Covid-19. The stigmatisation that coordinated digital misinformation campaigns spurred hostility towards migrants, including the risk of violence against them. There are increasing reports of harassment against migrant rights defenders and journalists covering migrant caravans in Central America, Mexico, and the US, including digital threats, searches of digital devicesFootnote 14 and even electronic surveillance by authorities in the US and Mexico (Jones et al., 2019).

Growing anti-migrant xenophobic sentiment, justification of surveillance and online media monitoring under the guise of security and public health, and the rise of extreme right and neo-fascist groups or political organisations globally impacts how migration management technologies function. From the Proud Boys in the US to the autocratic administration of Victor Orban in Hungary to the Maduro regime in Venezuela, people on the move and migrants have been linked to threats to national sovereignty (The World, 2019) that must be stopped at all costs – and most recently explicitly connected to the spread of the Covid-19 virus and characterised as ‘vermin’ and ‘biological weapons’ (Cohen, 2018).Footnote 15. Far-right groups have been engaged in violence again migrants around the world, including in the US, mainland Europe, and Canada. In border frontiers such as Greece, far-right extremism and anti-migrant sentiments reached boiling point, with the island of Lesbos becoming an epicentre for extreme right groups across the EU (Fallon, 2020). Most recently, ties have been discovered between far-right extremists and companies like Clearview AI and Palantir, which are responsible for the development and deployment of facial recognition technologies and algorithmic decision-making tools used for the detention and deportation of migrants (O’Brien, 2020). These groups have also been blatantly calling for the establishment of so-called ‘deportation squads’ using algorithms to identify potential targets (ibid.). The implications of Big Tech giants and corporate interests in these dangerous anti-migrant narratives is a very troubling development, all the more so because public-private partnerships lie at the heart of the development and deployment of migration management technologies.

The lack of adequate technical capacity within government and the public sector can lead to potentially inappropriate over-reliance on the private sector. Adopting emerging and experimental tools without in-house talent capable of understanding, evaluating, and managing these technologies is irresponsible and downright dangerous, as the companies and corporations hold the balance of power when determining what technology is developed and deployed, and subsequently procured by governments. Private sector actors have an independent responsibility to make sure technologies that they develop do not violate international human rights and domestic legislation. Yet much of technological development occurs in so-called black boxes, where intellectual property laws and proprietary considerations shield the public from fully understanding how the technology operates. Powerful actors can easily hide behind intellectual property legislation or various other corporate shields to launder their responsibility and create a vacuum of accountability.

These practices also give rise to the Border Industrial Complex (Miller, 2018) – the confluence of border policing, militarisation, and financial interest. States are seeking to leave people on the move, refugees, and undocumented people in particular beyond the duties and responsibilities enshrined in law through an overreliance on the private sector to ensure technological experimentation occurs outside of sovereign responsibility.Footnote 16 The growing role of the private sector in the governance of new technologies highlights the movement away from state responsibility to create governance structures in accordance with domestic and international principles under guise of proprietary technology, private interests, and discretion.

Increasingly, private companies are setting the migration control agenda. Whether through the automation of immigration and refugee applications, AI-powered ‘lie detectors’ at borders, or drone surveillance of refugee boats, the private sector is an integral player in the management of migration. States and government departments over-rely on private actors to develop and deploy technologies used to control migration. As a result, government liability and accountability are watered down and shifted to the private sector, where the legally-enforceable rights that allow individuals to challenge governments may not exist (see Statewatch, 2020). People on the move get caught in the middle, leading to grave human rights abuses and infringements on fundamental freedoms. The opaque, private technologies deployed at border zones to control migration desperately need to be regulated. Yet, the creation of legal black holes in migration management technologies is very deliberate to allow for the creation of opaque zones of technological experimentation that would not be allowed to occur in other spaces (among others, Molnar & Naranjo, 2020). While we are able to imagine mobilising around extreme issues such as the banning of killer robots in armed conflict, the grey spaces of migration management technology remain largely uncontested.

7 Concluding Remarks: Surveillance Panacea or Safety and Health for All?

They really have no respect for us, not today and not ever. – Negasi, living without status in Brussels, Belgium

While technology can offer the promise of novel solutions for an unprecedented global crisis, Covid-led innovation should not unfairly target refugees and people on the move, racialised communities, Indigenous communities, and other marginalised groups, or make discriminatory inferences that can lead to detention, family separation, and other irreparable harms. While often presented with a lot of promise, too often technological tools easily become tools of oppression and surveillance, denying people agency and dignity, and contributing to a global climate that is increasingly more hostile to people on the move. The pandemic is already being used to curtail access to people living in informal settlements or securitised refugee camps on Greek islands and detention centres for undocumented populations across Europe and throughout the world. However, instead of increased tracking and surveillance, a redistribution of resources, dignified living conditions, and access to medical care are paramount to stopping the spread of the pandemic.

As this chapter has highlighted, drawing on conversations with people on the move in Belgium and Greece, affected communities must be directly involved in technological development and discussions around proposed interventions, shaping and driving the conversation, not consulted as an afterthought or in a tokenising extractive way While conversations around the ethics of AI are taking place, ethics do not go far enough. What is needed is a sharper focus on oversight mechanisms grounded in fundamental human rights and context-specific accountability that recognises the particular lived experiences of people on the move and their experiences in the time of exception that is a global pandemic in an increasingly securitised and surveilled world.

Yet there is also a hopeful promise in the proliferation of new technologies in migration management. Policymakers, academia, and the public are being forced to reckon with fundamental normative ideas around what constitutes intelligence, how to manage and regulate new systems of cognition, and who should be at the table when designing and deploying new tools that can be used to either dismantle or reinforce the status quo. Culture, politics, institutions, and technology all iteratively shape one another. Ultimately, technology is a social construct (Franklin, 1990), a mirror to reflect the positives and negatives inherent in our societies, forcing us to rethink ideas of privilege and power in the current global crisis that is reshaping our world at every level. It remains to be seen whether the current global push towards fervent technological innovation will result in robust global governance, centred on the experiences of people on the move, or whether it will further dehumanise people on the move as they are increasingly having to contend with the sharp edges of harmful technological development in the time of a global health crisis.