What if being-fugitive bears the possibility of a recalibration of the human, a reopening of, rather than an opening to, the not open? New world. New world. Fred Moten, The Universal Machine (consent not to be a single being) (2018), 39.
Abstract
This article starts from the observation that practices of ‘algorithmic governmentality’ or ‘governance by data’ are reconfiguring modes of social relationality and collectivity. By building, first, on an empirical exploration of digital bordering practices, we qualify these emergent algorithmic categories as ‘clusters’—pulsing patterns distilled from disaggregated data. As fluid, modular, and ever-emergent forms of association, these ‘clusters’ defy stable expressions of collective representation and social recognition. Second, we observe that this empirical analysis resonates with accounts that diagnosed algorithmic governance as a threat to legal subjectivity and socio-political cohesion, and called for a reinvigoration of democratic values and their re-alignment with new ‘infrastructural publics’. Against this backdrop, however, we explore alternatives avenues of legal imagination by pushing in a different (somewhat opposite) direction. Against the re-inscription of liberal categories, we linger with the promise and prospect of illegibility as resistance against the foreclosure of future potentialities in algorithmic forms of subject-making. Instead of falling back on the projection of autonomous human agency and liberal subjectivity to counteract the ‘cluster’, we imagine emancipatory expressions of resistance that are enacted through fugitive, opaque, and experimental collectivities.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Data doubles, algorithmic subjects, digital selves (Haggerty and Ericson 2000; Aradau and Blanke 2018; Lupton 2020): as these new cybernetic figures enter the terrain of our thinking, the ‘spiral dance’ of the cyborg (celebrated by Haraway as an escape from the grasp of modernity) seems to have become a somewhat more anxious routine (Haraway 1991 pp. 149, 181).Footnote 1 If, for Haraway, the ‘hybrids of machine and organism’ suggested a way out of the ‘maze of dualisms’ in which the modern subject is trapped (Haraway 1991 p. 181), contemporary accounts of the more-than-human, data-driven avatars that are enacted in practices of algorithmic governance signal not the ontological emancipation but the epidermalization, dissolution, and displacement of the subject (Browne 2010 p. 131, Rouvroy 2012). Building on these strands of literature, this contribution traces how ‘algorithmic governmentality’ or ‘governance by data’ thereby disrupt some central tenets of social relationality and collectivity embedded in modernist ideals of the liberal legal subject (Rouvroy and Stiegler 2016). We show this by exploring digital bordering practices, to demonstrate how at the ‘virtual border’, the human subject is splintered into a ‘cluster’ of pulsing patterns distilled from disaggregated data. Yet, while diagnosing a disruption of the liberal ideals of the subject, we keep a distance from precisely these ideals—ideals that flourish in attempts to reinvigorate the ‘rule of law’ and to attune its values of human autonomy, publicness, or democratic inclusion to new computational settings (Hildebrandt 2020). We trace the algorithmic reconfiguration of sociality without reifying modernist notions of liberal subjectivity.
The argument unfolds in three steps. First, we recount how practices of subjectivation have been problematized in critical theory—with sites of subject-making being subversively (re)appropriated—to counter modes of neoliberal governmentality. We explore how such projects are complicated by emergent practices of ‘governance by data’ where the conditions of collective recognition and being-in-common—on which the activation of ‘subversive subjectivities’ hinges—are eroded. Drawing on an analysis of algorithmic risk assessment at the ‘virtual border’, the article identifies the formation of ‘clusters’ as objects of governance and modes of relationality that increasingly displace alternative enactments of (individual or collective) subjectivity. Second, we observe that this displacement can be read together with accounts on the degradation of sociality resulting from algorithmic governance and the crisis for the ‘rule of law’ that this entails. Finally, however, we resist reaffirming modernist modes of representation that are called for in these strands of scholarship and propose a style of ‘legal imagination’ that stays with fugitive, opaque and experimental modes of being and becoming. This be(com)ing does not strive for more transparency, autonomy, and inclusion under the guise of formal equality but seeks to activate the indeterminacies of algorithmic calculation—its forks, thresholds, weights and apertures—as openings toward different expressions of sociality and being-in-common. With caution and awareness of our positionality as White scholars, this attempt at inviting a new ‘legal imagination’—in the theme of this special issue—draws on different strands of critical Black studies.
2 Subjects and clusters: new modes of algorithmic be(com)ing
In social and critical theory, expressions of power and domination (as well as resistance, solidarity and emancipation) have long been perceived as entangled with the production of subjectivity—with the techniques of subjection or subjectification that shape social and legal agents in service of specific ideological formations or governance scripts (Althusser 1971; Foucault 1982; Butler 2000; Samaddar 2010).Footnote 2 In this vein, our analysis aligns with Foucault’s interest in ‘the different modes by which, in our culture, human beings are made subjects’ (Foucault 1982 p. 777)—in the relational and material processes of be(com)ing subject. The focus on materiality is important here: inspired by contemporary strands of new materialism and STS, we are particularly interested in how subjects are performatively enacted through sociotechnical (and increasingly data-driven) practices of governing.Footnote 3 While technologies of seeing, sensing, and sorting have long been perceived as central to the making and modulation of subjects in disciplinary or biopolitical regimes of rule (Hacking 2006), we think with those who argue that big data and associated algorithmic analytics are changing the conditions of visibility and compositions of sociality in these political formations—changes that spark specific legal anxieties (Aradau and Blanke 2018; Isin and Ruppert 2020; Johns 2021; Amaro 2021).
Lingering a little longer with these prior accounts, we share their perspective on power as productive: as an alignment of rationalities and techniques that engender new images of the political and social subject (Mezzadra and Neilson 2013, pp. 250ff).Footnote 4 Brown, for example, qualifies neoliberalism as a ‘constructivist project’ that is not primarily aimed at regulating economic relations but at cultivating the figure of the homo economicus as central vector of social and political life (Brown 2005, p. 40). This claim has powerfully been reframed by Pahuja (2011) in the context of international law’s enduring colonial commitments. To understand the ‘powerful technology of universalization’ at work in international law and development, Pahuja notes, we need to account for how the ‘poor person’ is ‘interpreted as a very specific (legal) subject, the homo œconomicus’: ‘[t]his proto-capitalist subject, possessed of an agency that is entrepreneurial and not radical’ (Pahuja 2011, pp. 219–221). In a similar vein, Urueña (2012) gives an account of the ‘fragmented subjectivities’ emerging from international law’s increasing preoccupation with the individual. Suspicious of those celebrating the empowering or emancipatory status of individuals as subjects of international law,Footnote 5 Urueña (2012) observes the alienation of those caught between the conflicting subjectivities of self-improvement and dependency—between the homo economicus of international economic law and the ‘empty souls’ of human rights law who can only ever be ‘spoken for’. The preoccupation of international lawyers with the productive power of subject-making is perhaps most explicit in critiques on quantification or governance by indicators. Situating these metrological techniques in a longer lineage of governmentality, Merry (2011, p. 90) traces how ‘the indicator comes to shape subjectivity’—engaging the person in ‘governing himself or herself in terms of standards set by others’.Footnote 6 The common threads in these critical accounts point to the production of a neoliberal subjectivity where discipline is internalized, multiplicity erased, and agency eroded.
Yet, ‘writing in the spirit of Marx’, Mezzadra and Neilson (2013, p. 252)Footnote 7 signal how these modes of subject-making are inevitably tied to the ‘theme of the liberation of subjectivity’—‘of revolutionary subjectivity’. The ‘subject’, they assert, is marked by an ‘excess that can never be fully expropriated’: a defiant and disruptive subjectivity emerging ‘in relation to and in tension with the figures … that are its correlates in the legal and political realms’ (Mezzadra and Neilson 2013, p. 252, 264). Rancière (1998, pp 16, 28), in this vein, situates the ‘political’ sphere (in contrast to the regimes of ‘police’) in the ‘destabilizing and subversive … subjectivation of the part with no part’—a reactivation of the ‘contingency of equality’.Footnote 8 These (collective) acts of disruption and defiance are often explicitly oriented around or against existing legal inscriptions or classifications. When the constitutional right to ‘human dignity’ devolved into managerial ‘metrologies of dignity’ in the South African ‘toilet wars’, Marks (2021) observed, affected communities mobilized around defiant expressions of indignity, consciously shedding the veil of their abstract legal subjectivity to underline the injustices it sustained. In a different setting, Mann (2021, p. 147) observed how migrants dwelling in Europe’s borderlands often engaged in ‘strategies of concealment and misrepresentation’, whereby the formal status of legal citizenship is discarded and traded for the ‘costumes of people who enjoy the individual protections granted to humans as such’ (Mann retells an encounter with Dominican asylum seekers disguised as Somalis).Footnote 9 These ‘border masquerades’, as Mann observes, are subversive and strategic substitutions between different sources of legal personhood (the ‘citizen’ and the ‘human’) that seek to disrupt the technologies of exclusion grafted on the former with (consciously fraudulent) imaginations of defiant universality expressed in the latter (Mann 2021, p. 154).Footnote 10 With both Marks and Mann, we thereby encounter expressions of the excess to which Mezzadra and Neilson (2013, p 275) have referred: emergent forms of subjectivity that target and (re)appropriate existing legal inscriptions. Importantly, for such expressions to obtain a durable and organized political form, particular material conditions need to be in place for the translation and creation of a collective subject in the making. These can relate to experiences of social recognition, commonality of purpose, or a shared sense of duress on which the formation of subversive political subjectivities hinges.Footnote 11 In relation to the biopolitical or neoliberal regimes of rule to which we have briefly alluded above (those built, for example, around the statistical techniques of classifying and calibrating populations) (Foucault 2008), such conditions of collective action were always implicit. As Desrosières (2002, p. 401) argued, statistics can be politicized and polemicized precisely because they work with and enact ‘stable collective objects, or the production of categories that can become evaluated and contested publicly’.Footnote 12 Subversive subjectivities, in this sense, are always latent in the statistical classifier.
With the rise of ‘algorithmic governmentality’,Footnote 13 we want to argue, these practices of classification and subjectivation have taken novel forms. Big data does subjects differently. Our interest in tracing such new forms of subject-making is sparked first of all by the invitation of this special issue to reflect on how new digital technologies transform and trouble key legal concepts and categories, which we observe in the reconfiguration of the legal subject.Footnote 14 This inquiry is also motivated by the observation that the material making of subjects—how people are seen, sensed and sorted in specific governance regimes—entails dynamics of power, inequality and exclusion that we consider to be insufficiently addressed in current attempts to regulate, proceduralize, or fix new digital technologies. Finally, we are particularly interested in how the use of big data and data analytics alters (and potentially erodes) the conditions for collective action and political subjectivation briefly touched upon above.
One prominent site where these new digital division materialize is the ‘virtual border’: the assemblage of databases, algorithmic processes, and sensors that surveil and sort people on the move (many more sites could be identified, of course, from online advertising to tools of military targeting) (Van Den Meerssche 2022). In a recent strategy on ‘the use of artificial intelligence in border control, migration, and security’, the European Commission identified nine areas of opportunity for artificial intelligence (AI) in this domain, ranging from ‘vulnerability assessments’ in asylum applications or the use of data analytics to detect ‘irregular travel patterns’ to the algorithmic screening and ‘triaging’ of visa applications and border crossings (European Commission 2020). Displaying the pre-emptive logic of contemporary security practices, the stated objective of these AI systems, as a recent report of the European Parliament noted, is not just to ‘verify[] and identify[] known persons’, but ‘to identify unknown persons of interest based on specific data-based risk profiles’ (European Parliament 2021).Footnote 15 In this sense, the strategy states, AI would enable processes of ‘[r]isk assessment performed on a group of individuals with the general aim to find patterns and cluster individuals for further investigation’ (European Commission 2020 p. 10 emphases added).
This orientation towards ‘patterns’ and ‘clusters’ is elaborated in the technical sections of the strategy, which clarify that the ‘classification categories’ dividing people at the border ‘could be defined based on a risk threshold or specific indicators’ or could be ‘less pre-defined where applications are grouped based on some “learned” similarity’ (European Commission 2020 p. 89). In the latter case, the use case description states, unsupervised AI models would be employed, using ‘vector space models’ to ‘partition data into clusters’ (European Commission 2020).Footnote 16 The key ‘benefit of using AI’, the strategy notes, is a flexibility in ‘uncovering correlations between input data and classification outcomes’ while also ‘allowing for more variety in input’ (European Commission 2020).Footnote 17 The extension in what can count as a feature is also an advantage of such vector space models. As Mackenzie (2015, p. 434) explains: ‘[t]he difference between classical statistics, which [seek] to explain associations between variables, and machine learning, which seeks to explore high-dimensional patterns, arises because vector spaces juxtapose almost any number of features’.Footnote 18 The inclusion of ever-more heterogenous data is central in the strategy’s aim to ‘identify patterns which were not observed as “strange” before’—a distillation of meaningful attributes and features through the ‘unsupervised uncovering of correlations’ (European Commission 2020, p. 90).Footnote 19 It is the ‘uncovered correlation’ and not (only) the legal classification that is dividing people at the digital border.
These strategic ambitions have led to a range of EU-funded experiments, such as the iBorderCtrl and Tresspass projects, that developed AI systems to ‘compress all data into actionable risk scores’ (Van Den Meerssche 2022). The experimental character of these changes is significant, and signals a broader shift in policy. Aradau’s claim that ‘experimentality’ has become ‘a mode of governance in borderzones’—an experimentality attuned to the space of play of machine learning models—resonates in the recent UK Border Strategy, which sets out that ‘the private sector must take the lead on border innovation, with government … creating an environment that encourages experimentation and technology adoption’ (HM Government 2020; Aradau 2022). At the ‘virtual border’, the power to divide and distribute is increasingly wielded in experimental form.
Yet, we also already observe the workings of this logic in the legal and infrastructural development of concrete border control programs, both in the EU and the UK. The European Travel Information and Authorisation System (ETIAS) that is currently under construction, for example, has developed its decision-making process around the identification of ‘security risks’, ‘illegal immigration risks’, and ‘high epidemic risks’ (Council Regulation 2018/1240, OJ L 236/1, Article 4). The ETIAS regulation provides that these risk assessments will be based on ‘an algorithm enabling profiling’ based on ‘specific risk indicators’ (Council Regulation 2018/1240, OJ L 236/1, Article 33). The latter are to be created by Frontex, based on the specific ‘risks’ defined by the European Commission. In a recent delegated act, the Commission clarified that such ‘risks’ would be determined on the basis of ‘particular sets of characteristics’ distilled from data that signal a heightened propensity to patterns of refusal of entry, overstaying, identified security threats or observed disease outbreaks (European Commission 2021, articles 3–6).Footnote 20 This observation of particular ‘characteristics’—defined as a ‘distinguishing sets of observable qualities or properties’ that could be ‘attributed’ to ‘specific groups of travelers’—would also be enriched with the identification of ‘any correlation with information collected through their application files’ (European Commission 2021). In this sense, the practice of pattern detection will necessarily be a ‘live’ process of iterative review and adaptation through which observed attributes and correlations become part of the open-ended taxonomy of ‘risk’ (European Commission 2021). The role of Frontex is to develop ‘risk indicators’ on this basis—an ‘algorithm enabling profiling’—which would allow for the automated processing of applications (Council Regulation 2018/1240, OJ L 236/1, Articles 22 and 33). In doing so, it is assisted by the recently created EU Innovation Hub for Internal Security, which is involved in exploring how AI will be deployed to optimize the performance of profiles (EU Innovation Hub for Internal Security 2021).Footnote 21 These ‘profiles’ signal the particular practice of subject-making as a new mode of algorithmic be(com)ing at the ‘virtual border’: decisions are not based on stable legal signifiers but on ‘characteristics’ and ‘correlations’ temporarily tied to observed patterns and propensities in data. In the computational calculus of ‘risk assessment’ envisaged here, the subject that is being enacted is a cluster of inferred attributes: a relational association between dispersed data points (or in more technical terms: a spatial proximity between emergent features in a vector space).
The cluster is an increasingly prevalent form of algorithmic subject-making that appears far beyond the spaces of the ‘virtual border’.Footnote 22 In line with the account above, several salient (and intertwined) characteristics of the cluster can be distilled, which display how it ‘troubles the fundamentals of legal and social representation and relation’ (Johns 2021, p. 66). The cluster is a fluid, relational subject. The ‘specific groups of travelers’ referred to above are only tentatively and temporarily tied together based on attributes and patterns extracted from data and always open to revision.Footnote 23 This underlines how important the continuous collection of data is in the composition of clusters: in the ever-expanding vector space, sensory tools are key to figure and refigure a subject that is inherently incomplete and continuously in (re)composition.Footnote 24 In the case of the experimental iBorderCtrl and Tresspass projects referred to above, for example, this includes on-site observations, biometrics and biomarkers (emotion AI), and social media scraping. This subject is not defined by stable and identifiable features but emerges from relational ties and associations between dispersed data points—from patterns and propensities. This fluid, relational character of the cluster limits the forms of political subjectivation mentioned above: the ‘translation of the common’ is troubled by the impossibility of encounter and mutual recognition in the calculus of the cluster (Aradau and Blanke 2018, p. 21, Fourcade and Johns 2020, p. 824). As Amoore (2021, p. 6) observes: ‘[t]o belong to a cluster is not about resemblance, [or] common characteristics’, but merely reflects ‘a spatialized proximity or distance’.
The cluster is an emergent, composite subject. In contrast to the rigid categories and classifications that are at the core of critiques of indicators (and more general concerns on biopolitics) (Rose 1998; Merry 2011; Hacking 2002), the ‘specific groups of travelers’ referred to above are composed inductively through inferred ‘characteristics’: the taxonomy of risk is itself enacted in the learning process. The subject emerging in this algorithmic calculus does not correspond to the stable statistical classifiers or fixed figures of alterity and enmity that have long been targets of critique, but is rendered visible through shifting correlations detected in data—pulsing patterns temporarily assembled in actionable form (Steyerl 2019). Emerging outside and against fixed (legal or political) categories and criteria, these composite clusters could be perceived precisely as an expression of the excess referred to by Mezzadra and Neilson: a surplus, or virtuality, captured and coded through new sensory technologies.Footnote 25 This disjunction has important consequences for the application of legal standards on equality and non-discrimination (Van Den Meerssche 2022). As Amoore (2021, p. 7) argues: the ‘logic of the cluster, thus, affords the state a means of pursuing racist borderwork while circumventing the social and juridical rights’ associated with ‘twentieth century transcripts of characteristics and categories’. The composition of the cluster not only entails new technologies of racialization (as we will return to below),Footnote 26 but also seems to evade and exceed structural categories on which legal frameworks are grafted.
The cluster is an inferential, correlated subject. The aim of the envisaged projects described above is not to construct causal knowledge about specific phenomena but to reveal and infer new patterns, probabilities and propensities. In this sense, the composition of the cluster is not a representation (of pre-existing entities) or evaluation (of past actions) but a speculation or projection (of potential future behaviors). The future, in other words, is seized in advance through the creation of an ‘augmented reality’—or what Rouvroy and Berns (2013, p. 182) describe as a simultaneous actualization of ‘a memory of the future’ and ‘a systematized serendipity’—an ‘actualization of the virtual’.Footnote 27 This entails a troubling temporal bind where the subject’s agency and actions are conditioned by algorithmic anticipation (a concern which figures at the heart of the liberal responses canvassed in the next section). At the same time, this also entails a reconfiguration of how state agencies are organized and public decisions are made, explained or reviewed—a reconfiguration that is often seen to erode possibilities of human judgment and the standards of accountability tied to those.Footnote 28 Yet, it is important to note that the changes we describe do not displace prior modes of governance (and the forms of subjectivation these imply). Just as tools of algorithmic inference tend to be nested into rule-based systems, correlational patterns are routinely discarded when conflicting with pre-existing norms or segue into more formal causal codes. Logics and temporalities of rule co-exist and conflict. In light of the significant socio-legal challenges that the study of such composite system poses, our aim here was modest: we sought only to signal, through the example of the (re)making of the subject at the ‘virtual border’, some salient changes in the modes of subject-making that are enacted algorithmically. We located these changes in the fluid and relational, emergent and composite, inferential and correlational nature of the cluster—a fragmented ‘subject’ made up of dispersed data points in a vector space, whose behaviors are anticipated and predicted. The ideal of an identifiable and individuated ‘human being’ pertaining to (equally identifiable and individuated) social groups or collectives, whose actions can be traced and assessed, is here displaced by an open-ended data-based risk profile—a speculative category to which beings are associated for reasons often unknown to them but also to the agents overseeing control of the ‘virtual border’. The emergence of algorithmic subjects, we argue in the next section, has led to a novel register of critique, driven by anxieties in seeing the liberal subject being threatened.
3 Algorithmic subjects and liberal anxieties
The preceding section explored how practices of algorithmic governance both enact and foreclose particular modes of subject-making. There is nothing new, of course, in arguing that regimes of power produce the subjects on and through which they act.Footnote 29 Yet, we observed how the combined use of big data and AI entails a mode of subject-making that differs from prior forms of biopolitical or neoliberal governance. We traced these changes through a brief account of current border control strategies and technologies, which are oriented towards the composition of ‘clusters’ as objects of classification and intervention. Our analysis of the ‘cluster’ as a specific subject of contemporary ‘governance by data’ also considered how the conditions of possibility for legal protection, collective action or subversive forms of subjectivation implicit in prior regimes of rule now risk being eroded. Practices of mutual recognition—the social articulation of a collective ‘we’—from which political alliances and values emerge are displaced by speculative ‘clusters’ with which no subjective identification is possible. Prospects of commonality and publicness—as well as certain specific legal safeguards—increasingly appear as affordances of technologies and infrastructures that are being repurposed or replaced.Footnote 30 In this context, concerns are raised on how algorithmic governance threatens the treasured ‘rule of law’.
In this section, we briefly canvass three recurrent concerns and associated normative interventions: the observed assault on the autonomous human subject, the problem of publicness and inclusion, and the prevalence of algorithmic biases and associated inequalities. Combined, these different positions display a problematization of digital technologies from the perspective of particular liberal ideals of the subject and the public—ideals of human autonomy, inclusion, and equality. While we consider these perspectives as crucial in capturing what is at stake in the rise of ‘governance by data’, we also signal important limitations of this liberal lexicon, on which we expand in the final section.
In the encounter with practices of algorithmic governance, perhaps the primary concern of critical scholars is with the erosion of the autonomous, rational and self-reflexive (legal and political) subject, which risks being reconfigured and displaced by the disembodied proxy of the cluster. As we argued, indeed, the composition of clusters alters how individuals are conceived, recognized, and interpellated as subjects.Footnote 31 Rouvroy and Berns (2013, pp. 173–174) lament, in this light, that both governmental and corporate entities are not interested in the individuated person as such, but merely in assessing the traces of its mirrored ‘data double’. These phenomenologically void projections, they argue, are disconnected from the objects of their abstraction in troubling and debilitating ways. This mode of critique therefore seeks to safeguard the ‘person’ against the projections of the ‘profile’—to protect the agential space of the former from the incursions by the later. Displacing the ‘person’ by the ‘profile’ cuts short processes of self-recognition and self-affirmation. Rouvroy contends: ‘being profiled in this or that way affects the opportunities available to us and the space of possibilities that defines us: not only what we have done or are doing, but what we could have done or could do in the future’ (Rouvroy et al. 2022, p. 130). This is seen to threaten the prototypical legal or political subject: ‘algorithmic governmentality deflects concerns or attention away from … previously privileged perspectives of causality and intentional agency or individual and collective “authority” (that is … the capability to “author” one’s actions, to have the “authority” to give account of one’s actions meanings)’ (Rouvroy 2012, p. 7).Footnote 32 For Rouvroy (2012, p. 12), ‘governance by data’ thereby ‘bypasses individual consciousness and rationality (not only because operations of data-mining are invisible, but also because its results are unintelligible for the instruments of modern rationality)’. This, she claims, undermines the self-reflexivity, critique and deliberation essential for a subject to politically identify as such: it erodes the ‘inactual, potential dimensions of human existence’ where ‘processes of subjectivation and individuation’ unfold (Rouvroy 2012, pp. 12–13). Or, framed most directly: ‘[t]here is no longer any subject’—in the recursive (re)composition of ‘infra-individual data into supra-individual profiles’, the ‘notion of subject is itself being completely eliminated … [Y]ou no longer ever appear’ (Rouvroy and Stiegler 2016, p. 12). We observe a similar line of argumentation with Hildebrandt (2019, p. 105), who diagnoses a ‘shrinkage of the inner self’ as a result of our ‘overdependence on computational decision-systems’. In the example of the ‘virtual border’, we noted how decisions are based upon the production of data-based risk ‘profiles’, which neither those being ‘profiled’ nor those processing risk indicators are capable of representing, contesting or reflecting upon. As such, individuals being profiled cannot narrate, rationalize or justify their own actions, since these actions are not their own (yet). Critical self-reflexivity—‘the capability to develop a mind of one’s own’ (Hildebrandt 2019, p. 106)—is perceived in these perspectives as intimately intertwined with the modern rule of law and the democratic liberal order. In this light, the performance of pre-emptive processes of algorithmic governance ‘corrupts’ and ‘reduces’ our ‘autonomy’ and ‘human agency’, which ‘begins where we engage in critical reflection’ (Hildebrandt 2016a, b, c, p. 7; Hildebrandt 2020, p. 254). Hildebrandt (2019, p. 106), therefore, ‘warn[s] against the attempt to subvert our capability to reflect on how we navigate our world’, and calls upon lawyers to invent new ways to ‘accommodate human action’—to ‘safeguard the fundamental uncertainty and indeterminacy it assumes, and to protect the pinch of freedom and autonomy that defines us’ (Hildebrandt 2016c, p. 30). While the modern rule of law afforded such expressions of ‘free will and deliberation’, the challenge now, in tune with Arendt’s critique on behavioralist modes of governance, is ‘to protect the incomputable nature of the human self, its foundational indeterminacy and the natality it expresses’ (Hildebrandt 2016c, p. 29; Hildebrandt 2019, p. 105).
While these few paragraphs cannot do justice to the richness and complexity of the claims developed by Rouvroy and Hildebrandt,Footnote 33 we can recognize a shared mode of critique that aims to protect forms of human subjectivity and autonomy against the dissolution of the self, inflicted by nascent practices of algorithmic governance. While this mode of critique convincingly canvasses and counters many of the problems sketched in the previous section, we are wary of how it continues to figure around ideals of the human as a rational, free and sovereign subject—around notions of the autonomous and reflective ‘inner self’. In doing so, this mode of critique reaffirms political and ontological boundaries between the human and nonhuman, between mind and matter.Footnote 34 This, we argue, presents a number of risks and limitations. It is, first of all, crucial to account for how abstractions of autonomy are fraught with histories of assimilation and annihilation where boundaries of human subjectivity are drawn.Footnote 35 For Wynter (2003, pp. 260 and 264), in this sense, the emergence of the sovereign and autonomous human subject was possible ‘only on the basis of the dynamics of a colonizer/colonized relation that the West was to discursively constitute and empirically institutionalize’. Who is the ‘us’ which Hildebrandt sees as ‘defined’ by ‘freedom and autonomy’? How are such affordances of autonomy attributed? Which lineages and legacies do they import, and which forms of political action do they enable and sustain in the encounter with algorithmic governance? What becomes visible and possible—as we explore in the next section—if we desediment this rational sovereign subject as site of our normative aspirations?
Secondly, we question to what extent attempts to protect and restore the primacy of human agency—in the encounter with what Hildebrandt describes as ‘mindless’ ‘data-driven agency’—is materially tenable and normatively desirable.Footnote 36 Which political possibilities do we invite and foreclose by trying to reinstate what Rouvroy (2012, p. 7) cherishes as ‘privileged perspectives of causality and intentional agency’? Which registers of critique or contestation open up if, instead, we consider the human and nonhuman, mind and matter, to be inherently entangled and agential forms not to pre-exist but only to result from such relational compositions? (Barad 2007, p. 141).Footnote 37 Perhaps, this might allow us to move beyond sterile normative tropes—such as the ‘human in the loop’, which, as Amoore (2020, pp 58–66)Footnote 38 notes, is an ‘impossible figure’—and allow us to craft more powerful and imaginative responses to the dissolution and displacement of the subject in practices of algorithmic clustering described above. While, as Hohmann (2021, p. 595, 598) observes, many debates on ‘responsibility, accountability, and legality’ in the context of algorithmic governance, ‘center back in on the human as a stable and accepted point of reference’, the subversive forms of subjectivation that we seek to foreground require us precisely to ‘unseat the sovereign, rational human being as the location of agency and subjectivity’.Footnote 39 This is not a renunciation of agential possibilities but a move beyond the rationalist, representationalist assumptions underlying invocations of autonomous human agency set out above—an opening to more-than-human and distributed forms of political agency (Petersmann 2021).
In addition to these concerns about human agency and autonomy, (international) legal scholars have also been preoccupied with problems posed by algorithmic governance for prospects of publicness and democratic inclusion. Kingsbury and Maisley (2021, pp 354, 365), in this vein, convincingly point to the possibility of a disjunction between ‘infrastructural publics’ and ‘legal publics’—between those collectively affected and mediated by a particular material environment and those tied together as subjects under specific (and potentially overlapping) legal regimes.Footnote 40 Interestingly, they observe how infrastructures—such as the emerging ‘virtual borders’ briefly described above—produce heterogenous, unstable and sometimes subliminal publics (Kingsbury and Maisley 2021, pp. 359–360)Footnote 41 which have varying degrees of affinity, cohesion and mobilization.Footnote 42 ‘Infrastructures’, they observe, may also ‘work to prevent the emergence of certain publics’, which they exemplify with the ‘politics of nonpublics’ enacted by the material making of South Africa under apartheid (Kingsbury and Maisley 2021, p. 361).Footnote 43 The analysis by Kingsbury and Maisley provides a powerful framework to problematize the composition of the cluster as a subject of algorithmic governance. We noted above how the digital infrastructures of bordering only produce temporary and fleeting bonds of association which preclude mutual recognition and collective action—in line with the ‘nonpublics’ that Kingsbury and Maisley point to and problematize.Footnote 44 ‘Non publics’ can meaningfully be defined (and distinguished from the forms of relationality that shape ‘publics’), in this light, as forms of association without the affordance for mutual recognition or collective action. The detachment of these ‘nonpublics’ from categories of legal protection, recognition, and representation displays the urgency and importance of understanding the frictions and disconnections between infrastructural and legal forms of sociality (Kingsbury and Maisley 2021, pp. 365–366).Footnote 45 In their attempt to realign infrastructural with legal ‘publics’ (under normative conditions of ‘publicness’), Kingsbury and Maisley (2021, p. 367) assert that the contextual, contingent bonds of belonging together infrastructurally have to be translated in legal inscriptions of subjectivity and collectivity. Law is a ‘vessel for normativity’ that leads to the ‘emergence of a new encompassing … public, a “public of publics”’—a mode of inclusion that gives a coherent constitutional form to scattered subjects only tied together by fluid, ephemeral and disempowering material conditions (Kingsbury and Maisley 2021; Kingsbury and Donaldson 2011). In reaction to the degradation of sociality that results from data-driven infrastructures of rule, this is a plea for a re-alignment and re-attachment of the fragmented subject into a ‘representative system’ that safeguards inclusion, voice and visibility (Cf. Habermas 1998). The ‘phantom public’ that dwells within the algorithmic calculus has to be given legal form (Lippmann 1993, referred to by Kingsbury and Maisley).
In line with Kingsbury and Maisley, Bechmann (2019) (studying social media platforms) is also concerned with how the processing of ‘data as humans’ threatens ‘democratic values of representation (including participation), accountability, and equality’. As with Kingsbury and Maisley, the noted problem is one of ‘underrepresentation’—the absence of (individual and collective) recognition with(in) the data and lack of democratic agency of those sensed and sorted algorithmically (Bechmann 2019, pp. 75, 87).Footnote 46 In this vein, Bechmann (2019, p. 75) demands ‘all humans’ to be ‘properly and equally represented’. The key concern is the inclusion in the ‘public’ sphere—the democratic domain—of those who are digitally divided or disregarded (Bechmann 2019, p. 78). In the light of these accounts, the cluster—this ephemeral bond of association without representational equivalent or pretence to publicness—poses a particular and urgent democratic problem.
While we consider these accounts to be promising and powerful problematizations of the algorithmic practices described above, we see value in the warning by Mezzadra and Neilson (2013, p. 274) against ideals that ‘search for legitimacy in the languages of inclusion and exclusion, the jargon of part and whole, or the horizon of a “pure politics” that plays itself out in the demos or the state’. The ‘becoming public’ that is called for by Kingsbury and Maisley—with reference to the liberal political theory of Arendt, Habermas and Fraser—also entails an enrollment in the systems of subjectivation that the language of inclusion both expresses and demands. Yet, as Tendayi Achiume (2022a) has powerfully argued, our critical focus should be on ‘marginalization upon inclusion’—on how inclusiveness can operate as yet another technology of domination. As elaborated in the next section, we are inspired here by Moten (Moten and Harney 2021) who observes that this ‘public sphere’—the domain where ‘putatively individual subjects act and speak in public in so-called collectivities or coalitions’—not only ‘exclude[s] [black folks] from modalities of citizenship, personhood, subjecthood’, but is ‘predicated on that exclusion, which is to say: predicated on the regulation and exclusion of that insurgency, which “blackness” instantiates’—an insurgency that ‘manifests itself as the refusal of the regulative force that has to be exerted in order for subjects to come into their own as subjects’.Footnote 47 In the following section, we therefore reflect on modes of sociality that work against this ‘regulative force’—a defiant fugitivity that operates against the (en)closure of inclusion and dwells within the indeterminacy of the ‘algorithmic aperture’.Footnote 48
Finally, a recurrent preoccupation among critical scholars concerns the biases that plague algorithmic governance. Conceptually close to aspirations for autonomy and inclusion, the focus here is on the values of equality. While gender, class, and multiple other biases are algorithmically activated and amplified, our brief analysis here focuses on racial biases as only one instantiation. Machine biases and systemic racism are deeply entangled. As Amaro (2022) argues, the ‘black technical object’—used to refer to the psychic fragmentation of non-White subjectivities who are the victims of intersectional digital biases—is ‘always-already pre-conditioned by an affective prelogic of race’ (Amaro 2020b, a, p. 304). This issue is often framed (and, we argue, reduced) by reference to the bias of training data or individual coders. To counter racial biases, computer scientists such as Buolamwini work to widen the scope of machine perception by gathering, including and coding more data of non-White individuals, for machines to learn to detect behavioral patterns beyond the ideal type of the liberal (White) human subject (Buolamwini 2020).Footnote 49 This attempt speaks directly to the erasure of certain subjects from the ‘public sphere’ hinted at above, and the resulting calls for a ‘greater’ inclusion and ‘better’ representation of racialized and marginalized groups. For Buolamwini, indeed, the issue is a representational one: ‘a lack of diversity in the [data] training set [that] leads to an inability to … characterize faces that do not fit the normal face derived from the training set’ (Buolamwini 2020). The situated gaze of the (potentially) biased data or coder needs to be traded for a ‘universal gaze’. This is a familiar concern for (international) legal scholars, who point to the ‘risk that biased, discriminatory, or otherwise unjustified outputs may result if data sets lack integrity, accuracy, and reliability’ (Endicott and Yeung 2021) and call for algorithms to be made ‘transparent’ and ‘inclusive’ (Benvenisti 2018). The problem of inequality, in this framing, is one of statistical error and polluted data—a deviation from the prevailing standards of equality and clean computational decision-making to be found and fixed. The normative objective of such legal interventions—equality through objective digital representation and neutral calculation—is intimately intertwined with liberal values of autonomy and inclusion.
There is no doubt that, faced with decision-making systems such as those used at the ‘virtual border’, these interventions are crucial in highlighting the ‘immediacy of racism and racialization’ (Amaro 2020b, a, p. 312). Yet, the ‘universal gaze’ proposed by Buolamwini (and implicit in the legal demands described above) can be seen as its own regime of perception and subjection. This is registered, in one sense, in the need for increased data extraction to ‘correct’ the skewed algorithm. Yet, as Amaro (2021, p. 155–156) argues, what it ‘takes for granted is the simultaneous effect of making visible those whose aspirations are to remain unseen by power’. As we elaborate in the next section, those whose lives are threatened by inclusion within the system that violently excluded them in the first place, do not necessarily demand to be recognized as equal ‘subjects’. By ‘reversing the coded gaze’—as Buolamwini does—and including ‘diverse’ faces into facial recognition training sets, the inclusivity seeks to make transparent an opacity that may serve as ‘the only container for safety’ for those populations who are now again rendered visible (in line with Glissant’s clamor for a right to opacity and Moten’s articulation of the ‘excess of living that itself becomes the production of life’) (Amaro 2021). The ‘universal computational gaze’, in this sense, limits the ‘self-determination of those that have little or no desire for inclusion in machine perception’ (Amaro 2020b, p. 307).
Such attempts to render visible or transparent other modes of being through a ‘universal gaze’ capture the ‘excess’ of Black life from within a confined computational milieu, which already ‘positioned the white object as the prototypical characteristic’ (Amaro 2020b, p. 307). ‘Whiteness’, Phan and Wark (2023) argue, ‘persists as a grounding norm’. Attempts to clean or de-bias this computational milieu—by safeguarding against the racialized development of risk profiles at the ‘virtual border’, for example—‘catalyzes disruption only on the level of superficiality’: ‘the white object remains whole, while the object of difference is seen as alienated, fragmented, and lacking in comparison’ (Amaro 2020b, p. 307). In this sense, Amaro (2021, p. 155) observed, calling out algorithmic biases is ‘already predicated on the reduction of chance and contingency: namely, to regress, to clean, to normalize the unexpected, and therefore normalize and exclude the opportunities that live outside of the lines of political engagement’. To fix algorithmic ‘biases’, in other words, erases the anomalies that exceed ‘normal’ behaviors moulded against the backdrop of an idealized (White) human figure (Phan and Wark 2021). The insistence on de-biasing, then, might serve, as Amaro (2020b, p. 302) argues, to further ‘reduc[e] the operation of individuation, and primarily the differences amongst the living, to no more than an assemblage of contradictions that are negated and subsumed into a higher, more homogenous, unity of existence’. The classificatory operation of risk analysis at the ‘virtual border’, in this sense, enacts its racializing logic through the construction of a coherent, homogenous computational milieu where, as Amaro (2022, p. 56) notes, the ‘operation of individuation is … relegated to a series of representations among a falsely unified species’. The racializing working of machine learning can thereby be seen as ‘the power that responds’, in Moten’s terms, to the ‘very emanation and the condition of possibility of becoming-common … of living in the world of fugitive and common things’ (Moten 2018, p. 24).Footnote 50 The normativity of Whiteness, in short, is reflected not in statistical bias or error but in the representational assertion of a ‘universal computational gaze’ which can be perceived as a ‘state of homogeneity’ that calls for ‘accurate’ and ‘inclusive’ data sets precisely tend to reinforce (Amaro 2022, p. 57). Whiteness, in this vein, is not (only) an ontological, biological, or normative attribute but a process of homogenization. Calls from legal scholars and critical AI experts urging to render algorithms less biased, more inclusive and equal, risks disavowing this important point, by reinscribing or recommitting themselves to a pre-existing substance of racial difference that underlies both the ‘overriding logic of correlation and hierarchy’ in machine learning processes and the attempts to retrace the inequalities enacted through algorithmic (risk) assessments to pre-existing racial categories (Amaro 2022, p. 61).Footnote 51 The problematic of bias, seen this way, is a trap: an erasure rather than an affirmation of difference, an enclosure rather than an opening.
Rather than folding ‘seamlessly into the desire for representation’ (Amaro 2020b, p. 304), it is perhaps on the terrain of error, dissonance, illegibility, and invisibility that new openings can be found: opportunities to ‘shift the pathological perspective from one of entropy and lack [or the loss of computational coherence as such] to a more affirmative process of psychic generation’ (Amaro 2020b, p. 311). In this sense, the misrecognition of machine perception might entail possibilities for Black life ‘outside of phenotypical calculation, prototypical correlation, and the generalization of category’ (Amaro 2020b, p. 307). What we need, Amaro and Khan (2020) note, is ‘a purposeful misrecognition of the dominant ontogenetic perspective of racial individuation’, to ‘bring forth new ideas of what it means to be Black in a world regulated by the substance of race’. This entails a ‘fundamental revaluation of the values that form individual and collective perception’, for otherwise to ‘make black technical objects [legible by and] compatible to computer vision algorithms’ risks the further reduction of the ‘lived potentiality of black individuals’ (Amaro 2020b, p. 304). How, then, can we contest algorithmic inequality without reifying the normalizing and reductive ideals of representational equivalence? Which spaces could be safeguarded for the ‘black technical object’ that ‘does not “want to be correct” or “corrected”’? (Amaro 2020b, p. 311). How can we interrupt or fracture the computational milieu rather than reaffirming presumptions of coherence and detectability?
Against this backdrop, and to make sense of modes of be(com)ing in algorithmic times, we draw on works from critical Black studies that engage with the possibility for social life beyond racial re-inscriptions that are enacted by and through the (return towards) representation of the subject and its collective. In contrast, we want to think of a mode of individual and collective be(com)ing beyond the liberal notion of human autonomy, beyond the logic of part and whole, beyond the procedural concern with bias. As Amaro and Khan claim, ‘[w]e must bring to light a notion of political subjectivity that does not organize at the threshold of existing perceptions of difference, but instead releases the energy from this interaction to form a potentially new individual and collective being’ (Amaro and Khan 2020). As one way of releasing this energy, we turn to notions of Black sociality.
4 Undoing the anti-blackness of be(com)ing in algorithmic times
In this section, we turn to works from critical Black studies to grapple with the questions set out above. We are aware that such works are not necessarily and definitely not primarily addressed to us, given our positionality and lived experience in this world.Footnote 52 There is a risk here of appropriating experiences of suffering that are not ours. It is important to stress, therefore, that we are neither speaking for or on behalf of ‘black folks’, nor trying to capitalize on or co-opt critiques from Black radical thinkers, which we see as also formulated against ‘us’.Footnote 53 Rather, we take on their call to undo the anti-Black premises of the world and its subjects of which we form part. It is, as such, the anti-Blackness of the modernist world in which we live that we are after, rather than the lived experience of Blackness that evades us. We see this as a particularly urgent project in light of how key tenets of the liberal lexicon—such as autonomy, inclusion and equality—are being recovered and reified in facing issues of algorithmic governance.
The reader might be questioning our turn to critical Black studies at this stage of the argument. In the preceding sections, we argued, first, that AI-based decisions—as exemplified through emerging digital infrastructures of the ‘virtual border’—are giving rise to new forms of subject-making through ‘clusters’ of data, which displace the liberal autonomous, rational and self-possessed individual subject and enact, instead, relational, emergent and inferential algorithmic ‘subjects’. Second, we argued that in response to the pressure put on the autonomous, rational and self-possessed individual subject, legal scholars are advocating for distinct ways of addressing these points of pressure. This is evidenced by calls for a (re)turn or (re)attachment to values of autonomy, inclusion and equality in relation to algorithmic infrastructures. Yet, we also argued that the liberal perspectives and anxieties in these interventions risk reinscribing the modernist world and its subject. It is in the expressed need for undoing, for refusing, for de-worlding the anti-Black foundations that underpin this world and its subject, that works from critical Black studies appear as uniquely insightful to us (despite us not being the primary audience of such works).
This is so for at least two reasons. First, critical Black studies thematized a (Black) sociality that operates beyond or outside of the liberal category of the subject. This matters for our claim, since it is precisely the individual subject as such that ‘algorithmic governmentality’ is displacing by clusters of dispersed data points. Works from critical Black studies can, therefore, help thinking possibilities of being and becoming-in-common in algorithmic times that counter the anti-Black foundations of liberal notions of inclusion and the public sphere, and refuse ideals of the sovereign subject. This opens up to potentialities of collective living—of emergent forms of sociality—that we see threatened by both the practices of algorithmic governmentality and the reification of the sovereign subject in current regulatory responses.Footnote 54 Second, we retrieve the power of refusal and resistance in Black sociality to counter and escape the ‘universal gaze’ of computational perception and subjectivation.Footnote 55
As alluded to in the previous section, critical AI scholars and activists working to counter the racial biases of algorithmic systems have called for more inclusive data to better reflect social diversity. What is at stake in those critiques are the biases of coders and computer scientists who, in attempting to ‘objectively’ program the ‘reality’ of the ‘world’, reproduce the sedimented racial, gender and class-based inequalities and discriminations that constitute it. The ‘world’, here, is composed by the ideal-type figure of the liberal human subject—as a self-possessed, self-reflexive and autonomous individual—that acts as the ‘norm’. A normative subjectivity, in other words, is reserved to and moulded by the lived experience of liberal White human subjects. Countering racial biases in algorithms would then expand the category of the liberal human subject by recognizing and including more diverse modes of being into it, thereby correcting this category by ‘cleaning’ it from its biases. But what if, instead of perfecting the category of the liberal human subject, we would let go of it in order not to reproduce the violence it enacts by transcendentalizing the figure of the self-possessed, self-reflexive and autonomous subject?
Indeed, this figure of the subject originated by being reserved to White normativities, against nonhuman animals and chattel slaves that constituted objects of property and labor.Footnote 56 As such, the White subject cannot be disentangled from the Black object. The hegemonic figure of the White human being, in other words, overdetermines the sense of be(com)ing a subject. As Hartman (1997, p. 6) puts it, the experience of historical violence against (non)human objects is upheld each time the category of the liberal human subject is re-enacted, since this category was originally constituted by and through the exclusion of any-thing nonhuman, especially the enslaved as objectified and dispossessed property.Footnote 57 Against this backdrop, the ‘vision of equality forged in the law’—and more precisely in the White normativities forming the law’s subject— ‘naturalized racial subordination while attempting to prevent discrimination based on race or former condition of servitude’ (Hartman 1997, p. 9). Calls to counter racial biases by expanding or excluding specific data inputs of machine learning risks therefore to naturalize the racial subordination that informs the category of the human as White or rather as an anti-Black subject. As Hartman notes, invocations of non-discrimination of subjects risk disavowing the ‘racial domination and liberal narratives of individuality [that are] utterly enmeshed in … emancipatory discourses of rights, liberty, and equality’ (Hartman 1997, p. 116)Footnote 58 —discourses that are invoked against (racial) biases in AI today. Against this reproduction of violence, critical Black scholars attend to the generative desedimentation of the liberal human subject, instead of working at expanding or reworking it. Concretely, this implies suspending invocations of liberal values such as autonomy, freedom and equality to attend to and think creatively about what could happen if these values were refused and rejected.
This matters when considering data-based subject-formations. The critiques of governance by data traced in the previous section—on the observed assault on the autonomous human subject, the problem of publicness and inclusion, and the prevalence of algorithmic biases and associated inequalities—call for a re-inscription of modernist ideals of the liberal subject and its public. A return to the subject as such, however, perpetuates and prolongs the anti-Black foundations that underpin this category. The cluster—as a correlational, emerging, and composite mode of be(com)ing subject—attempts to govern in the present the excess, potentialities and speculative possibilities that are located in uncertain futures. By taking our cue from Du Bois—who accomplished to produce and dissolve, at once, the notion of ‘Whiteness’ by way of an account of the position of ‘Blackness’ in America, and insisted on an excessiveness of being that refuses to abide by simplistic logics and oppositional categories of racial divisions (Chandler 2013, p. 127)Footnote 59—we want to ask what forms of be(com)ing in algorithmic times could be envisaged if, instead of retrieving the liberal subject and its collective, we linger with the excess, the opaque and the de-individuation performed by algorithms?
First, we note how the excess of being that algorithmic logics attempt to tame and to govern—or what happens in excess of, in difference to and beyond the radar of AI-based analytics—seemingly echoes the lived experience of Black folks in anti-Black worlds. As Amaro (2019) argues, ‘black being as such actualizes as an experience that is lived from both within and in excess of artificial modes of perception and the fictive imaginary of race’. Blackness exceeds the possibility of being human qua (White) subjects—an impossibility also performed by way of data profiling and the displacement of the autonomous and self-possessed subject in favor of a cluster of data points. A political problematization of algorithmic governance through the notion of racial biases remains stuck with binary racial stereotypes—or a fixed ontology of race—that misdiagnoses the problem at hand by offering recognition and inclusion as a ‘solution’, instead of attending to and leaving space for the excess of possibilities of actions that can materialize in improvization. These possibilities evade the relational networks that are registered by algorithms to profile and model the behaviors they track.
Second, we note something emancipatory in this excess that evades algorithmic individuation. For Du Bois, the simultaneous production and dissolution of the subject ‘marked out the very space and possibility of desire and that which is yet to come’ (Chandler 2013, p. 120). The question of desire is important, here, in relation to algorithmic governmentality. As Rouvroy contends in this context, the automatic detection of intention taps into pre-conscious impulses and undermines the very possibility of desire as the capacity not to do everything one is capable of.Footnote 60 By foretelling desire, algorithms produce ‘acting outs’ that short-circuit one’s capacities of desiring the not yet—or with Du Bois the ‘yet to come’. For Du Bois, the ‘double consciousness’ of African-Americans opened up possibilities that were not given in advance, to let transpire a sense of ‘self-as-becoming-other than the given’ (Chandler 2013, p. 153)—a given premised on pre-conceived categorical distinctions between (white) subjects and (black) objects. Could algorithms be configured without enclosing forms of pre-figured subjectivation? Can modes of being and becoming in algorithmic times remain open to the not yet given? How could such modes of be(com)ing refuse and escape the pre-emptive logics of algorithmic anticipation? It is in this evasion, this escape, this fugitivity from the given that insights of Black sociality appear key.
These interrogations find inspiration in Glissant’s (1997) articulation of a ‘right to opacity’—a right not be rendered transparent, not to be measured against, and not to be related to already given political and legal norms. A ‘right to opacity’ demands instead to remain open towards ‘what cannot be reduced, which is the most perennial guarantee of participation and confluence’ (Glissant 1997, pp. 189-190). In tune with Glissant’s Poetics of Relation, Mezzadra and Neilson (2013, pp. 274–275) perceive the political as a ‘social practice of translation [that] creates a collective subject that must continually keep open, open in translation, and reopen the processes of its own constitution’.Footnote 61 The opacity of open-ended potentialities collides with the total transparency and enclosure of algorithmic pre-emption of possible futures brought into the present. How, then, can one escape the logics of algorithms, which fix, foretell, and reduce the possibility of be(com)ing in algorithmic times?
As Harney and Moten also ask, could a sociality be envisaged, then, that enables a ‘fugitivity from the universal history that would fix the fate of the self’—a fate of the self-fixed by algorithmic predictions?Footnote 62 In their latest book, All Incomplete, Harney and Moten (2013, p. 57) took specific issue with the rise of algorithms. The ‘logistics of algorithmic composition, and the rhythm of logical capitalism’ pulse into everything, becoming the beat that rhythms life everywhere, all the time. ‘Algo-rhythmic’ logistical capitalism operates through a totalizing movement—a movement of total access to and putative transparency of the subject (its past, present, and future actions).Footnote 63 As exemplified by the ‘virtual border’, the algorithmic assessment and projection of ‘risk’ is aimed at rendering future possibilities actionable in the present. With Harney and Moten, we then wonder: ‘[c]an we dodge and blur an algorithmic syntax that straightjackets and atomizes us into total access, so we can get back to rebuilding our atrophied habits of assembly’? (Harney and Moten 2021, p. 171). What habits of assembly, of collective thinking, of being and acting can exist, without falling back on liberal invocations of the subject and its public?
What we learn from critical Black studies is that be(com)ing in algorithmic times without the subject bears potential to live in the break of modernist, logistical capitalist onto-epistemologies, where life is not reductively foretold in advance. Taking our cue from critical Black studies that problematize liberal reactions against racial biases and suggest distinct modes of be(com)ing, what transpires is the need to thrive for a living that attends to the potentialities, excess and possibilities that a computational life rhythmed by the pulses of algorithms is trying to control and govern—to dwell in the algorithmic aperture. In Amaro’s words: ‘[w]hat emerges is an optimistic view that actually says there is a term of resistance, there is a term of being that lives outside of governance’ (Hui and Amaro 2021, p. 60).
Why, then, would one call for an inclusion of Black beings into the category of the subject, from which they have always been excluded in the first place? As Harney and Moten (2013, p. 47) put it: ‘[W]e have to love our refusal of what has been refused (to us)’. This is a love for ‘subjectlessness’,Footnote 64 the purpose of which is not to correct the ‘biased’ constitution of the liberal subject but to dismantle the figure of the given subject as such. Against the logic of predictive analytics that anticipates and forecloses the be(com)ing of algorithmic subjects and their admission into existence, what if the ultimate aim, as Amaro (2020a) has argued, would be ‘to shift the racialized individual away from the stereoscopic image of categorical difference and move towards a non-representational mode of self-determinism from within or outside of the institutional structure’? It is in this possible mode of self-determinism that we see hope for the refusal of the perennial repetition of (algorithmic) governmentality and its onto-epistemology of race. Perhaps be(com)ing in algorithmic times by refusing, resisting and overcoming the ‘dominant ontogenetic perspective of racial individuation’ (Amaro and Khan 2020) bares the potential to reckon with the suffering of ‘those whose presence never counted, whose absence was never accounted for, their existence unable to be recounted—that which does not even leave a trace’ (Culp 2021, p. 113).
Yet how, concretely, could the opacity of be(com)ing be practiced and kept open against algorithmic foreclosures? Our intention here is not to offer an answer to this question, but to introduce a distinct way of making sense of the displacement of the liberal subject by the cluster. While most of critical scholarship that took issue with this matter did so by recovering liberal values of autonomy, inclusion and equality, our objective was to open up a different way of thinking the problem at stake—a different way to problematize the anti-Blackness of be(com)ing subjects in algorithmic times.
5 Conclusion
The article mapped out and explored how algorithmic technologies are reconfiguring modes of subjectivation today. By taking the example of the ‘virtual border’, we started—in the first part of the article—by demonstrating how once stable categories of both the individual and the collective(s) to which it belongs, are displaced by correlational and volatile forms of ‘subject’-making through the productions of fluid, emergent, inferential clusters—pulsing patterns and propensities distilled from distributed data points. This algorithmic mode of subjectivation, we argued, inhibits forms of mutual recognition, or being-in-common, and thereby erodes conditions of collective action. In response to this dissolution of the subject and the degradation of sociality it entails, as we analyzed in the second part of the article, legal scholars have challenged the threats posed by practices of ‘governance by data’ to liberal values of human autonomy, democratic inclusion and equality. While we see value in these approaches in countering the risk of an ‘algorithmic governmentality’ based on forms of projection and pre-emption, and in highlighting the ‘immediacy of racism and racialization’ in its operations (Amaro 2020a, p. 307) we pointed to some limits of this ‘legal imagination’ in counteracting specific forms of algorithmic violence, and argued that its projections of liberal subjectivity and inclusion are predicated on the erasure of difference and the disciplining of Black social life and its disruptive modes of being-in-common. In the third part of the article, we drew on contemporary strands of critical Black studies in speculating about what a ‘legal imagination’ could look like—in line with the theme of this special issue—that seeks to undo, refuse, and de-world the anti-Blackness of the computational milieu and lingers with the possibility to ‘catalyze future affirmative iterations of the self’ in algorithmic times (Amaro 2022, p. 62).
Data availability
Not applicable.
Notes
‘Cyborg imagery … means both building and destroying machines, identities, categories, relationships, space stories. Though both are bound in the spiral dance, I would rather be a cyborg than a goddess’.
As many have observed, the (international) legal form serves as a privileged technology of subject-making.
This is inspired by Foucault’s (1991, p. 194) famous quote that ‘power produces: it produces reality; it produces domains of objects and rituals of truth’.
We can think, for example, of Cassese, Gaeta, Zappalà 2008.
In the sense that the indicators thereby ‘promote self-management’, Merry perceives them as instantiating what Rose (1998 pp. 226ff) has qualified as ‘governance at a distance’. This links directly to the Foucauldian diagnosis of power as the internalization of discipline.
Reference is made to Samaddar 2010, p. 28.
In the sense that this exercise and emergence of collective agency disrupts the established legal and political order, Rancière’s argument aligns with what Lindahl (2013) has described as the constitutive moment of ‘a-legality’.
As Mann elaborates: ‘They have understood that the human subject of mediatized catastrophe does not fit as comfortably into bureaucracy as the citizen’ (Mann 2021, p. 147). Simone Browne shows how such ‘masquerades’ were already at use to counter the earliest forms of surveillance in slave plantations: the ‘ability to evade surveillance through makeup, wicked tricks, and hiding in plain sight exposes the one-drop rule as a social construction that, for some, could be subverted by performing whiteness … [and] hiding in plain sights—by identifying as white and using an alias—was a freedom practice to evade surveillance, and in so being a form of dark sousveillance’ (Browne 2015, p. 54). Browne also refers to contemporary artist Jillian Mayer, who ‘demonstrates how to use black lipstick, clear tape, scissors, white cream, some glitter, and black eyeliner to distort one’s face in order to make it indiscernible to [CCTV] cameras’, as well as artist Adam Harvey’s Computer Vision (CV) Dazzle project, which ‘explores the role of camouflage in subverting face-recognition technology’, at 163. Browne sees in these performances ‘productive possibilities that come from being unseen, where blackness [can] be subversive in its capacity to distort and interfere when it comes to machine readability and standard algorithms’, at 163.
Mann speaks of the ‘imagination of humanity as a stateless, peripatetic and globally homeless community’ (Mann 2021, p.154). The image of the mask resonates in Marx’s critique on the detachment of legal personhood from material relations of production. It also returns in Fanon’s engagement with the lived experience of blackness in Black Skin, White Masks (1952).
We do not invoke a ‘politics of identity’ here but, drawing on Mezzadra and Neilson, a ‘power of a common that is not given by nature, history, or culture but must be … invented and reinvented’ (Mezzadra and Neilson 2013, p. 274). We find inspiration in Moten’s notion of ‘black sociality’—an ‘undefinable, innumerable kind of gathering’ where ‘differences [are] cared for under … the duress of trying to reduce those differences into one thing’ (Abdurraqib and Moten 2021). On this process of reduction as a striving for transparency in Western thought see also Glissant 1997 and Sakai 1997.
See also Aradau and Blanke, (2018 p 21) and Johns 2021 p. 59 (‘[f]aced with the outputs of a statistical model (either directly or by their expression in law and policy), social and legal subjects may plot themselves or be plotted against available categories and values and understand more or less how those were generated’). The critique on and politicization of indicators is a case in point here.
We use Rouvroy’s concept of ‘algorithmic governmentality’ beyond the specific meaning that she grants it, merely to refer to modes of governance based on the combined use of big data and algorithms. In this basic sense, we think that the concept is to some extent interchangeable, for instance, with what Johns has described as ‘governance by data’(Johns 2021).
Cf. Johns 2021 at 66 (on how ‘data’s agglomeration troubles the fundamentals of legal and social representation and relation’).
A vector space model is an algebraic model that represents objects (such as a text, a word, a sentence, a document) as vectors of identifiers (such as index terms).
Claiming that AI models could be employed to ‘capture intuitive similarity’.
As Mackenzie elaborates: ‘This expansive inclusion of features vectorizes many data sources into one high-dimensional space that spans and subsumes all contextual, indexical, symbolic or lived differences in data’ (Mackenzie 2015, p. 434)..
This is indeed the essence of deep learning models, which, as Amoore notes, ‘seek[] to generate the rules from features that are not pre-programmed in advance’ (Amoore 2020 p. 6). This orientation towards the detection of ‘patterns’ is central to the collection of PNR (Passenger Name Records) data by the EU. As the Commission stated: ‘analysis of PNR data can provide the authorities with important elements from a criminal intelligence point of view, allowing them to detect suspicious travel patterns and identify associates of criminals and terrorists, in particular those previously unknown to law enforcement’. See European Commission (DG for Migration and Home Affairs), Passenger Name Record (PNR) (available online).
The Delegated Act clarifies that enhanced data collection and analysis would make it possible to ‘continuously identify sets of characteristics of specific groups of travellers’.
In a similar vein, the 2025 UK Border Strategy aims to ‘develop advanced new risking systems to target interventions more effectively, using … AI-driven decision making’ (HM Government 2020 p. 41). This ambition to ‘maximis[e] data driven, automated decision making’ resulted in a cooperation with BAE (British Aerospace Engineering) Systems: a multinational arms, security, and aerospace company that is in charge of creating and maintaining ‘Cerberus’, which is designed to provide ‘real time risk assessment’ ingesting a variety of data sources. BAE, ‘Using Data to Secure the UK Border’, https://www.baesystems.com/en/cybersecurity/feature/using-data-to-secure-the-uk-border.
Along similar lines, Amoore (2013) has written about the ‘mobile border’.
This inclusion of ever-more heterogenous data (‘more variety in input’, as the EU strategy states) is tied to the ‘process of experimentation’ that defines deep learning. A process in which ‘all space is feature space’ (Amoore 2020, p. 4).
Cf. Phan and Wark (2021, p. 2) (signaling an emergence of ‘modes of classification [that] figure racialized bodies neither as single, coherent subjects, nor as populations stratified according to characteristics that find their genesis in bodily difference, but instead as shifting clusters of data’).
As they hold: ‘une actualité dotée d’une “mémoire du futur” [qui] donne aussi consistance au rêve d’une sérendipité systématisée’. Cf. Rouvroy (2012) p. 157 on how algorithmic governmentality works with a ‘unique, supra-individual, constantly reconfigured “statistical body” made of the infra-individual digital traces of impersonal, disparate, heterogeneous, dividualized facets of daily life and interactions’.
For a splendid ethnographic approach to the study of such changes in the administrative process of border control in Canada, see Raso (forthcoming 2023).
For an interesting typology of different modes of subject-making, see Isin and Ruppert (2020).
On ‘interpellation’ as a mode of subjectification, see Althusser (1971). A telling example of subjectification through interpellation is Fanon’s realization of his ‘over-determination from the outside’ when upon his arrival from Martinique to mainland France, a white child points to Fanon on the streets and declares ‘Look, a Negro!’ (Fanon 1952).
Similar anxieties are voiced by Hildebrandt, for whom ‘what makes us subjects or agents is our capacity to will not to will something [thereby] enabling us to address and redress our self as the author of our actions’ (Hildebrandt 2016a, b, c p. 12). In these accounts, the expression of human autonomy under conditions of uncertainty is a key affordance of the rule of law.
Both bodies of scholarship move beyond romanticized ideal of liberal subjectivity, linking the notion of autonomy and subjectivity—in sophisticated and convincing ways—to the ‘virtuality’ or ‘double contingency’ that marks the human condition and precedes possibilities of politics or ‘commoning’. (Rouvroy 2012, p 10ff; Hildebrandt 2016c, p. 12).
The latter dichotomy is at the core of Hildebrandt (2016c). Rouvroy anticipates and pre-empts this critique to some extent, arguing that ‘[t]he rational, liberal, individual subject, or the autonomous legal subject have never been anything else than useful or even necessary functional fictions’ (Rouvroy 2012, p. 12).
There is, of course, a rich tradition of TWAIL, feminist and Marxist literature on the violent and exclusionary processes by which the boundaries of the sovereign human subject are drawn, and how particular modes of being are thereby both universalized and erased. See, for example, Baxi (2002); Anghie (2004); Parfitt (2019); Tzouvala (2020).
We see value in new materialist projects which ‘insists that agency is a matter of relation, and a process that entails a potentially infinite spectrum of matter around and within us’. This entails a refusal of modernist perspectives that ‘relocate agency firmly in the human’ (Hohmann 2021, p. 595; Barad 2007, Bennett 2009). The artificial and unproductive nature of the mind/matter distinction is at the heart of Latour (1993).
As Barad notes: ‘[a]gency is not an attribute but the ongoing reconfigurings of the world. The universe is agential intra-activity in its becoming’.
Pointing to the ‘multiple and distributed selves’ that ‘dwell within the calculus’.
This attempt at unseating the autonomous, rational, logocentric and phallocentric subject has, of course, a long legacy. For a particularly powerful expression, see Haraway (1991).
As Kingsbury and Maisley elaborate: ‘[b]ecause of their relational nature, infrastructures—like laws—always have publics … Sometimes infrastructural and legal publics overlap, but in extreme opposite cases laws entirely fail to give any regard to an infrastructural public’ Kingsbury and Maisley (2021, pp 365).
Kingsbury and Maisley (2021, p. 357) define infrastructures quite elaborately as ‘dynamic, systemic networks and assemblages, built at varying scales, with some spatial fixity and durability. They instantiate and precipitate changing social and technical relations, are embedded in and productive of specific legal-organizational forms and socio-technical practices, and are experienced differently by differently situated people’.
They refer to von Schnitzler ‘Infrastructure, Apartheid Technopolitics, and Temporalities of “Transition”’, in N. Anand, A. Gupta and H. Appel (eds.), The Promise of Infrastructure (2018), 133.
This erosion of collective agency and recognition was also signaled in Van Den Meerssche (2022). See also, Dijstelbloem and Broeders (2015) observing that the associations drawn in data-driven border surveillance ‘become so fragmented and shattered’ that those affected become ‘lost in categorization’. Dijstelbloem and Broeders’ concept of ‘non-publics’ aligns closely with the problematisation of Kingsbury and Maisley.
As Kingsbury and Maisley specify: ‘the gap between infrastructural publics and legal publics … has deleterious effects on substantive justice, economic efficiency, institutional legitimacy, and on publicness’ (Kingsbury and Maisley 2021, p. 366).
Interestingly, Bechmann (2019, pp 75–76)—in line with Kingsbury and Maisley—finds inspiration with Lippmann and Dewey. The philosophical tradition of liberal pragmatism seems to be a productive terrain for legal scholars engaged with the problems posed by new digital technologies. Hildebrandt (2020, p. 261) equally notes that her proposal for ‘agonistic machine learning is directly related to Dewey’s constitution of concerned publics’.
This refers to the notion of ‘black sociality’ invoked in Abdurraqib and Moten (2021) i.e., an ‘undefinable, innumerable kind of gathering’ where ‘differences [are] cared for under … the duress of trying to reduce those differences into one thing’). ‘Blackness’ here should not be reduced to a darker skin pigmentation but refers to modes of living that exceed the normative White subjectivity of the liberal subject. We take our cue from critical Black studies that disediment the category of the liberal ‘human subject’ to retrieve modes of being that exhaust, exceed and escape this category. In this sense, Blackness is an ongoing refusal of standards imposed from elsewhere – a constant escape from the fixity of racial ontology that structures White supremacy.
Cf. Amoore (2020) on the ‘aperture’ as ‘an opening or uncovering of attributes and relations’.
This call resonates in Benjamin, Race After Technology: Abolitionist Tools for the New Jim Code (2019).
Amaro (2022, p. 56) describes this as the ‘freezing of dynamic life into a homogeneous milieu’.
In a keynote Sharpe offered at the 2022 Venice Biennale of Art, she stated in reference to visual artist Leigh that ‘[f]or many years at the beginning of every talk, Simone Leigh would state two things: that her work centered black women, and that we—black women—are her work’s primary audience. It is important to hear that, and to know it. That attentive primacy is Leigh’s practice of care, even if, even as, the spaces of care that she holds open and ushers us into are fleeting’ (Biennale Arte 2022). If we are aware not to be the ‘primary’ audience of these practices of care, we engage with them with no intention to appropriate them.
As Sharpe poignantly notes, the suffering of Black Death, ‘to the degree that it is recognized, is pornotropically exploited in the service of rehabilitating anti-black and colonial institutions’, to then denounce: ‘Don’t use our death, our suffering, our lives and our work to re-generate your projects. Don’t turn our flesh into gold again. We do not consent to this’ (Harman and Sharpe 2021, p. 125–126).
As Rouvroy (2016) puts it: ‘Ce qui me semble donc surtout menacé, aujourd’hui, ce n’est pas la sphère privée (elle est, au contraire, hypertrophiée), mais l’espace public, l’“en commun”’.
The notion of ‘universal computational gaze’ comes from Amaro (2020b, p. 307).
As Chandler (2013, p. 65) insists: ‘[o]ne must strategically elaborate the way in which the figure of the subordinate, the figure of the other, gives rise in the movement of—instead of on the basis of—its production to the figure of the hegemon’. Chandler (2013, pp. 13, 24 and 32) retraced how, with Du Bois, the ‘ontological question concerning the ground of the being called the Negro’ presupposes ‘a Euro-American norm of what it means to be human’, which acts as a premise that is never interrogated in and of itself. As Chandler further notes: ‘the grounds of historical and social existence and identification were placed in question for “Africans”, or “Negroes”, or “Blacks”, configured in this vortex [of the Middle Passage], [but] what is not typically remarked is the way in which a fundamental questioning of the roots of identification and forms of historical existence for “Europeans” or “Whites” was also set loose at the core of this historical problematization’.
This exclusion was also endured after the abolition of slavery. As Hartman (1997, p. 10) puts it: ‘[w]hile the inferiority of blacks was no longer the legal standard, the various strategies of state racism produced a subjugated and subordinated class within the body politic, albeit in a neutral or egalitarian guise’.
As McKittrick (2013, p. 9) saliently emphasizes: ‘[t]his carries with it an expectation that the road to recovery is an evolution toward a mode of humanness that is produced through inequities’.
Du Bois saw in the doubling associated to the ‘double consciousness’ of African-Americans in the US a ‘[p]ossibility of a new way or new ways of social being, ways that would be excessive to the simple and oppositional divisions that would dominate the social field in which the Negro American arises’. Du Bois experienced this doubling as a ‘sense of always looking at one’s self through the eyes of others’ (Chandler 2013, pp 119–121).
As Rouvroy (2016) specifies: ‘Ce que cela met à mal – et on pourrait se rapporter pour cela à Deleuze et Spinoza – c’est la puissance des sujets, c’est-à-dire, leur capacité à ne pas faire tout ce dont ils sont capables. Du fait de la détection automatique de l’intention, le processus de formalization et d’expression du désir est court-circuité’.
As they further emphasize: ‘[i]n these practices of opening and translation, there emerges a figure of power that does not search for legitimacy in the language of inclusion and exclusion, the jargon of part and whole, or the horizon of a “pure politics” that plays itself out in the demos or the state’.
The quote comes from Judy’s engagement with Harney and Moten’s notion of the ‘undercommons’. R. A. Judy (2020, p. 423). See also Harney and Moten (2013).
On the play of words on ‘algo-rhythm’, see also the concept of ‘algo-ritmo’ that Dixon-Román (2016, p. 5) develops in relation to Derrida’s concept of ‘iterability’ as a ‘repetition of data assemblages with alterity’ or ‘repetition with differentiation and not uniform sameness’. Dixon-Román’s ‘algo-ritmo’ combines the ‘algo’ that translates in English to ‘something’ and ‘ritmo’ that translates in Spanish to ‘rhythm’, thus referring to ‘regular repeated pattern of something’.
As Moten (2017, p. vii) puts it: ‘our resistant, relentlessly impossible object is subjectless predication, subjectless escape, escape from subjection, in and through the paralegal flaw that animates and exhausts the language of ontology’.
References
Abdurraqib H, Moten F (2021) Building a stairway to get us closer to something beyond this place. Millennials are Killing Capitalism Podcast. https://soundcloud.com/anticapitalist-millennials/hanif-abdurraqib-fred-moten. Accessed 30 June 2022
Achiume T (2021) Digital racial borders. AJIL Unbound 115:333–338
Achiume T (2022a) Lecture at the Annual Conference of the European Society of International Law
Achiume T (2022b) Racial borders. Georgetown Law J 110(3):446–508
Althusser L (1971) Ideology and ideological state apparatuses (notes towards an investigation). In: Althusser L (ed) Lenin and philosophy and other essays. Monthly Review Press, New York
Amaro R (2019) AI as an act of thought. Sonic Acts Festival. www.youtube.com/watch?v=ys9gCR3PFF4 (at min 32:43). Accessed 30 June 2023
Amaro R (2020a) Threshold value. e-flux. www.e-flux.com/architecture/education/322664/threshold-value. Accessed 30 June 2023
Amaro R (2020b) As if. In: Jaque A, Otero Verzier M and Pietroiusti L (eds) More-than-Human. Het Nieuwe Instituut Rotterdam.
Amaro R (2021) Machine learning, surveillance and the politics of visibility. In: Vickers B, Allado-McDowell K (eds) Atlas of Anomalous AI. Lawson Publishing Limited Montgomery, Powys.
Amaro R (2022) The black technical object: on machine learning and the aspiration of black being. MIT Press, Massachusetts
Amaro R, Khan M (2020) Towards black individuation and a calculus of variations. e-flux. www.e-flux.com/journal/109/330246/towards-black-individuation-and-a-calculus-of-variations. Accessed 30 June 2023
Amoore L (2013) The politics of possibility: risk and security beyond probability. Duke university Press, Durham
Amoore L (2020) Cloud ethics: algorithms and the attributes of ourselves and others. Duke University Press, Durham
Anghie A (2004) Imperialism, sovereignty and the making of international law. Cambridge University Press, Cambridge
Aradau C (2022) Experimentality, surplus data and the politics of debilitation in borderzones. Geopolitics 27(1):26–46
Aradau C, Blanke T (2017) Politics of prediction: security and the time/space of governmentality in the age of big data. Eur J Soc Theory 20(3):373–391
Aradau C, Blanke T (2018) Governing others: anomaly and the algorithmic subject of security. Eur J Int Secur 3(1):1–21
Baxi U (2002) The future of human rights. Oxford University Press, Oxford
Barad K (2007) Meeting the universe halfway: quantum physics and the entanglement of matter and meaning. Duke University Press, Durham
Bennett J (2009) Vibrant matter: a political ecology of things. Duke University Press, Durham
Benjamin R (2019) Race after technology: abolitionist tools for the new Jim code. Polity, Cambridge
Bechmann A (2019) Data as humans: representation, accountability, and equality in big data. In: Jørgensen RF (ed) Human rights in the age of platforms. MIT Press, Massachusetts, pp 73–94
Benvenisti E (2018) Upholding democracy amid the challenges of new technology: what role for the law of global governance? Eur J Int Law 29(1):9–82
BAE, using data to secure the UK border. https://www.baesystems.com/en/cybersecurity/feature/using-data-to-secure-the-uk-border. Accessed 30 June 2023
Buolamwini J, Aspire mirror. Homepage. http://www.aspiremirror.com/ Accessed 30 June 2023
Buolamwini J (2020) Coded bias. www.codedbias.com. Accessed 30 June 2023
Brown W (2005) Edgework: critical essays on knowledge and politics. Princeton University Press, Princeton
Browne S (2010) Digital epidermalization: race, identity and biometrics. Crit Sociol 38(1):131–150
Browne S (2015) Dark matters: on the surveillance of blackness. Duke University Press, Durham
Butler J (2000) Antigone’s claim. Kinship between life and death. Columbia University Press, New York
Biennale Arte (2022) Meetings on art: what could a vessel be? www.youtube.com/watch?v=4BYPvgRJGQg&t=4108s. (33:17–33:42mins). Accessed 30 June 2023
Cassese PG, Zappalà S (eds) (2008) The human dimension of international law: selected papers of Antonio Cassese. Oxford University Press, Oxford
Chandler ND (2013) X—the problem of the Negro as a problem for thought. Fordham University Press, New York
Chandler D (2015) A world without causation: big data and the coming of age of posthumanism. Millennium J Int Stud 43(3):833–851
Cohen J (2017) Affording fundamental rights. Crit Anal Law 4(1):1–13
Cohen J (2020) The emergent limbic media system. In: Hildebrandt M, O’Hara K, Lie (eds) The law in the era of data driven agency. Edward Edgar, Cheltenham, pp 60–79
Culp A (2021) Afro-Pessimism and non-philosophy at the zero point of subjectivity, history, and aesthetics. In: Ferguson K (ed) The big no. Minnesota University Press, Minneapolis
Desroisières A (2002) The politics of large numbers: a history of statistical reasoning. Harvard University Press, Cambridge
Deleuze G (2001) Pure immanence: essays on a life. Princeton University Press, Princeton
de Goede M, Simon S, Hoijtink M (2014) Performing preemption. Secur Dialogue 45(5):411–422
Dijstelbloem H, Broeders D (2015) Border surveillance, mobility management and the shaping of non-publics in Europe. Eur J Soc Theory 18(1):21–38
Dixon-Román E (2016) Algo-Ritmo: more-than-human performative acts and the racializing assemblages of algorithmic architectures. Cultural Studies Critical Methodologies 16(5):482–490
Endicott T, Yeung K (2021) The death of law? Computationally personalized norms and the rule of law. Univ Toronto Law J 72(4):373–402
European Commission (DG for Migration and Home Affairs), Passenger Name Record (PNR)
European Commission, Delegated Decision on Further Defining Security, Illegal Immigration or High Epidemic Risks, 23 November 2021, Articles 3–6
European Commission (2020) (DG for Migration and Home Affairs), Opportunities and Challenges for the Use of Artificial Intelligence in Border Control, Migration and Security.
European Parliament (2021) Artificial intelligence at EU borders: overview of applications and key issues
EU Innovation Hub for Internal Security, Report to COSI, 24 February 2021 (copy on file with authors)
Foucault M (1982) The subject and power. Crit Inq 8(4):777–795
Foucault M (1991) Discipline and punish. Vintage Books, New York
Foucault M (2008) The birth of biopolitics: Lectures at the Collège de France 1978–1979 (transl. G. Burchell), Picador, New York
Fourcade M, Johns F (2020) Loops, ladders and links: the recursivity of social and machine learning. Theory Soc 49(5–6):803–832
Fanon F (1952) Black skin, white masks. Editions de Seuil, France
Glissant E (1997) The poetics of relation. The university of Michigan Press, Ann Arbor
Habermas J (1998) Between facts and norms. MIT Press, Massachusetts
Hacking I (2002) Making up people. In: Hacking I (ed) Historical ontology. Harvard University Press, Cambridge, pp 99–114
Hacking I (2006) Making up people. London Review of Books 28(16). https://www.lrb.co.uk/the-paper/v28/n16/ian-hacking/making-up-people. Accessed 30 June 2023
Haggerty KD, Ericson RV (2000) The surveillant assemblage. Br J Sociol 51(4):605–622
Haraway D (1991) A cyborg manifesto: science, technology, and socialist-feminism in the late twentieth century. In: Haraway D, Simians (eds) Cyborgs and women: the reinvention of nature. Routledge, New York
Hartman S (1997) Scenes of subjection: terror, slavery, and self-making in 19th century America. Oxford University Press, New York
Harney H, Moten F (2013) The undercommons: fugitive planning & black study. Minor Compositions, Wivenhoe
Harney H, Moten F (2021) All incomplete. Minor Compositions, Wivenhoe
Hui Y, Amaro R (2021) Designing for intelligence. In: 6In Vickers B and Allado-McDowell K (eds) Atlas of Anomalous AI, Lawson Publishing Limited Montgomery, Powys.
Hildebrandt M (2016a) Smart technologies and the end(s) of law: novel entanglements of law and technology. Edward Elgar, Cheltenham
Hildebrandt M (2016b) Law as affordance: the devil is in the vanishing points. Crit Anal Law 4(1):116–128
Hildebrandt M (2016c) Law as information in the era of data-driven agency. Mod Law Rev 79(1):1–30
Hildebrandt M (2019) Privacy as protection of the incomputable self: from agnostic to agonistic machine learning. Theoret Inquiries Law 20(1):83–121
Hildebrandt M (2020) Life and the law in the era of machine agency. In: Hildebrandt M, O’Hara K (eds) Life and the law in the era of data-driven agency. Edward Elgar, Cheltenham, pp 253–264
HM Government (2020) 2025 UK border strategy, London
Hohmann (2021) Diffuse subjects and dispersed power: new materialist insights and cautionary lessons for international law. Leiden J Int Law 34(3):585–606
Isin E, Ruppert E (2020) The birth of sensory power. Big Data Soc. https://doi.org/10.1177/2053951720969208
Johns F (2021) Governance by data. Annu Rev Law Soc Sci 17:53–71
Judy AR (2020) Sentient flesh: thinking in disorder, Poiēsis in Black, Duke University Press, Durham
Kingsbury B, Donaldson M (2011) From bilateralism to publicness in international law. In: Fastenrath U, Geiger R, Khan D-E, Paulus A, von Schorlemer S, Vedder C (eds) From bilateralism to community interest: essays in honour of Bruno Simma. Oxford University Press, Oxford, pp 79–89
Kingsbury B, Maisley N (2021) Infrastructures and laws: publics and publicness. Annu Rev Law Soc Sci 17:353–373
Latour B (1993) We have never been modern. Harvard University Press, Cambridge
Lindahl H (2013) Fault lines of globalization: legal order and the politics of a-legality. Oxford University Press, Oxford
Lupton D (2020) Data selves: more-than-human perspectives. Wiley, New York
Lippmann W (1993 [1927]) The phantom public. Transaction Publishers, New Brunswick
Mol (2003) The body multiple: ontology in medical practice. Duke University Press, Durham
McKittrick K (2013) Plantation futures. Small Axe 17(3(42)): 1–15
Mackenzie A (2015) The production of prediction: what does machine learning want? Eur J Cult Stud 18(4–5):429–445
Merry SE (2011) Measuring the world indicators, human rights, and global governance. Curr Anthropol 52(53):83–95
Mezzadra S, Neilson B (2013) Border as method: or, the multiplication of labor. Duke University Press, Durham
Marks S (2021) Dignity and indignity in the South African toilet wars’, Hersch Lauterpacht Memorial Lectures, March 2021. https://www.lcil.cam.ac.uk/lectures-events/hersch-lauterpacht-memorial-lectures. Accessed 30 June 2023
Mann I (2021) Border masquerades. Berkeley J Int Law 39(1):127–160
Moten F (2017) Black and blur. Duke university Press, Durham
Moten F (2018) The universal machine. Duke university Press, Durham
Moten F, Harney S (2021) Wildcat the totality. (Millennials are killing capitalism podcast), https://soundcloud.com/anticapitalist-millennials/wildcat-the-totality-fred. Accessed 30 June 2023
Massumi B (2002) Parables for the virtual: movement affect sensation. Duke University Press, Durham
Rancière J (1998) Dis-Agreement: Politics and Philosophy. (transl. J. Rose), Minnesota University Press, Minneapolis
Rose N (1998) Governing the soul: the shaping of the private self. Free Association Books, London
Rouvroy A (2012) The end(s) of critique: data-behaviourism vs. due-process. In: Hildebrandt M, De Vries E (eds) Privacy, due process and the computational turn. Philosophers of Law Meet Philosophers of Technology. Routledge, Abingdon, pp 143–168
Rouvroy A (2016) ‘Big data: l’enjeu est moins la donnée personnelle que la disparition de la personne’ (Le Monde, 22 January 2016). www.lemonde.fr/blog/binaire/2016/01/22/le-sujet-de-droit-au-peril-de-la-gouvernementalite-algorithmique/. Accessed 30 June 2023
Rouvroy A (2022) Athanasiadou and Klumbyté, ‘Re-Imagining a “We” Beyond the Gathering of Reductions: Propositions for the Three Ecologies. Footprint 16(1):121–134
Rouvroy A, Berns T (2013) Gouvernementalité Algorithmique et Perspectives d’Emancipation: Le disparate comme condition d’individuation par la relation? Réseaux 177:163–196
Rouvroy A, Stiegler B (2016) The digital regime of truth: from the algorithmic governmentality to a new rule of law. La Deleuziana—Online J Philos 3:6–29
Raso J (Forthcoming 2023) (Searching for) Agencies of the State. In: Johns F, Sullivan G and Van Den Meerssche D (eds) Global Governance by Data: Infrastructures of Algorithmic Rule (on file with authors).
Sakai N, Morris M (1997) Translation and subjectivity. Minnesota University Press, Mineapolis
Samaddar R (2010) The emergence of the political subject. Sage, New Delhi
Scheel S, Ruppert E, Ustek-Spilda F (2019) Enacting migration through data practices. Environ Plan D: Soc Space 37(4):579–588
Star, Bowker (2005) How to infrastructure. In: Lievrouw LA, Livingstone S (eds) The handbook of new media: social shaping and social consequences of ICTs. Sage, London
Parfitt R (2019) The process of international legal reproduction: inequality, historiography, resistance. Cambridge University Press, Cambridge
Pelizza A (2021) Identification as translation: the art of choosing the right spokespersons at the securitized border. Soc Stud Sci 51(4):487–511
Pahuja S (2011) Decolonising international law: development, economic growth and the politics of universality. Cambridge University Press, Cambridge
Petersmann M (2021) Response-abilities of care in more-than-human worlds. J Hum Rights Environ 12(1):102–124
Phan T, Wark S (2021) Racial formations as data formations. Big Data Soc. https://doi.org/10.1177/20539517211046377
Phan T, Wark S (2023) Race, By Proxy. In: Johns F, Sullivan G, Van Den Meerssche D (eds) Global governance by data: infrastructures of algorithmic rule (forthcoming 2023, on file with authors)
Steyerl H (2019) A sea of data: pattern recognition and corporate animism (forked version). In: Apprich C, Cramer F, Chun WHK, Steyerl H (eds) Pattern discrimination. Meson Press, Lüneburg, pp 1–22
Tzouvala N (2020) Capitalism as civilisation: a history of international law. Cambridge University Press, Cambridge
Urueña R (2012) No citizens here: global subjects and participation in international law. Martinus Nijhoff Publishers, Cheltenham
Van Den Meerssche D (2022) Virtual borders: international law and the elusive inequalities of algorithmic association. Eur J Int Law 33(1):171–204
von Schnitzler A (2018) Infrastructure, apartheid technopolitics, and temporalities of “transition”. In: N. Anand N. Gupta A. and Appel H. (eds) The promise of infrastructure. Duke University Press, Durham, pp 133–154
Wynter S (2003) Unsettling the coloniality of being/power/truth/freedom: towards the human, after man, its overrepresentation—an argument. New Centen Rev 3(3):257–337
Acknowledgements
We thank Leila Brännström, Markus Gunneflo, Gregor Noll, and Amin Parsa for the invitation to the workshop on ‘Tech and the Transformation of Legal Imagination’ as well as all the participants for generous engagement, especially Daniel Joyce and Geoff Gordon. We also thank the two anonymous reviewers for their constructive comments.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
This work was supported by a Dutch NWO ‘Veni’ Grant (VI.Veni.211R.026) on ‘Anthropocene Legalities: Reconfiguring Legal Relations Within More-than-human Worlds’.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Petersmann, M., Van Den Meerssche, D. On phantom publics, clusters, and collectives: be(com)ing subject in algorithmic times. AI & Soc 39, 107–124 (2024). https://doi.org/10.1007/s00146-023-01728-8
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00146-023-01728-8