Abstract
Artificial intelligence plays an increasingly important role in legal disputes, influencing not only the reality outside the court but also the judicial decision-making process itself. While it is clear why judges may generally benefit from technology as a tool for reducing effort costs or increasing accuracy, the presence of technology in the judicial process may also affect the public perception of the courts. In particular, if individuals are averse to adjudication that involves a high degree of automation, particularly given fairness concerns, then judicial technology may yield lower benefits than expected. However, the degree of aversion may well depend on how technology is used, i.e., on the timing and strength of judicial reliance on algorithms. Using an exploratory survey, we investigate whether the stage in which judges turn to algorithms for assistance matters for individual beliefs about the fairness of case outcomes. Specifically, we elicit beliefs about the use of algorithms in four different stages of adjudication: (i) information acquisition, (ii) information analysis, (iii) decision selection, and (iv) decision implementation. Our analysis indicates that individuals generally perceive the use of algorithms as fairer in the information acquisition stage than in other stages. However, individuals with a legal profession also perceive automation in the decision implementation stage as less fair compared to other individuals. Our findings, hence, suggest that individuals do care about how and when algorithms are used in the courts.
1 Introduction
Artificial Intelligence (“AI”) is swiftly becoming a relevant component in judicial decision-making processes around the globe (see, e.g., Reiling 2020). In China, “internet courts” already provide an online dispute resolution mechanism, also involving AI components (Fang 2018; Shi et al. 2021). In the US state of Wisconsin, judges utilize algorithms to derive recommended criminal sentences (Beriain 2018). Assessments of the defendant’s risk of engaging in violent acts are increasingly used in many countries (Singh et al. 2014) with varying degrees of accuracy (Tolan et al. 2019; Greenstein 2021).Footnote 1 Such technologies are typically referred to as Algorithmic Decision Making (“ADM”; see, e.g., Newell and Marabelli 2015; Araujo et al. 2020).Footnote 2
Straightforwardly, judges can benefit from the presence of ADM in judicial proceedings (Reichman et al. 2020; Winmill 2020).Footnote 3 First, advanced automation can potentially reduce the effort cost required to search through the documents, seek out the relevant legal provisions, or apply the law to the facts of the case.Footnote 4 Second, the well-established benefit of ADM lies in its ability to provide predictions that humans find difficult to generate, e.g., because the human capacity required to detect patterns in complex cases is limited (Alarie et al. 2018; De Mulder et al. 2022). Hence, ADM can potentially increase judicial accuracy by providing new information that cannot be detected by the naked eye or by improving the analysis process.
Nevertheless, it is currently somewhat difficult to study how ADM affects judicial decision-making, as the inner-working of judicial decision-making can be opaque, so it is not easy to observe which technology has been used. More importantly, the use of technology is not binary. Instead, judges may turn to ADM in different stages of adjudication, albeit the exact features vary across judicial systems. For instance, a highly technological court may take the following form: in the early stages of the judicial process, parties may be asked to upload their statements (e.g., the lawsuit and the statement of defense) onto a computerized system, while entering some general details about the lawsuit (sum, type of lawsuit, details on the parties). The judge can then observe the documents and verify whether they are consistent with the relevant procedural rules. After the initial submission, judges can use a computerized system to keep track of the process, send out automated reminders on deadlines, or run any number of preliminary analyses. Algorithms can then play a continuous rule in the adjudication process, e.g., by generating predictions, providing assessments and issuing electronic decisions. The more the technology develops, additional automated functions might be added to such a process. In the extreme case, a digitized dispute resolution might eventually be used to replace the judge and thereby circumvent the courts altogether (see, e.g., Ortolani 2019). However, we will restrict attention to a human judge who delegates some (but not all) of the judicial functions to ADM.
There have been extensive discussions on the implications of using automated procedures in the judicial process. These include aspects such as transparency and accountability, judicial independence, equality before the law, diversity, right to a fair trial, and efficiency (Matacic 2018; Zalnieriute and Bell 2019; Morison and Harkens 2019; Wang 2020; English et al. 2021). In particular, a salient concern for using technologies by judges is unfairness, e.g., algorithms can generate discriminatory outcomes based on race, ethnicity, or age (e.g., Jordan and Bowman 2022; Köchling and Wehner 2020). Moreover, some studies suggest that judges might use technology selectively, as they tend to rely more on extralegal factors in severe cases (Cassidy and Rydberg 2020).Footnote 5 At the same time, existing studies from behavioral economics suggest that judges may be susceptible to various cognitive biases (e.g., Guthrie et al. 2000, 2007; Winter 2020),Footnote 6 raising the question of whether the use of ADM can de-bias judges (see, e.g., Chen 2019).
Nonetheless, the concern of unfairness may well harm public trust in the judicial system. As one recent example, the use of software known as “COMPAS” to assess individual risk of reoffending has led to public outrage following the discovery that the algorithm led to racially discriminatory outcomes (Zhang and Han 2022).Footnote 7 Another famous example includes the fraud detection system “SyRI” in the Netherlands, which collected large amounts of personal data. SyRI was challenged by civil rights organizations in the District Court of The Hague, which then ruled that the technology violates the right to privacy (van Bekkum and Borgesius 2021; Buijsman and Veluwenkamp 2022).
These examples could be viewed as a special case of a more general issue: the relationship between AI and trust. In a recent review, Glikson and Woolley (2020) survey over 200 papers published in the last 20 years and identify different dimensions that determine whether individuals trust AI,Footnote 8 such as tangibility, transparency, reliability, and the tasks’ degree of the tasks’ technicality. These determinants are then found to have different effects, depending on how the AI manifests itself (as a physical robot, a virtual agent, or an embedded component). For instance, low reliability seems to decrease trust when the AI is embedded but may or may not decrease trust when the AI manifests itself as a robot. Such complexity makes it difficult to speculate on how judges respond to advice generated by ADM and, by extension, how this affects the public’s trust in those judges.Footnote 9
There is extensive writing on the importance of public trust in the courts (see., e.g., Burke and Leben 2007; Gutmann et al. 2022; Jamieson and Hennessy 2006), on the one hand, and the importance of trust in technology (Madhavan and Wiegmann 2007; Lee 2018; da Silva et al. 2018; Felzmann et al. 2019), on the other hand. However, the intersection is (at least empirically) under-explored,Footnote 10 with a few exceptions. Hermstrüwer and Langenbach (2022) use a vignette study to elicit perceptions of fairness in three contexts (predictive policing, school admissions, and refugees) on a scale ranging from “fully human” to “fully automated”. They find that purely algorithmic analysis is considered the least fair, but purely human decision-making is also considered somewhat unfair. Conversely, their study finds that combining automated processes with high human involvement yields a higher fairness perception. Yalcin et al. (2022) conducted a vignette experiment over MTurk and found that subjects care whether the judge is a human or an algorithm, finding evidence of higher trust in human judges.
Our study asks a related question: do individuals care about the stage in which technology is used by judges (rather than the overall degree of automation)? This question is crucial because it allows refining the conclusion as to what individuals (dis)like about the combination of human and machine adjudication. Specifically, we utilize a taxonomy by Parasuraman et al. (2000) that differentiates between four different stages of decision-making: (i) information acquisition, (ii) information analysis, (iii) decision selection, and (iv) decision implementation.
For each of these, we elicit beliefs about the Level of Automation (“LOA”) most likely to ensure the fairest outcome using an online exploratory survey of 296 participants. Our analysis yields two main findings. First, we find that individuals believe that an intermediate LOA generates the fairest results in the information acquisition stage, which is consistent with the study by Hermstrüwer and Langenbach (2022). However, we also find that lower levels of automation are believed to generate fairer outcomes in the remaining stages. This result suggests that individuals' preferences for combining human decision-making and algorithms is driven by AI’s relative advantage in acquiring information rather than its advantage in analyzing it. In other words, people trust judges to apply their legal expertise but less so to gather the relevant information. This conclusion seems particularly relevant for the distinction between inquisitorial and adversarial systems, as the judge's role in evidence collection is more passive in the latter than in the former.
Second, we find evidence that individuals with a legal profession believe that lower levels of automation are fairer in the implementation stage. This suggests that lawyers, unlike laypeople, are even more skeptical toward the AI’s ability to execute judicial decisions, i.e., lawyers tend to trust judges more strongly when it comes to implementation.Footnote 11
Our results seem important both for institutional design (e.g., how much technology to allow in judicial decision-making) and for judges who operate within those institutions.
The remainder of the paper is organized as follows: Sect. 2 situates our study within the existing literature. Section 3 describes our study’s design, with results reported in Sect. 4. Section 5 discusses the results, highlights some limitations, and concludes.
2 Related literature
Our paper is related to several streams of literature, including existing attempts to classify legal technologies, perceived procedural fairness of algorithms (in particular, in judicial decision-making), and work on the relative advantage of technology versus humans in judicial processes. We summarize the relevant points in turn.
2.1 A taxonomy of judicial decision-making automation
There are several existing attempts to create some classification for legal technologies. The Stanford University Codex TechindexFootnote 12 (see, e.g., McMaster 2019) categorizes existing legal technologies into nine categoriesFootnote 13 but does not discern between technologies intended for laypeople and for experts. These categories also do not easily lend themselves to researching decision-making or automation. A different attempt is contained in a report by “The Engine Room” (Walker and Verhaert 2019), which focuses on legal-empowerment technologies. Unfortunately, this attempt mainly revolves around technological applications and does not strive to provide a comprehensive categorization. The Law Society of England and Wales launched another attempt (Sandefur 2019), distinguishing between two “waves of AI in law”: a first (“rules-based”) wave, comprised of document automation, legal diagnostics, and legislative analysis tools, and a second wave, which embodies attempts to predict outcomes of disputes, analyze documents, and perform risk assessments. Much like the others, this attempt aims to describe current technological solutions rather than provide a clear taxonomy. More recently, Whalen (2022) proposed to classify legal technologies according to their “legal directness” and “legal specificity”,Footnote 14 whereas Guitton et al. (2022a, b) suggest mapping regulatory technologies along three different dimensionsFootnote 15: the project’s aim, divergence of interests between stakeholders, and the degree of human mediation. A different approach was taken by Tamò-Larrieux et al. (2022), who propose the concept of Machine Capacity of Judgment (MCOJ). According to this concept, classification should be derived by the artificial agent’s autonomy (i.e. freedom from outside influence),Footnote 16 decision-making abilities (including understanding the impact of decisions and balancing different options), and degree of rationality. Tamò-Larrieux et al. (2022) propose assigning a score to these parameters and leveraging those scores to determine how much to rely on the AI in question.
While these recent proposals seem useful for identifying what constitutes LegalTech and who it influences, they are less suitable for capturing how and when the technology is used by judges in their decision-making process.
We, therefore, take a different approach: combining elements from existing taxonomies and adjusting them to classify legal technologies. The starting point follows Parasuraman et al. (2000),Footnote 17 which break down the process of decision-making into four stages: (i) information acquisition—gathering, filtering, prioritizing, and understanding the data; (ii) information analysis—analyzing, interpreting, and making inferences and predictions; (iii) decision selection—prioritizing/ranking decision alternatives; and, (iv) decision implementation—executing the choice (e.g., writing-up and submitting the relevant document). While the categorization is not specific to legal decisions, it applies to judicial decision-making. For example, to make a ruling in the case, a judge must acquire relevant case law, analyze the information, generate alternatives, choose the best one, and implement the decision (e.g., write up a verdict). Notably, the stages are mutually inclusive, as to select the most relevant argument, one has to identify it and analyze it. Nevertheless, this taxonomy seems helpful for analyzing legal decision-making (see, e.g., Petkevičiūtė-Barysienė 2021).
This decision-making categorization is also closely related also to a paper by Proud et al. (2003), which describes the stages slightly differently (as “observe”, “orient”, “decide”, and “act”), but is nonetheless helpful for our purposes. They propose an LOA scale along the four stages of decision-making, with the underlying assumption that the preference may differ for each stage. We follow this assumption and elicit the beliefs of our survey respondents on fairness generated by the LOA in each stage of the judicial decision-making progress on a 5-point (Likert) scale ranging from “Manual” (i.e., no automation) to “full” (fully automated).Footnote 18 We describe the precise definitions in Table 1.
2.2 Perceived (procedural) fairness of ADM
Procedural fairness has long been applauded as means to keep litigants satisfied, cooperative, and trusting in the courts (see, e.g., Burke 2020; Burke and Leben 2007; MacCoun 2005). The presence of ADM in court procedures may, intuitively, affect both procedural fairness and the public perception of fairness. Studies on the perception of procedural fairness when AI is involved (not necessarily in the context of courts; see, e.g., Woodruff et al. 2018; Lee 2018; Lee et al. 2019; Saxena et al. 2019)Footnote 19 yield some mixed results. Some studies find that algorithms are seen as less fair than humans (Newman et al. 2020; Hobson et al. 2021), e.g., because they lack intuition and subjective judgment capabilities. Other studies, however, find that that the difference in perceived procedural fairness of human decision-makers versus algorithmic decision-makers is task-dependent (e.g., Lee 2018).
In the specific context of judicial decision making, recent studies show that people tend to trust human judges more than algorithms (see Yalcin et al. 2022) or that they, at least, do not trust a fully-automated judicial process (see Hermstrüwer and Langenbach 2022). Kim and Phillips (2021) further argue that “a robot would need to earn its legitimacy as a moral regulator by demonstrating its capacities to make fair decisions”.Footnote 20
Legal scholars seem to be divided on their attitude toward the use of technology in the course. While some seem supportive of such technologies (e.g., Reiling 2020; Winmill 2020), others take a more conservative approach (see, e.g., Sourdin and Cornes 2018; Ulenaers 2020). Often expressing concern of algorithmic bias (for a recent discussion, see Kim 2022). Such conservatism seems consistent with procedural fairness concerns, but may also driven by other reasons, e.g. dissatisfactory experience with technologies (Barak 2021), a fear of becoming redundant (Sourdin 2022), concern of being pressured into using more technologies (Brooks et al. 2020), or simply “Automation Bias” (Cofone 2021).
2.3 Relative advantage, compatibility, and personal innovativeness in information technology
Individuals might prefer a different LOA due to various reasons, including their attitudes towards technology. However, the existing literature on the determinants of technology acceptance suggests that these depend on social context and are subject to heterogeneity. For example, “innovators”, as Rogers calls them (Rogers 2003; for a summary of the theory, see, e.g., Kaminski 2011), are willing to take risks, are the first to develop new ideas, and are easy to persuade to accept new technologies. Other groups follow different patterns. “Laggards”, for example, will remain conservative and skeptical even after the implementation of the technology. Most people, however, fall into two other (and more moderate) groups—“early majority” and “late majority”. The early majority demands evidence about the usefulness of the technology before they are willing to adopt it. The late majority needs more than that—they demand information on the technology’s success among other people. Accordingly, personal innovativeness in information technology is an often-used construct in the context of technology acceptance (see Ciftci et al. 2021; Patil et al. 2020; Turan et al. 2015).
These studies identify several factors that influence whether a person will belong to a group that is quicker to accept technology. We focus on two such factors—Relative Advantage and Compatibility—which seem the most closely related to judicial decision-making.Footnote 21 The first factor, Relative Advantage, refers to the degree to which an innovation is seen as better than the idea, program, or product it replaces. The second factor, Compatibility, refers to whether the technology is consistent with the potential adopters' values, experiences, and needs. In other words, the first factor deals with whether the technology used by judges is a proper substitute for human decision-making, whereas the second factor deals with personal preferences.
Our study controls for these factors, along a few others (e.g., general trust and knowledge in legal technology) in order to isolate the question of interest—whether or not individuals care about the stage in which the judge turns to automation for assistance.
3 Study design
The following sections describe our study design. Section 3.1 describes our participants. Section 3.2 explains the method and procedure.
3.1 Recruitment of participants
We designed an online survey to elicit people’s beliefs about the fairness of using varying LOA for the different stages of the judicial process. Lithuania was chosen for the study,Footnote 22 for two main reasons. First, Lithuania has relatively average legal technologies: there is no AI used directly by judges in courts but there is a history (long before COVID-19) of using technology generally in the courts, e.g. e-services portal for courts,Footnote 23 judicial information systems,Footnote 24 and audio recordings of court hearings (Bartkus 2021).Footnote 25 The generally increasing level of automation in Lithuania is beneficial, as it makes it more likely that individuals will have varying degrees of awareness of at least some automation in courts. Second, Lithuania was chosen also for reasons of convenience, given that we had a logistical comparative advantage, which allowed us to recruit participants from several relevant groups (e.g., lawyers and other people with court experience) in this country with greater ease.
Participants were invited to participate using several recruitment methods. Given the different levels of knowledge and experience within legal systems, we wanted to recruit people both within and outside of the legal profession, and both with and without court experience (e.g., litigants, observers, defense attorneys, judges). Individuals from the legal community were invited to participate in the study by posting an invitation to a popular legal news site,Footnote 26 emailing scholars from several universities,Footnote 27and “snowballing”, i.e., reaching out to attorneys, judges, and legal scholars and asking them to share the invitation to participate with their colleagues. The snowballing sampling was also used to recruit people who have been to courts in any role, e.g., litigant, judge, observer, as court experience may greatly influence how people comprehend court work. The non-lawyer portion of the sample was approached by posting in various other Facebook groups.Footnote 28 In order to reach a wider range of ages among participants, emails with invitations to participate in the study were shared with elderly people attending Medard Čobot’s Third Century University (distributed by the university’s administration).Footnote 29 Given the variety of methods used, we cannot guarantee that our sample is representative of the entire population in Lithuania (or of the legal community in Lithuania). To mitigate this issue, we added several control variables (see the following section), which allow us to account for the heterogeneity among the participants. Overall, a convenience sample of 269 Lithuanian respondents participated in the study from May to June of 2021.
3.2 Method and procedure
The survey consists of several steps. First, we elicited information used to generate control variables. These include some demographics (age and gender) but, more importantly, measures for specific attributes that may influence the respondents’ attitude toward the use of ADM in the legal sphere. Specifically, respondents were presented with sets of 7 statements about various legal technologies in courts alongside Likert scales to measure the relevant feature (for the sources used to derive the statements, see Table 9 in Appendix B)Footnote 30:
-
Knowledge about Legal Technologies (“Knowledge in Tech”)—statements concerning the respondent’s general knowledge of legal technology. For instance, one statement was “In some countries, judges have access to a program that provides the judge with a detailed analysis of the case, evaluates arguments, and identifies possible outcomes of the case”. Then they are asked to indicate their level of knowledge on a Likert scale from 1 (“I know absolutely nothing about this”) to 5 (“I have tried this or a similar technology”).Footnote 31
-
Trust in legal technologies technology (“Trust in Legal Tech”), a scale consisting of three revised/adopted items, e.g., “Overall, I could trust legal technologies in courts”.
-
Relative advantage of legal technologies in courts (“Relative Advantage”)—a scale consisting of five revised/adopted items, e.g., “I think legal technologies would help save time for court clients and staff compared to how courts operate now”.
-
Compatibility—a scale consisting of four revised/adopted items from existing papers, e.g., “I think legal technologies would be well in line with my beliefs about how courts should operate”.
-
Personal innovativeness in information technology (“Personal Innovativeness”)—a scale consisting of four revised/adopted items, e.g., “Among my peers, I am usually the first to explore new information technologies”).Footnote 32
We pooled each set of statements using a simple mean, so that each feature is captured by one variable in the analysis.
The second stage of the survey involved the elicitation of the LOAs, i.e., the belief about which level of automation in the four stages (information acquisition, information analysis, decision selection, and decision implementation) is most likely to generate the fairest outcome. Respondents first read a few general sentences (e.g., “judges need a wide range of information to decide a case—legislation, decisions in similar cases, and legal arguments. The alternatives for outcome and arguments depend on the information found.”). Then, they were asked to “Choose the option you think would best ensure the fairest verdict in most cases.” As mentioned above, the answer was elicited on a 5-point Likert scale, ranging from a “Manual" to a “Full” automation level.Footnote 33 This was done with each of the four stages separately—i.e., each decision-making stage was described on a separate page, and the participant chose a level of automation for each decision making-stage separately. Tables 9 and 10 in Appendix B provide a complete translation of the questions given to the respondents.
3.3 Descriptive statistics
Descriptive statistics for our independent variables are presented in Table 2 (see also Table 4 in Appendix A for bivariate correlations). The sample has a mean age of approximately 41 years but also includes younger and older participants, which is essential given the possible generational gap regarding technology acceptance. There are slightly more females (60.7%) in the sample, and there are both participants with and without court experience.Footnote 34 Given the importance of awareness about how courts operate, our sample contains a substantial number of people with court experience: over 50 percent of the participants (N = 143, see Table 2) had court experience with a variety of roles during the proceedings, i.e., 16 people observed the process, 38 were witnesses, 23-legal representatives, 24 litigants, 5 experts, 4 defendants in criminal proceedings, 13 victims of a crime, 5 prosecutors and 2 judges.
Respondents reported relatively low levels of knowledge in legal technologies but rather high trust in such technologies. Furthermore, the descriptive statistics show a relatively strong belief in the relative advantage of legal technologies.
Table 5 in Appendix A compares the descriptive statistics between those with and without a legal profession, showing that lawyers in our sample have more knowledge about legal technologies (p < 0.001), but do not differ on trust, personal innovativeness, or compatibility.
4 Results
4.1 Perceived fairness generated by levels of automation
We begin our analysis by presenting descriptive results of LOA compared across the four stages of the judicial process in our taxonomy. The descriptive results are presented in Fig. 1 (a more detailed version is provided as Table 6 in Appendix A). The figure displays the percentage of responders who indicated that a particular LOA is most likely to produce a fair outcome.Footnote 35 The figure clearly shows that the density of subjects choosing an intermediate level is high in the information acquisition stage (40.15%), but lower in other stages (ranging from 16.85 to 21.19%). Respectively, the density of respondents choosing “low” is the highest in the other stages.
A Person chi-squared test reveals that LOAs differ across the stages (p < 0.001). As we elicited multiple LOAs from each participant, we also checked for within-subject differences using a Repeated Measures ANOVA, confirming that there are statistically significant differences between the beliefs regarding fairness generated by LOAs for different stages of decision-making (p < 0.001). Overall, this check reveals two key insights. First, the LOA for the information acquisition stage differs from the LOA for all other decision-making stages (p < 0.001 in all cases). Second, the LOAs for the other three stages do not differ from each other (p > 0.05 in all cases).
Next, Fig. 2 breaks down the data by different characteristics: Age (comparing older and younger respondents), gender, legal profession, and court experience (exact numbers are provided in Table 7 in Appendix A). The figure demonstrates that LOAs seem to be quite similar across different characteristics on a descriptive level (the only significant differences are between age groups). However, this does not yet constitute a full-blown analysis, as the variables capturing the characteristics are correlated (see Table 4 in Appendix A, which presents bivariate correlations). Hence, we proceed by using a regression model and control for these features simultaneously.
4.2 Linear regressions
Linear regressions (OLS) were conducted to predict preferred levels of automation in judicial decision-making stages. The regression model is
where the first three variables are dummies for the stage of the judicial decision-making (so that information acquisition is the baseline category), X is a vector of varying controls, and \(\epsilon\) is the error term. Results are reported in Table 3 (a full table, including the coefficients of the controls, is provided as Table 8 in Appendix A). As each observation represents one decision of one respondent (each respondent provided an answer for the four stages, so that there is a maximum of 1076 observations),Footnote 36 we cluster the standard errors by respondent. Column (1) excludes controls. Column (2) adds demographics (gender, age, age-squared, legal profession, court experience). Column (3) adds the elicited Knowledge in LegalTech, Trust in LegalTech, Personal Innovativeness, and Compatibility. Column (4) adds an interaction term between the stages and legal profession. Column (5) replaces the controls with respondent fixed effects, in order to control for any feature that varies by subject but is, for whatever reason, unaccounted for by our controls.
Table 3 reveals several key insights. First, the coefficients of LOA in the later stages of the judicial decision-making process (analysis, selection, implementation) are all negatively significant (p < 0.001). This suggests that the initial stage of information acquisition is believed to generate fairer results with a higher level of automation. In other words, individuals believe that the use of technology in later stages of the process is more likely to lead to unfair outcomes.
Second, respondents with a legal profession only differ from others in their perception of using automation in the implementation stage, with a negative coefficient for the interaction term (− 0.29, p < 0.005), suggesting they perceive automated implementation as relatively unfair.
Third, when comparing the coefficients of the three stages listed in the table (analysis, selection, and implementation), the coefficients are of similar size (and in fact, they are not significantly different from one another, as confirmed by a Wald test). This reaffirms that the results are driven by a distinction between the information acquisition and other stages (and not the differences across the three other stages).
5 Conclusion
Our study of beliefs about the level of automation in different stages of the judicial decision-making process reveals several interesting findings.
Firstly, people seem to believe that low levels of automation would ensure the fairest outcomes in judicial decision-making. The intermediate level of automation was preferred only in the first-information acquisition-stage. These results are consistent with the emerging literature on perceived algorithmic fairness within the law, which indicates that people might generally trust judges more than algorithms (Hermstrüwer and Langenbach 2022; Yalcin et al. 2022). However, it also suggests that a binary view of judges vs. algorithms might be insufficient for capturing how people perceive automation in the court, as perceptions change depending on the stage of the decision-making process. This finding might also be explained by the more general distinctions on trust in AI, as captured by the aforementioned paper by Glikson and Woolley (2020). Namely, ADM used in courts neither manifests itself to the public in a tangible way nor is it transparent, both of which tend to lead to lower trust.
Secondly, the evidence suggests that automation is perceived as most likely to generate fair outcomes in the first decision-making stage (information acquisition), which seems counter-intuitive: usually, one might expect individuals to trust algorithms in the analysis of the information more than its acquisition.Footnote 37 At the same time, people might be simply more familiar with the concepts of automatic information retrieval due to e.g., their day-to-day use of online search engines. As search engines have arguably gotten better over time, individuals may anticipate a similar process in ADM, which increases trust in information acquisition. However, other explanations for our finding may be found by turning to concepts from behavioral law and economics. For instance, suppose that judges fall pray to the so-called “confirmation bias” (see, e.g., Jones and Sugden 2001; for experimental evidence on adjudication, see Eerland and Rassin 2012), where they first form an opinion and then collect only the information that is consistent with that opinion. Individuals who anticipate the bias might then prefer to let an algorithm collect the evidence. Moreover, the presence of biases in the first stage might spillover to the following stages, e.g. because judges may turn to heuristics already from the beginning and this will form the basis for the subsequent stages.Footnote 38 An alternative explanation would be distrust in the current algorithm’s ability to perform analysis, selection and implementation, e.g., due to the usual aforementioned concerns of fairness and potential bias of ADM (for a discussion, see Kim 2022). In particular, implementation, unlike analysis, might be perceived as an inherently human process, requiring capabilities that are simply irreplaceable by a computer (see, e.g., Kasy and Abebe 2021). This is particularly true if implementation involves emotions (Yalcin et al. 2022; Xu 2022; Ranchordas 2022), e.g., allowing a human judge to incorporate equity concerns or compassion. Recall, however, that individuals with a legal profession in our study have an even stronger belief that automation in implementation is likely to yield unfair outcomes. This might be driven by either a true belief (e.g., due to conservatism, or to personal experience in representing clients in front of human judges) or political economy: if lawyers believe that their added value is in influencing implementation (e.g., by submitting written arguments to the judge before verdicts are written), they might object to automation in order to protect their stream of income.
Although our study is exploratory, the findings may potentially hold several important policy implications. First, our finding that the perceived fairness of automation is higher in the information collection stage implies that judges who are interested in maintaining their public support might take more liberty in using technology in the earlier stages of the process but avoid technology in later stages. From the perspective of judicial administration, one might even consider actively restricting judges from using automation for some actions, if judges prefer to save on effort costs and neglect the cost of lowering public support. Second, the stronger perception of lawyers regarding the unfairness of automated implementation may be especially important, as discontentment with automation might lead lawyers to communicate their criticism of the court to the clients. In other words, one might need to assign more weight to the preferences of lawyers because they might spill over to the clients.
Thirdly, the different stages we consider might be more relevant in inquisitorial systems—where the judge actively collects data—than in adversarial systems. The same is true for appeal systems in civil law systems, in which new evidence can be collected more easily in the appeal compared to common law systems (see, e.g., Feess and Sarel 2018). Of course, one should take this distinction with a grain of salt, as there may be second-order effects (e.g., if judges use automation to collect data, litigants may anticipate this and respond by hiding some information).
Lastly, we asked respondents to specify what they believe would yield a fairer outcome, i.e. we elicited their beliefs about fairness. This means that we do not directly ask whether they also prefer to have a fair process. As it is hard to imagine that people dislike fairness, it is plausible that respondents who believe that a certain process will yield a fair outcome also prefer to have that process in place. Therefore, our findings may well reflect the public’s preferences and not merely beliefs. Nonetheless, further research is needed to clarify this point, as it is possible that some specific sub-groups actually prefer unfair outcomes (e.g. guilty criminal defendants who would rather be unfairly exonerated).
Our study is subject to a few limitations. First, it is an exploratory study and, as such, uses simplified questions that aim at general opinions toward automation rather than specific opinions regarding adjudication fields. Nonetheless, it seems sufficient to illustrate the general point that differences exist between perceptions of technology at different stages. We leave the exploration of differences between legal fields for future studies. Second, our study builds on the existing literature, the highly-cited paper by Parasuraman et al. (2000), and applies discrete levels of automation, which has the advantage of keeping things simple for the subjects. However, future studies may well benefit by considering a more intricate distinction, such as the one proposed by Tamò-Larrieux et al. (2022). Third, as our pool of participants includes also non-professionals, one may question whether their views are essential. However, attitudes of the general public about court processes are no less critical to court legitimacy than experts’ opinions. Fifth, our use of Lithuanian respondents means that we can only capture opinions made against the background of judicial processes in Lithuania. A cross-country follow-up study may help establish whether our results generalize to other countries (in particular, given the possibility that inquisitorial systems differ from adversarial ones). Sixth, as discussed above, our convenience sampling is subject to limitations as well. Finally, our choice to restrict attention to fairly general questions might overlook the fine-grained details of the technology. Consequently, a different design, in which further details are provided on the technology’s capacity or functioning might yield different results. Future work would benefit from such attempts and shed further light on the ever-evolving issues discussed in this paper.
Notes
In addition to judicial tools, AI is indirectly involved in litigation through software used by litigants, e.g. automated document searches, prognosis of case outcomes (Kluttz and Mulligan 2019; Suarez 2020), and other tools that try to substitute for lawyers (Sandefur 2019; Sourdin and Li 2019; Janeček et al. 2021).
Newell and Marbelli (2015, p. 4) describe ADM as the case where “data are collected through digitized devices carried by individuals such as smartphones and technologies with inbuilt sensors—and subsequently processed by algorithms, which are then used to make (data-driven) decisions”. Note that some papers replace the word “Algorithmic” with “Automated”, yielding the same acronym of “ADM”.
Reichman et al. (2020) discuss the development of a technological system in Israel (“Legal-Net”), identifying both advantages for judges (e.g. it became easier to track the case’s progress) and disadvantages (e.g. it potentially created incentives to “conform” by issuing quick decisions at the expense of other considerations). Winmill (2020) describes how the use of technology in an Idaho court substantially enhanced efficiency.
For a similar argument regarding lawyers, see Alarie et al. (2018).
There are other critiques of using AI in legal decision-making. e.g., that certain types of AI over-rely on the assumption that past data can be used to predict the future (Fagan and Levmore 2019), that AI cannot properly judge whether the burden of proof is met (see Aini 2020), that AI outputs might be incomprehensible or alienated (Re and Solow-Niederman 2019), or that AI will unlikely produce high-quality decisions when data is insufficiently granular (Xu 2022).
However, there is also recent experimental evidence suggesting that judges may be less susceptible to framing effects than law students and laypeople (van Aaken and Sarel 2022).
Glikson and Woolley (2020) also differentiate between “cognitive trust”—a rational evaluation of the trustee—and “emotional trust”—a form of trust driven by feelings, affect, or personal connections.
Trust may also be affected by the degree of awareness to the technology, as many people may be under-informed about how technology is used, e.g. because they lack expertise. Still, even without perfect knowledge, people may hold opinions and preferences over the usefulness of—and need for—technologies in the courts (see, e.g., Barysė 2022a, b).
We discuss some possible explanations for why lawyers’ perceptions may differ in Sect. 5.
Namely, Marketplace, Document Automation, Practice Management, Legal Research, Legal Education, Online Dispute Resolution, E-Discovery, Analytics, and Compliance.
Legal directness refers to whether the technology engages directly with the law, whereas legal specificity asks whether the technology is generic or specifically applied only to legal issues.
The two papers by Guitton et al. use slightly different terms but are conceptually similar. The first dimension, Aim (or “primary benefit”) refers to whether the technology is meant to improve the accessibility to regulation or rather its efficiency.Divergence of interests concerns those sponsoring the technology, those implementing it, and its users. Human mediation concerns the degree of human (dis)integration, considering factors such as domain, code, and data.
Tamò-Larrieux et al. (2022) discern different types of autonomy, namely: autonomy in the choice of goal, autonomy in its execution, and autonomy from the environment, other agents, and organizations.
Parasuraman et al. (2000) is an influential paper (with over 4000 citations as of August 29, 2022 according to Google Scholar) and has been applied to conceptualize various decision-making processes. We build on this paper, as well as a working paper by (Petkevičiūtė-Barysienė 2021), which applies the framework also to a legal context..
Woodruff et al. (2018) studies qualitatively how knowledge on algorithmic fairness affects individuals’ perception of firms who use algorithms, finding a negative affect of unfairness on trust in those firms. Lee (2018) finds that algorithmic decisions are perceived as less fair when the task typically requires human skills. Lee et al. (2019) contrast different features of algorithmic fairness and find higher perceived fairness in response to some features of transparency (e.g., when the outcome of the algorithm was explained) but not others. Saxena et al. (2019) finds that algorithms that express “calibrated fairness”, which is based on meritocracy, are perceived as relatively fair.
Note that our study does not ask respondents directly about procedural fairness. Instead, it elicits beliefs about the LOA that is most likely to ensure the fairest outcomes. As such, the study is asking people to construe a fair process.
Examples of other factors include complexity (how difficult the innovation is to understand or use), trialability (the extent to which the innovation can be tested before adoption), and observability (the extent to which the innovation provides tangible results).
The survey was conducted in the Lithuanian language, but is presented henceforth using an English translation.
Teise.pro.
Vilnius University, Mykolas Romeris University.
E.g., “Apklausos” (“Surveys”), “Tyrėjai-tiriamieji” (“Researchers-respondents”).
Cronbach’s alpha, which is a test used to verify that the statements are indeed related, is as follows: Knowledge in Tech–0.897; Trust–0.872; Relative Advantage–0.911; Compatibility–0.927; Personal innovativeness 0.842. This suggests that the statements in each set are indeed related.
The questions for this point were restricted to knowledge about what kind of technologies for courts exist, without discussing how they work or any specific details. Our choice to use a general question is related to the issue underlying our study—that the existing attempts to classify AI lead to imperfect results We hence thought it is best to simply captures one’s general acquaintance with the technology, without going into details. While this yields a tradeoff, as we might miss the subtleties of precisely how one uses the technology, it seems preferable for two reasons. First, our sample does include legal professionals, so that there is a possibility that they encountered the technology directly (yielding a need to control for such acquaintance). Second, some aspects of the technology could, in principle, be used by individuals (e.g. if any collection of the evidence is done electronically).
Recall that Personal Innovativeness might be determined by the relative advantage and compatibility, but also by other factors. Hence, we controlled for this separately, to ensure that all the features are covered by our controls.
We also included an option of “This stage of the decision is not relevant to ensure a fair final decision in the case”. However, only a few people chose this option (see Table 6 for descriptive statistics).
Participants were asked to indicate whether they had court experience and name their role in the process, e.g., judge, litigant, attorney, observer.
In columns (2)–(5) there are missing observations for some controls, as we permitted subjects not to provide information for some questions.
In particular, the analysis of the information is probably viewed as more technical task, which generally has been found to increase trust (Glikson and Woolley 2020).
This is consistent with the well-established idea of a dual-process system (see, e.g., Bago and De Neys 2020), where individuals first use heuristics (“System I”) and only then turn to deliberate thinking (“System II”).
References
Agag G, El-Masry AA (2016) Understanding consumer intention to participate in online travel community and effects on consumer intention to purchase travel online and WOM: an integration of innovation diffusion theory and TAM with trust. Comput Hum Behav 60:97–111. https://doi.org/10.1016/j.chb.2016.02.038
Aini G (2020) A summary of the research on the judicial application of artificial intelligence. Chin Stud 09:14–28. https://doi.org/10.4236/chnstd.2020.91002
Alarie B, Niblett A, Yoon AH (2018) How artificial intelligence will affect the practice of law. Univ Tor Law J 68:106–186
Araujo T, Helberger N, Kruikemeier S, de Vreese CH (2020) In AI we trust? Perceptions about automated decision-making by artificial intelligence. AI Soc 35:611–623. https://doi.org/10.1007/s00146-019-00931-w
Bago B, De Neys W (2020) Advancing the specification of dual process models of higher cognition: a critical test of the hybrid model view. Think Reason 26:1–30. https://doi.org/10.1080/13546783.2018.1552194
Barak MP (2021) Can you hear me now? Attorney perceptions of interpretation, technology, and power in immigration court. J Migr Hum Secur 9:207–223
Bartkus J (2021) The admissibility of an audio recording in Lithuanian civil procedure and arbitration. Teisė. https://doi.org/10.15388/Teise.2021.120.3
Barysė D (2022a) Do we need more technologies in courts? Mapping concerns for legal technologies in courts. SSRN Electron J. https://doi.org/10.2139/ssrn.4218897
Barysė D (2022b) People’s attitudes towards technologies in courts. Laws 11:71. https://doi.org/10.3390/laws11050071
Beriain IDM (2018) Does the use of risk assessments in sentences respect the right to due process? A critical analysis of the Wisconsin v. Loomis Ruling Law Probab Risk 17:45–53. https://doi.org/10.1093/lpr/mgy001
Brooks C, Gherhes C, Vorley T (2020) Artificial intelligence in the legal sector: pressures and challenges of transformation. Camb J Reg Econ Soc 13:135–152
Buijsman S, Veluwenkamp H (2022) Spotting when algorithms are wrong. Minds Mach. https://doi.org/10.1007/s11023-022-09591-0
Burke K (2020) Procedural fairness can guide court leaders. Court Rev 56:76–79
Burke K, Leben S (2007) Procedural fairness: a key ingredient in public satisfaction. Court Rev 44:4–25
Cassidy M, Rydberg J (2020) Does sentence type and length matter? Interactions of age, race, ethnicity, and gender on jail and prison sentences. Crim Justice Behav 47:61–79. https://doi.org/10.1177/0093854819874090
Chen DL (2019) Judicial analytics and the great transformation of American Law. Artif Intell Law 27:15–42. https://doi.org/10.1007/s10506-018-9237-x
Ciftci O, Berezina K, Kang M (2021) Effect of personal innovativeness on technology adoption in hospitality and tourism: meta-analysis. In: Wörndl W, Koo C, Stienmetz JL (eds) Information and communication technologies in tourism 2021. Springer, Cham, pp 162–174
Cofone I (2021) AI and judicial decision-making. Artificial intelligence and the law in Canada. Lexis Nexis Canada, Toronto
Conklin M, Wu J (2022) Justice by algorithm: are artificial intelligence risk assessment tools biased against minorities? Past injustice, future remedies: using the law as a vehicle for social change. South J Policy Justice 16:2–11
Daugeliene R, Levinskiene K (2022) Artificial intelligence in the public sector: mysticism, possibility, or inevitability. In: New challenges in economic and business development 2022. University of Latvia, pp 90–95
da Silva JE, Scherf EDL, da Silva MVV (2018) In tech we trust? some general remarks on law in the technological era from a third world perspective. Rev Opinião Juríd Fortaleza 17:107. https://doi.org/10.12662/2447-6641oj.v17i25.p107-123.2019
De Mulder W, Valcke P, Baeck J (2022) A collaboration between judge and machine to reduce legal uncertainty in disputes concerning ex aequo et bono compensations. Artif Intell Law. https://doi.org/10.1007/s10506-022-09314-x
Eerland A, Rassin E (2012) Biased evaluation of incriminating and exonerating (non)evidence. Psychol Crime Law 18:351–358. https://doi.org/10.1080/1068316X.2010.493889
English S, Denison S, Friedman O (2021) The computer judge: expectations about algorithmic decision-making. In: Proceedings of the annual meeting of the cognitive science society, pp 1991–1996
Fagan F, Levmore S (2019) The impact of artificial intelligence on rules, standards, and judicial discretion. South Calif Law Rev 93:1–36
Fang X (2018) Recent development of internet courts in China part I: courts and ODR. Int J Online Dispute Resolut 5:49–60
Feess E, Sarel R (2018) Judicial effort and the appeal system: theory and experiment. J Leg Stud 47:269–294
Felzmann H, Villaronga EF, Lutz C, Tamò-Larrieux A (2019) Transparency you can trust: transparency requirements for artificial intelligence between legal norms and contextual concerns. Big Data Soc 6:205395171986054. https://doi.org/10.1177/2053951719860542
Glikson E, Woolley AW (2020) Human trust in artificial intelligence: review of empirical research. Acad Manag Ann 14:627–660. https://doi.org/10.5465/annals.2018.0057
Greenstein S (2021) Preserving the rule of law in the era of artificial intelligence (AI). Artif Intell Law. https://doi.org/10.1007/s10506-021-09294-4
Guitton C, Tamò-Larrieux A, Mayer S (2022a) A Typology of automatically processable regulation. Law Innov Technol. https://doi.org/10.1080/17579961.2022.2113668
Guitton C, Tamò-Larrieux A, Mayer S (2022b) Mapping the issues of automated legal systems: why worry about automatically processable regulation? Artif Intell Law. https://doi.org/10.1007/s10506-022-09323-w
Guthrie C, Rachlinski JJ, Wistrich AJ (2000) Inside the judicial mind. Cornell Law Rev 86:777
Guthrie C, Rachlinski JJ, Wistrich AJ (2007) Blinking on the bench: how judges decide cases. Cornell Law Rev 93:1
Gutmann J, Sarel R, Voigt S (2022) Measuring constitutional loyalty: evidence from the COVID-19 pandemic. SSRN Electron J. https://doi.org/10.2139/ssrn.4026007
Hermstrüwer Y, Langenbach P (2022) Fair governance with humans and machines. SSRN Electron J. https://doi.org/10.2139/ssrn.4118650
Heydari S, Fattahi Ardakani M, Jamei E, Salahshur S (2020) Determinants of completing the medication reconciliation form among nurses based on diffusion of innovation theory. J Res Health. https://doi.org/10.32598/JRH.10.3.1491.1
Hobson Z, Yesberg JA, Bradford B, Jackson J (2021) Artificial fairness? Trust in algorithmic police decision-making. J Exp Criminol. https://doi.org/10.1007/s11292-021-09484-9
Hübner D (2021) Two kinds of discrimination in AI-based penal decision-making. ACM SIGKDD Explor Newsl 23:4–13. https://doi.org/10.1145/3468507.3468510
Jamieson KH, Hennessy M (2006) Public understanding of and support for the courts: survey results. Georget Law J 95:899–902
Janeček V, Williams R, Keep E (2021) Education for the provision of technologically enhanced legal services. Comput Law Secur Rev 40:105519. https://doi.org/10.1016/j.clsr.2020.105519
Jones M, Sugden R (2001) Positive confirmation bias in the acquisition of information. Theory Decis 50:59–99. https://doi.org/10.1023/A:1005296023424
Jordan KL, Bowman R (2022) Interacting race/ethnicity and legal factors on sentencing decisions: a test of the liberation hypothesis. Corrections 7:87–106
Kaminski J (2011) Diffusion of innovation theory. Can J Nurs Inform 6:1–6
Kasy M, Abebe R (2021) Fairness, equality, and power in algorithmic decision-making. In: Proceedings of the 2021 ACM conference on fairness, accountability, and transparency. ACM, Virtual Event Canada, pp 576–586
Kim B, Phillips E (2021) Humans’ assessment of robots as moral regulators: importance of perceived fairness and legitimacy. https://doi.org/10.48550/ARXIV.2110.04729
Kim PT (2022) Race-aware algorithms: fairness, nondiscrimination and affirmative action. Calif Law Rev 110:1539
Kluttz DN, Mulligan DK (2019) Automated decision support technologies and the legal profession. Berkeley Tech LJ. https://doi.org/10.15779/Z38154DP7K
Köchling A, Wehner MC (2020) Discriminated by an algorithm: a systematic review of discrimination and fairness by algorithmic decision-making in the context of HR recruitment and HR development. Bus Res 13:795–848. https://doi.org/10.1007/s40685-020-00134-w
Kumpikaitė V, Čiarnienė R (2008) New training technologies and their us in training and development activities: survey evidence from Lithuania. J Bus Econ Manag 9:155–159. https://doi.org/10.3846/1611-1699.2008.9.155-159
Lee MK (2018) Understanding perception of algorithmic decisions: fairness, trust, and emotion in response to algorithmic management. Big Data Soc 5:205395171875668. https://doi.org/10.1177/2053951718756684
Lee MK, Jain A, Cha HJ et al (2019) Procedural justice in algorithmic fairness: leveraging transparency and outcome control for fair algorithmic mediation, pp 1–26
Lu J, Yao JE, Yu C-S (2005) Personal innovativeness, social influences and adoption of wireless Internet services via mobile technology. J Strateg Inf Syst 14:245–268. https://doi.org/10.1016/j.jsis.2005.07.003
MacCoun RJ (2005) Voice, control, and belonging: the double-edged sword of procedural fairness. Annu Rev Law Soc Sci 1:171–201. https://doi.org/10.1146/annurev.lawsocsci.1.041604.115958
Madhavan P, Wiegmann DA (2007) Similarities and differences between human–human and human–automation trust: an integrative review. Theor Issues Ergon Sci 8:277–301. https://doi.org/10.1080/14639220500337708
Matacic C (2018) Are algorithms good judges? Science 359:263–263. https://doi.org/10.1126/science.359.6373.263
McMaster C (2019) Is the sky falling for the Canadian artificial intelligence industry? Intellect Prop J 32:77–103
Min S, So KKF, Jeong M (2019) Consumer adoption of the Uber mobile application: insights from diffusion of innovation theory and technology acceptance model. J Travel Tour Mark 36:770–783. https://doi.org/10.1080/10548408.2018.1507866
Moore GC, Benbasat I (1991) Development of an instrument to measure the perceptions of adopting an information technology innovation. Inf Syst Res 2:192–222. https://doi.org/10.1287/isre.2.3.192
Morison J, Harkens A (2019) Re-engineering justice? Robot judges, computerised courts and (semi) automated legal decision-making. Leg Stud 39:618–635. https://doi.org/10.1017/lst.2019.5
Newell S, Marabelli M (2015) Strategic opportunities (and challenges) of algorithmic decision-making: a call for action on the long-term societal effects of ‘datification.’ J Strateg Inf Syst 24:3–14. https://doi.org/10.1016/j.jsis.2015.02.001
Newman DT, Fast NJ, Harmon DJ (2020) When eliminating bias isn’t fair: Algorithmic reductionism and procedural justice in human resource decisions. Organ Behav Hum Decis Process 160:149–167. https://doi.org/10.1016/j.obhdp.2020.03.008
Ortolani P (2019) The impact of blockchain technologies and smart contracts on dispute resolution: arbitration and court litigation at the crossroads. Unif Law Rev 24:430–448
Parasuraman R, Sheridan TB, Wickens CD (2000) A model for types and levels of human interaction with automation. IEEE Trans Syst Man Cybern Part Syst Hum 30:286–297. https://doi.org/10.1109/3468.844354
Patil P, Tamilmani K, Rana NP, Raghavan V (2020) Understanding consumer adoption of mobile payment in India: extending meta-UTAUT model with personal innovativeness, anxiety, trust, and grievance redressal. Int J Inf Manag 54:102144. https://doi.org/10.1016/j.ijinfomgt.2020.102144
Petkevičiūtė-Barysienė D (2021) Human-automation interaction in law: mapping legal decisions, cognitive processes, and automation levels
Proud RW, Hart JJ, Mrozinski RB (2003) Methods for determining the level of autonomy to design into a human spaceflight vehicle: a function specific approach
Ranchordas S (2022) Empathy in the digital administrative state. Duke Law J 71:1341–1389
Re RM, Solow-Niederman A (2019) Developing artificially intelligent justice. Stanf Technol Law Rev 22:242–289
Reichman A, Sagy Y, Balaban S (2020) From a panacea to a Panopticon: the use and misuse of technology in the regulation of judges. Hastings Law J 71:589–636
Reiling AD (2020) Courts and artificial intelligence. Int J Court Adm 11:1
Rogers E (2003) Diffusion of innovations, 5th edn. Free Press, USA
Sandefur RL (2019) Legal tech for non-lawyers: report of the survey of US legal technologies
Saxena NA, Huang K, DeFillips E, et al (2019) How do fairness definitions fare?: examining public attitudes towards algorithmic definitions of fairness. In: AIES ’19: proceedings of the 2019 AAAI/ACM conference on AI, ethics, and society, pp 99–106
Shi C, Sourdin T, Li B (2021) The smart court—A new pathway to justice in China? Int J Court Adm 12:4. https://doi.org/10.36745/ijca.367
Singh JP, Desmarais SL, Hurducas C et al (2014) International perspectives on the practical application of violence risk assessment: a global survey of 44 countries. Int J Forensic Ment Health 13:193–206. https://doi.org/10.1080/14999013.2014.922141
Sourdin T (2022) What if judges were replaced by AI?. Turk Policy Q
Sourdin T, Cornes R (2018) do judges need to be human? The implications of technology for responsive judging. In: Sourdin T, Zariski A (eds) The Responsive judge. Springer, Singapore, pp 87–119
Sourdin T, Li B (2019) Humans and justice machines: emergent legal technologies and justice apps. SSRN Electron J. https://doi.org/10.2139/ssrn.3662091
Suarez CA (2020) Disruptive legal technology, COVID-19, and resilience in the profession. S C Law Rev 72:393–444
Tamò-Larrieux A, Ciortea A, Mayer S (2022) Machine Capacity of Judgment: an interdisciplinary approach for making machine intelligence transparent to end-users. Technol Soc 71:102088. https://doi.org/10.1016/j.techsoc.2022.102088
Tolan S, Miron M, Gómez E, Castillo C (2019) Why machine learning may lead to unfairness: evidence from risk assessment for juvenile justice in Catalonia. In: Proceedings of the seventeenth international conference on artificial intelligence and law. ACM, Montreal QC Canada, pp 83–92
Turan A, Tunç AÖ, Zehir C (2015) A theoretical model proposal: personal innovativeness and user involvement as antecedents of unified theory of acceptance and use of technology. Procedia Soc Behav Sci 210:43–51
Ulenaers J (2020) The impact of artificial intelligence on the right to a fair trial: towards a robot judge? Asian J Law Econ. https://doi.org/10.1515/ajle-2020-0008
van Aaken A, Sarel R (2022) Framing effects in proportionality analysis: experimental evidence. SSRN Electron J
van Bekkum M, Borgesius FZ (2021) Digital welfare fraud detection and the Dutch SyRI judgment. Eur J Soc Secur 23:323–340
Vimalkumar M, Sharma SK, Singh JB, Dwivedi YK (2021) ‘Okay google, what about my privacy?’: user’s privacy perceptions and acceptance of voice based digital assistants. Comput Hum Behav 120:106763. https://doi.org/10.1016/j.chb.2021.106763
Walker T, Verhaert P (2019) Technology for legal empowerment. https://library.theengineroom.org/legal-empowerment/
Wang N (2020) “Black Box Justice”: robot judges and AI-based Judgment processes in China’s court system. In: 2020 IEEE international symposium on technology and society (ISTAS). IEEE, Tempe, AZ, USA, pp 58–65
Wang X, Yuen KF, Wong YD, Teo CC (2018) An innovation diffusion perspective of e-consumers’ initial adoption of self-collection service via automated parcel station. Int J Logist Manag 29:237–260. https://doi.org/10.1108/IJLM-12-2016-0302
Whalen R (2022) Defining legal technology and its implications. Int J Law Inf Technol 30:47–67. https://doi.org/10.1093/ijlit/eaac005
Winmill L (2020) Technology in the judiciary: one judge’s experience. Drake Law Rev 68:831–846
Winter CK (2020) The value of behavioral economics for EU judicial decision-making. Ger Law J 21:240–264. https://doi.org/10.1017/glj.2020.3
Woodruff A, Fox SE, Rousso-Schindler S (2018) A qualitative exploration of perceptions of algorithmic fairness. In: Conference on human factors in computing systems—Proceedings. Association for Computing Machinery
Xu Z (2022) Human Judges in the era of artificial intelligence: challenges and opportunities. Appl Artif Intell 36:2013652. https://doi.org/10.1080/08839514.2021.2013652
Yalcin G, Themeli E, Stamhuis E et al (2022) Perceptions of justice by algorithms. Artif Intell Law. https://doi.org/10.1007/s10506-022-09312-z
Yuen KF, Wang X, Ng LTW, Wong YD (2018) An investigation of customers’ intention to use self-collection services for last-mile delivery. Transp Policy 66:1–8. https://doi.org/10.1016/j.tranpol.2018.03.001
Zalnieriute M, Bell F (2019) Technology and the judicial role. SSRN Electron J. https://doi.org/10.2139/ssrn.3492868
Zhang J, Han Y (2022) Algorithms have built racial bias in legal system-accept or not? Sanya, China
Zhang T, Tao D, Qu X et al (2020) Automated vehicle acceptance in China: social influence and initial trust are key determinants. Transp Res Part C Emerg Technol 112:220–233. https://doi.org/10.1016/j.trc.2020.01.027
Acknowledgements
This study has received funding from the European Social Fund (Project No. 09.3.3-LMT-K-712-19-0116) under a grant agreement with the Research Council of Lithuania (LMTLT). We thank Pedro Christofaro, Abishek Choutagunta, Dylan Johnson, Mahdi Khesali, Hashem Nabas, Katharina Luckner, Stefan Voigt, and participants of the ILE Hamburg Jour Fixe for useful comments.
Author information
Authors and Affiliations
Contributions
The study conception, design, and data collection were performed by DB. Data analysis and final manuscript were performed by DB and RS. Both authors approved the final manuscript.
Corresponding author
Ethics declarations
Conflict of interest
The authors have no competing interests to declare.
Ethical approval
The study design and procedures were approved by the Psychological Research Ethics Committee of Vilnius University (250000-KP-25).
Informed consent
Informed consent was obtained from all individual participants included in the study.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendices
Appendix A
See Tables
4,
5,
6,
7 and
8.
Appendix B
See Tables
9 and
10.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Barysė, D., Sarel, R. Algorithms in the court: does it matter which part of the judicial decision-making is automated?. Artif Intell Law (2023). https://doi.org/10.1007/s10506-022-09343-6
Accepted:
Published:
DOI: https://doi.org/10.1007/s10506-022-09343-6
Keywords
- Judicial decision-making
- Human-automation interaction
- Legal technologies
- Levels of automation
- Perceived fairness