1 Introduction

Less than a month after Russia’s invasion of Ukraine in early 2022, a video began spreading on social media in which Ukrainian President Volodymyr Zelensky appeared to be standing at a podium, urging Ukrainians to surrender. Although the video looked real enough, at least to a casual observer, it was a fake—a deepfake. A typical deepfake video is a digital video in which, using deep neural network-powered artificial intelligence techniques, one person’s likeness has been at least partially superimposed over the likeness of another person originally appearing in a video.Footnote 1 Deepfake techniques also can be used to produce realistic audio fakes of a person’s speech. In the Zelensky deepfake, Russian disinformation agents (or their proxies) presumably recorded an actor, of roughly the same bodily proportions as Zelensky, standing at a podium giving a speech. Then, leveraging extant (authentic) video and audio recordings of Zelensky, they trained a deep learning system to model Zelensky’s facial movements, and mapped a simulation of Zelensky’s face overtop the actor’s face in the video, adding in similarly simulated and synchronized voice audio.Footnote 2

Obviously, one immediate goal of the producers and initial distributors of the Zelensky deepfake was dishonest: to deceive Ukrainian citizens and soldiers into believing that their president directed them to cease their resistance to the Russian invasion. Deepfake technology is even more well known, and notorious, for its use in producing pornographic videos in which famous actresses or actors appear to feature (Cole, 2017). In at least some—perhaps many—of these cases, too, the deepfakers’ intentions were surely dishonest, at least insofar as they intended their audiences to believe that their deepfakes were authentic.

In both the political and pornographic cases of deepfakes just mentioned, there are numerous reasons for concern: risks of political destabilization and even the fall of a sovereign nation (e.g., Ukraine), explicit depiction of an actor or actress without consent, erosion of trust in video and audio as reliable sources of evidence, and more. These concerns have been the focus of recent work in the philosophical literature on deepfakes.Footnote 3 However, there has been almost no sustained philosophical analysis of deepfakes from the perspective of concerns about honesty and dishonesty. That deepfakes are potentially deceptive is not surprising and has been noted.Footnote 4 But under what conditions does the use of deepfakes fail to be honest? And which human agents, involved in one way or another in a deepfake, fail to be honest, and in what ways? If we are to better understand the morality of deepfakes, these questions need answering. Our first goal in this paper, therefore, is to offer an analysis of paradigmatic cases of deepfakes in light of the philosophy of honesty.

Even though it is clear that many uses of deepfakes are morally problematic, there has been a rising counter-chorus claiming that deepfakes are not essentially morally bad, since there might well be uses of deepfakes that are not morally wrong, or even that are morally salutary, for instance, in education, entertainment, activism, and other areas.Footnote 5 However, while there are surely reasons to think that deepfakes can supply or support moral goods, it is nevertheless possible that these uses of deepfakes are dishonest. Our second goal in this paper, therefore, is to apply our analysis of deepfakes and honesty to the sorts of deepfakes hoped to be morally good or at least neutral. We conclude that, perhaps surprisingly, in many of these cases the use of deepfakes will be dishonest in some respects.

Of course, there might well be cases of deepfakes for which verdicts about honesty and moral permissibility do not line up. It might be that the use of one deepfake is honest but nevertheless morally wrong. And it might be that the use of another is dishonest but nevertheless morally permissible, all things considered. While we will sometimes suggest reasons why moral permissibility verdicts might diverge from honesty verdicts, we will not aim to settle the question of overall moral permissibility, which would require consideration of potentially many factors unrelated to honesty.

But if the dishonesty of dishonest deepfaking is neither necessary nor sufficient for moral wrongness, why does it matter if we have an analysis of its dishonesty? There are at least two reasons why having this analysis is philosophically significant. First, it is one of the standard goals of moral philosophy to provide philosophical explanations of, and grounds for, commonly shared moral intuitions. As noted above, one of our primary goals is to do precisely this, for the intuition that various deepfakes are (or are not) dishonest. Second, even if dishonesty might not be necessary or sufficient for moral wrongness, nevertheless it is an important factor for determining overall verdicts concerning moral permissibility or wrongness.Footnote 6

How important a factor? We are inclined to think there is at least a pro tanto moral reason against dishonest actions—perhaps even a prima facie moral obligation to be honest—in which case a good, overriding reason would be required for thinking that dishonesty is morally permissible. Moreover, how honesty ought to be weighed surely will depend on readers’ background moral theoretic commitments, which we cannot hope to address here. But we believe that the following analysis and arguments will be useful in helping to arrive at more philosophically clear and comprehensive all-things-considered moral decisions concerning the use of deepfakes.

In the following section, we sketch the most prominent recent account of honesty in the philosophical literature. In § 3, we give an analysis of the honesty of paradigmatic uses of deepfakes, distinguishing between the main phases of a deepfake’s lifecycle (production, distribution, and viewing) and the main agents involved in each stage. In § 4, we extend our analysis to several proposals for what might be taken to be morally salutary uses of deepfakes.

2 Honesty

Our goal is to examine whether and how the use of deepfakes might fail to be honest. But what is honesty? In this paper we will take on board Christian Miller’s (2021) recent account of honesty, since it is the most prominent recent account, and since we are inclined toward it. Fortunately, most of our discussion could be adapted for other recent accounts of honesty, if readers find specific aspects of Miller’s account problematic.Footnote 7

Miller’s focus is on honesty as a moral virtue, understood along traditional Aristotelian lines as involving cognition, motivation, and outward behavior.Footnote 8 Hence an honest person is disposed to think honest thoughts, have honest motives, and behave honestly, both across a variety of situations relevant to honesty and stably over time. In this paper, however, we are focused on evaluating a person’s actions as honest or as dishonest, but not their character.Footnote 9 With respect to actions specifically, Miller’s core approach can be captured as follows:Footnote 10

  1. (i)

    An agent acts honestly when she does not intentionally distort the facts as she sees them, in contexts she understands to be factual.

  2. (ii)

    An agent acts dishonestly when she does intentionally distort the facts as she sees them, in contexts she understands to be factual.

We shall try briefly to unpack a few features of this approach. First, “intentionally”, for Miller, is to be understood as the opposite of “accidentally.” Thus, behavior that arises from subconscious mental states can still count as dishonest. Second, Miller does not have an account of what “distorting the facts” amounts to, but he does offer as a close synonym “misrepresenting.” If Sam tells his teacher that his dog ate his homework, he would be misrepresenting or “distorting the facts” in communicating to the teacher what happened to his homework. Third, Miller’s account ties honest and dishonest behavior to subjective construals of the facts. So, on his approach, people five thousand years ago were not acting dishonestly when they reported that the Earth was flat, even though they were seriously mistaken about the objective facts.

Finally, by “factual contexts”, we mean contexts in which the agent believes that representing the facts is relevant and likely to be expected by many people. This condition is important, since in some contexts distortions of facts do not plausibly amount to dishonesty. For instance, an author writing historical fiction is representing historical times and places to her readers, but representing non-factual events as occurring in those times and places. The author reasonably believes that readers will not take the context—i.e., reading historical fiction—to be a purely factual context.Footnote 11 Similarly, when filmmakers use CGI to represent fictional superheroes flying through space or tossing train cars with ease, they reasonably believe that viewers will not take the context—i.e., watching superhero movies—to be a purely factual context.Footnote 12

What forms do dishonest actions take? We note first that we are concerned here with communicative actions that result in the formation of propositional attitudes (e.g., beliefs, doubts, etc.). Often these kinds of actions take linguistic forms (e.g., speech, writing), but sometimes not. To use one of Miller’s examples (2021: 13), if a shady salesman applies a fresh coat of paint atop a car’s rusty roof, the action communicates propositional content about the car’s condition to prospective buyers. Or if a corrupt police officer plants drugs in a person’s pockets during a raid, the action communicates propositional content to other law enforcement officials. For convenience, however, our examples will typically involve linguistic actions.Footnote 13

While there are a number of ways to be dishonest, perhaps the ways most likely to be relevant to discussions of deepfakes are lying, misleading, and bullshitting.Footnote 14 On the traditional account of lying, a person lies in communicating that P, in typical cases, when she intends her audience to believe that P is true, while herself believing that P is false.Footnote 15 For instance, if we told you we loved free climbing sheer rock faces, and intended for you to believe us, we would be lying.

A person misleads in communicating that P, in typical cases, when he believes that P is true, but intentionally communicates it in such a way that, he hopes, his audience will form a false belief about Q.Footnote 16 For instance, in the 1990 film, Quigley Down Under, the protagonist, Quigley, is asked by the antagonist whether he has any skill with a pistol. Quigley replies, “I never had much use for one,” which was true, despite his being quite skilled indeed with a pistol.

A person bullshits in communicating that P, in typical cases, when in doing so he is mainly trying to get his audience to take some course of action, but is not concerned about whether P is true. “He does not care,” as Harry Frankfurt put it, “whether the things he says describe reality correctly. He just picks them out, or makes them up, to suit his purpose” (Frankfurt, 2005: 56). For instance, a politician might tell his audience, “My opponent doesn’t care a whit about inflation!” If the politician’s concern is gaining his audience’s votes, and he does not care one way or the other about the truth of his claim, he is bullshitting.Footnote 17

We will take a broad view of dishonesty as encompassing all those sorts of actions, though we do not claim that producing and distributing deepfakes could not fail to be honest in other ways as well.

Finally, it is important to note again that, in our view, from an action’s being dishonest it does not immediately follow that the action is morally wrong. For instance, most people would judge that families who harbored Jews during the Nazi occupation, but lied to the Gestapo about doing so, did nothing morally wrong, despite the fact that their doing so was dishonest. Similarly, from an action’s being honest it does not immediately follow that the action is morally permissible, as telling the truth to the Gestapo in this case illustrates.

3 Paradigmatic Cases

In this section, we will use Miller’s account of honest action to introduce a basic model for evaluating deepfakes vis-a-vis honesty, and then we will apply that model to a paradigmatic kind of deepfake. Not all deepfakes are paradigmatic ones, of course, but for simplicity and clarity’s sake, we will begin by considering such a case. Our conclusion about the paradigmatic case will be unsurprising: those who produce and distribute these sorts of deepfakes engage in dishonesty. But, in examining the paradigmatic case, we give a more precise explanation of why it is that the use of paradigmatic deepfakes is dishonest, even if there is little doubt that dishonesty was afoot. Moreover, in so doing we introduce our general method of examining the use of deepfakes with an eye to honesty and dishonesty, which lays the groundwork for examining a number of other deepfakes in the following section, some of which are not so obviously dishonest and perhaps even seem morally above board in all respects.

3.1 A Paradigmatic Deepfake

Although nothing critical turns on what qualifies as a “paradigmatic deepfake”, and although not all deepfakes will satisfy fully the following description, we will understand a paradigmatic deepfake to be a digital video; based on an original video recording; intended to deceive some viewers; produced without the consent of the people featuring in the original recording, or the people falsely represented as featuring in the deepfake, or the people who produced the original video; and for which the act of production and distribution is morally dubious (or worse). For the sake of convenience, we will typically talk about deepfakes using visual language, and thus use video deepfakes as the paradigmatic medium of deepfakes. But deepfakes need not be videos; they can be, for instance, digital audio files. Our analysis of deepfakes is intended to apply as well to non-video deepfakes.

A number of different sorts of deepfakes could qualify as paradigmatic, but most of the scholarly and popular attention has concerned pornographic deepfakes. This is not because deepfakes have any intrinsic affinity with pornographic media. Rather, this is largely because of the sad truth that, so far, the vast majority of deepfakes are pornographic (Simonite, 2019), and also because these sorts of deepfakes raise glaring moral concerns about the treatment of and attitudes toward women.Footnote 18 We will use a case of a pornographic deepfake as a paradigm case. However, since we suspect that many of those falsely represented as featuring in real pornographic deepfakes would prefer the existence of those deepfakes not to be advertised, we will not cite such cases and instead use a fictional case, including only enough detail needed for our analysis. Consider the following fictional case, which certainly has real analogs:

Adult Film: Adult Productions produces a pornographic film featuring Amy and Bob engaging in an explicit sexual encounter. [Call this recording Adult FilmA, where the subscripted “A” stands for “authentic”.] Both Amy and Bob are adult film actors, and both consent to being recorded for the film. Some time later, Chuck, a proficient user of deepfake software, sets out to produce a deepfake that will use Adult FilmA as its primary source, and that will falsely represent Dana, a public figure having no association with pornographic films, as the female lead. Chuck’s purposes for producing the deepfake are to see what it would look like for Dana to do what Amy does in Adult FilmA, and also to distribute the deepfake to others, with at least some desire that others will believe it to be an authentic recording. Chuck gathers a number of authentic video and audio recordings of Dana, trains the deepfake software to model her facial likeness and voice dynamically, and replaces Amy’s face with Dana’s, resulting in a deepfake video that appears to feature Dana and Bob. [Call this deepfake Adult FilmD, where the subscripted “D” stands for “deepfake”.] Chuck stores Adult FilmD on his personal computer, views it, and then uploads it to Shady Forum, a popular and unregulated online discussion forum. Ed, a Shady Forum user, views Adult FilmD and then posts a link to it on Social Site, a popular social media platform, exposing the deepfake to a broader audience.

3.2 Phase-agent Analysis

We can think of a typical deepfake’s lifecycle as having three very general phases: the production phase, during which the deepfaker gathers source materials and produces the deepfake; the distribution phase, during which the deepfaker (or others) distribute the deepfake, making it available to people for viewing; and the viewing phase, during which people view the deepfake. Of course, in some cases phases might overlap (e.g., a deepfaker might view the deepfake while producing it), recur (e.g., a deepfake might be distributed multiple times), or be missing (e.g., a deepfake might never be distributed or viewed).

We can also think of any deepfake as involving agents who fill (in some cases unwittingly) some or all of the following roles, which we will outline using the Adult Film case above. Chuck is the deepfaker, that is, the person who produced the deepfake. Dana is the faked-in person, that is, the person whom the deepfaker artificially inserted into the deepfake (Adult FilmD), and thus who was misrepresented as featuring in the deepfake, but who was never actually recorded for the primary source recording (Adult FilmA) or the deepfake (Adult FilmD). Amy is the faked-out person, that is, the person who was recorded for and featured in the primary source recording (Adult FilmA), but whose likeness was removed or overwritten in the deepfake (Adult FilmD). Bob is a faked interactant, that is, the person who was recorded for and featured in the primary source recording (Adult FilmA), and during which he interacted with the faked-out person. Adult Productions is the faked-from party, that is, the party having a legitimate authorial right to the authentic content in the primary source recording (Adult FilmA). Chuck and Ed are both distributors, that is, both made the deepfake (Adult FilmD) available to additional audiences. Shady Forum and Social Site might also be considered distributors, even if only unwittingly. Chuck and Ed are both consumers of the deepfake (Adult FilmD), that is, they both view the deepfake.

Note, however, that not for every deepfake will there be agents who fill all these roles. For instance, in the Zelensky deepfake noted above, only the faked-in Zelensky appears, so there is no faked interactant. Also, in the case of audio deepfakes in which there was no original audio recording serving as the base media onto which faked content is overlayed, there is a faked-in party, but no faked-out party, since no other person’s auditory likeness is removed.Footnote 19

While surely not every case of a pornographic deepfake involves dishonesty, we take it to be obvious that dishonesty was afoot in the Adult Film case. But using our phase-agent analysis, along with Miller’s account of honest actions, we can, in a more precise way, try to answer the following questions about the Adult Film case: were there any actions that failed to be honest? If so, whose actions were these, and why were they not honest? When discussing other deepfakes in subsequent sections of this paper, we will not step through each and every phase and agent involved. But it is important to do so for this first, paradigmatic case, to make our general method clear.

3.2.1 Distribution Phase

Consider the distribution phase first, since it is the phase during which deepfakes are shared with a broader audience. In the Adult Film case, clearly none of the faked-in (Dana), faked-over (Amy), faked interactant (Bob), or faked-from (Adult Productions) parties failed to be honest with respect to actions in this phase, since none of these parties was aware of the distribution of the deepfake, or even that the deepfake existed or would exist. So, none of these parties intentionally distorted any of the facts relevant to what appears in Adult FilmD, though of course all these parties were misrepresented in some way by other parties in the case.

Clearly, the deepfaker, Chuck, failed to be honest. After creating the deepfake, Chuck distributed it by posting it in an online forum. This was a communicative act, since he knew that posting the deepfake would make it viewable to other users. In the case as described, Chuck intended for other users to believe that the deepfake was an authentic recording. Since Chuck knew that the deepfake was inauthentic, he thus intended to distort the facts as he understood them. Thus, Chuck’s act of distributing the deepfake was dishonest.

But what facts, precisely, did Chuck distort for his intended audience? Several. Most obviously, Chuck misrepresented Dana as engaging in sexual acts with Bob, acts which Dana never engaged in. But Chuck misrepresents more than this. To the extent that Dana’s bodily appearance differs from Amy’s, Chuck misrepresents Dana’s bodily appearance, since it will appear that Amy’s body is Dana’s.Footnote 20 Chuck also misrepresents Dana as being recorded engaging in these acts with Bob. Further, since Adult FilmA was recorded with Amy’s and Bob’s knowledge and consent—let us suppose it is evident that the film was not recorded using hidden cameras—Chuck also misrepresents Dana as both knowing that she was being recorded and consenting to being recorded.Footnote 21 These are all distinct facts. Having a particular bodily appearance is a fact about one’s physical features. Doing something is a fact about one’s actions. Being recorded doing something is a fact about one’s actions being captured by a recording device. Knowing that one is being recorded is a fact about one’s epistemic mental states. Consenting to being recorded is a fact about one’s volitional mental states. Chuck intentionally conveyed all these distorted facts about Dana to viewers of Adult FilmD, and thus was dishonest about all these matters.

Chuck also intentionally distorted facts about Bob, Amy’s sexual partner in Adult FilmA and the faked interactant in Adult FilmD. Since Chuck’s deepfakery did not target Bob, Bob’s authentic likeness still appears in Adult FilmD; and so Chuck did not distort the facts about Bob’s likeness itself, nor about his engaging in sexual acts. But Chuck did misrepresent Bob as engaging in sexual acts with Dana, which Bob did not do, and so Chuck’s misrepresentation was dishonest. Chuck did not distort the facts about Bob’s being recorded, knowing he was being recorded, or consenting to being recorded. But Chuck did misrepresent Bob as being recorded with Dana, as knowing about being recorded with Dana, and as consenting to this.

Chuck also may have intentionally distorted facts about Adult Productions, the company that produced the original film, Adult FilmA. If Chuck’s deepfake, Adult FilmD, includes the display of “Adult Productions” in the opening or closing credits of Adult FilmA, then Chuck will have misrepresented Adult Productions as having recorded Dana and Bob (rather than Amy and Bob), knowing they recorded Dana and Bob, and perhaps other distortions as well. Or, if Chuck removed all mentions of Adult Productions, he will have distorted the facts about who produced the video, perhaps making it seem as if he was responsible for all the content of the video (not just the faked parts).Footnote 22

We said above that Chuck intentionally conveyed a number of distorted facts to viewers, and thus was dishonest about a number of facts. But is Chuck really dishonest about all the facts he distorts? While he surely was aware that he was misrepresenting Dana as doing certain things—namely, engaging in sexual interactions with Bob—it is unlikely that he thought expressly about each of the other facts he distorted. If dishonest actions are intentional distortions of the facts as one understands them (in contexts believed to be factual), how could Chuck be dishonest in representing Dana as having a certain bodily appearance, knowing she was being recorded, and so on, if he had no specific intention to convey those particular distorted facts?

Miller’s account of honesty alone does not settle this question.Footnote 23 But since we think it is worth seeing how this question might be settled, we will sketch a couple of approaches for further specifying the account. On either approach, Chuck fails to be honest to some extent. On the first approach, one can intentionally distort a fact only if one expressly intends to distort that particular fact.Footnote 24 The distortion of that particular fact must be part of the content of one’s intention. On this way of specifying intentional distortion, Chuck intentionally distorted the fact about Dana’s interaction with Bob—since Dana did not interact with Bob—and so Chuck is dishonest for doing so. Similarly, he intentionally distorted facts about Dana’s not being recorded interacting with Bob, about Amy’s interactions with Bob, and so on. But since Chuck did not intentionally distort some other facts, he was not dishonest in distorting those particular facts.

On the second approach, one acts intentionally with respect to outcome O1, if either (a) O1 is part of the content of one’s intention (broadly construed)Footnote 25 in acting, or (b) O1’s happening follows from the outcomes O2-On that are part of the content of one’s intention, and one reasonably ought to have known so (e.g., if one were to have paused for a moment to think, one would have seen the entailment), and if one had been aware of the entailment, one still would have done the action. Consider an example unrelated to distorting facts. Suppose Gary is with a group of people, all of whom are his friends except for Harry, whom Gary does not care for. Gary wants to make fun of Harry, to make Harry look like a fool. Gary succeeds. Harry looks like a fool. Gary’s friends laugh, and as a result, Harry is embarrassed. While Gary did not explicitly consider that making fun of Harry also would embarrass him, had Gary thought for a moment longer, he would have seen that making fun of Harry also would embarrass him. Moreover, had Gary taken that extra moment to consider, thus having both outcomes in view—viz., that Harry would look like a fool and that he would be embarrassed—Gary would have acted no differently. Both outcomes would have been part of the content of Gary’s intention.

On this second approach to specifying “intentionally”, Gary intentionally both made Harry look like a fool and embarrassed him, which at least seems plausible to us. Similarly, it seems plausible that, had Chuck taken a moment to consider what facts about Dana he would be distorting with his deepfake, he would have seen that the deepfake misrepresents not just Dana’s behavior, but also some or perhaps all of the following: her bodily appearance, her being recorded, her knowledge of being recorded, and her consenting to it. And it seems plausible that Chuck nevertheless would have produced the deepfake and distributed it. So, on this second approach, Chuck intentionally distorted some or perhaps all these further facts about Dana, and thus was dishonest on each count. We prefer this second approach, but perhaps others will prefer the first approach. On either approach, Chuck dishonestly distorts at least some facts about Dana.

Finally, what about Ed, the other distributor of Adult FilmD in this case? Did he act dishonestly? If Ed believed Adult FilmD to be a deepfake, but posted a link to it on Social Site at least in part so that others would believe the deepfake to be authentic, then what he does parallels lying. Or if, while believing Adult FilmD to be a deepfake, Ed posted a link to it without concern for its distortion of the facts—e.g., if he simply wanted people to enjoy a video he enjoys, and the video’s authenticity did not matter to him—then what he does parallels bullshitting. Either way, by promoting the deepfake, he, too, intentionally distorted the facts (mis)represented in the video to his audience, and thus acted dishonestly. If instead Ed believed Adult FilmD to be authentic, then, while his distributing it might well be a moral failure on other grounds, it would not be a failure of honesty.

3.2.2 Production Phase

Again, it seems clear that none of the faked-in (Dana), faked-out (Amy), faked interactant (Bob), or faked-from (Adult Productions) parties failed to be honest with respect to any actions relevant to the production phase. None was aware of the existence of the deepfake or, we can suppose, of any intent to produce a deepfake based on Adult FilmA.

While it is clear, we think, that Chuck acted dishonestly in distributing Adult FilmD to others, did his act of producing this deepfake fail to be honest? After all, producing the deepfake media file is distinct from posting it online, just as writing a (physical) letter is distinct from mailing it to someone. But because, in this case, distributing the deepfake with at least some hope of viewers believing it to be authentic was one of Chuck’s intentions for producing it, Chuck’s act of producing it is dishonest. This is because, it seems to us, the correct account of Chuck’s action of producing the deepfake involves his intentions for doing so. And since his intention to get others to believe the deepfake to be authentic is, in part, an intention to distort the facts as he understands them, Chuck’s act of producing the deepfake is dishonest.Footnote 26,Footnote 27

But suppose we alter the case slightly. Suppose that, while Chuck did intend to distribute Adult FilmD (e.g., by posting it on Shady Forum), he had no specific intention to deceive anyone. Rather, he was simply unconcerned about whether anyone would believe that Adult FilmD was authentic. Perhaps Chuck relished the thought that others would view his digital creation, or that others would enjoy it as he does. Would Chuck, in this version of the case, be dishonest? While the case in the previous paragraph parallels lying, this version of the case parallels bullshitting: the fact that his deepfake distorts the facts is of no concern to him with respect to distributing the deepfake to others. So it is bullshit. But bullshit is not honest.Footnote 28 If Chuck were to distribute his deepfake just as he did in the original version of the case—i.e., without any indication that the video is a fake—he would be intentionally distorting the facts as he understands them, even if he had no specific intention to deceive.Footnote 29

Consider another variation of the Adult Film case, in which Chuck produced Adult FilmD, but without any intention to distribute it, at least not at the time he produced it. Would Chuck’s act of producing the deepfake, in this variation of the case, be dishonest?Footnote 30 Although we are uncertain about this, we lean toward answering “no”. On the one hand, some people might have the intuition that simply producing a deepfake like the one Chuck produced amounts to some sort of failure of honesty. On the other hand, Chuck did not, in this variation of the case, distort any facts to anyone. So it is unclear to us how merely producing a deepfake—despite its having fictitious or distorted representations of real events—can be, strictly speaking, dishonest, absent any intent to distribute it to anyone. But if one still seems to have the lingering intuition—as at least one of us might have—that Chuck is dishonest in producing the deepfake even if he did not intend to distribute it, perhaps this merely reflects the conflation of a more general intuition of moral wrongness with the recognition that deepfakes involve distortions of facts. But again, since we are uncertain about this sort of case, we leave it to readers for further consideration. And, of course, we are not at all suggesting that Chuck’s act of producing the deepfake was not a moral failure for other reasons.Footnote 31

3.2.3 Viewing Phase

In the Adult Film case, Chuck and Ed each view Adult FilmD. But could the mere viewing of a deepfake be a failure of honesty? We think this is very unlikely. Chuck, the deepfaker, is dishonest both in distributing the deepfake and in producing it—so long as his reasons for producing it involved distributing it—but not in merely viewing it. Similarly for Ed: even if he exhibits other moral failings in his activity on Shady Forum and Social Site, merely viewing the deepfake does not itself seem dishonest. We are not suggesting, however, that viewing a deepfake is never morally wrong, just that it is does not seem in itself to be a failure of honesty.

3.3 Intentions and Non-Factual Contexts

According to our stipulated characterization of “paradigmatic deepfakes” above, the deepfaker intends to distribute the deepfake, and intends for at least some viewers to believe the deepfaked content to be veridical, which implies that the deepfaker believes that he is distributing the deepfake into a factual context (i.e., a context in which representing the facts is both relevant and expected). But if this is not the case, then would producing or distributing a deepfake be dishonest?Footnote 32 This will come up again when we consider special cases of deepfakes in the following section. But it will be helpful to address the question in a general way here.

Adapting a term from Viola and Voto (2023), let us use “overt deepfake” to refer to a deepfake whose content most reasonable viewers would judge as obviously both not representing reality and not so intended by the deepfaker. For instance, in a viral video in 2019, Steve Buscemi’s face was deepfaked onto Jennifer Lawrence’s body (Kelleher, 2019). It is obvious that the video’s content was not authentic, not so much because of the quality of the deepfake, but rather because of the clear mismatch of Buscemi’s face with Lawrence’s body; and it is thereby also obvious that it was not intended to be taken as genuine. It was intended to generate laughs, and succeeded. On our analysis, the deepfaker most likely was not dishonest in distributing the Buscemi/Lawrence deepfake. This is because the overtly non-genuine content of the deepfake converts the context of distribution into a non-factual context, i.e., a context in which people would not reasonably expect the content to represent the facts. And, as we noted in § 2, an agent acts dishonestly when, among other things, she believes the context to be a factual one.

It is also possible to distribute a deepfake whose content is not overtly fake, and yet do so without dishonesty. In yet another variation of the Adult Film case, suppose instead that Chuck posted his deepfake—the content of which is not overt—on a rather unsavory online forum called “Deepfake Celebrity Videos”, which everyone knows is a place for posting deepfakes rather than authentic videos.Footnote 33 If Chuck indeed knew that those who frequent this forum expect the videos disseminated there to be deepfakes, then Chuck believed the context to be non-factual, in which case Chuck did not intentionally distort the facts as he understood them in a context he believed to be factual. And so Chuck’s posting of the deepfake, while perhaps morally problematic for other reasons, most likely was not dishonest. It is still possible that Chuck’s action was dishonest, however, depending on other factors concerning the content of his intention. For instance, if he hoped that his deepfake would be further distributed beyond the unsavory forum and into factual contexts where people would likely believe the deepfake’s content to be genuine, then his posting of the deepfake was dishonest.

Some authors—e.g., Viola and Voto (2023), Cavedon-Taylor (2024), and Fallis (2020)—have argued (or at least suggested) that, if deepfakes become widespread and people become widely aware of this development, it might well “[shift] our default attitude toward…photographs and videos, possibly including genuine ones, toward a skeptical and (possibly) detached attitude” (Viola & Voto, 2023: 30). And, if this happens, they reasonably suggest that perhaps most people will not be deceived by deepfakes. Adapting this line of thought to our position in this paper, if that happens, then someone who produces and distributes a deepfake in such an environment will not be dishonest in doing so. Why? Because, so long as the deepfaker is aware that most people have a general skepticism toward digital recordings, she will not satisfy the factual context condition for dishonesty. Her intent would be similar to that of filmmakers who use CGI: they reasonably expect that people will not believe the CGI to be representing reality. We agree that, if this sort of general skepticism about digital recordings takes root, dishonesty in deepfaking would be far less common. However, we think it is clear that, at present, most people do not have this sort of general skepticism. Moreover, we think it is unlikely that most people will adopt this form of skepticism any time soon, if ever. But time will tell.

The above sort of analysis can be deployed easily to examine the honesty of uses of other paradigmatic kinds of deepfakes as well, for instance, the Zelensky political deepfake noted in the introduction. But now we want to extend our analysis to some interesting kinds of non-paradigmatic deepfakes.

4 Special Cases

What makes the following kinds of deepfakes interesting, from our point of view in this paper, is that each has been (or could be) claimed to be, on the whole, morally salutary, or at least not morally objectionable. Rather than taxing the reader with, for each example, the more extended sort of analysis undertaken in the previous section, we will focus on the special features of these cases of deepfakes. We hope it will be clear how our verdicts in these cases might extend to a wider range of other deepfakes with similar features. Since our focus is honesty, we do not aim to give an all things considered verdict concerning the bigger-picture question of the moral permissibility of using these deepfakes. But since considerations of honesty ought to be at least part of—probably a significant part of—all-things-considered-verdicts about moral permissibility, verdicts about the honesty of using deepfakes will be important for answering that bigger-picture question.

4.1 Deepfakes Communicating Important Truths

What features of a deepfake might render its production and distribution honest, or at least not dishonest? Perhaps a deepfake, in virtue of its content, can communicate a message that is important and true.Footnote 34 For such a deepfake, communicating this message might well be the primary goal of the deepfaker and distributor(s). For instance, a number of deepfakes have been used to communicate important humanitarian or political messages. Malaria No More UK released a deepfake aimed at raising awareness about the death toll caused by malaria each year, and challenging world leaders to act.Footnote 35 David Beckham delivers the message in the video—presumably because his fame on the pitch would attract international attention—and in doing so he appears to speak nine different languages, one after the other. But Beckham himself spoke only in English; the rest was faked.Footnote 36 Taking a different approach, Solidarité Sida, a French charity organization, released a deepfake in which it appears that Donald Trump gives a short speech dismissing concerns about AIDS. Their stated goal was to move world leaders to act, “by broadcasting a piece of fake news. The first piece of fake news that might eventually become true.” (Skinner, 2019)

We are happy to grant that these deepfakes’ primary intended messages are true and communicate important facts. Insofar as the deepfakers’ and distributors’ intentions were to represent these facts, they were, to that extent, honest. However, in producing and distributing these deepfakes, they communicated more than just their primary messages. They also distorted other facts as they understood them, and in factual contexts. Malaria No More UK was aware, of course, that Beckham did not—and, we assume, cannot—speak all nine of the languages he appeared to speak. In misrepresenting him as if he did speak those languages, they intentionally distorted the facts about what Beckham did speak and about what languages he was able to speak.Footnote 37 Thus, Malaria No More UK was partially dishonest in producing and distributing their deepfake. Solidarité Sida, too, acted dishonestly, at least if they intentionally distorted the facts about Trump, even if there was no failure in honesty in their underlying message about AIDS.

Deepfakes also have been proposed and already used for broadly educational purposes. For instance, for an exhibit called “Dalí Lives”, the Dalí Museum in St. Petersburg, Florida developed impressive deepfake videos of the already-deceased Salvador Dalí appearing to introduce himself, welcome visitors, and interact with them, all with his distinctive flair. The purpose of the deepfake is “to have visitors empathize with Dalí as a human being” (Lee, 2019), and to help introduce them to Dalí’s art.Footnote 38 Deepfakes have been proposed to realistically represent events that did not actually happen, but could very well have happened, and that tell us something worth knowing about history. For instance, CereProc, a technology company specializing in text-to-speech software, produced and distributed an audio deepfake in which it sounds as if John F. Kennedy is delivering the speech that his assassination prevented.Footnote 39 Even more creatively, it might be possible to create deepfake videos in which historical authors appear to deliver lectures, using only their texts, photographs, or perhaps other authentic videos of them (Fallis, 2020: 627).

Again, we are happy to grant that these kinds of deepfakes would communicate worthwhile messages, and insofar as communicating those messages is the intent behind producing and distributing these deepfakes, to that extent doing so would not be dishonest. However, producing and distributing deepfakes of this kind communicates more than just the intended historical facts. Dalí Museum decision-makers knew, of course, that Dalí himself never said or did what he appears to be saying or doing in their deepfakes, and so their conveying these distorted facts was intentional. Thus they were partially dishonest in producing and distributing their deepfake. Similarly, even though JFK wrote the speech he did not end up giving, the deepfakers knew he did not give it, and so the deepfaked audio is a distortion of the facts, and thus producing and distributing it was partially dishonest. A similar verdict would result for any deepfake representing someone as giving a lecture they did not give.

One might reply that, for some or all the deepfakes noted above, using a deepfake was the only way—or at least the most effective way—to communicate the relevant truths about those important facts.Footnote 40 And so, surely communicating them via deepfakes was not dishonest. First, for many deepfakes, we doubt that this is true. For instance, Malaria No More UK might have communicated their message about malaria with an authentic video of Beckham speaking in English, with subtitles in the other languages. Or they might have taken the time to coach Beckham to speak a few sentences in those other languages. Or they might have brought on additional celebrities who are native speakers of the other languages to deliver those lines. In the case of Solidarité Sida’s deepfake, we doubt there are many people who would honestly claim that a deepfaked Trump dismissing the seriousness of AIDS is the most effective method of spurring world leaders into action. And while the Dalí Museum’s deepfakes of their namesake are undeniably impressive, their attraction seems more in their AI-assisted artistic achievements than in their ability to convey the facts about Dalí’s art and personality. After all, it is not as if there are no extant genuine videos of Dalí one can watch.

Second, however, even if it is true that a given deepfake is the most effective way to convey the intended facts, it is also nevertheless true that any other facts intentionally distorted by the deepfake are intentionally distorted. For instance, it might well be true that Beckham speaking those nine languages indeed would be the most effective way—which is not to say the only effective way—to convey to an international audience particular facts about malaria. Nevertheless, this deepfake misrepresents Beckham as having spoken those languages, and clearly Malaria No More knew this when producing it and distributing it. Perhaps there were outweighing moral reasons in favor of using this deepfake, but even if so, that does not mean the conditions for dishonesty were not met.

Similarly, watching a deepfake video of a past occurrence might well be a superior way to learn certain facts about that occurrence. For instance, to our knowledge there is no video recording of the origin of American baseball’s “seventh inning stretch” tradition, when, in 1910, then-President William Taft stood up to stretch midway through the seventh inning.Footnote 41 But, despite video cameras being relatively rare at the time, there are extant video recordings of Taft. So it seems possible to produce a reenactment of the first seventh inning stretch using deepfake technology to fake Taft’s likeness into the video. But even if doing so would provide a unique insight into that historical event, thus in some sense conveying the facts about that event, doing so would also involve intentionally distorting other facts. Taft would be misrepresented as being video recorded; as having a particular precise body shape—it is unlikely a perfect body double can be found, after all—as standing in a particular place, with a particular pose; as standing next to a person with the likeness of the actors used in the reenactment; and so on. And, depending on the particulars of production and distribution, the deepfake might foreseeably wind up appearing in contexts in which many believe they are viewing a factual representation of the past. Thus, to the extent that considerations of honesty are important when engaging in moral deliberation about producing and distributing any of these kinds of deepfakes, those considerations might well count against moral permissibility.

4.2 Deepfakes for Entertainment

While the deepfakes considered above were intended primarily to communicate important facts to audiences, not all deepfakes are meant for this sort of communication. Some deepfakes are meant merely for entertainment. For instance, Sway, a small technology startup, recently developed a smartphone app enabling users to star in their own deepfaked dance videos.Footnote 42 The app instructs users to record themselves from different angles, using the smartphone’s built-in camera. The app then fakes the user’s facial likeness into a number of pre-recorded videos in which an actor performs a dance routine. With a tap and a swipe, users can then share their dance deepfakes directly to a number of social media platforms. Some have also proposed that deepfakes might equip a wider range of media producers (e.g., small companies without the big budgets of Hollywood studios, or even individuals) with inexpensive but powerful tools for creative visual storytelling (Kerner & Risse 2021: 103–104, Dhillon 2019).

If one produces and distributes a deepfake for these or other sorts of entertainment purposes, but not to convey important information, could doing so involve any failure of honesty?Footnote 43 We think so. We agree that using creative media to entertain ourselves and others is often laudable. But even in producing and distributing media intended for entertainment and not information sharing, we can still convey facts to our audiences, and thus we can also distort facts. For instance, even when one uses Sway’s app to post on social media a self-deepfaked dance video for the purpose of having some fun, one nevertheless intentionally distorts the facts—the video represents one as performing those dance moves, when one in fact did not—often in contexts in which viewers would, by default, expect videos to be authentic. This would be a form of bullshitting: one intentionally distorts the facts about one’s dance performance for one’s audience, but without particular concern for the facts themselves. Rather, one’s goal—not a bad goal in itself—is to entertain. But in the process one does indeed convey distorted facts. Similarly, if a movie studio, large or small, deepfakes actors’ likenesses into scenes in which these actors themselves did not act, the studio (mis)represents the actors as having performed in those scenes, and thus the studio intentionally distorts the facts to the audience.Footnote 44 We do not, of course, claim that these sorts of uses of deepfakes always would stem from bad intentions. But failures of honesty sometimes result from harmless or even worthy goals.

4.3 Labels and Disclosures: Packaging Deepfakes to Avoid Deception

Perhaps a deepfake, during its production or distribution phase, can be packaged in such a way that its audiences will not be deceived by it, thus knowingly converting the context into a non-factual one (i.e., one in which most viewers would not expect a representation of the relevant facts). Some writers have suggested, for instance, that requiring that deepfakes be presented with some sort of warning label or disclosure might go some distance toward preventing audiences from being deceived.Footnote 45 Indeed, the DEEP FAKES Accountability Act, first proposed in the U.S. House of Representatives in 2019, would require, among other things, that many deepfakes be distributed with watermarks and/or audible disclosures alerting audiences that the video and/or audio contained manipulated content.Footnote 46 Would producers and distributers of deepfakes avoid dishonesty, so long as they packaged up their deepfakes with disclosures? It is possible, but not as easy as it might seem.

By distributing a deepfake, or by producing one with the intent to distribute it, one is intentionally distorting the facts for an audience. But if in doing so one also intentionally takes steps to remedy the distorted facts for the audience—if one intentionally distorts the facts but also informs the audience about the distortions—perhaps one does not fail to be honest. We see three conditions here: if a deepfake producer or distributor packages a deepfake with a disclosure, does so for the purpose of preventing audiences from being deceived by the deepfaked content, and believes that the disclosure will prevent audiences (of sound mind) from being deceived, then probably the producer or distributor does not fail to be honest.Footnote 47 If any of those three conditions is not met, the act of producing or distributing fails to be honest.

Recall the Adult Film case from the previous section. Suppose Chuck, the deepfake producer and distributor in that case, adds a disclosure to his deepfake and merely believes that the disclosure will prevent audiences from believing that, for instance, Dana in fact starred in the film. But if Chuck did not add the disclosure in order to prevent deception, he does not satisfy the second condition noted just above, and thus his action still fails to be honest. Suppose the DEEP FAKES Accountability Act had become law, and Chuck added the disclosure only to stay out of trouble, but all the while still hoping that at least some viewers would be deceived. If so, Chuck would have intentionally distorted the facts for his viewers without really intending to remedy the distorted facts. Or, suppose instead that Chuck added the disclosure, and did so for the purpose of preventing deception, but believed it likely that many viewers still would be deceived. In that case, Chuck would be intentionally distorting the facts, but not intentionally remedying the distorted facts to a significant degree by his own lights, since one cannot intentionally φ (e.g., remedy the distorted facts for one’s audiences) while believing that one is not φ’ing.

How feasible is it that one could meet the conditions above, and thus produce or distribute deepfakes with disclosures and do so honestly? Although we cannot fully address this question here, we suggest that it is not as feasible as might be assumed. Consider some ways in which a deepfake might be packaged with a disclosure.Footnote 48 If a video with deepfaked content is distributed in a physical medium (e.g., a disc), a disclosure label might be added to the video’s physical packaging. This would be an ineffective method of disclosure, however, since many viewers would not see the label. Since this is easily foreseeable, it is unlikely the distributor would believe that the disclosure would prevent viewers from being deceived. But then the distributor would fail to be honest in distributing the deepfake, since, in distributing it with a disclosure believed to be ineffective, they would be intentionally distorting the facts but not intentionally remedying the distortion.

Alternatively, a disclosure might be added either to the beginning or the end of a deepfake, perhaps in the opening or closing credits in the case of movies or television shows. It takes but a moment’s thought, however, to foresee that many, perhaps most, viewers of the video will not see a disclosure added to the closing credits. Similarly, many viewers will likely not see a disclosure at the beginning of a video, whether due to inattentiveness, fast-forwarding past the usual piracy warnings and credits, or even viewing edited clips of the deepfaked content on streaming services such as YouTube. If the deepfake distributor is aware of these likelihoods, then they will have, in distributing the deepfake, intentionally distorted facts for their audience without intentionally remedying the distortions.

A more targeted approach might involve displaying a disclosure only during the segment of a video in which deepfaked content appears. For instance, the 2016 film, Rogue One: A Star Wars Story, included a scene in which the already-deceased actor Peter Cushing appeared to play a role, but which was accomplished using a stand-in actor and deepfake technology (Lincoln, 2016). No disclosures were given in the film at all, and so probably some viewers assumed that Cushing was alive during the filming of Rogue One. If the filmmakers had added a watermark or other visual notice appearing during that scene and disappearing afterward—e.g., a simple “Deepfake!”, or a more informative “Dear Audience, be aware that Peter Cushing did not actually appear in this scene. Guy Henry performed for this scene, which was then digitally altered so that Cushing’s likeness appears instead of Henry’s”—viewers would be less likely to assume they were watching a performance by Cushing, and thus the filmmakers might have avoided a failure of honesty.

But these sorts of disclosures are unlikely to be used much in film or other visual arts. Watermarks or other disclosures covering a significant portion of the screen would be at least somewhat visually jarring, detracting from the aesthetic experience intended by many filmmakers. And for other sorts of visual media less focused on viewers’ aesthetic experience, deepfakers with overt intentions to deceive would be motivated to avoid visual disclosures, since it would defeat the communicative purpose of a dishonest deepfake.

What about watermarks covering only a small portion of the video or image, so as not to be significantly visually jarring? As Viola and Voto (2023) note in their discussion of the app, DeepNude—which takes an image of a clothed person, and outputs an AI-generated nude image of that person—less obtrusive watermarks can be cropped away easily. Clearly, doing so before distribution would likely be dishonest. But even if the producer and initial distributor do not crop the watermark, it is easy for them to foresee, at least for some types of deepfakes, that someone else is likely to redistribute the deepfake with the watermark cropped away. And for deepfakes intended to be distributed broadly to a general audience, or even to a niche audience in widespread or indeterminate locations, those who will see the redistributed version of the deepfake lacking the watermark will indeed still be the intended audience. Such a case would be similar to adding a disclosure to the opening or closing credits in a film, discussed above, where the distributor does not actually believe that the disclosure will prevent the intended audience from being deceived.Footnote 49 So, at least in a range of cases, visual disclosures either would not be employed or would not clearly prevent the one who added the watermark before distribution from being dishonest.Footnote 50

Finally, another interesting method is to include the disclosure as part of the deepfaked content itself. For instance, in 2018, the Flemish Socialist Party published a deepfake in which Donald Trump appears to call on Belgium to exit the Paris Climate Accords (von der Burchard, 2018; Rini & Cohen 2022). At the end of the video, the deepfaked Trump says, “We all know climate change is fake, just like this video.” The Flemish Socialist Party’s deepfake was fairly crude, so most viewers would be unlikely to believe it to be a genuine recording of Trump.Footnote 51 But had the deepfake been realistic, it seems plausible that at least some viewers would have thought the video authentic, and the disclosure merely a joke. In addition, a disclosure of this sort would not be a practical solution for many deepfakes.

More could be said about how the use of disclosures might prevent dishonesty in producing and distributing deepfakes, but we hope it is clear at least that adding a disclosure to a deepfake does not automatically render honest the production and distribution of the deepfake.

5 Conclusion

In this paper we have offered an analysis of a range of deepfakes in light of the philosophy of honesty. While most people would readily agree that producing and distributing deepfakes of the paradigmatic sort—e.g., pornographic deepfakes—is dishonest, we have tried to explain in some depth why this is so. And while we accept that one can have morally good or at least neutral reasons for using other sorts of deepfakes, nevertheless, if what we have argued is correct, it is more difficult than one might have expected to produce and distribute even these sorts of deepfakes honestly.