Introduction

To combat financial fraud, public and private organizations regularly run awareness campaigns (see, e.g., campaigns by the Swedish Bankers’ Association (2023), U.S. Immigration and Customs Enforcement (2023), and UK Finance (2022)). Despite good intentions, there is little empirical evidence as to whether such campaigns work (Gotelaere & Paoli, 2022; Button & Cross, 2017; Cross & Kelly, 2016; Prenzler, 2020). However, there could be multiple reasons to suspect that they do not. For one thing, the behavior targeted by campaigns may be strongly influenced by situational and emotional factors, including relationship status and mental health problems (Poppleton et al., 2021). Furthermore, criminals may imitate bank or police officers (Choi et al., 2017) and, as one anti-fraud expert from the bank considered in our paper puts it, “(...) most awareness campaigns contain a message á la; don’t trust people that call you and say they are from the police or a bank... But there is something weird about the police or a bank promoting that message.” In other words, the people influenced by an awareness campaign, launched by a public authority or bank, might disregard the campaign’s message when a fraudster claims to be from a public authority or bank. This could especially be the case when we consider people susceptible to fraud.

In this paper, we consider an awareness campaign by Spar Nord, a systemically important Danish bank. In September 2020, the bank sent out a warning message through its online banking portal, targeting clients over 40 years of age. We consider this message a concrete example of an awareness campaign and utilize it as a quasi-experiment. To this end, we employ choice-based data collection; we collect data on all of the bank’s known fraud victims and a sample of “baseline” clients (i.e., clients assumed not to be fraud victims). The strategy is motivated by the rarity of fraud cases and how some of our features are expensive to collect. To account for the fact that we undersample baseline clients, we weigh observations inversely proportional to the probability that they are in our data. For analysis, we use linear probability models with difference-in-differences and regression discontinuity designs. Our features include information about age, sex, relationship status, financial funds, and residence. For a subset of observations, we also have access to and employ educational information. None of our models find that the message had a significant effect, indicating that awareness campaigns relying on mass messaging have little effect on financial fraud.

The rest of our paper is organized as follows. The “Related Literature on FinancialFraud” section reviews related literature on financial fraud. The “Institutional Setting” section presents the institutional setting of our study. The “Data” section presents our data. The “Methodology” section presents our methodology and difference-in-differences and regression discontinuity models. The “Results” section presents our results. Finally, the “Discussionand Conclusion” section ends the paper with a discussion and conclusion.

Related Literature on Financial Fraud

Despite generally declining crime rates, financial fraud remains a major concern across the globe (Prenzler, 2020; Button & Cross, 2017). To address the issue, some scholars argue for a “public health approach,” recommending, among other things, public campaigns targeting high-volume types of fraud (Levi et al., 2023). Button and Cross (2017), in particular, advocate for campaigns targeted to those most likely to be victimized by fraudsters. In this connection, several studies have identified key risk factors, including (age-related) decline in cognitive ability (Gamble et al., 2014; Han et al., 2016; James et al., 2014), emotional responses (Kircanski et al., 2018), overconfidence about financial knowledge (Gamble et al., 2014; Engels et al., 2020), and depression and lacking social needs-fulfillment (Lichtenberg et al., 2013, 2016). A different approach, as explored by Wang et al. (2020), targets criminals with warnings about the consequences of their actions; see Prenzler (2020) for a comprehensive review. In general, however, there exists little empirical evidence as to whether awareness campaigns actually work, i.e., help reduce financial fraud (Gotelaere & Paoli, 2022; Button & Cross, 2017; Cross & Kelly, 2016; Prenzler, 2020).

A study by the AARP Foundation (2003) found that direct peer counseling may help reduce telemarketing fraud. The study considered people deemed to be “at-risk” as they appeared on fraudster call sheets seized by the U.S. Federal Bureau of Investigation. Trained volunteers (i.e., peer counselors) called and reached 119 subjects, randomly delivering a control message (asking about TV shows) or a message warning about telemarketing fraud. Within five days, subjects were then recalled and given a fraudulent sales pitch. Results showed that the warning message significantly reduced fraud susceptibility.

Scheibe et al. (2014) showed that warning past victims may reduce future fraud susceptibility. The study considered 895 people identified as fraud victims by the U.S. Postal Inspection Service. Volunteers called the subjects and delivered either (i) a specific warning (about a type of fraud emulated later), (ii) a warning about a different type of fraud, or (iii) a control (asking people about TV preferences). Two or four weeks later, the subjects were recalled and exposed to a mock fraud. Both types of warning messages reduced fraud susceptibility, though outright non-cooperation (instead of mere skepticism) was more prevalent for participants who had gotten the specific warning message. This warning message, however, appeared to lose effectiveness over time as opposed to the warning about a different type of fraud.

Burke et al. (2022) showed that brief online educational interventions with video or text may reduce fraud susceptibility. The study invited 2600 panelists (selected from a representative U.S. internet panel) to randomly receive either (i) a short video warning about investment fraud, (ii) a short text warning about investment fraud, or (iii) no educational information. Subsequently, participants were presented and asked to consider three investment opportunities, two of which exhibited signs of fraud. Results showed that intervention recipients were significantly less likely to express interest in the fraudulent investment schemes. Furthermore, a follow-up experiment showed that while the effect decayed over time, it persisted at least 6 months after the initial intervention if a reminder was provided after 3 months.

Smith and Akman (2008) examined a comprehensive New Zealand and Australian campaign to raise awareness about consumer fraud. The month-long campaign involved radio and TV appearances and the dissemination of 2600 printed posters and 282,000 flyers. The study concluded that the campaign was effective in raising public awareness about fraud and increased reporting. However, evaluation in terms of preventing and reducing fraud was limited.

Looking at the literature, there are several reasons why we might suspect fraud awareness campaigns to have a limited effect. Firstly, if the behavior targeted by fraud awareness campaigns, by its very nature, is strongly influenced by situational and emotional factors, it might be unreasonable to expect that potential victims remember campaigns when they are contacted by a fraudster. Secondly, as argued by Cross and Kelly (2016), awareness campaigns may overwhelm clients with details. Thirdly, there is no guarantee that clients will even notice an awareness campaign. In this sense, awareness campaigns are more “passive” than direct educational interventions as considered, for example, by Burke et al. (2022). Finally, we note that banks may have other reasons, besides combating financial fraud, to run awareness campaigns. Campaigns may, for instance, serve as an opportunity to strengthen client relationships (Hoffmann & Birnbrich, 2012).

Institutional Setting

Spar Nord (henceforth denoted as “the bank’’), is a systemically important Danish bank providing both retail and wholesale services. The bank is headquartered in Northern Jutland and has approximately 60 branches across Denmark. It estimates its own market share to equal about 5% of the Danish market (Spar Nord, 2023); a market characterized by a high degree of digitization (Danmarks Nationalbank, 2022). In particular, (Statistics Denmark, 2023) estimates that 94% of Danes used mobile or online banking in 2022.

In late September 2020, the bank sent out a warning message through its online banking portal. An English translation of the message is given in Fig. 1 (the original message, in Danish, is given in Appendix A). The message warned clients to safeguard their NemID, a Danish government, and online banking log-on solution (see Medaglia et al. (2017) and the Danish Agency for Digital Government (2023) for more information). The message also warned about fraud in a broad sense, asking clients not to share their personal information over the phone. We note that the message had a focus on criminals contacting clients. In this regard, some of the fraud types considered in our paper may differ from the rest, motivating robustness checks (see the “Data” section and Appendix G). We also note that the message was short and, in line with suggestions by Cross and Kelly (2016), did not overwhelm clients with details. At the same time, however, the message did not warn about particular types of fraud, a strategy that may be more effective as shown by Scheibe et al. (2014). The bank targeted the message to clients over 40 years of age. However, we note some treatment non-compliance, i.e., some clients under 40 received it while some over 40 did not (see Fig. 3 and the “Explorative Analysis” section). We are not aware of other warning messages being sent out to the bank’s clients during the data collection period.

Fig. 1
figure 1

Warning message sent by the bank in late September 2020. NemID (literally translating to “EasyID”) is a Danish government and online banking log-on solution; see Medaglia et al. (2017) and the Danish Agency for Digital Government (2023) for more information

Following Gotelaere and Paoli (2022), we define financial fraud as “intentionally deceiving someone in order to gain an unfair or illegal (financial) advantage.” The definition excludes robberies (where threats or physical violence are used). We do, however, allow credential theft (e.g., stealing and misusing a password) to be encompassed by the definition, although victims may not be deceived in a direct sense.

Data

Our study employs choice-based data collection (i.e., endogenous stratification); we collect data on all known fraud victims at the bank and a sample of “baseline” clients (i.e., clients assumed not to have been defrauded). The strategy is motivated by two factors: (i) fraud victims are rare and (ii) some of our features are encoded by hand and thus expensive to collect. We only consider private individuals (i.e., not firms or businesses) over 18 years of age. In total, we have 1447 observations, including:

  1. 1.

    247 fraud victims, and

  2. 2.

    1200 baseline clients.

As fraud victims, we consider clients with a case registered in the database of the bank’s fraud department 6 months before and up to 6 months after the warning message was sent (denoted as our data collection period). If a client was subject to multiple fraud cases in the data collection period, we only consider the case that was registered first; yielding a one-to-one relationship between fraud cases and (unique) victims. This choice is motivated by the observation that it can be difficult to accurately determine if the fraud in a case is new or a continuation of previous fraud. Notably, the choice affects less than five clients in our data. However, it means our study does not account for revictimization. We record each case (along with victim features) relative to the date where the first fraudulent transaction in the case occurred, denoted as our time of (feature) recording.Footnote 1 This may be a few days before a case is opened.Footnote 2 We also restrict ourselves to cases where the first fraudulent transaction lies in the data collection period.Footnote 3 Motivated by a desire to discard petty fraud, we only consider cases involving more than 1000 Danish Kroner (approximately €135 or $150). In some cases, the bank successfully recovers funds (immediately) after a fraud is discovered. By our definition, however, fraud has already taken place. Table 1 lists the different types of fraud in our data. Cases of non-delivery fraud may differ from the rest, as these (often) are client initialized (imagine, e.g., a client that finds a “too-good-to-be-true” offer online). This motivates us to do a robustness check (see the “Results” section and Appendix G).

Table 1 Types of fraud in our data (247 fraud cases in total). We stress that the bank internally uses another, more detailed, typology. If a fraud case can be described by multiple types of fraud, we assign it to the most specific (listed in decreasing order)

Our baseline clients are sampled as follows. We consider all clients with an active account at any point in time 6 months before and up to 6 months after the warning message was sent.Footnote 4 For each client, we sample a random date in the data collection period (used for feature recording) where the client had an active account. The procedure is motivated by a desire to mimic how our fraud victims are collected over a time period. We further remove any client with a fraud case and draw a random sample of 1200 clients. Any of these could, in principle, be a fraud victim; they may have failed to report or notice a fraud. Consider, for example, an embarrassed victim of romance fraud. We do, however, argue that victims have strong incentives to report frauds, hoping to recover funds. Furthermore, fraud is relatively rare (although the probability that a client is mislabeled naturally increases with the size of our sample).

Table 2 Levels in the qualifications framework of the Danish Ministry of Higher Education and Science (2021), education examples, and corresponding encoding of \(\textit{Educ}\)

Features

For any bank client i (including both fraud victims and baseline clients), we record five main features:

  • \(\textit{Age}_{i}\); client i’s age in years (i.e., number of days lived divided by 365.25),

  • \(\textit{Female}_{i}\); equal to 1 if client i is female (otherwise 0),

  • \(\textit{Partner}_{i}\); equal to 1 if client i is (i) married, (ii) in a registered civil partnership, or (iii) shares a bank account with a life partner (otherwise 0),

  • \(\textit{Funds}_{i}\); summed investment and pension funds held by client i at the bank (before a potential fraud case; measured in (i.e., divided by one) million Danish Kroner), and

  • \(\textit{Urban}_{i}\); equal to 1 if client i resides in a major Danish municipality; Copenhagen, Aarhus, Odense, or Aalborg (otherwise 0).

Our \(\textit{Partner}\) feature is encoded by hand, using manual look-ups in multiple banking systems. A client failing all three listed criteria is assumed to be single. This includes people in relationships unknown to the bank (e.g., people married abroad).

Table 3 Feature statistics over all observations. In agreement with the data-providing bank, some statistics are kept confidential. The features Warn and Time are defined in the “Methodology” section. “Std.” is used as an abbreviation for ’standard’
Table 4 Feature correlations over all observations. Information about financial funds is kept confidential in agreement with the data-providing bank. The features Warn and Time are defined in the “Methodology” section

For 669 observations, we also have educational information. Whenever we use the information in a model, we drop observations where it is missing. The information is encoded in a feature:

  • \(\textit{Educ}_{i}\); denoting client i’s educational level on an ordinal scale \(\{0,1,2\}\).

The feature is encoded by hand, using free-text descriptions of occupational and educational histories, primarily written by bank advisors. Encoding is done using the Qualifications Framework of the Danish Ministry of Higher Education and Science (2021), placing educations on ordinal levels 1–8 (see Table 2). We apply a grouping such that levels 1–3 (anything less than a high school degree) yield \(\textit{Educ}=0\), levels 4–5 (including a high-school degree) yield \(\textit{Educ}=1\), and levels 6–8 (including a university degree) yield \(\textit{Educ}=2\). A detailed description of the encoding process is provided in Appendix B. Recognizing that it might not be meaningful to run a regression directly on (nor take the mean of) an ordinal feature, we one-hot encode \(\textit{Educ}\) in all our regression models. Using standard indicator functions, we let

$$\begin{aligned} \textit{Educ}0_{i}=\mathbbm {1}_{\{\textit{Educ}_{i}=0\}},\text { } \textit{Educ}1_{i}=\mathbbm {1}_{\{\textit{Educ}_{i}=1\}}, \text { and } \textit{Educ}2_{i}=\mathbbm {1}_{\{\textit{Educ}_{i}=2\}}, \end{aligned}$$
(1)

dropping the former (i.e., \(\textit{Educ0}\)) in all models to avoid multicollinearity.

We provide statistics on our full data set in Tables 3 and 4. Statistics on observations recorded after the warning message was sent (relevant to our regression discontinuity setup) can be found in Appendix C.

Fig. 2
figure 2

Fraction of fraud victims relative to all observations (under and over 40 at the time of the warning message) per month. Bins are made per running month. For instance, the bin between −1 and 0 contains observations recorded 1 month before and up until the warning message was sent

Fig. 3
figure 3

Client age (when the warning message was sent) and whether or not the client received the warning message

Explorative Analysis

Figure 2 illustrates the fraction of fraud victims recorded in our data set per month relative to when the warning message was sent. Prior to the message, clients below and above 40 years of age followed roughly similar trends. Thus, we believe the figure supports that a difference-in-differences setup is reasonable. After the warning message, we see a disproportionately large jump in the proportion of fraud victims over 40, suggesting that the warning message did not reduce fraud susceptibility.Footnote 5

Fig. 4
figure 4

Client age (at the time of feature recording) and observation type

Figure 3 illustrates the relationship between client age and whether a client received the warning message. Non-compliant clients amount to approximately 8.9% of all observations. If we only consider clients recorded after the message was sent (relevant to our regression discontinuity setup; see Fig. 6), non-compliant clients constitute approximately 9.0%. We believe that the observed non-compliance is due to several factors: (i) wrong database entries, (ii) clients not being registered in the bank’s messaging system, and (iii) the fact that some clients may only have opened a bank account after the message was sent. Figure 4 further illustrates the relationship between client age and observation type at the time of feature recording.

In preparation for our regression discontinuity setup, we follow recommendations by Imbens and Lemieux (2008) and provide a number of statistics and plots in Appendices C and D, only considering observations recorded after the warning message was sent. Based on these, we have no concerns about discontinuity at the treatment limit (i.e., being 40 years old).Footnote 6

In agreement with the data-providing bank, we keep the distribution of losses associated with fraud cases confidential. However, we note that neither a standard Student’s nor a Welch’s t-test indicates a difference in the average loss associated with fraud cases recorded before and after the warning message was sent (regardless of whether one considers (i) all fraud cases, (ii) cases associated with clients over 40 years of age when the warning message was sent, or (iii) cases associated with clients that actually received/would receive the warning message).Footnote 7

Methodology

We employ two types of linear probability models, relying on (i) difference-in-differences and (ii) regression discontinuity designs. All models are implemented in Python with the statsmodels module (Seabold & Josef Perktold, 2010), use weighting as described in the “Weighting” section, and employ robust standard errors.Footnote 8

Difference-in-differences

A difference-in-differences design compares the outcomes of two groups over time; one group receiving a treatment and one group not receiving it. Assuming that outcomes in the two groups, absent any treatment, would have followed parallel trends, any difference over time between the relative group differences can be ascribed to the treatment. The approach has been heavily shaped by works within labor economics (see, e.g., works by Card and Krueger (1994); Ashenfelter (1978); Ashenfelter and Card (1985)). For a general introduction, we refer to Angrist and Krueger (1999).

Let \(y_{i}\in \{0,1\}\) denote if client i is a fraud victim (with \(y_{i}=1\)) or baseline client (with \(y_{i}=0\)). Our basic difference-in-differences model is given by

$$\begin{aligned} y_{i}=\beta _{0}+\beta _{1}{} \textit{Warn}_{i}+\beta _{2}{} \textit{Time}_{i}+\beta _{3}{} \textit{Warn}_{i}\times \textit{Time}_{i}, \end{aligned}$$
(2)

where \(\beta _{0}\in \mathbb {R}\) is an intercept, \(\beta _{1},\beta _{2},\beta _{3}\in \mathbb {R}\) are regression coefficients, and

  • \(\textit{Time}_{i}\in \{0,1\}\) denotes if client i is observed before (\(\textit{Time}_{i}=0\)) or after (\(\textit{Time}_{i}=1\)) the warning message was sent, and

  • \(\textit{Warn}_{i}\in \{0,1\}\) denotes if client i is in the group that received the warning message (\({\textit{Warn}_{i}=1}\)) or not (\({\textit{Warn}_{i}=0}\)). Notably, we record \(\textit{Warn}\) without regard for the time of feature recording (see the “Data” section).

We progressively add control variables (i.e., features from the “Features” section) to the model in Eq. 2. In connection to this, a correlation of 0.66 between \(Age_{i}\) and \(Warn_{i}\) might lead to worries about multicollinearity. To address concerns, we report variance inflation factors in Appendix E (see James et al. (2021) for an introduction). Disregarding interaction terms (and features directly involved in interaction terms), our variance inflation factors indicate that multicollinearity is not a cause of concern for any of our models. We note that models employing our \(\textit{Educ}\) feature (with the one-hot encoding given in the “Features” section) use fewer observations than our other models (as we drop observations with missing data). Our primary interest is always the treatment effect on the treated group, i.e., the coefficient associated with \(\textit{Warn}\times \textit{Time}\).

Regression Discontinuity

A regression discontinuity design employs the idea that observations within some bandwidth of a threshold are similar; except that those above it receive a treatment while those below do not. The approach has been heavily shaped by works on education (see, e.g., Thistlethwaite and Campbell (1960); Angrist and Krueger (1991); Angrist and Lavy (1999)). For a general introduction, we refer to Imbens and Lemieux (2008).

Our regression discontinuity design is heavily inspired by Imbens and Lemieux (2008) and employs local linear estimation without a kernel (but uses weighting to account for undersampling). We discard all observations recorded before the warning message was sent. Furthermore, we only consider clients within a bandwidth \(h^{*}\) of being 40 years old (see the “Bandwidth Selection” section). Due to non-compliance (see Fig. 6), we employ a fuzzy design (Roberts & Whited, 2013). Consider client i’s age when the warning message was sent and subtract the message threshold (i.e., 40). Denote the result as \(\textit{Chron}_{i}\).Footnote 9 Recall that \(\textit{Warn}_{i}\in \{0,1\}\) denotes if client i received the warning message. We first fit

$$\begin{aligned} \textit{Warn}_{i} = \mu _{0}+\mu _{1} \textit{Chron}_{i}+\mu _{2}\mathbbm {1}_{\{\textit{Chron}_{i}\ge 0\}}+\mu _{3}{} \textit{Chron}_{i}\mathbbm {1}_{\{\textit{Chron}_{i}\ge 0\}}, \end{aligned}$$
(3)

where \(\mu _{0}\in \mathbb {R}\) is an intercept, \(\mu _{1},\mu _{2},\mu _{3}\in \mathbb {R}\) are regression coefficients, and \(\mathbbm {1}_{\{\textit{Chron}_{i}\ge 0\}}\) is a standard indicator function. Recall, furthermore, that \(y_{i}\in \{0,1\}\) denotes if client i is a fraud victim. We secondly fit

$$\begin{aligned} \textit{y}_{i} = \rho _{0}+\rho _{1} \textit{Chron}_{i}+\rho _{2}{} \textit{Chron}_{i}\times \textit{Warn}_{i}+\rho _{3}{} \textit{WarnPred}_{i}, \end{aligned}$$
(4)

where \(\rho _{0}\in \mathbb {R}\) is an intercept, \(\rho _{1},\rho _{2},\rho _{3}\in \mathbb {R}\) are regression coefficients, and \(\textit{WarnPred}_{i}\) is a prediction calculated from the fitted model in Eq. 3.

As in our difference-in-differences setup, we progressively add control variables to both Eqs. 3 and 4. This includes our \(\textit{Educ}\) feature, though it means using fewer observations (as we drop observations with missing data). In all specifications, our primary interest is the local average treatment effect (at 40 years of age), i.e., the coefficient associated with \(\textit{WarnPred}\).

Bandwidth Selection

To select a bandwidth for our regression discontinuity models, we employ a slightly modified version of the cross-validation procedure by Imbens and Lemieux (2008). Let \(\textit{WarnPred}_{i}\) and \(\hat{y}_{i}\) denote predictions from Eqs. 3 and 4, respectively. Our procedure considers symmetric bandwidths \(h=11,\dots ,20\),Footnote 10 uses \(K=10\) fold cross validation, and goes as follows:

  1. 1.

    consider clients with \(-h\le \textit{Chron}_{i}\le h\) (i.e., within h years of being 40 years old),

  2. 2.

    construct K folds of the data,

  3. 3.

    for each fold, \(k=1,\dots ,K\),

    1. 1.

      estimate Eqs. 3 and 4 using all observations but those in fold k,

    2. 2.

      using observations \(i=1,\dots ,n_{k}\) in fold k, calculate the sums of squared errors,

      $$\begin{aligned} CV_{W}(h,k)=\frac{1}{n_{k}}\sum _{i=1}^{n_{k}}(\textit{Warn}_{i}-{\textit{WarnPred}_{i}})^{2} \text { and } CV_{Y}(h,k)=\frac{1}{n_{k}}\sum _{i=1}^{n_{k}}(y_{i}-\hat{y}_{i})^{2}, \end{aligned}$$
      (5)
  4. 3.

    finally, calculate the cross validation estimated sums of squared errors associated with h,

    $$\begin{aligned} CV_{W}(h)=\frac{1}{K}\sum _{k=1}^{K}CV_{W}(h,k) \text { and } CV_{Y}(h)=\frac{1}{K}\sum _{k=1}^{K}CV_{Y}(h,k). \end{aligned}$$
    (6)

We select the smallest bandwidth \(h^{*}\) implied by the cross validation procedure, i.e.,

$$\begin{aligned} h^{*}=\min \left\{ \underset{h}{\arg \min }\ \text { } CV_{W}(h), \underset{h}{\arg \min }\ \text { } CV_{Y}(h)\right\} . \end{aligned}$$
(7)

Results from our procedure yield an optimal bandwidth \( h^{*}=17\). For simplicity, we do not change the bandwidth as we add control variables. In all tables with model results, we report the number of clients used to run our analysis (i.e., the number of clients falling within the bandwidth).

Weighting

We use weighting to account for the fact our baseline clients are undersampled. The approach is inspired by a maximum-likelihood interpretation of linear regression; we weigh each observation inversely proportional to the probability that it should be in our sample. To be specific, we weigh each baseline observation by

$$\begin{aligned} \omega _{0} = \left( \frac{n-f}{N-f}\right) ^{-1}, \end{aligned}$$
(8)

where n denotes the total number of observations in a sample, f denotes the number of fraud cases in said sample, and N denotes our population size (i.e., the total number of clients at the bank).Footnote 11 All fraud observations are simply weighted by \(w_{1}=1\).

We use the weighting scheme in all our regression models, i.e., both in Eqs. 2, 3, and 4 (including when we calculate our regression discontinuity bandwidth \(h^{*}\), as per “BandwidthSelection” section). Notably, both n and f will change depending on the model specification we consider (i.e., how many observations are available).

Table 5 Weighted difference-in-differences coefficient estimates (standard errors in parenthesis) and total number of observations (n) and fraud cases (f) used to fit each model. We use * to denote confidence at the 0.05 p-level and ** to denote confidence at the 0.01 p-level. Coefficients of primary interest are emphasized with bold font
Table 6 Weighted regression discontinuity coefficient estimates (standard errors in parenthesis) and total number of observations (n) and fraud cases (f) used to fit each model. An optimal bandwidth \(h^{*}=17\) is used. We use * to denote confidence at the 0.05 p-level and ** to denote confidence at the 0.01 p-level. Coefficients of primary interest are emphasized with bold font

Results

Results from our difference-in-differences setup are given in Table 5. As a robustness check, we also run all models on observations recorded just 3 months before and 3 months after the warning message was sent. As one might expect (see the “Related Literature on FinancialFraud” section), age generally appears to increase fraud susceptibility. On the other hand, having a partner generally appears to decrease fraud susceptibility. The latter result holds for all our difference-in-differences models except Model 14, incorporating all features and considering data recorded only 3 months before and after the warning message was sent (with a corresponding reduction in the number of employed observations). The result might indicate that people with a partner talk to them before falling victim to fraud (though more research is needed to support this). In relation to our treatment of interest, i.e., the bank’s warning message, all models display insignificant treatment effects on the treated group (i.e., insignificant \(\textit{Warn}\times \textit{Time}\) coefficients). Disregarding interaction terms (and features directly involved in interaction terms), our variance inflation factors indicate that multicollinearity is not a cause of concern for any of our models (see Appendix E). Results using unweighted models (all showing insignificant treatment effects) are included in Appendix F.

Table 6 displays results from our regression discontinuity setup. In all models, the local average treatment effect associated with the warning message (i.e., the coefficient associated with \(\textit{WarnPred}\)) is insignificant. Disregarding interaction terms (and features directly involved in interaction terms), our variance inflation factors indicate that multicollinearity is not a cause of concern for any of our models (see Appendix E). Results using unweighted models (all showing insignificant treatment effects) are included in Appendix F.

As argued in the “Data” section, non-delivery and romance fraud may differ from the other types of fraud in our data. As a robustness check, we run our models on data where (i) all victims of non-delivery fraud and (ii) all victims of non-delivery and romance fraud are dropped (see Appendix G). All models, still, show insignificant treatment effects.

Discussion and Conclusion

We consider a fraud awareness campaign by a systemically important Danish bank, targeting clients over 40 years of age with a mass warning message. To evaluate the campaign’s effect, we use two different quasi-experimental designs. Our difference-in-differences design compares clients before and after the warning message was sent (assuming parallel trends over time). Our regression discontinuity design compares clients within a bandwidth of being 40 years old (assuming that these are comparable). Our study employs data on all known fraud victims at the bank and a sample of “baseline” clients (i.e., clients assumed not to have been defrauded). We consider a multitude of models, controlling for age, sex, relationship status, financial funds, residence, and (for a subset of our data) education. No model finds any evidence that the message had a significant effect. Robustness checks, excluding non-delivery and romance fraud (as these might differ from other types of fraud), also fail to find any significant effect.

Our results show that the considered campaign, relying on mass messaging, had little effect in terms of reducing financial fraud. This can be due to the particular message or how it was delivered. The message is, however, in line with recommendations from Cross and Kelly (2016), arguing that warnings should not be too specific or overwhelm clients with details. At the same time, though, the message did not warn about particular types of fraud, a strategy that may be more effective as shown by Scheibe et al. (2014). Our results might prompt one to question whether mass messaging campaigns are worthwhile endeavors for banks and public authorities. Moreover, skepticism about other types of awareness campaigns (e.g., TV spots) might be warranted. Limitations of our study include an absence of information regarding whether clients opened and read the warning message. This, however, illustrates a fundamental challenge of fraud awareness campaigns (as opposed to direct interventions): it is impossible to guarantee that people will pay attention to, or even notice, awareness campaigns. It might also explain why our results contrast with those showing that direct counseling or educational interventions can reduce fraud susceptibility, such as AARP Foundation (2003); Scheibe et al. (2014); Burke et al. (2022). We stress that our study does not take revictimization into account. Furthermore, we note that the warning message in our study was sent in late September 2020, resulting in data collection spanning from March 2020 to March 2021 during the COVID-19 pandemic. This might influence our results, though Denmark saw varying degrees of restrictions both before and after the warning message was sent.