1 Introduction

Recommender systems (RS) are automated decision support tools designed to provide custom-made advice on items to facilitate people’s navigation in large product or information spaces. RS offer personalized suggestions based on presumed needs and preferences of a user and other people’s behavior. They help people overcome information overload, either by providing recommendations on request, or by delivering them proactively (Jannach et al. 2010; Ricci et al. 2015). The accuracy of predictions about users’ needs and preferences is a fundamental principle in RS research (Jannach et al. 2012). In RS based on the collaborative filtering (CF) paradigm, for instance, the accuracy of recommendations derives from behavioral patterns previously disclosed by all users. A potential side-effect of accuracy as a quality metric is that the RS may recommend items too similar to what the user liked before. Such recommendations may no longer satisfy the user (Konstan and Riedl 2012; McNee et al. 2006). To deal with this issue, scholars explore other quality metrics that tap into the novelty, diversity, and unexpectedness of recommended items. Notwithstanding this, users generally appreciate recommendations based on accurate user profiles, whereas they don’t like recommendations based on impersonal algorithms. Accuracy as a quality metric, therefore, remains a necessary condition; if the RS is not accurate, the user will not trust it (Jannach et al. 2016). Precisely this element of trustworthiness potentially turns accurate RS into powerful tools to deal with people’s concerns with information privacy during navigation in large product or information spaces. Many online customers mistrust service providers, and fear that commercial organizations will disclose personal information without their explicit consent. Already before the advent of the smartphone, for instance, 72% of the people indicated to be afraid that external parties recorded and profiled their online behavior (Smith et al. 2011), while 85% of the people expressed the desire to obtain more control over the ways in which firms used their personal data online (Bélanger and Crossler 2011). With the emergence of large-scale business intelligence and data analytics aimed at mobile devices, the information privacy concerns of the public at large may even further increase (Chen et al. 2012). Within the RS community, scholars increasingly take into consideration personality traits when developing user-centered RS – among others, to tailor the technology to cognitive, affective and other differences that influence user interaction with RS (Tkalcic and Chen 2015). In this work, we bring together three important factors for the design of recommendation systems: accuracy of recommendations, information privacy concerns, and personality traits. In our study, we invited participants to work on several travel planning tasks. Participants had to download these tasks via a mobile application, and to execute them on their own smartphone. Results of our experimental study show that all of the three factors positively influenced user engagement.

2 Theory

2.1 Information privacy research and proactive recommendation delivery

Information privacy refers to “the ability of the individual to personally control information about one’s self” (Smith et al. 1996). Advances in information and communication technology, especially regarding the increasing pervasiveness of Web 2.0 and Web 3.0 technologies (Chen et al. 2012), have fueled a debate about the ethical aspects of information privacy (Westin 2003). People wish to be connected with external information of outstanding quality provided by external parties, yet aspire to be or remain in control of their personal information (Bélanger and Crossler 2011). The normative aspects of information privacy have been discussed in literature reviews: like physical privacy, information privacy is often regarded a fundamental human right (i.e., people ought to be in control over their personal information) (Smith et al. 2011; Bélanger and Crossler 2011).

There is a tension between the fundamental human right to be in control over personal information and the commercial trend towards personalization of product and service offerings. Personalization refers to “the ability to provide content and services tailored to individuals based on knowledge about their preferences and behaviors” (Adomavicius and Tuzhilin 2005). It is a crucial factor in the development and proper execution of a vendor’s Customer Relationship Management (CRM) (Xu et al. 2011). In e-commerce transactions, customers often make a privacy calculus to determine, if they perceive the tradeoff between keeping or sharing their personal data with a vendor and what they would receive in return, as (un)fair. Intriguingly, customers are more willing to give away personal information when it feels convenient to them, and when they trust the vendor (Dinev and Hart 2006; Chellappa and Sin 2005). This so-called privacy-personalization paradox has been widely documented in the literature (Smith et al. 2011; Norberg et al. 2007; Dinev and Hart 2006; Chellappa and Sin 2005; Li et al. 2011; Xu et al. 2012; Ho and Tam 2006; Awad and Krishnan 2006).

A substantial part of the research into information privacy focuses on individual differences in people’s concerns regarding the disclosure of personal information online (Bélanger and Crossler 2011; Li et al. 2011). This emphasis on the individual is largely due to the success of the Concern for Information Privacy (CFIP) Instrument, a 15-item scale based on four dimensions (collection, errors, secondary use, and unauthorized access), which is usually treated as an overall proxy for privacy concerns (Smith et al. 1996; Stewart and Segars 2002). The instrument measures people’s concerns regarding the disclosure of personal online information without the need to actually have them give out any personal information. The CFIP thus measures ways in which people in general differ in their apprehension regarding information privacy, and such attitudes are independent from context and domain; see (Tkalcic and Chen 2015) for a similar point. The CFIP has become the gold standard in IS research for measurement of individual differences in privacy concerns (Smith et al. 2011; Bélanger and Crossler 2011).

More recently, the Internet Users’ Information Privacy Concerns (IUICP) Instrument was put forward as an alternative measurement instrument. Like the CFIP, the IUIPC on an aggregate level taps into overall information privacy concerns (Malhotra et al. 2004), but from the angle of social contract theory and justice theory (Tyler 1994). It specifically assesses the perceived trade-offs between the individual, who discloses his/her own online data, and the external party, who uses the online data. An outcome is perceived as fair when the amount of online data collected by the other party is regarded as just (collection), when the information exchange can be approved or opted-out at will (control), or when it is clear that the information exchange took place, and with what effect (awareness). Similar to the CFIP, the IUIPC is a context- and domain-independent self-report measure of information privacy concerns. Where the CFIP taps into corporate-centered information exchange, the IUIPC more clearly focuses on the extent to which people value perceived fairness of information exchange (Xu et al. 2012). The IUIPC explains more variance than the CFIP (Malhotra et al. 2004), see also Bélanger and Crossler (2011), Li et al. (2011), and Xu et al. (2012).

Thus far, most research on information privacy concerns has been conducted from the perspective of technology acceptance models. In this well-known paradigm, surveys are collected to link information privacy concerns to a participant’s behavioral intention to use a specific information technology (Davis 1989; Davis et al. 1989). However, results have been inconclusive as to the existence and direction of an association of information privacy concerns and behavioral intention to engage with technology. Many studies have predicted and found a reluctance among people high on such concerns to use information technology (Dinev and Hart 2006; Malhotra et al. 2004; Li et al. 2011; Xu et al. 2012), yet other studies have failed to replicate this negative effect (Son and Kim 2008); for a recent review, see Kokolakis (2017). For instance, a study that explored people’s behavioral intention to engage on the social networking site Facebook failed to establish evidence for a negative effect of high levels of information privacy concerns on app engagement (Wisniewski et al. 2015). Under reference to King et al. (2011), the authors reasoned that most participants in their student sample probably did not know how mobile applications work; see also Kokolakis (2017).

Unfortunately, conclusions in the literature on technology acceptance models derive from self-reported behavioral intentions, not from measurement of actual user behavior (i.e., log data). Intriguingly, the limited number of studies that did collect log data, showed that users high on information privacy concerns, when downloading and engaging with a mobile application, in reality do not care about privacy issues (Buck et al. 2014; Kelley et al. 2013). Actual engagement with an IS system captures many user activities (such as searching and browsing, text writing and editing, information storage and retrieval, and engagement duration). Not all those activities appear equally troublesome from an information privacy point of view (Kokolakis 2017). Interestingly, Wisniewski et al. (2015) propose that a transparent flow of information on mobile applications may matter more for the engagement of users high on information privacy concerns. This observation is relevant to the present discussion: proactive RS may be effective, precisely because the user acknowledges the link between disclosed information and the reception of accurate recommendations. The transparency between the two will lead the user to trust and appreciate the recommendations received (Jannach et al. 2016). RS with maximum accuracy in everyday practice not always make perfect recommendations (Knijnenburg et al. 2010; Konstan and Riedl 2012; McNee et al. 2006; Jannach et al. 2016), but it nonetheless makes sense that, in some RS scenarios, a high accuracy of recommendations will lead to higher user engagement, also for people highly concerned with their privacy. Like (Wisniewski et al. 2015), this position is grounded in the literature on the role of information transparency in information exchange (Awad and Krishnan 2006).

Based on the above, we posit that the high (vs. low) information privacy concerns of a user will positively impact user engagement with a mobile application. Moreover, under reference to Awad and Krishnan (2006) and Wisniewski et al. (2015), we predict that this effect will depend on the high (vs. low) accuracy of proactively delivered recommendations. The added value of accurate RS will be so obvious to the user that also those with enhanced information privacy concerns will stop worrying, and engage in the RS. We thus hypothesize:

  • Hypothesis 1a: The user’s high (vs. low) information privacy concerns will positively influence user engagement.

  • Hypothesis 1b: The effect of high (vs. low) recommendation accuracy on user engagement is moderated by a user’s high (vs. low) information privacy concerns.

2.2 The five factor model of personality and proactive recommendation delivery

Scholars in personality research explore individual differences in personality and ability. They focus on the relatively enduring ways, in which individuals differ from each other regarding the ways in which they think, feel, and act. Some personality psychologists develop measurement instruments to tap into theoretically grounded aspects of personality (Mischel et al. 2007). In recent years, the domain of personality psychology has come under the attention of many other academic disciplines, including those working on information and communication technologies. Among others, the measurement of individual differences in trait personality holds promise for the development of user-centered RS (Tkalcic and Chen 2015).

One of the most widely used personality theories is the five-factor model of personality, also known as the Big Five (McCrae and Costa 1987). This model assumes that a person’s personality can be broken down into five hierarchical dimensions that remain stable across a lifespan. Each dimension has a psychobiological grounding, and is independent from people’s age, gender, ethnicity, and social economic status (Costa and MacCrae 1992). The five factors in the model are: Openness to Experience, Conscientiousness, Neuroticism, Extraversion, and Agreeableness. Each of these overarching Big Five personality traits is composed of more specific traits. For instance, trait Openness to Experience broadly captures the extent to which someone is original, has broad interests, and is open to novel experiences and ideas. Specifically, someone high on Openness is creative, curious, imaginative and unconventional (or rather conventional, cautious). Likewise, trait Conscientiousness describes the extent to which someone is precise, careful and reliable (or rather imprecise, careless, and undependable), while trait Neuroticism captures a person’s tendency to experience negative emotional states and distress (or is rather calm and even-tempered) (McCrae and John 1992).

The personality traits Agreeableness and Extraversion are different from the other three dimensions, as they represent the social dimensions of the five-factor model. Agreeableness taps into the tendency to get along with other people; it specifically captures the extent to which someone is altruistic, caring, and emotionally supportive (or rather indifferent, self-centered, and hostile). People high on Agreeableness are often regarded as team players, as they display high levels of trust and cooperation in their social interactions (Bell 2007; Driskell et al. 2006). Extraversion, finally, is a person’s tendency to experience positive emotional states and to feel good about oneself and others. It specifically taps into the extent to which someone is outgoing, cheerful, and warm (or rather quiet, timid, and withdrawn) (Costa and MacCrae 1992). Research in offline (i.e., social, work, and entrepreneurial) settings has found that people high on Extraversion more readily engage in social interactions with other people than those low on Extraversion (Bell 2007; Driskell et al. 2006; Zhao and Seibert 2006).

In RS research, scholars have become interested in the role of Big Five personality traits in user engagement. As summarized in Tkalcic and Chen (2015), the research activities roughly break down into: (1) the implicit extraction of Big Five personality traits from user data, and (2) the explicit measurement of Big Five personality traits via questionnaires. Overall, however, the evidence for, and direction of, effects of Big Five personality traits on engagement with RS is inconclusive. Regarding trait Extraversion, for instance, early studies reported negative effects of trait Extraversion on engagement, under the assumption that people high on Extraversion would favor offline social interaction over online social interaction (Amichai-Hamburger et al. 2002). The negative effect of Extraversion on user engagement is sometimes replicated (Karumur et al. 2016), but more often contradicted. Among others, people engaging in online games tend to be higher on trait Extraversion (Teng 2008). On Facebook, users high on trait Extraversion interact more with others, share more content and objects, engage in group membership more often, and significantly click more on ’likes’ (Bachrach et al. 2012). Likewise, Ross et al. (2009) already identified that people high on Extraversion engage more heavily in group membership on Facebook, even though they did not find an impact on the number of clicks, number of Facebook friends, or time spent online. Also from a NeuroIS perspective, evidence has been generated for a causal relationship between trait Extraversion and information search behavior via an IS platform. Participants, who had filled out an extensive Big Five questionnaire first, went through a sequence of information-seeking tasks. With the help of an eye-tracker, it was found that people high on Extraversion displayed heightened engagement on a series of information search tasks (Al-Samarraie et al. 2017).

Interestingly, scholars have expressed interest in the behavioral underpinnings of information privacy concerns (Bélanger and Crossler 2011). It has, specifically, been observed that “[o]ther individual differences could be studied, such as the effects of […] personality traits […] on information privacy concerns. There is also a need to study the moderating effects these individual differences could have on the linkages between information privacy and dependent variables like e-business adoption” (Bélanger and Crossler 2011). Along such lines, a single correlational study tentatively explored the relation between Big Five personality traits and individual differences in privacy concerns. This study reported that individuals with enhanced self-reported Extraversion, due to their social orientation, were more concerned about the wellbeing and relationship with other people than about what happens to their personal information — that is, as long as sharing or disclosing personal information was thought to facilitate social interaction (Junglas et al. 2008). It should be noted, however, that these conclusions were merely based on associations, not on causality testing. Also, measurements on behavioral intention to use a technology, or actual user interaction with RS were not included. Still, the study points to the possibility that trait Extraversion and information privacy concerns may have an impact on user engagement in interaction.

Building on these loose threads of evidence, we maintain that Extraversion will moderate the effects of recommendation accuracy on user engagement. That is, a user high on Extraversion will display more user engagement with the RS out of a wish to make a positive contribution, whereas a user low on Extraversion will do so only when engagement seems to make sense (i.e., when the RS provides highly accurate recommendations). Moreover, under reference to Junglas et al. (2008), we posit that this two-way interaction effect will be further moderated by a user’s high (vs. low) self-reported information privacy concerns. We thus hypothesize:

  • Hypothesis 2a: The user’s high (vs. low) Extraversion will positively influence the effect of low (vs. high) recommendation accuracy on user engagement.

  • Hypothesis 2b: The user’s high (vs. low) information privacy concerns will moderate the interaction between high (vs. low) Extraversion and high (vs. low) recommendation accuracy on user engagement.

3 Method

3.1 Participants and design

We randomly assigned participants from the business school of a Dutch university to the experimental conditions of a Recommendation Accuracy (high, low) factorial design on User Engagement, to which self-reported scores for Information Privacy Concerns and personality traits (i.e., Extraversion) were added as covariates. The initial sample consisted of 156 undergraduate students (87 men and 69 women; Mage = 21.17 years, SD = 1.49) enrolled in an applied statistics course. Participation was on a voluntary basis (a case assignment in the absence of monetary or course-related rewards). It was made clear from the outset that the generated data would be used for educational purposes to practice statistical techniques in aggregated, anonymized, form. The data of participants (N = 25; 16.03%) who reportedly did not receive proactive recommendations while working on the experiment (see below) as well as the data of participants with missing values (N = 4; 2.56%) were omitted. As a result, the final sample consisted of the data of 127 participants (69 men and 58 women; Mage = 21.22 years, SD = 1.49). The latter sample was used for the analyses reported here.

3.2 Materials and procedure

Participants received an email in which they were invited to participate in a voluntary case assignment dedicated to “responsible e-tourism”. In this email, they were pointed to a PDF document with a detailed description made available on the course website. The email further contained the link to an online pre-survey assessing Information Privacy Concerns and personality traits (see below). Recipients were also informed how to download the smartphone application to be used in the case assignment from the Google Play or Apple App Store, depending on their mobile operating system.

The PDF document stated the research objective of the case assignment as: gaining insight into the ways in which students would use and interact with an intelligent e-tourism application on their own smartphone. To that end, the document presented the business school participants with a concise introduction into e-tourism: the practice of offering complicated, tourism-related, product-service bundles via intelligent information and communication technologies. The text discussed recommender systems in non-technical terms as intelligent software tools that proactively offer personalized suggestions how to solve a problem or need, often based on a user’s known and unknown preferences (Werthner et al. 1999; Werthner and Ricci 2004). Ethical aspects and information privacy concerns were briefly addressed in relation to the potential invasiveness of recommender systems on smart devices such as the smartphone. The Five Factor Model of personality (the “Big Five”) was introduced as one way in which researchers and practitioners nowadays derive preferences based on personality traits. The PDF document concluded that intelligent software tools, such as recommender systems, need to be designed with a keen eye on the ethical aspects involved in order to arrive at responsible e-tourism.

The participants were invited to download and work with a mobile e-tourism application that was developed specifically for the course. The application was entitled e-Tourism Challenge, and it presented participants with the following plausible – yet (at the moment of data collection) imaginary – scenario and a first challenge (labeled Task 1):

(1) Prince Harry of England has publicly announced his forthcoming wedding at Buckingham Palace, London, and set the date for …. This surely will be the event of the year, and you really wish to spend the weekend in London when this royal wedding takes place. Therefore, your task will be to find a way to travel from Delft, the Netherlands, to London, UK, on …. For Task 1 we will ask you: (1) to explore what modes of transportation (e.g., bus, train, et cetera) to select for a trip to England, and (2) how much money you are prepared to spend on this trip.

While working on Task 1 (as mapped out in the scenario above), participants had the possibility to search for information via a Search & Browse tab in the in-application browser (IAB). They could write down, store, and edit findings in a Notes section before submitting their final solutions. Thus, for the first challenge (Task 1), they could, for instance, record in writing what modes of transportation they considered for their trip to London, and within what price range they were making inquiries (see Fig. 1). Also, participants had the possibility to switch between foreground and background – for instance, to search for information or a Website in another browser app, or to engage in any other activity, instead. The second challenge (Task 2) in the application built on the first one, and was introduced as follows:

(2) When in London, you need a place to stay. Therefore, your next task will be to find a place that meets the following requirements: (1) it must be located as close as possible to Buckingham Palace, and (2) you must book it from … until …. We will ask you: (1) to report the name of the location you selected, (2) the number of stars for this location, and (3) how much money it costs you to spend the weekend at your location of choice.

Fig. 1
figure 1

Screenshot of the in-application dashboard

Again, participants could leave and edit notes in the designated section before submitting a final solution. This time, the records in writing would probably capture the type of accommodation they considered for their stay in London, and the price range, in which they were carrying out their inquests.

Importantly, while working on Task 1, but also while working on Task 2, participants started to receive proactive recommendations in the form of a series of potentially useful hyper-links, exactly 60 sec after their first exploration of the scenario. The accuracy of these proactive recommendations was put under experimental control, meaning that it was rendered technically impossible for participants to be exposed to a mixture of and/or switches between the two experimental conditions. Participants either received a series of highly relevant, or a series of not so relevant, recommendations (see Fig. 2).

Fig. 2
figure 2

Screenshots of recommendations, top left: high accuracy for Task 1, top right: low accuracy for Task 1, bottom left: high accuracy for Task 2, bottom right: low accuracy for Task 2

Once a participant had finished working on Tasks 1 and 2, she could submit results, after which the e-Tourism Challenge application was rendered dysfunctional. That is, the participant could no longer access the IAB, nor reinstall the application. Instead, participants were linked to an online post-survey assessing manipulation checks, their experiences with mobile application usage in general and with the current application in particular, and demographics. Participants were debriefed regarding the purpose of the study in writing and during the lectures that followed.

3.3 Measures

Manipulation of recommendation accuracy

The accuracy of a recommender system in online experiments is typically measured by the Precision metric (signifying the share of relevant items among all items recommended) or related rank accuracy measures such as Mean Average Precision (MAP) and Normalized Discounted Cumulative Gain (nDCG) (Jannach et al. 2012; Rossetti et al. 2016). Given that study participants were anonymous first-time users of the app the relevance of recommended items was operationalized by manually selecting items that were relevant to the (static) task context of participants (Adomavicius and Tuzhilin 2015). Manipulation was induced by means of the proactive delivery of a set of recommended hyperlinks that were either very relevant or only marginally relevant for the task at hand. Specifically, participants in the high recommendation accuracy condition received a set of very relevant hyperlinks that were directly useful for solving the challenges of mode and means of travel (in Task 1) and finding an accommodation (in Task 2). People in this experimental condition thus received very handy and to-the-point links only (very precise recommendations). In contrast, participants in the low recommendation accuracy condition received a set of hyperlinks that were too generic to solve the challenges of mode and means of travel (in Task 1) and accommodation (in Task 2). People in this experimental condition thus received links that ostensibly referred to a stay in London, but they could not really contribute to solving the challenges posed since they were not relevant (i.e. low Precision) for the participants’ task context (see Fig. 2).

Information privacy concerns

We used the original Internet Users’ Information Privacy Concerns (IUIPC) scale (Malhotra et al. 2004) to measure self-reported differences in our participant’s Information Privacy Concerns. The IUIPC construct derives from a ten-item scale, which is based on three first-order dimensions labeled Control, Collection, and Awareness. Even though this leaves room for an analysis of each of these first-order dimensions in isolation, the scale was developed to assess IUIPC as an overarching second-order construct. Table 1 provides each of the ten items in the IUIPC, and shows the results of a Principal Component Analysis. Consistent with the literature (Malhotra et al. 2004), a free Promax Rotation acknowledged the multidimensionality of the scale, while clearly confirming that all items load on a single, overarching, dimension. Each item was assessed on a 7-point scale anchored at 1 (strongly disagree) and 7 (strongly agree). The second-order scale for overall IUIPC was reliable; Cronbach’s α = .89).

Table 1 Factor loadings for principal component analysis with promax rotation (N = 127)

Extraversion

Extraversion is one of the social dimensions in the Five Factor Model of personality (or Big Five). Many assessment instruments have been developed to measure individual differences in Big Five personality traits. In the present research, we opted for the Mini-IPIP Scales (Donnellan et al. 2006) – a widely used, highly condensed and psychometrically valid version of the larger International Personality Item Pool — Five Factor Model (Goldberg 1999). Table 1 provides each of the four items in the Mini-IPIP Scales that assess trait Extraversion. Each item was measured on a 5-point scale anchored at 1 (very inaccurate) and 5 (very accurate). The overall score was computed after reverse coding of the respective items. Again, the results of a Principal Component Analysis confirmed the internal consistency of the Extraversion scale. The scale was reliable; Cronbach’s α = .77).

User engagement

It is well-documented in the literature on organizational behavior that people characterized by higher levels of task engagement typically refrain from submitting bottom-line results, but spend time and energy in handing in decent solutions to a challenge (Rich et al. 2010), also in computer-mediated settings (Ray et al. 2014). In accordance to this literature, we operationalized user engagement in the present study in two different ways:

Engagement clicks

In the IAB, participants could (re-)read the scenarios and descriptions for Tasks 1 and 2, search and browse for information within the application environment or via another search engine, and write down, store, and edit their findings. Engagement Clicks was therefore operationalized as the sum of all editing, description reading, and searching activities, including all switches between foreground and background.

Engagement duration

Engagement Duration was operationalized as the total duration of a participant’s interaction with the IAB. Engagement Duration derived from the time it took a participant (in sec) to complete Tasks 1 and 2 – thus, from the first exploration of the scenarios in the IAB until the moment in time when the solutions were submitted, and the e-Tourism Challenge was rendered dysfunctional.

Manipulation check

So as to capture the extent to which participants had perceived the accuracy of the proactively delivered recommendations in congruence with experimental conditions, we used the following four items: “The travel solutions I produced for the e-Tourism Challenge were of good quality”, “Overall, I found it helpful to receive the recommended set of links for the e-Tourism Challenge”, “I was generally pleased with the recommended set of links I got for the e-Tourism Challenge” and “The recommended set of links for the e-Tourism Challenge enabled me to submit high-quality travel solutions”, on a 7-point scale anchored at 1 (not at all true for me) and 7 (very true for me; Cronbach’s α = .83).

4 Results

4.1 Manipulation check

We conducted a linear regression analysis on the four items of the manipulation check for recommendation accuracy. The analysis yielded a significant main effect of low versus high recommendation accuracy, β = .60, t(126) = 2.67, p < .01, indicating that participants indeed had experienced the quality of the proactive recommendations they had received in congruence with their experimental condition. The manipulation therefore had been effective.

4.2 Sample characteristics, correlations and overall distributions

Table 2 shows that the majority of smartphone users in the present study worked with iOS rather than Android as their operating system. The other descriptives in the table derive from a list of covariate items that are more often collected in studies on information privacy concerns (Malhotra et al. 2004; Smith et al. 1996). We modified these items to tailor them to mobile app usage – so as to explore the familiarity of our student sample with mobile applications in real life. Our participants turned out to be seasoned app users: the vast majority of users in our sample had many years of experience in the usage of mobile applications on their smartphones, and were familiar with the phenomena of privacy invasions and information misuse that are often associated with information privacy concerns. They heard or read about it, but overall encountered such privacy invasions (very) infrequently in their own life (Awad and Krishnan 2006).

Table 2 Sample descriptions

Table 3 shows the descriptive statistics and a traditional correlation matrix for the factors in the experimental design, including the arithmetic means and standard deviations for Engagement Clicks and Engagement Duration. First, Extraversion appeared not to significantly correlate with the other variables in the study. Second, Recommendation Accuracy was significantly and positively correlated with overall IUIPC. Also, overall IUIPC was significantly and positively correlated with Engagement Clicks and Engagement Duration. Finally, Engagement Clicks and Engagement Duration were significantly and positively correlated – in the absence of any multicollinearity.

Table 3 Descriptive statistics and correlations

However, it should be emphasized that the two dependent variables in the present study, Engagement Clicks and Engagement Duration, were count variables. Count variables, among others, refer to a number of occurrences or an enumeration of events, or to series of events that occur over a period of time. They often contain zeros and low values (i.e., low arithmetic means), and tend to be skewed in terms of distribution (Cameron and Trivedi 2013; Hilbe 2014; Coxe et al. 2009). It has been argued that arithmetic means, then, may lead to biased interpretation, something that can be compensated for by plotting the observed distribution of the count variables in question (Coxe et al. 2009; Aiken et al. 2015). Figures 3a and 4a, therefore, show the overall count distributions of Engagement Clicks and Engagement Duration, clearly confirming that our dependent variables must be treated as count variables. Also provided are boxplot count distributions of Engagement Clicks and Engagement Duration, split by low and high levels of Recommendation Accuracy (see Figs. 3b and 4b).

Fig. 3
figure 3

Histogram and box plot for the engagement measure based on users’ clicks

Fig. 4
figure 4

Histogram and box plot for the engagement measure based on duration

4.3 Engagement clicks

We used the General Linear Models (GLMs) module in the SPSS statistics software package to evaluate our hypotheses for Engagement Clicks, the first dependent variable in our study. As outlined in Coxe et al. (2009) and Aiken et al. (2015), the general linear model (GLM) is a larger set of regression models specifically developed for the analysis of count data. The GLM module allows for analyzing various types of count data with standard Poisson regression and alternatives, depending on the nature and extent of skewness of the distribution. As visible in Fig. 3a, a negative binomial regression was the most appropriate analysis, due to a variance of the dependent variable Engagement Clicks larger than the mean (i.e., overdispersion; Coxe et al. 2009). For our analyses, we ran a negative binomial regression, with an adjusted overdispersion parameter ϕ = .17, SE = .05, 95% CI [0.35, 0.55].

Following the procedure outlined in Aiken et al. (1991), Coxe et al. (2009), and Aiken et al. (2015), we dummy-coded our experimental factor, Recommendation Accuracy, into – 0.5 and + 0.5 (for low vs. high recommendation accuracy, respectively), and centered the measures of IUIPC and Extraversion at the grand mean. Next, we computed the cross-products for the higher-order interactions between those predictors on Engagement Clicks, and entered all of these interaction terms into the equation together with the main effects to test for causality. Thus, we put to the test a full-factorial negative binomial regression model with seven effects.

We tested our full-factorial model against an intercept-only model. The Likelihood Ratio chi-square for the full negative binomial model was χ2 (7) = 15.61, p < .03, with a model deviance D = 127.09, which showed that our fitted model led to a significant reduction in deviance, with goodness of fit values for the Log Likelihood function = − 432.93, Akaike’s Information Criterion (AIC) = 883.85, the Bayesian Information Criterion (BIC) = 909.45.

Table 4 summarizes the results of the negative binomial regression analysis. Confirming our Hypothesis 1a, overall IUIPC had a significant positive effect on the log odds of Engagement Clicks, Wald χ2 (1) = 11.71, p < .001, indicating that people high on information privacy concerns produced a higher number of clicks in the IAB than those low on information privacy concerns. That is, people high on overall IUIPC engaged more heavily in editing, reading, and searching activities, as well as in switching between foreground and background. No other significant effects were revealed.

Table 4 Final model estimates of fixed and interaction effects on engagement clicks

4.4 Engagement duration

For Engagement Duration, the second dependent variable in our study, we again relied on the GLM module in SPSS. As shown in Fig. 4a, the count distribution for Engagement Duration was characterized by a partly exponential shape, which deviated from the partly linear form of standard generalized linear models. It is more appropriate to analyze such exponential dispersion models or Tweedie distributions (Jorgensen 1987) with the help of mixed count regressions. We ran a Tweedie mixed count regression with a maximum likelihood estimate 𝜃 = 207.63, SE = 14.29, 95% CI [181.43, 237.61] to evaluate our hypotheses for Engagement Duration.

We applied the same procedure for building regression models with fixed and interaction effects as before (Aiken et al. 1991; Coxe et al. 2009; Aiken et al. 2015), meaning that we tested the same full-factorial regression model with seven effects. Our full-factorial model was tested against an intercept-only model. The Likelihood Ratio chi-square for the full model was χ2 (7) = 137.96, p < .0001, with a model deviance D = 337.23, which showed that our fitted model led to a significant reduction in deviance, with goodness of fit values for the Log Likelihood function = − 1481.04, Akaike’s Information Criterion (AIC) = 2980.07, the Bayesian Information Criterion (BIC) = 3005.67.

Table 5 summarizes the regression results. Again, in confirmation of Hypothesis 1a, overall IUIPC had a significant positive effect on the log odds of Engagement Duration, Wald χ2 (1) = 11.71, p < .001; people high on IUIPC interacted longer with the IAB than those low on IUIPC. In support of Hypothesis 1b, the log odds of Engagement Duration as a function of IUIPC were dependent on Recommendation Accuracy, Wald χ2 (1) = 6.58, p < .01. People high (vs. low) on IUIPC interacted longer with the IAB in general, but especially when Recommendation Accuracy was high. This set them apart from the lack of engagement of people low on IUIPC. Confirming our Hypothesis 2a, the log odds of Engagement Duration as a function of Extraversion were dependent on Recommendation Accuracy, Wald χ2 (1) = 5.02, p < .03. People high on Extraversion generally displayed high Engagement Duration, whereas those low on Extraversion did so only when Recommendation Accuracy was high. We did not find any evidence that this interaction effect of Recommendation Accuracy and Extraversion on Engagement Duration was further moderated by IUIPC. Hypothesis 2b, therefore, had to be rejected. No other significant effects were revealed.

Table 5 Final model estimates of fixed and interaction effects on engagement duration

5 Discussion

The present study explored to what extent the engagement of users in proactive RS was influenced by the accuracy of recommendations, their concerns with information privacy, and differences in trait personality. The contributions of our research are as follows:

Scientific relevance

First, our work offers refinement to the already vast literature on the role of information privacy concerns in user engagement with IS (Smith et al. 2011; Bélanger and Crossler 2011; Dinev and Hart 2006). Our results especially contribute to the growing awareness that information privacy concerns may exert a positive impact on user engagement (Buck et al. 2014; Kelley et al. 2013; Kokolakis 2017). That such a positive effect indeed exists for clicks and duration, two different measures of user engagement, is the main insight from our study. This empirically confirms a possibility raised in Wisniewski et al. (2015), and adds to the growing popularity of the Internet Users’ Information Privacy Concerns (IUIPC) measure (Malhotra et al. 2004) for information privacy concerns (Li et al. 2011; Xu et al. 2012). Our work further hints towards the possibility that accurate RS can compensate for a lack of user engagement due to high information privacy concerns. Even though recommendation accuracy did not lead to higher user engagement for all participants, our results revealed that people high on information privacy concerns engage even more with the RS. This result squares with a growing literature on the benefits of information transparency on engagement with IS (Awad and Krishnan 2006; Jannach et al. 2016). A similar trend due to recommendation accuracy was observed for users low on trait Extraversion, who also engaged longer with the RS when recommended items were highly accurate. Theoretically, this is in line with prior separate findings on trait personality in IS settings (Junglas et al. 2008; Al-Samarraie et al. 2017). However, it remains to be seen to what extent each of the observed trends counter or rather confirm criticism from within the RS community, that users typically display more engagement and satisfaction with RS, not because the recommendations they receive are objectively of higher quality, but because they are merely subjectively perceived as such (Knijnenburg et al. 2010).

Limitations

We should, in that respect, hasten to acknowledge the limitations of this research. First of all, the results derive from a single study. Of course, generalization should not be based on the results of single studies (Driskell and Salas 1992). Second, the small size of our experiment did not allow for the provision of more robust evidence for our findings. We specifically lacked the statistical power to provide a thorough analysis of the trends in personality traits that were observed and reported in this research (see also our discussion of future work below). Moreover, the participants in our sample were taken from the classroom. Even though college student samples are typically used in research on the Big Five, cf. John and Srivastava (1999), it has been widely acknowledged in general (Cozby and Bates 2012) and for Big Five personality traits in particular (Rammstedt et al. 2010) that findings from college student samples do not fully generalize to the wider population. Also related to the issue of external validity, a large percentage of participants in our sample reported to have infrequently or very infrequently experienced privacy invasions during mobile application usage (see Table 2). On the one hand, this confirms a trend observed elsewhere that younger people are less concerned with privacy issues than older generations (Hassing 2015). However, it also illustrates that our findings may not entirely generalize to other user groups such as older demographics. Further, we kept the factor recommendation accuracy under experimental control, but the same cannot be said for our participants. They had full freedom to download, explore, interact, or ignore the software at their convenience – i.e., as long as no task results had been submitted. This may have led some of our participants to procrastinate more than typically allowed for in the heavily controlled setting of the laboratory. Naturally, this procedure does not meet the requirements of a full-fledged laboratory experiment, but added to the ecological validity of our study (Mook 1983). Finally, user engagement in the setting of proactive RS seldom is a one-shot event, but rather a series of interactions at various intervals. The present setup did not fully accommodate users in that respect. Users could not really learn from experience, and come to trust the RS on its merits, as the mobile application was rendered dysfunctional, once the participant had submitted results to all challenges. This is something that should be taken into consideration in future iterations of our RS.

Future work

We did not explore all Big Five personality dimensions in our study. This constitutes both a weakness and an opportunity for future research, as the personality dimensions Openness, Conscientiousness, Neuroticism, and Agreeableness may also impact user engagement. The RS literature reveals mixed, and often contradictory, evidence for the role of these other Big Five dimensions on user engagement. For instance, the positive effects of Openness and Conscientiousness on online game engagement (Teng 2008) are only partially confirmed for user engagement in Facebook (for Openness, but Conscientiousness produces mixed effects) (Bachrach et al. 2012), while they yield (negative) effects for Conscientiousness and no effects for Openness in a study on engagement in diverse RS (Chen et al. 2013). It would be interesting to explore in future research how Openness and Conscientiousness, in tandem with information privacy concerns, impact user engagement in proactive recommendations. Probably, users high on Openness, due to their dispositional curiosity and eagerness to explore novel stimuli, would engage with a system regardless of the accuracy of the recommendations or information privacy concerns. Users high on Conscientiousness (i.e, precision), on the other hand, would likely engage heavily in a system offering accurate recommendations (Al-Samarraie et al. 2017). This level of user engagement could be further enhanced by someone’s information privacy concerns; see Junglas et al. (2008). This association with information privacy concerns may not exist for people high on Neuroticism, however (Junglas et al. 2008). Such users, characterized by heightened negative emotional states and distress, would preferably not engage in a system offering recommendations that are inaccurate – hence, causing distress. On the other hand, and more counterintuitive, low-quality recommendations may also cause the neurotic user to engage more with the system. People high on Neuroticism often are anxious not to make mistakes and, as a consequence, invest substantial cognitive effort in exploring and correcting errors (Perkins 2014). Proactive suggestions of low quality may cause the neurotic user to spend more time in error detection and error correction. Future work could attempt to shed light on these and other issues related to the Big Five.

Moreover, Extraversion is not the only social dimension in the Big Five framework: Agreeableness is the second. We did collect trait Agreeableness scores for our sample, but, due to our small sample size, we could not include those scores in our regression models (we lacked statistical power to meaningfully produce a full factorial model with fifteen effects). Nevertheless, tentative exploration revealed that Agreeableness was positively and moderately correlated with Extraversion (r = .30, p < .0001). Full factorial GLM analyses of those fifteen effects on our two user engagement factors seemed to hint towards a pattern, in which the two social dimensions positively reinforce one another, and, in interaction, produce an uplifting effect on user engagement. These results probably stem from an over-fitted model, thus we should be cautious to draw conclusions. Still, this issue could be addressed in greater detail in future study. Prior work in non-IS domains has reported robust personality configurations with high Extraversion and Agreeableness clusters, as well as with low Extraversion and Agreeableness clusters (Rentfrow et al. 2013). Collection of much larger, and preferably independent, samples of people working with the e-Tourism Challenge would, therefore, offer the opportunity to model personality trait clusters together with information privacy concerns on a larger scale.

Yet, another avenue of future work would be to adopt in follow-up research endeavors some of the more recent insights from social justice research (Tyler 1994). So far, scholars on information privacy concerns primarily relied on self-report measures like the IUIPC to capture (un)fairness. Remarkably, the research domain of social justice has also proven successful in the development of experimental paradigms for business ethics that temporarily induce people with (un)fairness levels in the safety of the laboratory (De Cremer et al. 2011), as well as in a more realistic, ecologically valid, IS context (De Cremer et al. 2017). In that light, some promising work in the absence of RS has mapped the neural correlates of decision-making under fair vs. unfair circumstances. Especially aversively motivated people seem to respond strongly to situations in which they feel obliged to accept an unfair offer (Servaas et al. 2015). It thus makes intuitive sense to postulate that information privacy concerns in RS have a neural basis rooted in aversive motivation. Future study could seek to explicitly put this to the test, and to explore to what extent user engagement in proactive RS will be triggered by the anticipated (un)fairness of RS in some users.

6 Conclusions

In conclusion, the present research highlights the need to account for a user’s information privacy concerns and personality traits in the development of proactive, accurate, RS. More generally, this work points to the importance of controlling for anticipated fairness in the design of recommender systems.