Abstract
Social interactions are essential for well-being. Therefore, researchers increasingly attempt to capture an individual's social context to predict well-being, including mood. Different tools are used to measure various aspects of the social context. Digital phenotyping is a commonly used technology to assess a person's social behavior objectively. The experience sampling method (ESM) can capture the subjective perception of specific interactions. Lastly, egocentric networks are often used to measure specific relationship characteristics. These different methods capture different aspects of the social context over different time scales that are related to well-being, and combining them may be necessary to improve the prediction of well-being. Yet, they have rarely been combined in previous research. To address this gap, our study investigates the predictive accuracy of mood based on the social context. We collected intensive within-person data from multiple passive and self-report sources over a 28–day period in a student sample (Participants: N = 11, ESM measures: N = 1313). We trained individualized random forest machine learning models, using different predictors included in each model summarized over different time scales. Our findings revealed that even when combining social interactions data using different methods, predictive accuracy of mood remained low. The average coefficient of determination over all participants was 0.06 for positive and negative affect and ranged from − 0.08 to 0.3, indicating a large amount of variance across people. Furthermore, the optimal set of predictors varied across participants; however, predicting mood using all predictors generally yielded the best predictions. While combining different predictors improved predictive accuracy of mood for most participants, our study highlights the need for further work using larger and more diverse samples to enhance the clinical utility of these predictive modeling approaches.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
The social context plays a crucial role in mental health (Blanco et al., 2008; Teo et al., 2018). Whereas positive social interactions can enhance mood, a lack of social interactions is a risk factor for low mental well-being (Blanco et al., 2008; Teo et al., 2018). Moreover, the onset of psychological disorders is often characterized by lower engagement in social interaction (Krach et al., 2010; Perry & Pescosolido, 2012). Thus, understanding features of social interactions that predict mood and well-being is important to the assessment and treatment of psychopathology.
A person’s social context is made up of complex interactions between various components of social experience. The first component of social experience includes (daily) social situations and interactions (Phongsavan et al., 2006). These interactions form social relationships that develop over time and make up a person’s social environment (Asendorpf & Rauthmann, 2020), sometimes referred to as a personal network. A second component of social context includes the perceptions of these interactions and relationships, for example whether an interaction was perceived as pleasant or unpleasant (Asendorpf & Rauthmann, 2020). These perceptions are sometimes referred to as the psychological component or cognitive representation (Harpham, 2002; Phongsavan et al., 2006). Research has shown that interactions between these components are closely linked to an individual’s well-being (Brown et al., 2011; Krach et al., 2010; Perry & Pescosolido, 2012). To measure components of the social context, several technologies from various disciplines have been developed (Langener et al., 2023). Three commonly used methods are (a) digital phenotyping; which passively captures a person’s social behavior, (b) the experience sampling method (ESM); which measures perceptions of social situations, and (c) egocentric networks; which assess an individual's personal social network that form a person’s social environment.
Digital phenotyping refers to the collection of passive data via smartphones and wearable devices, such as data on phone calls, text messages, app usage, GPS, Wi-Fi, and movement, measured continuously throughout the day (Torous et al., 2016). These data are sometimes considered more objective than self-reported data and are collected with lower respondent burden. Previous studies have suggested that passive smartphone measures can provide insights into an individual’s social behavior (Burns et al., 2016; Eskes et al., 2016; Fulford et al., 2020; Stanislaus et al., 2020). However, results from previous work have been inconsistent, and the specificity of passive smartphone measures for capturing aspects of an individual’s social behavior remains unclear (Langener et al., 2023). As an example, the association between the distance traveled (measured per GPS) and number of interactions differed between people with schizophrenia (r = 0.07) and healthy controls (r = 0.6; Fulford et al., 2021) complicating the interpretation of ‘distance traveled’. Despite this limitation, we consider passive smartphone measures as potential indicators of an individual's social behavior, with the important constraint that we do not know which specific aspect they capture.
ESM can measure multiple components of social situations, including one’s daily social interactions, as well as their perception of these specific interactions (e.g., Čolić et al., 2020; Mills et al., 2014). Within the ESM, participants fill out brief questionnaires using their mobile phone several times a day after a push notification is sent (Kubey et al., 1996). Long ESM questionnaires are burdensome for the participants (Eisele et al., 2020) but can cover social situations more broadly than digital phenotyping.
To assess information on long-term social relationships, egocentric networks are often constructed. Researchers typically start by asking the participant to list the names of important social contacts with whom they are close with or have interacted with during a period of time. After those contacts are identified, further details about the characteristics of these contacts and the relationship with these contacts are assessed (Perry et al., 2018), including social role of the contact (e.g., partner, family, friend) or level of respondent closeness to members in the network. This approach can measure facets of support that are more difficult to assess through ESM or digital phenotyping.
Although these methods capture different aspects of the social context over different time scales that relate to well-being, they have rarely been combined in previous research (Langener et al., 2023). Combining data from multiple sources may help to increase the precision with which we measure the impact of social context on health. Previous research studies have often focused solely on using digital phenotyping to measure social situations and predict mood, but the predictive accuracy is often low to moderate (e.g., Abdullah et al., 2016; Jacobson & Chung, 2020; Jacobson et al., 2020). Complementing digital phenotyping with data from other self-reported aspects of the social context may be needed to improve predictive accuracy and thus increase clinical utility (Currey & Torous, 2022).
The Current Study
In this study, we integrated data from three methods: digital phenotyping, ESM, and egocentric networks to construct individualized random forest machine learning models to investigate how aspects of social context predict positive and negative affect. We used individualized random forest machine learning models, as they are shown to outperform non-individualized prediction models yet are rarely used in research on well-being (Abdullah et al., 2016; Benoit et al., 2020; Cai et al., 2018; Hart et al., 2012). Results from individualized prediction models also offer the potential to deliver just-in-time interventions, which are personalized interventions delivered directly to individuals when needed most (Nahum-Shani et al., 2017).
We measured social context over various time scales using digital phenotyping, ESM, and egocentric networks. Digital phenotyping measures social behavior near continuously throughout the day, while ESM measures social situations at specific time points (i.e., in reference to a short period of time before the questionnaire was completed). Egocentric networks focus on the constant characteristics of the social environment and are often assessed only once or twice within a period of time. To integrate data across different temporal resolutions, we enhanced ESM with data obtained from egocentric networks (Stadel et al., 2023). Additionally, we investigated the impact of a given time scale on prediction performance. We further examined the extent to which these different methods uniquely predicted mood to identify the most important measurements in predicting health outcomes.
Methods
Participants and Procedure
We used data from a student sample collected in 2022 (N = 15, female = 14, male = 1). Participants were on average 22 years old (SD = 4.10, min = 18, max = 35). A detailed description of the study procedure can be found in the respective Open Science Framework (OSF) repository (https://osf.io/jqdr9/). The study was approved by the ethics board of the University of Groningen (research code: PSY-2223-S-0018). The analyses were preregistered before being conducted (https://osf.io/738pr).
The participants filled out ESM questionnaires for 28 days. During those 28 days, participants received five semi-random questionnaires measuring positive and negative affect. Participants were also instructed to record any social interactions that lasted longer than 5 min. Social interactions were defined according to Hall's (2018a) definition of a focused social interaction, which includes “(1) mutual acknowledgment by both partners of a shared relationship, (2) conversational exchange, and (3) focused attention by both partners on that exchange”. Thus, mutual text messaging was also included in the assessment. We used two different assessment formats to capture social interactions—which we refer to as signal- and event-based reporting (Myin-Germeys & Kuppens, 2022). Each format was used for 2 weeks, with the order balanced between participants. In the signal-based reporting format, participants were asked to retrospectively report all their social interactions since the last questionnaire during the daily scheduled semi-random questionnaires. In contrast, during the event-based format, participants were required to report social interactions immediately after they occurred. In both conditions, respondents filled out the same scheduled questionnaires to evaluate their mood and daily activities. An overview of all questions asked can be found here: https://osf.io/5hmdz.
Alongside the ESM, participants installed an app (Behapp) that collected data passively for 28 days (Jagesar et al., 2021). On Android devices data on location, calls, texts, Wi-Fi, screen states, and app usage was collected. At the time of this data collection, only location data was collected on IOS devices (although Behapp now collects more than location data on IOS devices).
Participants completed an egocentric network questionnaire to assess their social relationships before and after the ESM period. We asked the participants to list names from their social network with whom they have contact with in their daily lives. Subsequently, further questions about the relationship characteristics were asked, for example, how close they are to a specific person. An overview of all the questions asked in the egocentric network can be found at https://osf.io/8zg3x. In addition, the nicknames of interaction partners were recorded during the ESM period, which allowed us to link names from the egocentric network with those from the ESM period to gain insight into characteristics of interaction partners (for more information see Stadel et al., 2023).
As stated in the preregistration three participants were excluded because they did not own an Android phone and/or had less than 75% of passively collected data via Behapp. Additionally, one participant was excluded because they filled out less than 75% of the scheduled ESM data. This results in a total sample of 11 participants (female = 10, male = 1, total included ESM measures, N = 1313).
Measures
Outcome Variable
We aimed to predict positive affect and negative affect as measured during the scheduled questionnaires. Positive affect was assessed by taking the mean of the answers to three items (“I feel happy”, “I feel energetic”, “ I feel relaxed”), and negative affect was assessed by taking the mean of the answers to four items (“I feel sad”, “I feel anxious”, “I feel stressed”, I feel irritated”). The questions were measured using a slider that was labeled with Strongly disagree (left side) and Strongly agree (right side) on an 11-point Likert scale, which we rescaled to one to 11 for ease of calculation.
Predictors
In Table 1, we summarize the predictors used to assess an individual’s social behavior. These behaviors were aggregated in 3, 6, and 24 h time windows before the scheduled ESM questionnaire was filled out. For example, data collected between 11:00 and 14:00 (i.e., a 3-h time window) could be used to measure social behavior that occurred before the ESM questionnaire was filled out at 15:00 (see also Fig. 1). We excluded data collected during the hour prior to ESM measurement (between 14:00 and 15:00 in the above example) to predict mood in advance (illustrated in Fig. 1). As a robustness check, we also completed analyses excluding data collected during the 30 min prior to ESM measurement.
We used time windows of 3 and 6 h to measure social behavior that might more immediately impact mood, such as a specific social interaction that occurred recently before the mood assessment. We used the 24 h time window to measure impactful social interactions that occur less frequently (e.g., meeting up with friends at night) or patterns of behavior that occur over extended periods (e.g., being at work for several hours). Longer time scales also can be used to measure irregularities in an individual’s behavior that may indicate changes in mood, for example, if a person suddenly does not leave the house for a whole day (Cai et al., 2018).
We derived several variables from raw passive smartphone measures to measure aspects of social interaction. Some measures, such as total minutes of calling, were directly summarized into a variable in the aggregated dataset at a specific temporal resolution. Other measures, such as GPS coordinates, required preprocessing steps before they can be aggregated. To derive meaningful location variables, we clustered GPS coordinates into staypoints, where a staypoint is any location within a radius of 350 m where participants stayed for at least 30 min (Zheng et al., 2009). The staypoint where the participant spent most nights is labeled as their home location. App usage variables were added by classifying apps based on their category in the Google Play Store (Google Play Store Team, 2020).
We used data from the ESM questionnaires to create social interaction variables. We created variables that indicate the duration that a participant spent in a specific social interaction. For example, the total minutes that someone spent in face-to-face interactions or in online interactions. Furthermore, participants could select the purpose or topic of their interactions, such as “catching up” or “joking around”. To define the content of an interaction we used Hall's (2018b) definitions of “communication episodes”, which describes four different types of content and behavior of an interaction (i.e., striving behavior, mundane and maintenance behavior, work or school talk, and negative content). Additionally, we created variables about the last interaction a person had before the scheduled ESM questionnaire and how long ago this interaction was. This means that the most recent interaction may have taken place before the chosen level of aggregation. For example, if the level of aggregation was 3 h, the last interaction could have occurred 4 h before the ESM questionnaire was filled out. We exclude interactions that occurred during the lag of 30/60 min to be able to predict mood in advance.
We created variables to indicate various interaction partners by matching the nicknames from the egocentric network with data from the ESM period. This allowed us to examine the impact of the characteristics of interaction partners on mood (Stadel et al., 2022). For example, we created a variable that indicated the length of time (in minutes) that someone spent in an interaction with a friend or a family member, as well as a variable that reflected the level of closeness with an interaction partner during a given period. During periods where multiple interaction partners were present, we calculated the majority or mean value for the given variable to provide an overall measure of relationship characteristics. For example, if participants had an interaction with three friends and one stranger, the interaction would count as having an interaction with friends and not with strangers. Unless indicated otherwise, all questions were measured on an 11-point Likert scale (see Table 1).
To prepare social interaction predictors for analyses, we employed several preprocessing steps using the caret package. First, we used the centering and scaling function to normalize continuous variables (Kuhn, 2008). Next, we transformed categorical variables into dummy variables. We excluded any predictor with zero or near-zero variance. Passive smartphone measures were considered missing if no data was recorded for 24 h. We imputed missing data based on the k-nearest neighbors in the training set.
Participants had the option to exclude individuals from their egocentric network if they did not have frequent contact with them or if they were not considered relevant. Additionally, “stranger(s)” were not included in the egocentric network. We assumed that those people were not close to the participant, the participant did not discuss personal issues with them, and they did not provide emotional and practical support. Therefore, scores on those variables were given a value of zero. We imputed selected social interaction variables (e.g., “Spending time with this person costs energy”, “Spending time with this person gives energy”, “I can be myself with this person”) as we had no clear hypothesis of how the participant would have rated these variables.
Prediction Model
Prediction Algorithm
We used random forest models to predict positive and negative affect. Random forest models are computationally efficient models that can capture nonlinear relationships and are commonly used in digital phenotyping studies to predict mental health outcomes (e.g., for a review see Benoit et al., 2020). Random forest models combine several decision trees to make predictions (Breiman, 2001). A decision tree splits the outcome variable into different subsets based on the predictor variables. The final nodes of the tree correspond to the predicted outcome value. A disadvantage of decision trees is that they are not very robust, as small changes in the data can cause large changes in the final estimated tree (Hastie et al., 2009). To address this limitation, a random forest generates several random decision trees. More specifically, a number of bootstrapped observations and randomly selected variables are used to build these random decision trees. The final prediction is based on the average over all trees (Hastie et al., 2009). Random forest requires the tuning of hyperparameters, such as the number of randomly selected predictors and the number of random decision trees, those hyperparameters are chosen by applying cross-validation. The optimal number of randomly selected predictors was chosen by excluding the last observation in the training set. We chose between three to 18 randomly selected predictors in each tree. For the other hyperparameters (e.g., number of trees), we used the default values from the caret package (Kuhn, 2008).
We used the caret package to run the machine learning models. Analyses were conducted using R version 4.2.1 (2022-06-23). The code used to clean and analyze the data can be found in the following GitHub repository: https://github.com/AnnaLangener/CombiningMethods_MoodPredictions.
Model Validation
Two main strategies exist for evaluating time-series prediction models, namely, fixed origin versus rolling origin evaluation (Hewamalage et al., 2022; Tashman, 2000). In the fixed origin the dataset gets split into two sets where, for example, the first 70% of observations is used for building a model and the last 30% for evaluation. In contrast, the rolling origin approach uses multiple training and evaluation sets to predict the next set of observations, repeating the process until the end of the dataset is reached (see Fig. 2 for an illustration).
A rolling origin evaluation has several advantages over a fixed origin approach. Rolling origin evaluation is considered more robust, less dependent on specific patterns, and it can recalibrate the model with new data (Tashman, 2000), and thus, is usually recommended (Hewamalage et al., 2022). The rolling origin evaluation setup is similar to another approach referred to as blocked cross-validation (Bulteel et al., 2018).
We expected that the window size would have an impact on the prediction performance (Gama et al., 2014). Therefore, we tested three rolling window sizes ranging from around 3 days of data to 6 days of data (15, 20, and 30 ESM measurements). We opted for relatively short moving window sizes in order to capture changes in the association between social behavior and mood. This decision was also influenced by our relatively small number of ESM measurements. Accordingly, we wanted to ensure that an adequate number of data points remained for inclusion in the test set. Each test set contains three data points (also called prediction horizon), which is roughly half a day of data. We combined all predictions made into a single test set to evaluate the prediction performance. Figure 2 illustrates the rolling origin approach with a window size of five data points and a test set of three data points.
To evaluate the prediction performance, we calculated the coefficient of determination (R2; which was defined as one minus the fraction between the sum of squared differences between the outcome and the model predictions and the sum of squared differences between the outcome and an intercept-only model), the mean absolute error (MAE), and the mean absolute percentage error (MAPE) in the test set. We decided to use multiple measures to evaluate the prediction performance because R2 is a relative measure of fit (as it depends on the total variance) and does not provide any information about the overall accuracy of the prediction model. In contrast, measures such as the MAE can be used to evaluate the prediction error and thus the accuracy of the model (Rocca & Yarkoni, 2021). We additionally report the MAPE since it makes the evaluation of prediction models comparable with each other. Both MAE and MAPE have been reported in previous papers that used passive smartphone measures for prediction (e.g., Jacobson & Chung, 2020; Shah et al., 2021). Lastly, we investigated whether our model performs better than a baseline model, which uses the mean in each training set as the primary predictor. Our model must perform better than this baseline model to be considered acceptable.
Variable Importance
We assessed the variable importance for the best-performing model for each participant, by calculating the (mean) Shapley value of each predictor in the train setFootnote 1 (similar to Jacobson & Bhattacharya, 2022; Shah et al., 2021). Shapley values are based on coalitional game theory and assume that each variable is a “player” in the prediction task and the prediction made is the “payout” of a game (Molnar, 2018; Štrumbelj & Kononenko, 2014). In other words, the “game” refers to predicting a single data point, and the “gain” is the difference between the actual and mean prediction. The variables, or “players”, work together to receive this gain by forming coalitions. Importantly, Shapley values measure a variable's contribution relative to the mean prediction, without directly assessing the actual prediction performance.
To determine the Shapley values we used the iml package to calculate the marginal contribution of a variable across all possible coalitions of variables (Molnar, 2018). Variables that are not in a coalition are replaced with random values from the dataset, which means that they should not be related to the outcome variable anymore. The prediction is then calculated both with and without the variable of interest, resulting in the marginal contribution of a variable (Molnar, 2018). To get an overall variable importance score we computed the mean absolute Shapley value across all data points as in previous research (e.g., Jacobson & Bhattacharya, 2022; Shah et al., 2021).
Deviations from Preregistration and Robustness Checks
Based on the collected ESM data we derived several variables (see Table 1). We used Hall's (2018b) definitions of so-called communication episodes to describe the content of an interaction (i.e., striving behavior, mundane and maintenance behavior, work or school talk, and negative content). In our preregistration, we included the content “catching up” in both the “mundane and maintenance behavior” and “striving behavior” communication episodes. However, “catching up” is only applicable to the “striving behavior” episode (Hall, 2018b). We have corrected this error in our data analysis.
During the study, participants recorded the start and end times of their social interactions, while the timestamp of when they completed the ESM questionnaire was recorded automatically. For instance, a participant may have reported an interaction starting at 11:00 and ending at 13:00, with the ESM questionnaire completed at 13:10 on 21.11.22. We manually added the date to the start and end times of each interaction to create variables. This led to some inconsistencies, for example, participants reported having an interaction from 20:15 to 00:00 and handing in the survey at 21:24. We dealt with inconsistencies between automatic timestamp and participant self-report by manually adjusting the end time of an interaction if we believed the participant had made an error in logging the time (N = 7 out of 1313 observations). For example, we replaced the self-reported end time (e.g., 00:00) with the survey completion time (e.g., 21:24).
To ensure the robustness of our data analysis, we conducted analyses using the original data (except for one interaction that lasted for 0 min). The results from this analysis are reported in the Supplementary Materials and discussed in the section Robustness Checks. Notably, the overall results exhibited minimal changes when comparing the different analyses. All changes made to the data can be found in the following R script: https://github.com/AnnaLangener/CombiningMethods_MoodPredictions/blob/master/data_matching.R.
Using different moving window sizes in the cross-validation set-up led to different test set lengths. For instance, let us consider a scenario where a participant fills out 100 ESM questionnaires. If we select a training size of 15 data points, the test set across all moving windows will include a total of 85 data points. This is because the first 15 data points are necessary to train the first model, while the other data points can be used to test the model. In contrast, when a moving window size of 30 data points is used, the test set will only include 70 data points, as more data points are utilized for training the model. Hence, the test sets will have different sizes. To investigate whether different test set lengths affect the prediction performance, we recalculated the performance measures using the same test set length, meaning that we only used data after the 30th ESM measurement to calculate the prediction performance. We completed multiple robustness checks for our prediction models.
Overview of All Models
Different models were compared to evaluate the additive predictive value of ESM, digital phenotyping, and egocentric networks in predicting positive and negative affect (similar to Sano et al., 2018; see Table 2). The predictors for each model varied. In the first model, data from all measures were used (i.e., digital phenotyping, ESM, egocentric networks). In the second model, only digital phenotyping data was used, and in the third model only ESM data was used. In the last model, only data from the egocentric network, combined with the names of interaction partners that were assessed during the ESM period, were used. We will refer to this model from now on as using only egocentric network variables. We further varied the level of aggregation for the predictors and the cross-validation rolling window size.
We completed three robustness checks. First, we varied the time window used prior to the mood assessment in which predictors were aggregated (see Fig. 1). We used a 30 min lag instead of a 60 min lag. Second, we tested the impact of our preprocessing steps to clean the timestamp data (i.e., manually adjusting the end time of one interaction vs. seven interactions). Lastly, we checked the impact of using an equal test set length for different moving window sizes.
Results
Descriptive
Participants completed on average n = 119.4 (SD = 8.44, min = 108, max = 135) scheduled ESM questionnaires. These assessments were used to measure participants’ positive and negative affect throughout the study. Participants had an average positive affect of M = 6.42 (SD = 1.89), and an average negative affect of M = 2.58 (SD = 2.12). Additionally, participants logged on average 77.73 (SD = 42.91, min = 8, max = 162) social interactions throughout the study (see Table 3). We created line plots to illustrate how each predictor aggregated on different scales changes over time for each participant (see Supplementary Material).
Overall Prediction Performance Using Different Time Scales
Positive Affect
We first evaluated the performance of different models predicting positive affect. First, we varied the predictors included in each model: data from all measures, digital phenotyping only, ESM only, or egocentric networks data only. Second, we constructed models at different levels of aggregation. Third, we varied the number of observations used to train the model in the cross validation set-up (i.e., 15, 20, 30 observations; see Table 2 for an overview).
We examined the accuracy of each model in predicting positive affect. We start by describing the overall prediction performance based on the best-performing model (i.e., model with the highest R2 for each participant). Figure 3 displays the predicted and observed values of positive affect measured during the scheduled ESM questionnaires for each participant, providing an overview of the model’s performance. Table 4 summarizes the accuracy of the best-performing model per participant.
Participant 3 and Participant 10 had several instances where a model could not make a prediction when using all predictors, as there was insufficient data to impute the missing ESM and egocentric network predictors. In these cases, we considered the model to be performing worse than a model that could generate predictions for all data points with a more limited set of variables.
For the majority of participants (9 out of 11) the best-performing model predictions outperformed predicting the mean in the training set (using the same window size, see Supplementary Material for the results). However, R2 values were low with an average R2 over all participants of 0.06 (min = − 0.05, max = 0.21, SD = 0.08), indicating that the model was not able to explain a substantial fraction of variance in the data (note that negative R2 can occur because our measure of R2 is relative to an intercept-only model; a negative R2 implies that our fitted model performs worse than an intercept only model). The models achieved a moderate average correlation of 0.33 (min = 0.21, max = 0.52, SD = 0.09) between predictions and outcomes and an average MAE of 1.23 (min = 0.66, max = 1.92, SD = 0.36). This means that, on average, the prediction differed by 1.23 points from the observed value, which was measured on an 11-point scale. This resulted in an average MAPE of 22.62% (min = 8.69%, max = 47.89%, SD = 12.29%).
We explored different levels of aggregation and moving window sizes to determine their impact on the model performance. Results demonstrated that a time frame of 6 h or 24 h produced the most accuracy results for the majority of participants (see Table 4; 5 out of 11 each). There was one participant whose model demonstrated the best performance when aggregated at 3 h. The optimal moving window size in the cross validation set-up also varied, with 15 being the best for three participants, 20 for five participants, and 30 for three participants. Overall, the performance within a participant was not very robust using different levels of aggregation and moving window sizes as indicated by the relatively large standard deviation of performance measures across different time scales (see Table 4).
To summarize, the overall prediction performance for positive affect was weak and varied per participant. Moreover, the prediction performance varied between different time scales that were used to summarize the predictor variables and to train the prediction model.
Negative Affect
Next, we examined how well we were able to predict negative affect. Figure 4 displays the predicted and observed values of negative affect. Table 5 shows the accuracy of the best-performing model. If the model was unable to make any prediction (as was the case for Participant 3 and Participant 10), the model was considered as performing worse than a model capable of making predictions for all data points.
Similar to the positive affect models, for most participants (10/11) trained models showed better predictions than using only the mean from the training set. Nevertheless, the R2 values were low, with a mean of 0.06 (min = − 0.08, max = 0.3, SD = 0.13), indicating that the variance in the observed data was not well explained by the predictions. The correlation was moderate, with a mean of 0.33 (min = 0.13, max = 0.59, SD = 0.15). The MAE was slightly lower than when predicting positive affect, with a mean of 1.01 (min = 0.54, max = 1.86, SD = 0.43). However, the MAPE was higher, with an average of 36.74% (min = 18.16%, max = 55.54%, SD = 12.38%).
In line with results from the positive affect models, participants had varying optimal levels of aggregation and moving window sizes for predicting negative affect. The optimal time frame of aggregation was 6 h for four participants and 24 h for seven participants. The moving window size with the highest accuracy was 15 for one participant, 20 for four participants, and 30 for six participants. The results show varying levels of prediction performance stability when changing the moving window size or level of aggregation.
Overall, the prediction performance for negative affect was weak and also varied per participant. Similar to results from the positive affect models, prediction performance varied between different time scales that were used to summarize the predictor variables and to train the prediction model.
How Much Does Each Method Add to Predicting Mood
We further investigated which combination of variables led to the highest prediction performance for positive and negative affect. Four different sets of predictors were examined, including ESM, passive smartphone measures, egocentric networks, or a combination of all. We start by investigating the set of predictors that performed best per participant.
Results demonstrated that the best set of predictors varied across participants (see Table 4). Including all predictors improved prediction accuracy of positive affect for five participants. Using only egocentric network variables resulted in the highest performance for two participants. For another three participants, using only Behapp outperformed the other models. However, two of these three participants were Participant 3 and Participant 10, who both had a low number of logged interactions, resulting in missing values in the predicted data points. Lastly, for one participant, using only ESM achieved the best results.
The best model's prediction performance was often similar to the next best-performing model (using a different moving window size, level of aggregation, or set of predictors). Therefore, we also investigated how often different sets of predictors lead to an acceptable prediction performance over different time scales and participants. We define an acceptable prediction as one that performs better than the mean prediction from each training set (i.e., higher R2 and lower MAE) and obtains a positive R2 value. Figure 5 shows the distribution of R2 and MAPE for the participants whose best-performing model had an acceptable prediction. We observe that there was variability between which predictors performed best between participants, but that models that performed well were often close to each other.
Among the selected participants, using all predictors resulted in acceptable predictions for 22.2% of the models (n = 22). This means that out of all 99 models (3 levels of aggregation × 3 moving window sizes × 11 participants) across all participants in which all predictors were used, 22 models gave predictions that outperformed an intercept-only model and the mean from each training set. This was followed by using only ESM (13.1%, n = 13) and only Behapp data (12.1%, n = 12). Using only variables from the egocentric network performed worse for the selected participants (10.1%, n = 10).
We observed similar results for negative affect models. Including all predictors improved model accuracy for four participants; whereas using only Behapp outperformed other models for three participants (including Participants 3 and Participant 10). For two participants, using only egocentric network variables achieved the best performance, and for another two participants, using only ESM produced the best predictions. Figure 6 shows that, similarly to predicting positive affect, there was variability between which predictors performed best across participants (whose best-performing model made an acceptable prediction), but that good-performing models were often close to each other.
Using all predictors yielded the best results and led to an acceptable prediction in 14.1% of the cases (n = 14). The second-best model used only ESM variables (12.1%, n = 12), followed by using only network variables (6.1%, n = 6). Among the selected participants, only in three cases using only Behapp data produced acceptable predictions (3%, n = 3).
Robustness Check
To ensure the robustness of our results, we used four different datasets summarized inTable 2.Footnote 2 First, we examined whether the R2 and MAPE were similar across the different datasets. We also checked whether the best predictors included in a model were robust by recalculating how often different sets of predictors (i.e., using all predictors, only egocentric network, only Behapp, or only ESM) resulted in acceptable predictions. A full overview of the robustness checks can be found in the Supplementary Material (section Robustness Checks). In this section, we provide a brief summary of the results.
For positive affect, we observed a strong association between the R2 values and the MAPE across the different datasets (all r > 0.97). Across all datasets, using all predictors performed best, followed by using only Behapp and only ESM. Using only egocentric network variables performed worst. This suggests that when predicting positive affect, the results are robust across different datasets.
For negative affect, the association between R2 values across different datasets revealed a high correlation for three out of the four datasets (all r > 0.98). However, a notable difference emerged when utilizing a dataset that included an equal test set length for different moving window sizes (r = 0.7). A similar pattern emerged when examining which predictors performed best. Upon closer examination, we found that this pattern was primarily due to one participant (Participant 5, see the Supplementary Material Fig. 51 for more information) who had large values of negative affect at the beginning of the study. By making the test length equal across different moving window sizes, we excluded those data points, leading to unstable results.
Variable Importance
We computed Shapley values to gain insights into the importance of different variables in predicting mood. This analysis was conducted on the best performing model for each participant, considering all predictors. However, it is important to note that we only computed Shapley values for participants whose best performing model using all predictors had a higher predictive accuracy compared to the baseline model, as none of the variables in models with poor predictive accuracy will be important.
Figure 7 illustrates the overall variable importance for each variable and participant. As an example, the highest variable importance was found for respondent 6 for positive affect: The total duration of app usage had an average mean absolute Shapley value of 0.22. This indicates that, on average, app usage contributed to a difference of 0.22 compared to the mean prediction across different coalitions and data points for this participant. Some values are zero, for example, the total minutes spent with a teacher or superior showed a value of zero for respondents 1, 2, 7, 8, and 9 for positive affect. This implies that this particular variable had no impact on the prediction, most likely because it was often removed from the model due to its (near-)zero variance. Moreover, our results suggest that certain variables are of little importance across participants. For example, text messaging was not a strong predictor of mood. Additionally, the content of conversations (e.g., about work or school versus small talk), seemed to have little impact (although this pattern could be due to limited variation as participants did not record many interactions throughout the day). Phone calls recorded via Behapp also appeared relatively unimportant for predicting mood.
Discussion
Insights and Reflections on Clinical Utility
In this study, we examined the extent to which various aspects of social context predicted mood. To capture the social context, we integrated data from digital phenotyping, ESM, and egocentric networks. Our three primary objectives were to: (1) to assess the accuracy of mood prediction using different methods that measured distinct aspects of the social context over different time scales, (2) to investigate the extent to which the chosen time scale impacted prediction performance, and (3) to explore whether a combination of different methods increases predictive accuracy.
Low Overall Predictive Accuracy and Variation Among Participants
Our findings indicated that the overall predictive accuracy is low, as reflected by an average R2 value of 0.06 for positive affect and negative affect. Additionally, the average MAE for positive affect is 1.23 points, and the average MAPE is 22.62%. Similarly, for negative affect, the average MAE is 1.01, and the average MAPE is 36.74%. These results suggest that our machine learning models are unable to explain a large portion of the variance. Furthermore, the overall predictive ability varied among participants. Some participants had a moderate predictive accuracy, with the highest R2 values of 0.21 and 0.3 for positive affect and negative affect, respectively. In contrast, for other participants, the predictions were worse than those of a baseline model. In practice, it will thus be challenging to determine in advance which participants’ personalized models will have acceptable predictive accuracy and can be used for personalized treatment.
Our results are similar to the findings of Asselbergs et al. (2016) who aimed to predict day–to–day mood using solely passive smartphone measures and individualized prediction models. Their goal was to replicate the research conducted by LiKamWa et al. (2013), who were the first to investigate the potential of predicting mood based on passive measures. Contrary to our results, LiKamWa et al. (2013) found a high predictive accuracy of 93% of the self-reported mood within a tolerated error margin of 0.5 around the observed scores. The findings of Asselbergs et al. (2016), however, were less promising, as their predictions were worse compared to a simple benchmark model, which is in line with our low predictive accuracy. Several factors may have contributed to these results. LiKamWa et al. (2013) conducted their study over a duration of 60 days, while both our study and that of Asselbergs et al. (2016) had considerably shorter durations of 28 and 42 days, respectively, which could have affected predictive accuracy (Asselbergs et al., 2016). Additionally, our study focused on a different population, consisting of Dutch students, while LiKamWa et al. (2013) conducted their research with Chinese students. These differences in population could also have influenced the predictive accuracy of the models.
Variation of Predictive Accuracy Across Time Scales
We observed that the chosen time scale for aggregating and analyzing the data influenced the prediction performance, leading to considerable variation in the predictive accuracy across different models. Generally, it appears that longer aggregation windows (e.g., 6 and 24 h) yielded superior results compared to shorter time windows (e.g., 3 h). One possible explanation for this could be that the participants did not interact frequently throughout the day; thus, longer time windows (e.g., 6 and 24 h) were able to capture infrequent, but impactful, social interactions.
A recent study investigated whether passive smartphone measures predicted negative affect using data from 50 participants who reported their negative affect 10 times a day (Niemeijer et al., 2022). Their aim was to predict the average negative affect of the next day using different machine learning models and they conducted a multiverse analysis to explore the impact of various preprocessing and analysis choices for each sensor on their results. Consistent with our own findings, this study exhibit a low predictive accuracy for negative affect, with the best model achieving an R2 value of 0.10 in the test set (Niemeijer et al., 2022).
Similar to our results, the multiverse analysis conducted by Niemeijer et al. (2022) revealed that different choices in methodology led to varying results in predictive performance. In practical applications, it is crucial to determine the appropriate preprocessing steps for each sensor and identify the optimal time scale in advance or to incorporate them into the cross-validation setup. Choosing preprocessing steps and the optimal time scale retrospectively can hinder real-time applications and potentially result in overfitting. Therefore, future research should focus on applying suitable methods to determine optimal preprocessing steps and time scales beforehand (Verachtert et al., 2022).
To summarize, in contrast to prior research that yielded promising results in mood prediction (LiKamWa et al., 2013), our findings align with studies indicating that it is challenging to accurately predict mood (Asselbergs et al., 2016; Niemeijer et al., 2022). Despite the combination of measures (i.e., passive smartphone measures, ESM, egocentric networks) the overall predictive accuracy for mood in our study remained low. Based on our results and following the findings by Asselbergs et al. (2016) and Niemeijer et al. (2022), the application of mood prediction models in real-world clinical settings is currently unlikely to be useful due to the low predictive accuracy and the observed variation among respondents and prediction models.
Different Parts of the Social Context Are Important to Predict Mood
Our results suggest that the optimal set of social context predictors varied among participants and that there is variation among participants in which specific variable was most important for predicting mood. This makes sense, considering that different parts of the social context may be more or less important for predicting the mood of different individuals. For instance, for people with few social interactions throughout the whole study period, passive measures, such as app usage, might be a better indicator of their mood than the total minutes they spent in interactions. However, we found that utilizing all predictors generally yielded the best predictions, which highlights that a combination of methods leads to higher predictive accuracy for the majority of participants and that different methods (digital phenotyping, ESM, egocentric networks) indeed tap into different aspects of the social context, even though the overall predictive accuracy was low.
We calculated Shapley values and found that phone calls recorded via Behapp are relatively unimportant for predicting mood. This is somewhat at odds with the finding that, for some participants (e.g., Participants 1 and 4), the total minutes spent calling recorded via ESM, is more important. This may be explained by the fact that Behapp only captures phone calls via the mobile network and no other forms of communication like (video) calls made via WhatsApp. Results support the integration of digital data sources to measure distinct aspects of social behavior that may otherwise be missed by relying solely on a single method.
Limitations and Future Outlook
A unique strength of our study is the demonstrated impact of aggregation window and number of cross-validation observations on variability in predictive performance. We evaluated our prediction performance by choosing the model that led to the highest predictive accuracy over different time scales. However, in practical applications, it is crucial to determine the appropriate level of aggregation and moving window size in advance or to incorporate it into the cross-validation setup. Choosing the optimal time scale retrospectively can hinder real-time applications and potentially lead to overfitting. Thus, in future research methods should be applied to determine the optimal time-scale beforehand (Verachtert et al., 2022).
One limitation of our study is that a significant portion of the ESM measures and passive measures employed were not validated, which introduces potential challenges when interpreting our findings (Flake & Fried, 2020; Langener et al., 2023). It remains uncertain which specific aspects of social behavior are captured by the selected passive measures as results are inconsistent across studies and participants (Langener et al., 2023). For example, the association between the distance traveled and the number of interactions that a person had differed between people with schizophrenia and healthy controls (Fulford et al., 2021). Therefore, it is crucial for future research to prioritize the development and validation of measures that characterize social situations and the social environment. Additionally, we decided to classify apps based on how they were categorized in the Google Play Store. Fortunately, more scientific classification systems have recently become available, such as the one developed by Schoedel et al. (2022), which are recommended for use in the future.
In our study, we used positive and negative affect as an indicator of mood. This decision was based on previous research indicating a link between short-term emotions, commonly measured through positive and negative affect scales, and psychological well-being (Houben et al., 2015). However, a recent study suggests that researchers should be more critical when using a score that consists of multiple items and that using single items might be superior (Cloos et al., 2023; McNeish & Wolf, 2020). Thus, it would be interesting to investigate how results would change if we would only aim to predict the score of a single item, such as “I feel happy” or “I feel sad”, instead of positive or negative affect. In addition, the selection of items to measure positive and negative affect or other constructs in ESM research is often arbitrary. This can reduce construct validity (Bringmann et al., 2022; Flake & Fried, 2020), which could potentially lead to lower predictive performance when attempting to predict these constructs.
Our study employed a relatively small sample size of students (N = 11), with a majority of participants being female students (n = 10). It is important to acknowledge the limitations of this sample composition, as it may restrict the generalizability of our findings to a broader population. To gain a more comprehensive understanding of the predictive accuracy and its potential variations, future research would benefit from utilizing larger and more diverse samples. This goes in line with the current call to have larger samples for studies using passive smartphone measures in order to get more robust findings. One potential solution could be to pool data across studies, even though this might come with privacy concerns (e.g., Davidson, 2022; Huckvale et al., 2019). Furthermore, it is challenging to combine data from different studies as measures might differ, for example, the frequency of ESM assessments and collection of passive data. Additionally, constructs across different studies are likely to be operationalized in different ways which makes it further hard to pool data across studies (e.g., Davidson, 2022; Huckvale et al., 2019).
The variability in predictive ability and importance of variables in predicting mood across participants indicates that it is difficult to generalize our findings, even in this sample with participants from fairly homogeneous backgrounds. Our findings also suggest that, despite our intensive data collection, the sample sizes for the prediction models may have been a limiting factor Thus, future research using individual-level tailored models to predict mood within individuals based using similar methods and larger sample sizes are recommended.
Another limitation of our study comes from the mixed design incorporating both event-based and signal-based ESM for self-reporting social behavior. In the signal-based format, participants retrospectively reported all social interactions since the last questionnaire during daily semi-random assessments. During the event-based format, participants reported social interactions immediately after they occurred. Each format was implemented for a duration of two weeks. It is possible that participants’ reporting behavior may have varied after two weeks, potentially changing the relationship between ESM predictors and mood. However, we employed a rolling window cross-validation strategy, which involved retraining the model. Thus, the model is likely to capture changes in participants' response patterns. Hence, while there might have been a temporary decline in performance during the transition period from one reporting format to the other, it is unlikely to have significantly impacted the overall predictive accuracy.
We employed various self-reported questionnaires to measure social behavior and mood. One limitation of these measures is their burden on participants, which could potentially impact the validity of the collected data (e.g., Asselbergs et al., 2016; Eisele et al., 2020). In contrast, passive smartphone measures are a less burdensome alternative. For future research, it would be interesting to investigate how the respondents’ motivation to complete questionnaires influenced the validity of the data and subsequently affected the predictive performance compared to using only passive measures.
One strength of our study is that we conducted various robustness checks using different datasets, which enhances the reliability of our findings. However, it is important to note that we applied certain preprocessing steps without investigating their impact on the results. Specifically, data quality and missing data are important considerations when using digital phenotyping data (Bähr et al., 2022; Niemeijer et al., 2023; Roos et al., 2023). For example, different devices than those tested while developing the app may lead to missing data or third-party apps may interfere with the app used to passively collect data (Bähr et al., 2022). In this study, we labeled passive smartphone measures as missing if the corresponding sensor was not recorded for 24 h (choosing a different time window may produce different outcomes). Nevertheless, the data coverage in our sample was rather high, with minimal missing data due to potential problems with the app (data coverage per hour above 85%, see Supplementary Material section Data Quality). Therefore, we do not anticipate any significant changes in the results. Furthermore, when using a larger level of aggregation (i.e., 24 h), it could be argued that the first day of data should be excluded since participants were unable to log their social behavior prior to the study start (i.e., the first ESM measurement). However, we made the decision to utilize all available data to not lose any valuable information. As a result, it is possible that the first mood assessments include unrecorded social interactions that took place before the study started. Nevertheless, considering that most participants reported few interactions during the study, we believe that excluding the first day of data would not have significantly impacted the predictive performance.
Conclusion
Our study highlights the challenges with accurately predicting mood based on social behavior and suggest that employing a combination of measures yields higher predictive accuracy. We demonstrated the substantial influence of the chosen time scale for aggregating predictors and analyzing data on predictive accuracy, which supports calls for future research to determine appropriate time scales for psychological constructs a priori. We encourage future research that integrates across multiple measures (i.e., ESM, egocentric networks, digital phenotyping) to improve our understanding of how social behavior impacts mood and well-being in everyday life.
Notes
We are using Shapley values due to their ability to handle categorical and noncategorical predictors without introducing significant biases (Henninger et al., 2023). Shapley values can be computed in both the training and testing sets, and a recent study found that both approaches produce similar results (Borup et al., 2022). A drawback of using Shapley values is that they only measure a variable's contribution relative to the mean prediction, without directly assessing the actual prediction performance. Recently, a new method has been developed, which calculates the Shapley value based on the actual prediction performance (Borup et al., 2022).
It is important to distinguish between robustness checks, such as using different datasets, and parameters that are expected to impact the model's performance, such as level of aggregation and window size. Ideally, the former should not result in different performances (and any improved performance based on a particular robustness check is not interpreted as it being a better model), but the latter may and should be considered as hyperparameters to be fine-tuned. In this section, we refer to the use of different datasets based on different preprocessing steps, which ideally should not change the performance of the models.
References
Abdullah, S., Matthews, M., Frank, E., Doherty, G., Gay, G., & Choudhury, T. (2016). Automatic detection of social rhythms in bipolar disorder. Journal of the American Medical Informatics Association, 23(3), 538–543. https://doi.org/10.1093/jamia/ocv200
Asendorpf, J. B., & Rauthmann, J. F. (2020). States and situations, traits and environments. In G. Matthews & P. J. Corr (Eds.), The Cambridge handbook of personality psychology (2nd ed., pp. 56–68). Cambridge University Press.
Asselbergs, J., Ruwaard, J., Ejdys, M., Schrader, N., Sijbrandij, M., & Riper, H. (2016). Mobile phone-based unobtrusive ecological momentary assessment of day-to-day mood: An explorative study. Journal of Medical Internet Research, 18(3), e72. https://doi.org/10.2196/jmir.5505
Bähr, S., Haas, G.-C., Keusch, F., Kreuter, F., & Trappmann, M. (2022). Missing data and other measurement quality issues in mobile geolocation sensor data. Social Science Computer Review, 40(1), 212–235. https://doi.org/10.1177/0894439320944118
Benoit, J., Onyeaka, H., Keshavan, M., & Torous, J. (2020). Systematic review of digital phenotyping and machine learning in psychosis spectrum illnesses. Harvard Review of Psychiatry, Publish Ah. https://doi.org/10.1097/HRP.0000000000000268
Blanco, C., Okuda, M., Wright, C., Hasin, D. S., Grant, B. F., Liu, S.-M., & Olfson, M. (2008). Mental health of college students and their non-college-attending peers: Results from the National Epidemiologic Study on Alcohol and Related Conditions. Archives of General Psychiatry, 65(12), 1429–1437. https://doi.org/10.1001/archpsyc.65.12.1429
Borup, D., Goulet Coulombe, P., Rapach, D., Schütte, E. C. M., & Schwenk-Nebbe, S. (2022). The anatomy of outof- sample forecasting accuracy.
Breiman, L. (2001). Random forests. Machine Learning, 45(1), 5–32. https://doi.org/10.1023/A:1010933404324
Bringmann, L. F., Elmer, T., & Eronen, M. I. (2022). Back to basics: The importance of conceptual clarification in psychological science. Current Directions in Psychological Science, 31(4), 340–346. https://doi.org/10.1177/09637214221096485
Brown, L. H., Strauman, T., Barrantes-Vidal, N., Silvia, P. J., & Kwapil, T. R. (2011). An experience-sampling study of depressive symptoms and their social context. Journal of Nervous and Mental Disease, 199(6), 403–409. https://doi.org/10.1097/NMD.0b013e31821cd24b
Bulteel, K., Mestdagh, M., Tuerlinckx, F., & Ceulemans, E. (2018). VAR(1) based models do not outpredict AR(1) models in current psychological. Psychological Methods, 23, 740–756. https://doi.org/10.1037/met0000178
Burns, R. J., Deschênes, S. S., & Schmitz, N. (2016). Associations between depressive symptoms and social support in adults with diabetes: Comparing directionality hypotheses with a longitudinal cohort. Annals of Behavioral Medicine: A Publication of the Society of Behavioral Medicine, 50(3), 348–357. https://doi.org/10.1007/s12160-015-9760-x
Cai, L., Boukhechba, M., Wu, C., Chow, P. I., Teachman, B. A., Barnes, L. E., & Gerber, M. S. (2018). State affect recognition using smartphone sensing data. In Proceedings of the 2018 IEEE/ACM international conference on connected health: Applications, systems and engineering technologies (pp. 120–125). https://doi.org/10.1145/3278576.3284386
Cloos, L., Ceulemans, E., & Kuppens, P. (2023). Development, validation, and comparison of self-report measures for positive and negative affect in intensive longitudinal research. Psychological Assessment, 35, 189–204. https://doi.org/10.1037/pas0001200
Čolić, J., Bassett, T. R., Latysheva, A., Imboden, C., Bader, K., Hatzinger, M., Mikoteit, T., Lieb, R., Gloster, A. T., & Hoyer, J. (2020). Depersonalization and derealization in embarrassing social interactions: An experience sampling study in social phobia, major depression and controls. Journal of Anxiety Disorders, 70, 102189. https://doi.org/10.1016/j.janxdis.2020.102189
Currey, D., & Torous, J. (2022). Digital phenotyping correlations in larger mental health samples: Analysis and replication. Bjpsych Open, 8(4), e106. https://doi.org/10.1192/bjo.2022.507
Davidson, B. I. (2022). The crossroads of digital phenotyping. General Hospital Psychiatry, 74, 126–132. https://doi.org/10.1016/j.genhosppsych.2020.11.009
Eisele, G., Vachon, H., Lafit, G., Kuppens, P., Houben, M., Myin-Germeys, I., & Viechtbauer, W. (2020). The effects of sampling frequency and questionnaire length on perceived burden, compliance, and careless responding in experience sampling data in a student population. Assessment. https://doi.org/10.1177/1073191120957102
Eskes, P., Spruit, M., Brinkkemper, S., Vorstman, J., & Kas, M. J. (2016). The sociability score: App-based social profiling from a healthcare perspective. Computers in Human Behavior, 59, 39–48. https://doi.org/10.1016/j.chb.2016.01.024
Flake, J. K., & Fried, E. I. (2020). Measurement schmeasurement: Questionable measurement practices and how to avoid them. Advances in Methods and Practices in Psychological Science, 3(4), 456–465. https://doi.org/10.1177/2515245920952393
Fulford, D., Mote, J., Gonzalez, R., Abplanalp, S., Zhang, Y., Luckenbaugh, J., Onnela, J.-P., Busso, C., & Gard, D. E. (2020). Smartphone sensing of social interactions in people with and without schizophrenia. Journal of Psychiatric Research. https://doi.org/10.1016/j.jpsychires.2020.11.002
Fulford, D., Mote, J., Gonzalez, R., Abplanalp, S., Zhang, Y., Luckenbaugh, J., Onnela, J.-P., Busso, C., & Gard, D. E. (2021). Smartphone sensing of social interactions in people with and without schizophrenia. Journal of Psychiatric Research, 137, 613–620. https://doi.org/10.1016/j.jpsychires.2020.11.002
Gama, J., Žliobaitė, I., Bifet, A., Pechenizkiy, M., & Bouchachia, A. (2014). A survey on concept drift adaptation. ACM Computing Surveys, 46(4), 1–37. https://doi.org/10.1145/2523813
Google Play Store Team. (2020). Google Play Store Team, 2020. Select a category and tags for your app or game [WWW Document]. Google Play Console Help. https://support.google.com/googleplay/android-developer/answer/9859673?hl=en&visit_id=637860381831546735-765826470&rd=1
Hall, J. A. (2018a). When is social media use social interaction? Defining mediated social interaction. New Media & Society, 20(1), 162–179. https://doi.org/10.1177/1461444816660782
Hall, J. A. (2018b). Energy, episode, and relationship: A test of communicate bond belong theory. Communication Quarterly, 66(4), 380–402. https://doi.org/10.1080/01463373.2017.1411377
Harpham, T. (2002). Measuring social capital within health surveys: Key issues. Health Policy and Planning, 17(1), 106–111. https://doi.org/10.1093/heapol/17.1.106
Hart, T., Hoffman, J. M., Pretz, C., Kennedy, R., Clark, A. N., & Brenner, L. A. (2012). A longitudinal study of major and minor depression following traumatic brain injury. Archives of Physical Medicine and Rehabilitation, 93(8), 1343–1349. https://doi.org/10.1016/j.apmr.2012.03.036
Hastie, T., Friedman, J., & Tibshirani, R. (2009). The elements of statistical learning: Data mining, inference, and prediction. Springer. https://doi.org/10.1007/978-0-387-21606-5
Henninger, M., Debelak, R., Rothacher, Y., & Strobl, C. (2023). Interpretable machine learning for psychological research: Opportunities and pitfalls. Psychological Methods.
Hewamalage, H., Ackermann, K., & Bergmeir, C. (2022). Forecast evaluation for data scientists: Common pitfalls and best practices. Data Mining and Knowledge Discovery. https://doi.org/10.1007/s10618-022-00894-5
Houben, M., Van Den Noortgate, W., & Kuppens, P. (2015). The relation between short-term emotion dynamics and psychological well-being: A meta-analysis. Psychological Bulletin, 141, 901–930. https://doi.org/10.1037/a0038822
Huckvale, K., Venkatesh, S., & Christensen, H. (2019). Toward clinical digital phenotyping: A timely opportunity to consider purpose, quality, and safety. NPJ Digital Medicine, 2(1), 88. https://doi.org/10.1038/s41746-019-0166-1
Jacobson, N. C., & Bhattacharya, S. (2022). Digital biomarkers of anxiety disorder symptom changes: Personalized deep learning models using smartphone sensors accurately predict anxiety symptoms from ecological momentary assessments. Behaviour Research and Therapy, 149, 104013. https://doi.org/10.1016/j.brat.2021.104013
Jacobson, N. C., & Chung, Y. J. (2020). Passive sensing of prediction of moment-to-moment depressed mood among undergraduates with clinical levels of depression sample using smartphones. Sensors (switzerland), 20(12), 1–16. https://doi.org/10.3390/s20123572
Jacobson, N. C., Summers, B., & Wilhelm, S. (2020). Digital biomarkers of social anxiety severity: Digital phenotyping using passive smartphone sensors. Journal of Medical Internet Research, 22(5), 1–10. https://doi.org/10.2196/16875
Jagesar, R. R., Vorstman, J. A., & Kas, M. J. (2021). Requirements and operational guidelines for secure and sustainable digital phenotyping: Design and development study. Journal of Medical Internet Research, 23(4), e20996. https://doi.org/10.2196/20996
Kirtley, O. J., Lafit, G., Achterhof, R., Hiekkaranta, A. P., & Myin-Germeys, I. (2021). Making the black box transparent: A template and tutorial for registration of studies using experience-sampling methods. Advances in Methods and Practices in Psychological Science, 4(1), 2515245920924686.
Krach, S., Paulus, F., Bodden, M., & Kircher, Ti. (2010). The rewarding nature of social interactions. Frontiers in Behavioral Neuroscience. https://doi.org/10.3389/fnbeh.2010.00022
Kubey, R., Larson, R., & Csikszentmihalyi, M. (1996). Experience sampling method applications to communication research questions. Journal of Communication, 46(2), 99–120. https://doi.org/10.1111/j.1460-2466.1996.tb01476.x
Kuhn, M. (2008). Building predictive models in R using the caret package. Journal of Statistical Software, 28, 1–26.
Langener, A. M., Stulp, G., Kas, M. J., & Bringmann, L. F. (2023). Capturing the dynamics of the social environment through experience sampling methods, passive sensing, and egocentric networks: Scoping review. JMIR Mental Health, 10(1), e42646. https://doi.org/10.2196/42646
LiKamWa, R., Liu, Y., Lane, N. D., & Zhong, L. (2013). MoodScope: Building a mood sensor from smartphone usage patterns. In Proceeding of the 11th annual international conference on mobile systems, applications, and services (pp. 389–402). https://doi.org/10.1145/2462456.2464449
McNeish, D., & Wolf, M. G. (2020). Thinking twice about sum scores. Behavior Research Methods, 52(6), 2287–2305. https://doi.org/10.3758/s13428-020-01398-0
Mills, J., Fuller-Tyszkiewicz, M., & Holmes, M. (2014). State body dissatisfaction and social interactions: An experience sampling study. Psychology of Women Quarterly, 38(4), 551–562. https://doi.org/10.1177/0361684314521139
Molnar, C. (2018). Interpretable machine learning (2nd edn.). Leanpub. https://christophm.github.io/interpretable-ml-book/
Myin-Germeys, I., & Kuppens, P. (2022). The open handbook of experience sampling methodology: A step-by-step guide to designing, conducting, and analyzing ESM studies. Center for Research on Experience Sampling and Ambulatory Methods Leuven.
Nahum-Shani, I., Smith, S. N., Spring, B. J., Collins, L. M., Witkiewitz, K., Tewari, A., & Murphy, S. A. (2017). Just-in-time adaptive interventions (JITAIs) in mobile health: Key components and design principles for ongoing health behavior support. Annals of Behavioral Medicine: A Publication of the Society of Behavioral Medicine, 52(6), 446–462. https://doi.org/10.1007/s12160-016-9830-8
Niemeijer, K., Mestdagh, M., & Kuppens, P. (2022). Tracking subjective sleep quality and mood with mobile sensing: Multiverse study. Journal of Medical Internet Research, 24(3), e25643. https://doi.org/10.2196/25643
Niemeijer, K., Mestdagh, M., Verdonck, S., Meers, K., & Kuppens, P. (2023). Combining experience sampling and mobile sensing for digital phenotyping with m-path sense: Performance study. JMIR Formative Research, 7(1), e43296. https://doi.org/10.2196/43296
Perry, B. L., & Pescosolido, B. A. (2012). Social network dynamics and biographical disruption: The case of “first-timers” with mental illness. American Journal of Sociology, 118(1), 134–175. https://doi.org/10.1086/666377
Perry, B. L., Pescosolido, B. A., & Borgatti, S. P. (2018). Egocentric network analysis: Foundations, methods, and models (Vol. 44). Cambridge: Cambridge University Press.
Phongsavan, P., Chey, T., Bauman, A., Brooks, R., & Silove, D. (2006). Social capital, socio-economic status and psychological distress among Australian adults. Social Science and Medicine, 63(10), 2546–2561. https://doi.org/10.1016/j.socscimed.2006.06.021
Rocca, R., & Yarkoni, T. (2021). Putting psychology to the test: Rethinking model evaluation through benchmarking and prediction. Advances in Methods and Practices in Psychological Science, 4(3), 25152459211026864. https://doi.org/10.1177/25152459211026864
Roos, Y., Krämer, M. D., Richter, D., Schoedel, R., & Wrzus, C. (2023). Does your smartphone “know” your social life? A methodological comparison of day reconstruction, experience sampling, and mobile sensing. Advances in Methods and Practices in Psychological Science, 6(3), 25152459231178736. https://doi.org/10.1177/25152459231178738
Sano, A., Taylor, S., McHill, A. W., Phillips, A. J., Barger, L. K., Klerman, E., & Picard, R. (2018). Identifying objective physiological markers and modifiable behaviors for self-reported stress and mental health status using wearable sensors and mobile phones: Observational study. Journal of Medical Internet Research, 20(6), e210. https://doi.org/10.2196/jmir.9410
Schoedel, R., Oldemeier, M., Bonauer, L., & Sust, L. (2022). Systematic categorisation of 3,091 smartphone applications from a large-scale smartphone sensing dataset. Journal of Open Psychology Data. https://doi.org/10.5334/jopd.59
Shah, R. V., Grennan, G., Zafar-Khan, M., Alim, F., Dey, S., Ramanathan, D., & Mishra, J. (2021). Personalized machine learning of depressed mood using wearables. Translational Psychiatry, 11(1), 338. https://doi.org/10.1038/s41398-021-01445-0
Stadel, M., Stulp, G., Langener, A. M., Elmer, T., van Duijn, M. A. J., & Bringmann, L. F. (2023). Feedback about a person’s social context—Personal networks and daily social interactions. Administration and Policy in Mental Health and Mental Health Services Research. https://doi.org/10.1007/s10488-023-01293-8
Stadel, M., van Duijn, M., Wright, A. G., Bringmann, L., & Elmer, T. (2022). Considering the ‘With Whom’: Differences between event-and signal-contingent ESM data of person-specific social interactions. https://doi.org/10.31234/osf.io/dhbw4
Stanislaus, S., Vinberg, M., Melbye, S., Frost, M., Busk, J., Bardram, J. E., Kessing, L. V., & Faurholt-Jepsen, M. (2020). Smartphone-based activity measurements in patients with newly diagnosed bipolar disorder, unaffected relatives and control individuals. International Journal of Bipolar Disorders, 8(1), 32. https://doi.org/10.1186/s40345-020-00195-0
Stevens, L. M., Mortazavi, B. J., Deo, R. C., Curtis, L., & Kao, D. P. (2020). Recommendations for reporting machine learning analyses in clinical research. Circulation Cardiovascular Quality and Outcomes, 13(10), e006556.
Štrumbelj, E., & Kononenko, I. (2014). Explaining prediction models and individual predictions with feature contributions. Knowledge and Information Systems, 41(3), 647–665. https://doi.org/10.1007/s10115-013-0679-x
Tashman, L. J. (2000). Out-of-sample tests of forecasting accuracy: An analysis and review. International Journal of Forecasting, 16(4), 437–450. https://doi.org/10.1016/S0169-2070(00)00065-0
Teo, A. R., Marsh, H. E., Liebow, S. B. L., Chen, J. I., Forsberg, C. W., Nicolaidis, C., Saha, S., & Dobscha, S. K. (2018). Help-seeking on Facebook versus more traditional sources of help: Cross-sectional survey of military veterans. Journal of Medical Internet Research, 20(2), e62. https://doi.org/10.2196/jmir.9007
Torous, J., Kiang, M. V., Lorme, J., & Onnela, J.-P. (2016). New tools for new research in psychiatry: A scalable and customizable platform to empower data driven smartphone research. JMIR Mental Health, 3(2), e16. https://doi.org/10.2196/mental.5165
Verachtert, R., Michiels, L., & Goethals, B. (2022). Are we forgetting something? Correctly evaluate a recommender system with an optimal training window. In Proceedings of the Perspectives on the Evaluation of Recommender Systems Workshop (PERSPECTIVES) at RecSys22, Seattle, WA, USA
Zheng, Y., Zhang, L., Xie, X., & Ma, W.-Y. (2009). Mining correlation between locations using human location history. In Proceedings of the 17th ACM SIGSPATIAL international conference on advances in geographic information systems (p. 475). https://doi.org/10.1145/1653771.1653847
Acknowledgement
We thank Marilyn Piccirillo for her comments on the manuscript and for proofreading it.
Funding
Funding was provided by Nederlandse Organisatie voor Wetenschappelijk Onderzoek (VI.Vidi.201.119, NWO-Veni 191G.037).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
We preregistered and conducted our study (https://osf.io/vp8ar/) based on the template for registering experience sampling method studies (Kirtley et al., 2021) and on the recommendations for reporting machine learning analyses in clinical research (Stevens et al., 2020). To make our research as open as possible, all of our code is available on GitHub (https://github.com/AnnaLangener/CombiningMethods_MoodPredictions). Due to the collection of highly sensitive and private information, including raw GPS data and the names of social interaction partners, data sharing is not feasible. We have no competing interests to declare.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Below is the link to the electronic supplementary material.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Langener, A.M., Bringmann, L.F., Kas, M.J. et al. Predicting Mood Based on the Social Context Measured Through the Experience Sampling Method, Digital Phenotyping, and Social Networks. Adm Policy Ment Health 51, 455–475 (2024). https://doi.org/10.1007/s10488-023-01328-0
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10488-023-01328-0