Introduction

Humans are a social species. Therefore, human memory formation, as much other cognitive activity, often occurs in social contexts. Yet, scientific memory research has paid limited attention to the social situatedness of memory encoding, taking memory operations devoid of contextual embedding as its point of departure1, thereby delineating many key principles. Meanwhile, contextual binding in space and time has proven crucial for successful episodic memory formation2. The social encoding context has, by comparison, received little attention, although recently, social influences on memory systems have attracted scientific interest. These range from investigations into brain mechanisms of social memory conformity3 to the finding that the hippocampus, a critical region for spatial navigation and episodic memory formation, also represents a "social space" of relationships4. Such findings lend momentum to the hypothesis that the social context in which items are encoded affects their subsequent remembering.

Memory research also initially focused on content-general mechanisms. However, human memory differs depending on the to be encoded material and several content-related factors have been shown to affect episodic memory. For instance, ample evidence indicates that emotionally relevant material is often remembered differently and often better than neutral material5,6, which may be due to both higher accuracy and more pronounced response bias for emotional contents7,8. Moreover, such emotion effects may be driven by arousal, equally affecting positive and negative contents, or by positive or negative valence5,9.

Next to emotional significance, the material's self-reference also plays an important role10,11. Whereas emotional significance is typically evaluated via normative ratings, self-reference is typically experimentally induced by a corresponding processing instruction on individually presented words, which is known to enhance recognition or recall. This "self-reference effect" is attributed to elaborative stimulus processing and deep encoding12, similar to other levels of processing manipulations, such as concreteness decisions13. Setting self-referential processing apart from other encoding manipulations, recent theories propose that the "self" acts as a hub, enabling efficient integration and storage of self-relevant information11. Self-referential processing in social contexts is so far hardly understood. However, being confronted with social evaluations from somebody else likely creates a powerful social context, inducing self-referential updating to detect discrepancies with one's self-concept, adapt it, or reject the information [see14]. In healthy people, such updating processes are often positively-biased, inducing more elaborative processing of self-serving (rather than merely self-relevant) information15. Self-reference and emotion appear to interact and engage at least partly overlapping neural systems16.

On the neural level, event-related brain potentials have been used to study mechanisms of stimulus encoding under various processing demands, ranging from free viewing and incidental encoding17, over levels of processing instructions18, and stimulus appraisals5, to self-referential processing19 or processing in evaluative social contexts20. Early posterior negative (EPN) potentials appearing over the visual cortex around 200 ms post-stimulus generally index feature-based attention and conceptual stimulus encoding, which can gate episodic memory encoding21. EPN potentials have been found to co-vary with task-driven attention deployment22, emotionally motivated attention23, self-reference19, but also contextually induced social relevance20. A series of later surface-positive potentials occurring over frontal and parietal brain areas index episodic memory encoding proper [e.g., see24]. The parietal part of these effects, typically known as late positive potential (LPP) and occurring from about 400 ms post stimulus-onset, is also sensitive to explicit emotional appraisal5 and self-reference19. Moreover, higher late positive potentials during emotional appraisal predict better subsequent memory for the respective material5. Whereas P3b-like late parietal positivities can vary in topography and functional significance, they often have overlapping generators, at least some of which are likely to reflect memory formation25,26. Fronto-parietal slow waves after about 800 ms are generally assumed to reflect strategic verbal memory encoding27.

Our own previous work showed that social-evaluative context considerably impacts amplitudes of both early negative and late positive brain potentials. When identical trait adjectives were presented as personality feedback in virtual interaction setups, EPN and LPP brain potentials were considerably larger when feedback seemed to come from more relevant interaction partners20,28. However, no previous studies investigated whether the putative social-evaluative context in which material was presented had any long-term memory consequences.

The present research aims to fill this gap, investigating the influence of perceived social context on memory formation. We study the effects of a social-evaluative encoding context, which previously yielded large electrophysiology effects, on long-term recognition memory of positive, negative, and neutral trait adjectives. We present the exact same word lists in three different encoding contexts, measuring event-related brain potentials at encoding and recognition memory one week later. In detail, we compare a group that incidentally encoded words in a social-evaluative feedback context with two other encoding groups. As in previous work20,28,29,30 the social-evaluative feedback context was established by asking participants to introduce themselves briefly in front of a camera and telling them that they would be later evaluated on the basis of their self-introduction, either by another person next door or by a computer system. To disclose the evaluations, the participants would later see trait adjectives on a computer screen and given the opportunity to agree or disagree with the evaluations. In a second group, participants were instructed to learn the same trait adjectives intentionally, the material being divided into two blocks, and in a third group, participants incidentally encoded the words via a levels-of processing manipulation. The levels-of-processing manipulation also consisted of two blocks, comprising self-referential processing, where participants could decide whether they found an adjective self-descriptive and a concreteness/abstractness decision without self-reference. Behaviorally, we expected self-reference judgments to be positively biased, with participants accepting positive adjectives more often as feedback (Experiment 1) and judging them more often as self-descriptive (Experiment 3) than negative or neutral adjectives. We also expected recognition memory to be better for emotional (in the social context and the self-reference condition particularly positive) than for neutral adjectives. On the electrophysiology level, we expected EPN and LPP amplitudes to be enhanced by emotional content as well as social context. Specifically, we investigated how the different encoding contexts, namely the social feedback manipulation, the learning instruction, and the levels-of-processing tasks, compared regarding recognition memory and EPN, LPP, and positive slow wave amplitudes.

Results

Behavior during the encoding task

Social-feedback group (feedback acceptance)

Positive feedback was more often accepted than negative and neutral feedback (ps < 0.001), while neutral feedback was also more often accepted by participants than negative feedback (p = 0.001) as reflected by a main effect of emotional content (F(2,46) = 62.99, p < 0.001, ηp2 = 0.733) and subsequent follow-up tests. A smaller main effect of feedback sender (F(1,23) = 5.12, p = 0.033, ηp2 = 0.182) revealed that participants were more likely to accept feedback from the more relevant (“human”) sender. No interaction was found (F(2,46) = 0.48, p = 0.622, ηp2 = 0.020). Reaction times were 913 ms on average. While they did not differ significantly between sender conditions (F(1,19) = 4.09, p = 0.057, ηp2 = 0.177), they were descriptively longer for decisions on human feedback (999 ms) compared to computer feedback (827 ms). There were no significant differences between emotion conditions and no interaction (Fs < 0.81, ps > 0.452).

Levels-of-processing group (self-descriptiveness and concreteness)

Here, effects of task (F(1,23) = 21.22, p < 0.001, ηp2 = 0.480), emotion (F(2,46) = 16.25, p < 0.001, ηp2 = 0.414), and their interaction were observed (F(2,46) = 23.31, p < 0.001, ηp2 = 0.503). Within the self-reference task participants selected significantly more positive than both neutral (p < 0.001) and negative (p < 0.001) words as self-descriptive, neutral and negative words not differing (p = 1.0), resulting in a large emotion effect (F(2,46) = 30.61, p < 0.001, ηp2 = 0.571). In the concreteness task, there was no effect of emotional content on concrete/abstract decisions (F(2,46) = 1.05, p = 0.359, ηp2 = 0.044). Reaction times were 1110 ms on average and differed between conditions (F(1,23) = 12.76, p = 0.002, ηp2 = 0.357), with longer reaction times during concreteness decisions (1258 ms) than during self-reference decisions (963 ms). There were no differences between emotion conditions and no interaction (Fs < 0.23, ps > 0.797).

Recognition memory (one week later)

Recognition memory data are shown in Fig. 1 and further detailed in Table 1 and Supplement A.

Figure 1
figure 1

Behavioral data for the social-feedback, verbal-learning, and levels-of-processing group. Discrimination accuracy (Pr = Hits – False alarms) and response bias (Br = false alarms/(1-Pr)) are shown per emotional category and group. Error bars depict ± 1 Standard Error of the Mean.

Table 1 Memory data for the three groups.

Discrimination accuracy

As shown in Fig. 1, left panel, experimental groups differed in discrimination accuracy (F(2,65) = 7.57, p = 0.001, ηp2 = 0.189). The social-feedback group was significantly more accurate than both the verbal-learning (p = 0.007) and the levels-of-processing group (p < 0.001). The latter two groups did not differ (p = 0.403). Group and emotion did not interact (F(4,130) = 0.45, p = 0.776, ηp2 = 0.014), but a three-way interaction between group, encoding condition, and emotion (F(4,130) = 20.53, p < 0.001, ηp2 = 0.387) indicated that depending on the experiment, emotion effects differed between conditions (see Table 1). Within the social feedback group, discrimination accuracy did not differ between encoding during human compared to computer feedback for negative (p = 0.120), neutral (p = 0.330), or positive words (p = 0.774). Likewise, for the verbal-leraning group, no encoding differences between block A and block B were found for all three emotion conditions (ps > 0.441). Within the levels-of-processing group, encoding in the self-reference task compared to the concreteness task led to higher accuracy for positive words (p = 0.003) but lower accuracy for neutral words (p = 0.001). For negative words, accuracy did not differ between the self-referential and concreteness tasks (p = 0.113).

Response bias

Response bias differed between the groups (F(2,65) = 10.31, p < 0.001, ηp2 = 0.241; see Fig. 1, right panel). Both the social-feedback and the verbal-learning group responded more liberally than the levels-of-processing group (p < 0.001 and p = 0.002, respectively). Group interacted with emotion (F(4,130) = 9.09, p < 0.001, ηp2 = 0.218; see Table 1), suggesting specific bias differences between the groups. We compared the bias per emotion category, showing group differences for positive words (F(2,65) = 25.367, p < 0.001, ηp2 = 0.438), with a larger positive-bias in the social feedback group compared to the verbal-learning (p = 0.020) and levels-of-processing group (p < 0.001; see Fig. 1, right panel). There was also a more liberal responding to positive words in the verbal-learning than in the levels-of-processing group (p < 0.001). Response bias differed also for neutral words (F(2,65) = 7.03, p = 0.002, ηp2 = 0.178), showing no difference between the social feedback and verbal-learning group (p = 0.208), but more conservative responding in the levels-of-processing group compared to the other two groups (ps < 0.05). Finally, response bias group differences were also found for negative words (F(2,65) = 4.94, p = 0.010, ηp2 = 0.132), again showing no difference between the social feedback and verbal-learning group (p = 0.595), but more conservative responding in the levels-of-processing group compared to the other two groups (ps < 0.05). Finally, there was a three-way interaction between group, condition, and emotion concerning response bias (F(4,130) = 6.42, p = 0.001, ηp2 = 0.165). Follow-up investigations within groups revealed for the social-feedback no significant bias differences between encoding words in the human or computer block for positive (p = 0.882) and negative words (p = 0.053), but more liberal responding to neutral words encoded as human feedback (p = 0.032). For the verbal-learning group, bias did not differ between the three emotion conditions (ps > 0.288). For the levels-of-processing group, encoding words in the self-relevance compared to the concreteness task led to more liberal responding to positive (p = 0.004) and negative words (p = 0.019), but not for neutral words (p = 0.126).

Encoding ERPs

EPN (200–400)

For the EPN, an effect of experimental group was found (F(2,69) = 8.50, p < 0.001, ηp2 = 0.198; see Fig. 2a). The social-feedback group had a larger EPN than the verbal learning (p < 0.001), but not than the levels-of-processing group (p = 0.286). The latter showed also a larger EPN than the verbal-learning group (p = 0.005). Further, there was a main effect of emotion (F(2,138) = 6.65, p = 0.002, ηp2 = 0.088; see Fig. 2b) that was due to larger EPN for positive (p < 0.001) and negative (p = 0.020) words compared to neutral ones. Positive and negative words did not differ (p = 0.272). No interactions emerged (Fs < 0.95, ps > 0.438).

Figure 2
figure 2

EPN effects of (A) experimental group and (B) emotion. Scalp topographies depict the mean amplitude differences for the respective interval. ERPs show the time course averaged from highlighted sensors. Between-group difference potentials contain 95% bootstrap confidence intervals. Between-emotion difference-potentials contain 95% bootstrap confidence intervals of intra-individual differences. For bar charts, error bars show 95% confidence intervals.

LPP (400–800 ms)

Experimental groups differed also for the LPP (F(2,69) = 5.65, p = 0.005, ηp2 = 0.141; see Fig. 3a). The social-feedback group had much larger LPP amplitudes than the verbal-learning (p = 0.001), and somewhat larger ones than the levels-of-processing group (p = 0.070), the latter groups did not differ (p = 0.134). A main effect of emotion (F(2,138) = 3.71, p = 0.027, ηp2 = 0.051; see Fig. 3b) showed a larger LPP for positive compared to neutral (p = 0.031) and compared to negative words (p = 0.019), the latter two not differing (p = 0.908). An interaction of emotion and sensor ROI (region of interest) (F(4,138) = 4.50, p = 0.002, ηp2 = 0.115) revealed that the emotion effect was more pronounced over the parietal ROI (F(2,138) = 3.88, p = 0.023, ηp2 = 0.103) than over the frontal ROI (F(2,138) = 2.53, p = 0.083, ηp2 = 0.035). Similar to the global emotion effect, for the parietal cluster a larger LPP for positive than negative words (p = 0.007) was found. Neither positive (p = 0.180) nor negative words (p = 0.156) differed from neutral ones. A further four-way interaction occurred (F(4,138) = 3.48, p = 0.010, ηp2 = 0.092) which is resolved and illustrated in the Supplement—Section B. Sensor ROIs are further described in the methods section below.

Figure 3
figure 3

LPP and slow-wave effects of (A) experimental group and (B) emotion. Scalp topographies depict the mean amplitude differences for the respective interval. ERPs show the time course from averaged highlighted sensors (upper panel frontal ROI, lower panel parietal ROI). Respective difference potentials contain 95% bootstrap confidence intervals of the group differences. Emotion difference waves contain 95% bootstrap confidence intervals of intra-individual differences. For bar charts, error bars show 95% confidence intervals.

Slow-wave (800–1200 ms)

Groups also differed in their slow-wave amplitudes (F(2,69) = 3.57, p = 0.033, ηp2 = 0.094; see Fig. 3a). Here, the verbal-learning group had a larger positive-going slow-wave than the social-feedback (p = 0.021) and the levels-of-processing groups (p = 0.026), who did not differ (p = 0.933). No other main effects (F(2,138) = 1.08, p = 0.342, ηp2 = 0.015) or significant interactions emerged (Fs < 2.10, ps > 0.084).

Run 2/passive viewing re-presentation of word stimuli

After debriefing, a passive viewing run was designed to test for the immediate effects of the preceding experimental setup. There were no effects of the experimental group, emotion, or any interactions between these factors for the EPN, LPP, and the slow waves (Fs < 1.40, ps > 0.236).

Correlations between encoding ERPs and memory performance

Correlations showed the strongest relationships between ERPs and memory indices for LPP amplitudes (see Table 2). For the social-feedback group, parietal LPP amplitudes were related negatively to the false alarm rate. Bayes factor analysis revealed moderate evidence of a relationship for positive words. Although the levels-of-processing group showed descriptively a similar relationship pattern over parietal sensors, Bayes Factors did not support such a relationship. For the verbal-learning group, the strongest relationships of LPP amplitudes were found over the frontal cluster. Here, strong evidence was found for a relationship between ERP amplitudes during encoding and hit rate during recognition for neutral words. A similar pattern was found concerning negative and positive words, but Bayes Factors showed only anecdotal evidence of a relationship.

Table 2 Relationship of ERPs during encoding and memory performance.

Discussion

Humans typically acquire information in social contexts. Therefore, we investigated its role for long-term recognition memory and encoding ERPs for positive, negative, and neutral words. We compared a group for whom a social context was induced via a self-introduction followed by evaluative feedback ("social-feedback" group) with a verbal-learning and a standard levels-of-processing group. Although the social-feedback group was not aware of the recognition session one week later, this group showed superior recognition accuracy, but also an increased response bias, particularly for positive feedback adjectives. Encoding ERPs revealed the highest EPN and LPP amplitudes in the social-feedback group. Only in the slow-wave window were ERPs most positive-going in the verbal-learning group. ERPs from the levels-of-processing group fell in between the other two. Together, these findings reveal a distinct effect of evaluative social context, enhancing long-term recognition memory and encoding ERPs beyond both explicit learning and standard levels of processing manipulations, while also inducing specific biases.

The highest memory accuracy in the social feedback group indicates that the social context facilitated memory encoding, likely via elaborative stimulus processing which is known to enhance memory [e.g., see11,12,31]. Since explicit learning tasks are typically easier and result in better memory than incidental ones32,33, these findings highlight the power of social context to increase stimulus encoding.

Groups also differed in their response bias, with less conservative responses in the social-feedback and the verbal learning groups. Moreover, the social-feedback group was specifically biased towards classifying positive words as "old". Healthy participants are known to have optimistic feedback expectations34, leading to distorted memory when information is incongruent with their positive self-view35. In fact, better than expected evaluations induce positively biased self-updating36 as well as a specific feedback-related positivity37. Self-serving tendencies also occurred in feedback acceptance (social-feedback group) and self-descriptiveness decisions (levels-of-processing group), the highest acceptance rates for positive adjectives corresponding with positive self-views of healthy participants [e.g., see34].

Concerning ERPs, words in the social feedback context elicited the largest EPN and LPP amplitudes, followed by levels-of-processing and, finally, instructed learning. Per se, these ERP modulations are broadly in line with findings of increased EPN and LPP amplitudes for various manipulations of self-reference and social feedback19,38,39. Given the EPN's role in early attentional selection22, results show a rapid prioritizing role of social-evaluative context beyond mere task-relevance. The highest LPP in the social feedback group is also broadly in line with the view that P3b-like components, such as the LPP, index successful memory encoding25,26. Dolcos and Cabeza5 found that positive and negative items that elicited larger LPP than neutral ones during emotional appraisal were subsequently better remembered. Such items also elicited a larger fronto-central subsequent memory positivity than neutral ones. Whereas we have no subsequent memory ERP data for the present set of experiments, the much larger encoding positivities we found in the social context condition in parallel with the considerably higher recognition rates are certainly consistent with the view that late positive potentials reflect, at least in part, episodic memory encoding, although the subsequent memory effect provides a purer measure for this. Exploratory correlations also support a role for the LPP in memory formation (see Table 2). Topographic differences may indicate that the verbal learning group relied more on frontally mediated strategic mechanisms, particularly to support encoding of neutral items. However, given the modest sample sizes of each group, these relationships should be interpreted cautiously.

Only in the slow-wave window did the verbal-learning group exhibit more positive-going ERPs than the other two groups. Given the role of positive-going slow waves in strategic verbal memory encoding27, this finding might explain why the verbal-learning group had similar recognition accuracy (and even more hits) than the levels-of-processing group, despite having smaller amplitudes on the earlier ERPs. Therefore, encoding in the social feedback and levels-of-processing groups seems to have relied more on automatic mechanisms, whereas the verbal-learning group recruited more strategic processes.

As expected, significant emotion effects on brain potentials were also observed, although they were smaller than the context effects. Emotion sensitivity of EPN and LPP components aligns with much previous research [e.g., see17,40,41,42]. Largest amplitudes for positive words have also been reported before [see43]. Self-referential processing might have contributed to a more substantial increase for positive content, whereas in other situations, a bias towards negative material might occur [e.g., see40,41].

Overall, we show that social context during encoding enhances long-term recognition memory beyond explicit learning or typical deeper incidental encoding tasks (self-descriptiveness or concreteness decisions). Notably, although both the social feedback and self-descriptiveness task probably recruited self-referential processing, effects in the feedback condition, where any self-reference was socially contextualized, by far exceeded those of task-driven self-referential processing alone. This finding may give rise to questions regarding the role of the initial structured interview and self-introduction, which were designed to make the entire social context more salient and credible. However, this manipulation also bears some similarities with common stress induction methods. We observed no specific increases in state anxiety ratings compared to the other groups. Trait anxiety ratings did not differ either. Still, the present implementation might be regarded as a mild form of stress induction and may have induced arousal or led to secretion of stress hormones which we did not measure. The existing literature on memory effects of pre-encoding stress suggests that the present effects are unlikely solely due to pre-encoding stress. Previously reported effects of pre-encoding stress 30 min before encoding (which would be quite similar to the present timing) on recognition memory the next day reveal modest44, ambiguous45,46, and often also impairing effects47 which contrasts with the quite large effects that we found after a weeklong retention interval. Nevertheless, any endocrine effects of the present context induction should be addressed in future studies. Although the present retention interval of a week leaves ample time for consolidation or spontaneous rehearsal, which we cannot directly assess, the immediate repetition run revealed no significant effects, suggesting that effects were either due to initial encoding, where corresponding ERP modulations occurred, or arose considerably later.

The fact that responses in the social feedback and the levels-of-processing groups were self-paced may give rise to concerns regarding processing times. However, processing times were quite similar in these two experiments, and if anything, they were longer in the levels-of-processing group, yet memory performance was much higher in the social feedback group. The levels-of-processing group returned for testing on average half a day after the social feedback group, which might be another source of concern. Yet, there was no difference in testing lag between the verbal learning group and the social feedback group. The verbal learning group was the only group aware of the test, and they could use the interstimulus interval for rehearsal while the other two groups were performing the decision task. Still, the social feedback group clearly outperformed the other two groups regarding recognition memory.

Our focus was on the between-groups analysis, which de-emphasizes potentially interesting effects in individual conditions. Still, in Experiment 1, some "sender" differences occurred at encoding (see Supplement B), replicating previous research [e.g., see28]. However, these had little effect on long-term recognition, suggesting that participants integrated both blocks from experiment 1 into one episode whose items were considerably more memorable than those from the other two experiments. Interactions in the levels-of-processing group indicate that encoding task modulates emotion effects in long-term recognition memory, only self-descriptiveness and not concreteness decisions resulting in an emotion effect on recognition accuracy, in line with the suggestion of interacting memory effects of self and emotion16.

In sum, although we used identical stimuli and presentation parameters across three experiments, the psychological encoding contexts elicited pronounced between-group differences in recognition memory and ERPs: The social-feedback context enhanced recognition memory beyond both explicit verbal-learning and self-descriptiveness/concreteness judgments. It also induced the largest EPN and LPP amplitudes, followed by levels-of-processing and verbal learning. The most positive-going slow wave amplitudes in the verbal learning group suggest that this group might have relied particularly on strategic memory processes. Our findings specify some important social factors in human long-term memory. They resonate with the fact that humans typically acquire information in social contexts [e.g., see48], social evaluation being a particularly salient factor in human life [e.g., see49]. Finally, these results have implications for memory formation in educational settings or memory assessment in legal contexts, where the extent to which individuals were exposed to an evaluative social context will affect their memory for elements of the initial episode.

Methods

Participants

Three groups of twenty-four participants each were recruited at Bielefeld University. Participants provided written informed consent and received 10 Euros per hour for participation. The Ethics Committee at Bielefeld University approved the study and the study was performed in accordance with the regulations of the Declaration of Helsinki. All subjects were right-handed, had normal or corrected-to-normal vision, and were free from a self-reported neurologic or psychiatric disorder. They were tested twice with a lag of about one week (T1–T2 difference M = 7.13 days, Min = 6 days, Max = 9 days, see Table 3). Experimental groups did not differ concerning BDI scores (F(2,68) = 0.03, p = 0.973, ηp2 = 0.001), or STAI state (F(2,68) = 1.34, p = 0.269, ηp2 = 0.038) and trait anxiety scores (F(2,64) = 0.72, p = 0.492, ηp2 = 0.022). There was a significant effect of the T1-T2 difference (F(2,64) = 7.30, p = 0.001, ηp2 = 0.186), showing a significant longer latency of about half a day for the levels-of-processing group compared to both the social-feedback group (p = 0.004) and verbal-learning group (p < 0.001). The social-feedback and verbal-learning groups did not differ (p = 0.464). Four participants in total did not attend the second session. One of these was in the levels-of-processing group and three in the verbal-learning group.

Table 3 Demographic information for the participants in each experiment.

Stimuli

The stimulus set had been rated by 22 other students on nine-point Likert-type scales in terms of valence and arousal using the Self-Assessment Manikin52 and a similarly constructed concreteness scale. The selected 270 adjectives (90 negative, 90 neutral, 90 positive) were assigned to three separate lists. Neutral adjectives were allowed to deviate from emotional adjectives on both arousal and concreteness since truly neutral trait adjectives are rare in an interpersonal evaluative context. The three lists did not differ in relevant emotional and lexical properties, and the list-condition assignment was counterbalanced (see Table 4).

Table 4 Comparisons of negative, neutral, and positive adjectives for three separate lists by One-Way-ANOVAs.

Procedure

An outline of the experimental setup for the three groups is presented in Fig. 4.

Figure 4
figure 4

Example experimental and main experiment trial structure for the (a) social-feedback group, (b) verbal-learning, and (c) levels-of-processing group.

Social-feedback group. Upon arrival, participants were told that another, unknown person, would evaluate them based on their self-presentation. In a second condition, a randomly operating computer algorithm was supposedly going to give them feedback. All subjects underwent both "sender" conditions. Condition sequence and word lists were counterbalanced. Importantly, random feedback was presented in both conditions. Participants were instructed to briefly describe themselves in a structured interview in front of a camera. They were informed that the video of their self-description would be presented to another person. During EEG preparation, participants completed a demographic questionnaire and the BDI50 and STAI51 to characterize the sample. A research assistant left the testing room 15 min ahead of the fictitious feedback to ensure face validity, guiding an "unknown person" to a laboratory room next to the testing room.

Stimuli were presented by software described as "Interactional Behavioral Systems", supposedly allowing instant online communication. Participants were told that the unknown other person would select adjectives describing the participant, which would be subsequently presented. In the other condition, feedback was denoted as random computer feedback. Feedback-adjectives were presented for 1500 ms, after which the participants had to indicate via button-press whether or not they accept this evaluation. After the response, a fixation cross was presented for 2000 to 3500 ms. For each sender, 30 negative, 30 neutral, and 30 positive adjectives (one of the three lists, see Table 4) were presented twice. The desktop environment and stimulus presentation were created using Presentation (www.neurobehavioralsystems.com). Afterwards, participants were debriefed that no social evaluation had taken place and presented a passive viewing run of all 60 negative, 60 neutral, and 60 positive adjectives again to test for any immediate post-processing effects of the manipulation.

Participants were informed that an unrelated test would be conducted in a second session one week later, where an unexpected recognition test took place. All adjectives from the first session and a set of new items were shown for 1500 ms each, followed by an old/new decision and a fixation cross for 1500 to 2000 ms.

Verbal-learning group. The very same stimuli and presentation parameters were used. In order to keep the visual input constant, the same background presentation setup (Interactive Behavioral Systems) was used but never referred to. The self-introduction phase was omitted, and no social significance was assigned to the stimulation. Instead, participants were instructed to memorize all adjectives for recognition testing in the next session one week later. Adjectives were presented for 1500 ms, followed by a variable fixation cross presented for 2000 to 3500 ms. To mimic the "social-feedback" experiment, two counterbalanced conditions were presented (block "A" and block "B"), using the same material and number of repetitions (see Fig. 4b, d, Table 4). Condition order was counterbalanced. A passive viewing run finished the experimental session.

Levels-of-processing group. The same materials and presentation parameters as in the previous two experiments were used. As in experiment 1, active word evaluation was instructed, but no social context was created. Participants were told to either decide via button press on the self-descriptiveness of the presented words or, in condition two, perform a binary concreteness judgment on the words, classifying them as either abstract or concrete. This closely reflected the encoding trial structure for the social-feedback group. Adjectives were presented for 1500 ms, after which the binary decision was requested. Material and experimental parameters were the same as in the social feedback group (see Fig. 4c, d, Table 4). Again, the session was finished with a passive viewing run.

EEG recording and analyses

EEG was recorded from 128 BioSemi active electrodes (www.biosemi.com) at 1024 Hz. Two separate electrodes were used as ground electrodes, a Common Mode Sense active electrode (CMS) and a Driven Right Leg passive electrode (DLR), which form a feedback loop to measure the average potential close to the reference in the A/D-box (www.biosemi.com/faq/cms&drl.htm). Four additional electrodes (EOG) measured horizontal and vertical eye movement.

Pre-processing and statistical analyses were performed using BESA (www.besa.de) and EMEGS53. Offline, data was re-referenced to an average reference and filtered with a high-pass forward filter of 0.16 (6 db/oct) and a 30 Hz low-pass zero-phase filter (24 db/oct). Filtered data were segmented from 200 ms before word onset until 1500 ms after stimulus presentation. The 200 ms before stimulus onset were used for baseline correction. Eye-movements were corrected using the automatic correction method implemented in BESA54. Remaining artifacts were rejected based on an absolute threshold (< 120 µV), signal gradient (< 75 µV/∂T), and low signal (i.e., the SD of the gradient, > 0.01 µV/∂T). Noisy EEG sensors were interpolated using a spline interpolation procedure.

On average, 9.5 percent (12.2 in total) of electrodes were interpolated in the social feedback group, and 12 percent of trials were rejected, leading to 53 trials on average. There was no difference in rejected trials concerning encoding or emotion conditions (Fs < 1.12, ps > 0.302). In the verbal learning group, 3.5 percent (4.5 in total) of the electrodes were interpolated, and 12 percent of trials were rejected, leading to 53 trials on average. There was no difference in rejected trials concerning encoding or emotion conditions (Fs < 0.15, ps > 0.861). In the LoP group, 7.4 percent (9.4 in total) of electrodes were interpolated, and 24 percent of trials were rejected, resulting in 46 trials on average. There was no difference in rejected trials concerning emotion conditions (F(2,46) = 1.54, p = 0.225), but a difference concerning encoding conditions (F(1,23) = 10.01, p = 0.004) with 48 trials on average in the self-reference condition compared to 44 trials in the concreteness condition.

Statistical analyses

For all data, we calculated three (group: social-feedback, verbal-learning, levels-of-processing) by two (condition: human sender/block A/self-reference, computer sender/block B /concreteness) by three (emotion: positive, negative, neutral) mixed ANOVAs. For memory data, the discrimination index (Pr = hits-false alarms), and the response bias (Br = false alarms/(1 − Pr)) were calculated according to Snodgrass and Corwin's two-high-threshold model55. Br values of 0.5 indicate no response bias, while higher values indicate a liberal and lower values a conservative response strategy. Raw data are presented in Supplementary Materials Section A. Furthermore, for the social-feedback and levels-of-processing groups, decision rates (agreement, self-descriptiveness, concreteness) were assessed.

ERP data were analyzed and visualized with EMEGS53 and in-house Matlab scripts. Effect sizes were estimated as partial eta-squared56. When Mauchly's test indicated a violation of sphericity, degrees of freedom were corrected according to Greenhouse–Geisser. We report uncorrected degrees of freedom but corrected p-values and effect sizes. Time windows of interest were segmented from 200 to 400 ms for the EPN. The LPP was scored from 400 to 800 ms and the slow-wave from 800 to 1200 ms. For the EPN, a parieto-occipital sensor region of interest (ROI) of eighteen electrodes was examined (P9, P9h, P7, PO9, PO9h, PO7, I1, OI1, O1, P10, P10h, P8, PO10, PO10h, PO8, I2, OI2, O2). The LPP and the slow-wave were measured at a frontal and a posterior sensor ROI of seventeen electrodes each (frontal: F1, Fz, F2, FFC1, FFC1h, FFCz, FFC2h, FFC2, FC3h, FC1, FC1h, FCz, FC2h, FC2, FC4h, FCC1h, FCC2h; parietal: CCPz, CP3h, CP1, CP1h, CPz, CP2h, CP2, CP4h, CPP1, CPPz, CPP2, P1, Pz, P2, PPO1, PPOz, PPO2).

Finally, exploratory correlations between ERP amplitudes during encoding and memory performance were performed. To this end, mean ERP amplitudes were correlated with different memory indices (Hits and False Alarms (FA)) using JASP (www.jasp.org). We calculated both Bonferroni-corrected inferential and Bayesian Pearson correlation coefficients. For Bayesian analyses, the null hypothesis was specified as a point-null prior (i.e., standardized effect size δ = 0). It defined the alternative hypothesis as a Jeffreys-Zellner-Siow (JZS) prior, i.e., a folded Cauchy distribution centered around δ = 0 with the scaling factor r = 0.707. This scaling factor assumes a roughly normal distribution. To assign verbal labels to the strength of evidence, we followed the taxonomy suggested by Jeffreys57, labeling Bayes Factors with a BF10 of 1 as no evidence, BF10 between 1—3 as anecdotal evidence, 3—10 as moderate evidence, 10—30 as strong evidence, 30—100 as very strong evidence, and larger BFs as extreme evidence in favor of the alternative hypothesis.