Abstract
Background
The use of computerized devices for neuropsychological assessment (CNADs) as an effective alternative to the traditional pencil-and-paper modality has recently increased exponentially, both in clinical practice and research, especially due to the pandemic. However, several authors underline that the computerized modality requires the same psychometric validity as "in-presence" tests. The current study aimed at building and validating a computerized version of the verbal and non-verbal recognition memory test (RMT) for words, unknown faces and buildings.
Methods
Seventy-two healthy Italian participants, with medium–high education and ability to proficiently use computerized systems, were enrolled. The sample was subdivided into six groups, one for each age decade. Twelve neurological patients with mixed aetiology, age and educational level were also recruited. Both the computerized and the paper-and-pencil versions of the RMT were administered in two separate sessions.
Results
In healthy participants, the computerized and the paper-and-pencil versions of the RMT showed statistical equivalence for words, unknown faces and buildings. In the neurological patients, no statistical difference was found between the performance at the two versions of the RMT. A moderate-to-good inter-rater reliability between the two versions was also found in both samples. Finally, the computerized version of the RMT was perceived as acceptable by both healthy participants and neurological patients at System Usability Scale (SUS).
Conclusion
The computerized version of the RMT can be used as a reliable alternative to the traditional version.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
Introduction
In the past few decades, the use of computerized neuropsychological assessment devices (CNADs) has received increasing attention in both clinical and research settings, because it is considered as a more advantageous alternative to conventional examiner-based tests [1]. The interest in CNADs has exponentially increased in response to the COVID-19 pandemic, which has forced significant changes in clinical practice. Indeed, one of the properties of CNADs is to allow remote encounters between clinicians and patients [see 2 for review]. Furthermore, focusing on the fact that neuropsychological tests on digital platforms are sensitive and cost-effective, several studies have already used these tools with many clinical populations [3]. However, to date, "paper-and-pencil" tests remain the gold standard for neuropsychological assessment [4, 5]. The American Academy of Clinical Neuropsychology and the National Academy of Neuropsychology, in a joint document, defined CNADs as any tool that administers, evaluates, or interprets tests of brain function using a computer, portable devices (such as digital tablets or smartphones), or any other digital interface instead of a human examiner [3]. However, such a position statement, that emphasizes the administration of cognitive tests by a computer "instead of" a human examiner, can be extended to all computerized neuropsychological platforms, including a broader spectrum of ever-growing technology and, therefore, include tools that may allow a better presentation of the stimuli and data recording through computer presentation (such as immersive and non-immersive Virtual Reality) [4]. As aforementioned, CNADs allow to present with extreme accuracy a wider range of tasks including, for example, those involving multi-tasking, multiple measurements, divided attention, speed processing and response times for multiple measurements. Furthermore, CNADs provide new parameters, more easily and quickly available [as, for instance, recording of response latencies, see 6], leading to greater accuracy in detecting cognitive changes [7,8,9], in both the clinical and research fields [10, 11]. Finally, computerized tests can be administered to many subjects in shorter times and with lower costs; the evaluation of the responses can be more precise, enabling automatic administration and scoring procedures, and thus reducing the probability of human error [3]. On the other hand, in a conventional neuropsychological assessment, there can be a great deal of variation between different examiners or between the same examiner along an assessment session, in the manual use of tools (such as a stopwatch), or in the manual presentation of material (such as stimulus cards) [12]. Furthermore, while in an examiner-centred paper-and-pencil assessment the patient interacts with an examiner who presents stimuli, records responses, notes key behavioral observations and also understands the patient’s level of effort and motivation, in CNADs, the patient interacts with a computerized test station or a tablet through one or more alternative input/output devices (keyboard, voice, mouse, joystick, touchscreen, head-mounted display), sometimes without any supervision or observation. Additionally, computerized assessment does not permit examiners to introduce flexibility into their evaluation or provide any structured encouragements to the examinee. For those reasons, the use of CNADs may require familiarity with technical skills that can significantly impact on relevant evaluation parameters. Patients, especially the elderly, may have limited familiarity with computer interfaces, as well as negative attitudes and anxiety about computers [13]: as a result, their performance may differ significantly from pencil-and-paper measurements [14,15,16]. On these assumptions, computerized tests are not simply the replacement of paper-and-pencil for a computer screen and electronic response capture and they could not be considered tout court directly comparable to an examiner-administered evaluation [17,18,19]. Rather, a traditional test administered by computer becomes a new test, different from the conventional one both qualitatively and technically. Indeed, several studies have proved that there are substantial differences between the computerized measurements and the corresponding examiner administered tests in different samples [20,21,22,23]. In addition, the "replication crisis" [24] has increased the need for appropriate and robust statistical inference: when examiner-based tests are proposed in a computerized version, new psychometric data should be provided according to the same standards of psychometric tests. It cannot be assumed a priori that the normative data for a test administered by the examiner apply equally well to a computerized version of the same test, due to changes in the administration method and patients’ familiarity with the computer. Thus, if a computerized test adapted from a conventional test is not simply a slightly different format for an existing test, it becomes essential that equivalence testing between the paper-and-pencil and the computerized versions of that test is provided. New normative data for computerized tests must be collected when no equivalence with the paper-and-pencil version of the tests is proved. Several tests have already been adapted for computerized presentation, such as the Montreal Cognitive Assessment – MoCA [25], the Clock Drawing Test, the Pentagon Drawing Test [26], the Rey Complex Figure copy task [27], the Bells Cancellation, Line Bisection and Five Elements Drawing Tests [28]: these tests are, of course, critical in the assessment of cognitive functioning.
Among neuropsychological patients, memory problems are one of the most reported symptoms and, due to growing life expectancy, differentiating early neurodegenerative disorders from normal aging is for sure one of the greatest challenges for the future. Notably, despite the extensive body of literature supporting the use of recognition memory paradigms to enhance differential diagnosis between neurological patients, depressed ones and healthy controls [29], its clinical assessment is still limited due to the paucity of validated, clinical tools available. As for the Italian population, only the verbal and non-verbal recognition memory test (RMT) [30] and the visual long-term recognition memory test [31] are available. The RMT represents a relatively fast screening tool, which may be useful for the early detection of memory disorders; two parallel versions are also available for multiple testing [32]. However, despite its promising features, the original examiner-centred version of the RMT may be prone to human error: clinicians are indeed required to control for several factors, namely the duration of the stimuli presentation and the registration of patients’ responses and behaviours [see 30, methods section]. For all these reasons, we decided to provide a validation of a computerized version of the verbal and non-verbal recognition memory test (RMT—Form A) for words, unknown faces, and buildings [32], comparing it to the results of the original paper-and-pencil test. To this end, a computerized version of the RMT was built and we administered both the pencil-and-paper and the computerized version to different age-balanced groups of healthy participants, with medium–high education. Furthermore, in order to evaluate the possible impact of different neurological conditions on patients' ability to interact with the computer interface, the two version of the RMT were also administered to a small sample of neurological patients with mixed aetiology.
Materials and methods
Materials
The previously published battery of verbal and non-verbal recognition memory tests (RMT) [32] and its computerized version were used. The battery evaluates different components of recognition memory for words, unfamiliar faces, and unknown buildings. Stimuli selection and test construction are described elsewhere [30]. Briefly, the original, pencil-and-paper version had the same administration procedure for the three subtasks (i.e., for words, faces and buildings): during the study phase, the 30 target stimuli were individually displayed at a 3-s. interval per item and, for each stimulus, participants were asked to judge whether they like it or not to enhance an adequate level of attention. In the recognition phase, performed immediately after the end of the study phase, three stimuli were presented simultaneously: the target stimulus and two distractors. Participants were asked to recognize, among the three items, the previously seen one. Targets were distributed randomly to the top, bottom, or centre of the sheet, ensuring an equal number of placements in the three positions. The number of correct answers corresponds to the total score of each subtask (range: 0–30). Overall, the test lasts about 20 min.
Here we present a computerized version of the RMT. The test was built on PsychoPy and the stimuli were shown to the participants using a MacBook Pro 13" (with a set screen resolution of 1680 × 1050). The stimuli were the same as in Smirni et al.’s [30] original version and, in the study phase, each stimulus was displayed at a 3-s. interval. To ensure an adequate level of attention on each stimulus, participants were asked to press "1" on the keyboard if it produced any emotion or feeling (e.g. happiness, sadness, etc.), "0" if it did not produce any particular emotion or feeling (Fig. 1). With respect to the original one, in this computerized version, administration and scoring procedures are fully automatic: indeed, instructions are shown on the PC/laptop monitor and participants are invited to provide responses using some keys on the keyboard. Furthermore, after each subtest completion, an electronic sheet with all the answers provided by the participant, the raw score and its conversion into adjusted and equivalent scores is available. The system automatically applies the regression formulas of Smirni et al. [30] to the raw scores obtained by the participants, even in the case of minimum or maximum raw score (0/30 and 30/30 respectively). In agreement with Novelli et al. [33], in the former case the equivalent score is always zero, in the latter, the equivalent score is equal to four, disrespectfully from the foreseen automatic correction.
In addition, the participants were administered the Montreal Cognitive Assessment (MoCA) [34] to assess the general cognitive level, and the System Usability Scale (SUS) [35], which evaluates the subjective opinion from a user regarding a particular software. The SUS consists of 10 items in which even-numbered questions are posed in negative form. All the items need to be rated on a Likert scale ranging from 1 (completely disagree), to 5 (completely agree).
Methods
The two versions of the RMT (i.e., the original pencil-and-paper and the computerized versions) were administered by one of the authors (BF) to a sample of 72 healthy participants, in two testing sessions over two separate days, at least one week apart and within 10 days. The examiner was always present in the same room during the computerized administration of the test, but no further instructions or clarifications were requested from the examinees. The administration order of the two versions (first pencil-and-paper; first computerized), was counterbalanced across subjects as well as the administration order of the three subtasks. Inclusion criteria for participants were: i) age between 20 and 79 years; ii) at least 5 years of education. Exclusion criteria were: i) a history or clinical evidence of neurological or neuropsychiatric disease; ii) subjective complaint of memory or other cognitive impairments; iii) a Montreal Cognitive Assessment (MoCA) adjusted score below the Italian cut-off score [34]. The sample was then subdivided into six groups, one for each age decade; the number of males and females was matched across the six groups.
Furthermore, a continuous series of 12 neurological patients with mixed aetiology, age and educational level was recruited at the Centre for Cognitive Rehabilitation (CeRiN) between May and August 2023. The two versions of the test RMT were administered by two psychologists expert in neuropsychology (EB and ST), to evaluate the possible impact of different neurological conditions on patients' ability to interact with the computer interface. The administration order of the two versions was counterbalanced across the patients as well as the administration order of the three subtasks.
Written informed consent was obtained from all participants and the study was conducted in accordance with the Declaration of Helsinki, with the approval of the local ethics committee.
Socio-demographic characteristics of both healthy participants and neurological patients are shown in Table 1, together with the SUS scores.
Statistical analysis
Learning effect
A series of linear mixed effects models were run to check for possible effects of learning. The raw scores at the three RMT subtasks were considered as a continuous dependent variable. As a factorial variable, the version of test (Version, two levels: paper-and-pencil; computerized), the administration order of the two versions (Version order, two levels: first paper-and-pencil; first computerized), and the administration order of the three subtasks (Task order; six levels: faces-words-buildings; faces-buildings-words; buildings-faces-words; buildings-words-faces; words-faces-buildings; words-buildings-faces), were added to the model. A by subject random intercept was also added to account for inter-subject variability. The same analysis was separately performed for healthy participants and neurological patients.
Inter-rater reliability
Inter-rater reliability between the computerized and the paper-and-pencil versions of the RMT was separately assessed for healthy participants and neurological patients through intra-class correlations. Results were interpreted according to Koo and Li’s guidelines [36].
Healthy participants
Equivalence between the two versions of the RMT for faces, buildings and words was assessed through a two one-sided test procedure (TOST) for dependent samples [37]. Indeed, while the lack of a significant difference between scores on a paired t test cannot be taken as an evidence of agreement, equivalence tests demonstrate that scores from different tests are equal within a predetermined boundary. Normality assumptions on raw variables were checked by the evaluation of skewness and kurtosis values (judged as abnormal if ≥|1| and |3|, respectively [38]). Given the non-normality of the raw scores, data were reflected and log transformed prior to analyses with equivalence tests. The TOST procedure involves an a priori determination of an equivalence boundary (± delta), within which score difference can be considered trivial. Delta was set a priori at 1.1 raw score points, which log transforms to 0.1 and results in an equivalence interval of − 0.1 and 0.1. α level was set at 0.005.
Neurological patients
The presence of any difference in the performance between the two versions of the RMT for faces, buildings and words was assessed through paired samples t-tests.
Results
Learning effect
When testing for possible effects of learning in healthy participants, ANOVA on the run final models did not showed a significant 3-way interaction effect (Version*Version order*Task order) for any of the three RMT subtasks (words: F(5, 60) = 1.102, p = 0.369; faces: F(5, 60) = 0.154, p = 0.987; buildings: F(5, 60) = 0.543, p = 0.743).
Similarly, no interaction effect (Version*Version order*Task order), was found when checking for a possible learning effect in the neurological patients for any of the three RMT subtasks (words: F(3,2) = 3.379, p = 0.237; faces: F(3,2) = 8.063, p = 0.112; buildings: F(3,2) = 0.573, p = 0.686).
Inter-rater reliability
In the healthy participants, inter-rater reliability between the two versions of the RMT was moderate for faces (ICC = 0.699), buildings (ICC = 0.644) and words (ICC = 0.746). In the neurological patients inter-rater reliability was moderate for buildings (ICC = 0.539), and good for faces (ICC = 0.756) and words (ICC = 0.836).
Healthy participants
When comparing the two versions of the RMT for faces, no significant differences were found (t(71) = 0.298, p = 0.767, d = 0.003). Consistently, they also showed statistical equivalence on the TOST procedure (both upper and lower equivalence bound yielding a p < 0.005; computerized version mean = 27.95 ± 2.11; paper-and-pencil version mean = 27.93 ± 2.20). Also when comparing the two versions of the RMT for buildings, no significant effect was found (t(71) = 895, p = 0.374, d = 0.105). Consistently, they also showed statistical equivalence on the TOST procedure (both upper and lower equivalence bound yielding a p < 0.005; computerized version mean = 27.87 ± 2.46; paper-and-pencil version mean = 28.23 ± 1.98). Finally, no significant differences were found even when comparing the two versions of the RMT for words (t(71) = 1.88, p = 0.064, d = 0.221). Consistently, they also showed statistical equivalence on the TOST procedure (both upper and lower equivalence bound yielding a p < 0.005; computerized version mean = 27.25 ± 2.56; paper-and-pencil version mean = 27.66 ± 2.26) (Fig. 2). Table 2 reports means and standard deviations scores obtained by the sample at the two versions of the RMT (paper-and-pencil, computerized), in all the three subtests (words, unfamiliar faces and unknown buildings), for each decade of age.
Neurological patients
When comparing the two version of the RMT, no significant differences were found for faces (computerized version mean = 21.41 ± 5.44; paper-and-pencil version mean = 22.83 ± 3.79; t(11) = 1.48, p = 0.167, d = 0.427), buildings (computerized version mean = 20.50 ± 4.10; paper-and-pencil version mean = 21.83 ± 4.96; t(11) = 1.07, p = 0.309, d = 0.308), or words (computerized version mean = 19.66 ± 6.80; paper-and-pencil version mean = 20.91 ± 5.50; t(11) = 1.29, p = 0.224, d = 0.372). (Fig. 2). Table 2 reports means and standard deviations scores obtained by the neurological patients at the two versions of the RMT.
System usability scale – SUS
The SUS total median value from healthy participants was 97.5 (IQR = 5.63), with a minimum rating score of 72.5 and a maximum rating score of 100 (Fig. 3). The mean score of 95.7 (± 6.23), shows that the system is perceived as "excellent" and "acceptable" according to the adjective scale developed by Bangor et al. [39]. When evaluating the user experience of the neurological patients, the SUS total median value was 90.0 (IQR = 10.0), with a minimum rating score of 57.5 and a maximum rating score of 100 (Fig. 3). The mean score of 84.4 (± 13.9), shows that the system is perceived as "good" and "acceptable" [39].
Discussion
In recent years, a large body of literature has shown an increasing interest in the use of computerized modalities in neuropsychological assessment both in clinical practice and in research [3, 7, 8, 12, 40]. CNADs have been proved to be highly effective, compared to the traditional assessment. On the other hand, this considerable enthusiasm is reduced by others studies arguing that the computerized tests cannot be considered tout court the equivalent of the conventional pencil-and-paper test [17, 21]. The same studies underlined that a computerized version requires the same psychometric validity of the tests directly administered by an examiner. In addition, these studies pointed out that the use of CNADs creates a specific evaluation setting in which the patient’s performance can be conditioned by the peculiar modality and then it may provide information that do not correspond to the real subject’s ability [17, 21].
According to this theoretical-applicative context, the present study aimed to create and validate a computerized version of the pencil-and-paper verbal and non-verbal recognition memory test (RMT—Form A) for words, unfamiliar faces and unknown buildings [32]. Both the pencil-and-paper and computerized versions were administered to different age-balanced groups of healthy participants, with medium–high education (mean education: 14.04 years; range: 8–25), and ability to use a high computerized system, proficiently and without performance slowdowns (SUS mean score: 97.52/100; range: 72.5–100). Our results showed a moderate inter-rater reliability between the two versions of the RMT for words, unknown faces and buildings. Furthermore, when comparing the two versions, no significant differences were found, and according to the TOST procedure, the three tests can be considered equivalent in both modalities (i.e., computerized and paper-and-pencil). We also enrolled a small sample of neurological patients with mixed aetiology, age (range: 20–76) and educational level (range: 8–18) to evaluate the possible impact of different neurological conditions on the patients’ ability to interact with the computerized version of the RMT. Results showed a moderate-to-good inter-rater reliability between the two versions of the RMT. Furthermore, no difference was found in patient’s performance between the two versions of the RMT for words, unknown faces and buildings. Finally, results on SUS showed that the system is perceived as "good" and "acceptable" [39] (SUS mean score: 84.4/100; range: 57.5–100). Moreover, none of the patients referred feeling of discomfort while interacting with the computer interface, nor did they need further instructions from the experimenter, others then those provided by the computerized test. These findings support the possibility to use, in clinical practice, the computerized version as an alternative to the traditional one.
The computerized version of the RMT can offer several advantages in the assessment procedures, such as, for example, the reduction of the examiner’s influence, higher consistency administration, presentation of the stimuli, and scoring. It can also support a higher number of screening and follow-up activities on large populations [7, 40, 41], due to the ease of data collection and the scoring procedure, the immediate availability of results and a significant reduction in costs linked to administering the tests. Furthermore, another strength is that, in the spirit of Open Science, test materials and instructions are freely available on the web under a Creative Common licence (https://osf.io/yuqbh/?view_only=1edfc3db31b74ac581f83dfbc4b7cced). Therefore, interested neuropsychologists can use them in both clinical and research settings. Finally, this computerized version of the RMT may be suitable for online assessment, which has recently been acknowledge as a reliable alternative to the "in-presence", traditional, neuropsychological assessment for diagnosing neurocognitive and degenerative disorders [42, 43], such as preclinical dementia [44], as well as in healthy population assessment and in Alzheimer’s disease [45]. Online assessment makes it certainly easier for patients to access medical care permitting to save time and money even if several potential limitations must be considered (e.g., technological devices malfunctioning during the session; video or audio quality variability during the assessment).
The computerized version of the RMT is not exempt from limitations. For example, this test cannot be administered to poorly educated people over 80 years, because they usually are not familiar with technological devices. In fact, our validation sample included participants aged from 20 to 79 years, with medium–high education, and their rating of usability for the computerized version of the RMT was almost at ceiling. However, technological devices are expected to become usual tools for the whole population.
Data availability
Datasets associated with the current study are available from the corresponding author upon request.
References
Rabin LA, Spadaccini AT, Brodale DL et al (2014) Utilization rates of computerized tests and test batteries among clinical neuropsychologists in the United States and Canada. Prof Psychol Res Pract 45:368. https://doi.org/10.1037/a0037987
Marra DE, Hamlet KM, Bauer RM, Bowers D (2020) Validity of teleneuropsychology for older adults in response to COVID-19: A systematic and critical review. Clin Neuropsychol 34:1411–1452. https://doi.org/10.1080/13854046.2020.1769192
Bauer RM, Iverson GL, Cernich AN et al (2012) Computerized neuropsychological assessment devices: joint position paper of the American Academy of Clinical Neuropsychology and the National Academy of Neuropsychology. Arch Clin Neuropsychol 27:362–373. https://doi.org/10.1093/arclin/acs027
Parsons TD (2016) Clinical Neuropsychology and Technology. Springer International Publishing, Cham
Terruzzi S, Albini F, Massetti G, Etzi R, Gallace A, Vallar G (2023) The neuropsychological assessment of unilateral spatial neglect through computerized and virtual reality tools: A scoping review. Neuropsychol Rev 1-39. https://doi.org/10.1007/s11065-023-09586-3
Aşkar P, Altun A, Cangöz B et al (2012) A comparison of paper-and-pencil and computerized forms of line orientation and enhanced cued recall tests. Psychol Rep 110:383–396. https://doi.org/10.2466/03.22.PR0.110.2.383-396
Woo E (2008) Computerized neuropsychological assessments. CNS Spectr 13:14–17. https://doi.org/10.1017/S1092852900026985
Crook TH, Kay GG, Larrabee GJ, Grant I, Adams K (2009) Computer-based cognitive testing. Neuropsychological assessment of neuropsychiatric and neuromedical disorders, pp 84–100
Russell EW (2000) The application of computerized scoring programs to neuropsychological assessment. Clinician’s guide to neuropsychological assessment, pp 483–515
Gur RC, Ragland JD, Moberg PJ et al (2001) Computerized neurocognitive scanning: I. Methodology and validation in healthy people. Neuropsychopharmacology 25:766–776. https://doi.org/10.1016/S0893-133X(01)00278-0
Schatz P, Browndyke J (2002) Applications of computer-based neuropsychological assessment. J Head Trauma Rehabil 17:395–410. https://doi.org/10.1097/00001199-200210000-00003
Sheppard LD, Vernon PA (2008) Intelligence and speed of information-processing: A review of 50 years of research. Pers Individ Dif 44:535–551. https://doi.org/10.1016/j.paid.2007.09.015
Iverson GL, Brooks BL, Ashton VL et al (2009) Does familiarity with computers affect computerized neuropsychological test performance? J Clin Exp Neuropsychol 31:594–604. https://doi.org/10.1080/13803390802372125
Cernich AN, Brennana DM, Barker LM, Bleiberg J (2007) Sources of error in computerized neuropsychological assessment. Arch Clin Neuropsychol 22:S39–S48. https://doi.org/10.1016/j.acn.2006.10.004
Hoskins LL, Binder LM, Chaytor NS et al (2010) Comparison of oral and computerized versions of the word memory test. Arch Clin Neuropsychol 25:591–600. https://doi.org/10.1093/arclin/acq060
Williams JE, McCord DM (2006) Equivalence of standard and computerized versions of the Raven Progressive Matrices Test. Comput Human Behav 22:791–800. https://doi.org/10.1016/j.chb.2004.03.005
Feldstein SN, Keller FR, Portman RE et al (1999) A comparison of computerized and standard versions of the Wisconsin Card Sorting Test. Clin Neuropsychol 13:303–313. https://doi.org/10.1076/clin.13.3.303.1744
French CC, Beaumont JG (1987) The reaction of psychiatric patients to computerized assessment. Br J Clin Psychol 26:267–278. https://doi.org/10.1111/j.2044-8260.1987.tb01360.x
Ozonoff S (1995) Reliability and validity of the Wisconsin card sorting test in studies of autism. Neuropsychology 9:491. https://doi.org/10.1037/0894-4105.9.4.491
Berger SG, Chibnall JT, Gfeller JD (1997) Construct validity of the computerized version of the Category Test. J Clin Psychol 53:723–726. https://doi.org/10.1002/(SICI)1097-4679(199711)53:73.0.CO;2-I
Campbell KA, Rohlman DS, Storzbach D et al (1999) Test-retest reliability of psychological and neurobehavioral tests self-administered by computer. Assessment 6:21–32. https://doi.org/10.1177/107319119900600103
Choca J, Morris J (1992) Administering the Category Test by computer: Equivalence of results. Clin Neuropsychol 6:9–15. https://doi.org/10.1080/13854049208404112
Tien AY, Spevack TV, Jones DW et al (1996) Computerized Wisconsin Card Sorting Test: comparison with manual administration. Kaohsiung J Med Sci 12:479–485
Gelman A, Geurts HM (2017) The statistical crisis in science: How is it relevant to clinical neuropsychology? Clin Neuropsychol 31:1000–1014. https://doi.org/10.1080/13854046.2016.1277557
Yu K, Zhang S, Wang Q et al (2015) Development of a computerized tool for the chinese version of the montreal cognitive assessment for screening mild cognitive impairment. Int psychogeriatrics 27:213–219. https://doi.org/10.1017/S1041610214002269
Garre-Olmo J, Faúndez-Zanuy M, López-de-Ipiña K et al (2017) Kinematic and pressure features of handwriting and drawing: preliminary results between patients with mild cognitive impairment, Alzheimer disease and healthy controls. Curr Alzheimer Res 14:960–968. https://doi.org/10.2174/1567205014666170309120708
Petilli MA, Daini R, Saibene FL, Rabuffetti M (2021) Automated scoring for a Tablet-based Rey Figure copy task differentiates constructional, organisational, and motor abilities. Sci Rep 11:1–19. https://doi.org/10.1038/s41598-021-94247-9
Massetti G, Albini F, Casati C et al (2023) Validation of “Neurit. Space”: Three Digital Tests for the Neuropsychological Evaluation of Unilateral Spatial Neglect. J Clin Med 12:3042. https://doi.org/10.3390/jcm12083042
Coughlan AK, Hollows SE (1984) Use of memory tests in differentiating organic disorder from depression. Br J Psychiatry 145:164–167. https://doi.org/10.1192/bjp.145.2.164
Smirni D, Turriziani P, Oliveri M et al (2010) Standardizzazione di tre nuovi test di memoria di riconoscimento verbale e non verbale: uno studio preliminare. G Ital di Psicol 37:227–248. https://doi.org/10.1421/32178
Barvas E, Mattavelli G, Meli C et al (2022) Standardization and normative data for a new test of visual long-term recognition memory. Neurol Sci 43:2491–2497. https://doi.org/10.1007/s10072-021-05642-z
Smirni D, Smirni P, Di Martino G et al (2018) Standardization and validation of a parallel form of the verbal and non-verbal recognition memory test in an Italian population sample. Neurol Sci 39:1391–1399. https://doi.org/10.1007/s10072-018-3433-z
Novelli G, Papagno C, Capitani E, Laiacona M (1986) Tre test clinici di memoria verbale a lungo termine: taratura su soggetti normali. Archivio di psicologia, neurologia e psichiatria
Conti S, Bonazzi S, Laiacona M et al (2015) Montreal Cognitive Assessment (MoCA)-Italian version: regression based norms and equivalent scores. Neurol Sci 36:209–214. https://doi.org/10.1007/s10072-014-1921-3
Brooke J (1996) SUS-A quick and dirty usability scale. Usability Eval Ind 189:4–7
Koo TK, Li MY (2016) A guideline of selecting and reporting intraclass correlation coefficients for reliability research. J Chiropr Med 15:155–163. https://doi.org/10.1016/j.jcm.2016.02.012
Lakens D, Scheel AM, Isager PM (2018) Equivalence Testing for Psychological Research: A Tutorial. Adv Methods Pract Psychol Sci 1:259–269. https://doi.org/10.1177/2515245918770963
Kim H-Y (2013) Statistical notes for clinical researchers: assessing normal distribution (2) using skewness and kurtosis. Restor Dent Endod 38:52–54. https://doi.org/10.5395/rde.2013.38.1.52
Bangor A, Kortum P, Miller J (2009) Determining what individual SUS scores mean: Adding an adjective rating scale. J usability Stud 4:114–123
Bilder RM (2011) Neuropsychology 3.0: Evidence-based science and practice. J Int Neuropsychol Soc 17:7–13. https://doi.org/10.1017/S1355617710001396
Parsey CM, Schmitter-Edgecombe M (2013) Applications of technology in neuropsychological assessment. Clin Neuropsychol 27:1328–1361. https://doi.org/10.1080/13854046.2013.834971
Adams JL, Myers TL, Waddell EM et al (2020) Telemedicine: a valuable tool in neurodegenerative diseases. Curr Geriatr reports 9:72–81. https://doi.org/10.1007/s13670-020-00311-z
Parikh M, Grosch MC, Graham LL et al (2013) Consumer acceptability of brief videoconference-based neuropsychological assessment in older individuals with and without cognitive impairment. Clin Neuropsychol 27:808–817. https://doi.org/10.1080/13854046.2013.791723
Chaytor NS, Barbosa-Leiker C, Germine LT et al (2021) Construct validity, ecological validity and acceptance of self-administered online neuropsychological assessment in adults. Clin Neuropsychol 35:148–164. https://doi.org/10.1080/13854046.2020.1811893
Loh P, Donaldson M, Flicker L et al (2007) Development of a telemedicine protocol for the diagnosis of Alzheimer’s disease. J Telemed Telecare 13:90–94. https://doi.org/10.1258/135763307780096159
Acknowledgment
We are grateful to all volunteers who took part in the study. We thank Dr. Tommaso Mastropasqua for his support with test creation.
Funding
Open access funding provided by Università degli Studi di Trento within the CRUI-CARE Agreement. The research was supported by a grant (RF-2019-12369567) from the Health Ministry to CP.
Author information
Authors and Affiliations
Contributions
Elena Baruzzo: Conceptualization, Investigation, Writing – Original draft. Stefano Terruzzi: Investigation, Formal analysis, Writing – Original draft. Beatrice Feder: Investigation, Data curation, Writing – Original draft. Costanza Papagno: Conceptualization, Writing – Review & Editing, Supervision, Project administration. Daniela Smirni: Conceptualization, Writing – Original draft, Supervision.
Corresponding author
Ethics declarations
Ethical approval
The study was conducted in accordance with the Declaration of Helsinki, with the approval of the local ethics committee.
Conflict of interest
All authors declare no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Baruzzo, E., Terruzzi, S., Feder, B. et al. Verbal and non-verbal recognition memory assessment: validation of a computerized version of the Recognition Memory Test. Neurol Sci 45, 1979–1988 (2024). https://doi.org/10.1007/s10072-023-07171-3
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10072-023-07171-3