People with severe to profound intellectual disabilities tend to be passive and detached from their environment (i.e., from objects and activities) when no direct staff supervision is available (Dairo et al., 2016, 2017; Dixon-Ibarra et al., 2013; Melville et al., 2017; Stancliffe & Anderson, 2017). The situation is even more serious and problematic when intellectual disability is combined with visual impairment (Desideri et al., 2021; Dijkhuizen et al., 2016; Hanzen et al., 2018, 2020). People with intellectual disability and visual impairment (a) may be largely unaware of objects and activities in their immediate surrounding, and (b) may also be unable to orient and move in their daily context thus experiencing widespread passivity and isolation (Jarjoura, 2019; Lancioni, Singh, et al., 2017; Lancioni et al., 2021a; Nair et al., 2020).

Extensive staff supervision to improve the situation may not be considered a realistic solution of the problem. In fact, staff supervision can hardly be ensured over protracted periods of time across different parts of the day within daily contexts. Moreover, the use of extensive supervision would interfere with the people’s development of initiative and self-determination and would also hinder their social image (Lancioni, O’Reilly, et al., 2017; Lancioni, Singh, et al., 2017; Mumbardó-Adam et al., 2020; Wehmeyer, 2020). A better approach to the problem may be represented by the use of technology-aided programs providing the necessary supervision/support (Cuturi et al., 2016; Goo et al., 2019; Lancioni et al., 2018a, b; Shih et al., 2013; Taylor et al., 2016).

Several technology-aided programs have been reported to increase activity engagement and mobility in people with intellectual disability and visual impairment without the need of staff supervision. For example, Lancioni, Singh, et al. (2017) developed a program in which electronic boxes fitted with an optic sensor and regulated by a remote control system were used to support the participants. The boxes were located on different desks that the participants were to reach to collect objects or put away the objects collected. Every intervention session started with the activation of the box on one of the desks containing objects, which emitted verbal cues for spatial orientation. The participant was to reach the desk and take one of the objects available there. Reaching the desk activated the optic sensor and caused the box to provide 5 s of preferred stimulation. Subsequently, the box at a storage desk where the objects were to be put away started to emit cues. Bringing the object to the storage desk triggered the optic sensor available there and led the box to provide 20 s of preferred stimulation. At the end of it, the process started again and continued in the same way until the session time had elapsed.

Lancioni et al. (2019) set up a program in which smartphones, mini speakers, and portable light sources were used. Three smartphones were placed in three boxes, which contained different groups/types of objects. The other three smartphones were located at the destinations where the objects of the three boxes were to be put away. Each smartphone was connected to one of the aforementioned speakers and light sources. At the start of a session, one of the smartphones inside the boxes presented via the related speaker auditory spatial cues. The participant was expected to walk toward those cues and search inside one of the boxes. This caused the smartphone inside the box to (a) ask the participant to take an object and (b) trigger the smartphone at the destination where the object was to be transported. This latter smartphone then started to call (providing spatial cues to) the participant. Reaching the destination and triggering the light sensor of the smartphone available there caused that smartphone to ask the participant to put away the object, and activate 15 s of preferred stimulation. The end of this stimulation triggered the smartphones at the boxes and enabled one of them to start calling the participant. The process continued as just described throughout the session.

Lancioni et al. (2021b) developed a program involving the use of a smartphone with special applications and indoor motion sensors and mini speakers to help participants collect objects from different areas and bring them to a storage desk. The smartphone was linked to the sensors and speakers available at the areas where the participant was to collect the objects and at the storage desk where the objects were to be put away. Every session started with the mini speaker of one of the areas calling the participant. When the participant reached the area and was detected by the sensor available there, the speaker presented praise and the request to take an object. This was followed by the activation of the speaker of the storage desk presenting calls. Once the desk was reached and the sensor detected the participant’s presence, the speaker presented praise, the request to put away the object and 15 s of preferred stimulation. The end of the stimulation was followed by the speaker of another area starting to call the participant, so that the participant could collect another object and transport it to the storage desk. The same process continued until the end of the session.

All three programs described above were reported to be successful in helping the participants engage in functional activity (collecting and putting away objects) and mobility. Notwithstanding the positive results, some questions about the different programs might be raised. For example, the first program was based on technology specifically built for the study, thus not directly accessible to others and relatively expensive. The second program had the advantage of relying on commercial technology and supporting activities that required differential use of material (i.e., different types of objects were to be transported to different destinations). Yet, the technology included several clusters of smartphones, mini speakers, and light sources that made it relatively complex and expensive. The third program relied on commercial technology, which was simpler and less expensive than the technology used for the second program, but it did not support differential use of objects (i.e., all objects collected were put away at the same storage desk).

The present study was to develop and assess a new program that (a) included a technology package even simpler than that used for the third program mentioned above and (b) supported activity engagement involving differential use of objects and mobility. In practice, the new program (a) relied on a smartphone linked to motion sensors and mini speakers and (b) was designed to help seven participants with severe to profound intellectual disability and blindness collect objects of two different groups and transport the objects of each group to a specific destination.

Method

Participants

Table 1 identifies the seven participants by their pseudonyms and reports their chronological age, and the age equivalents for their daily living skills on the second edition of the Vineland Adaptive Behavior Scales (Balboni et al., 2016; Sparrow et al., 2005). The participants (two women and five men) were between 23 and 49 years of age. One of the participants (Xavier) had a minimal residual vision that allowed him to see large obstacles in his way, while the others were totally blind. The Vineland age equivalents for their daily living skills (personal sub-domain) varied between 2 years and 4 months (Adan) and 3 years and 1 month (Jane). All participants attended rehabilitation and care centers. Their psychological records indicated that their level of intellectual disability had been estimated to be within the severe to profound range, but no specific tests were applied for their assessment.

Table 1 Participants’ pseudonyms, chronological age, and Vineland age equivalents for daily living skills (personal sub-domain)

The participants were included in the study on the following basis. First, they could use auditory spatial cues to orient and reach indoor destinations (within a room or across adjacent rooms). Two of them had also been involved in previous programs with the use of auditory spatial cues (i.e., Jane and Brady). Second, they were familiar with simple verbal instructions such as “take” and “put away” an object. Third, they seemed to enjoy (i.e., as indicated by behaviors such as alerting/orienting and smiling) various forms of reportedly preferred environmental stimulation including music, songs, praise, and familiar voices. Thus, it was thought that using those forms of preferred environmental stimulation contingent on their response performance could serve to motivate/strengthen such performance. Fourth, activities involving differential use of objects and mobility were considered relevant forms of functional occupation within the participants’ context. Fifth, staff supported the study (of which they were informed in advance), as they thought that participants would largely benefit from a technology system supporting their activity engagement.

While the availability of preferred stimulation during the intervention sessions (see below) was thought likely to make the participants’ involvement in the study a positive experience, it was impossible to reliably determine their assent to be involved. Thus, their legal representatives were asked to provide formal consent on their behalf before the start of the study. The study complied with the 1964 Helsinki Declaration and its later amendments and was approved by an institutional ethics committee.

Procedures

Setting, Activities, Stimulation, Sessions, and Research Assistants

Quiet rooms of the centers that the participants attended constituted the setting for the study sessions. Activities consisted of (a) collecting 20 objects from two different containers (each including a set of 10 objects) located in a specific area (on a specific desk) of the room, (b) transporting the objects to two different destinations (one set of objects to one destination and the other set to a second destination located within the same or adjacent room), and (c) depositing/putting away each of the objects at the destinations. Activities could vary across sessions with regard to the sets of objects to be collected, transported, and put away. The objects could involve kitchen tools, drink items, boxes, paper packs, and other sets of materials available within daily contexts. For each object to be collected, transported, and put away, the technology system provided the participants with verbal orientation cues, instructions, praise and other reportedly preferred stimulation events such as music and songs, and familiar voices (see the “Participants” and “Technology System” sections). Sessions consisted of the time periods the participants needed to complete an activity (i.e., collect and transport to different destinations the two groups of 10 objects). Sessions could also be interrupted before the activity was completed. Session interruption occurred when a 30-min time limit had elapsed or when research assistant’s guidance had been used on four consecutive activity responses. Research assistants, who carried out the sessions and recorded the participants’ data (see below), were familiar with the use of technology-aided interventions for people with intellectual and multiple disabilities as well as with data collection.

Technology System

The technology system used during the intervention phase of the study involved (a) a Samsung Galaxy A22 with Android 11 operating system that was equipped with Amazon Alexa, MacroDroid, and Philips Hue applications, (b) five Philips Hue indoor motion sensors, (c) a Philips Hue Bridge and Philips Hue smart bulb working via Bluetooth, (d) a 4G Long-Term Evolution Wi-Fi router, and (e) three Bluetooth mini speakers. The Philips Hue Bridge, smart bulb and application, and the router were instrumental for the functioning of the Philips Hue sensors.

The sensors were box-like devices with a 5.5-cm side and 3.5-cm height. One of the sensors was placed in front of the desk with the two containers and the two groups of objects to be transported to different destinations. Two sensors were placed in the containers (i.e., one sensor per container). The other two sensors were placed before the destinations to which the objects were to be transported (i.e., one sensor at each destination). Sensor activation (caused by the participant’s arrival at the desk or destination or by the participant’s hand being inserted in a container) was detected through the Amazon Alexa application and transmitted to the smartphone via the MacroDroid. The Bluetooth mini speakers, which were linked to the smartphone, were placed on the desk with the two containers, and at each of the two destinations where the objects were to be transported/deposited.

Figure 1 summarizes the working of the technology system. Switching on the system (i.e., starting a session) activated the mini speaker of the desk with the two containers with objects, which started to call the participant (i.e., one or two-word calls that could include the participant’s name) at intervals of 5 s. The calls were spatial orientation cues guiding and encouraging the participant to reach the desk. As soon as the participant reached the desk (i.e., triggered the sensor before the desk), the mini speaker on the desk presented verbal praise and the instruction to take an object. The instruction was repeated at intervals of 5–10 s until the participant triggered the sensor of one of the two containers with objects. The sensor was triggered as the participant explored (i.e., inserted the hand into) one of the containers to take an object. Triggering the sensor of a container led to the automatic activation of the mini speaker of the destination matching that container, that is, of the destination where any of the objects taken from that container was to be transported and deposited.

Fig. 1
figure 1

The flowchart summarizes the working of the technology system

The speaker at the destination kept on calling the participant until the participant reached it (i.e., triggered the sensor in front of the destination; as described above). Once the participant was at the destination, the mini speaker available there presented verbal praise, the instruction to deposit the object, and 15 s of a preferred stimulation event (e.g., a 15-s segment of a preferred song or music piece; see the “Participants” section). At the end of the stimulation event, the mini speaker of the desk with the containers was automatically activated (i.e., started to call the participant). When the participant arrived at the desk (i.e., triggered the sensor in front of it), the mini speaker presented praise and the instruction to take an object. If the participant took an object from the same container used before, the speaker of the destination previously reached was activated again. If the participant took an object from the other container, the speaker of the second destination (not targeted for the previous response) was activated. The process then continued as described above. The only exception occurred when the participant explored one of the containers, in which no more objects were presumably available (i.e., a container whose sensor had been activated on 10 previous response occasions). In that case, the system repeated the instruction to search and take an object at intervals of 5–10 s until the participant triggered the sensor of the other container. The session continued until (a) the system had provided support (i.e., spatial orientation cues, instructions, praise, and preferred stimulation events) for 20 activity responses, that is, for collecting, transporting, and putting away the two groups of 10 objects available or (b) a 30-min period had elapsed, whichever came first.

Experimental Conditions

The study was conducted according to a nonconcurrent multiple baseline design across participants (Barlow et al., 2009; Lancioni et al., 2021c). Consistent with the design requirements, the participants were provided with different numbers (i.e., between 5 and 10) of baseline sessions during which the technology system was not available. The baseline sessions were followed by intervention sessions, which were carried out with the support of the technology system. Between 46 and 80 intervention sessions were implemented for the different participants. The numbers of sessions varied across participants based on their availability. Video recordings of the sessions were regularly accessible to a study coordinator who would view them and provide feedback to the research assistants to promote their correct performance and thus ensure procedural fidelity (Sanetti & Collier-Meek, 2014).

Baseline

Before the start of a baseline session, the research assistant guided the participant physically and verbally to inspect the desk with the containers and objects (each container included a group of 10 objects) and the destinations where those objects were to be transported and deposited. Thereafter, the research assistant accompanied the participant near the desk with the containers (as it would occur during the intervention sessions with the technology system) and presented the participant with the instruction to take an object and transport it to a destination. If the participant remained inactive or showed no progress for 30–40 s, the research assistant intervened with guidance (i.e., guided the participant physically and verbally to take an object from one of the containers, transport the object to the right destination, and deposit it there). The right destination could include, among others, a table, a cupboard, or a combination of chairs. This was followed by a new instruction to take another object and transport it. The research assistant’s guidance was used as indicated above. The session continued until the participant had responded to all the 20 instructions scheduled or until research assistant’s guidance had occurred for four consecutive responses. Interrupting the sessions after four consecutive guidance instances was to minimize failure and frustration.

Intervention

The intervention phase was introduced by two or three practice sessions, which served to familiarize the participants with the functioning of the technology system. During these sessions, the research assistant could provide guidance to facilitate the participants’ successful use of the system’s support (i.e., spatial orientation cues and instructions). Yet, all participants had the prerequisites to manage such support independently (see the “Participants” section). During the regular sessions that followed the practice sessions, no research assistant’s guidance was available. The research assistant would only accompany the participant near the desk with the containers and related groups of objects and switch on the technology system. Once switched on, the system worked as described in the “Technology System” section and summarized in Fig. 1. In practice, the system presented spatial orientation cues, instructions, praise, and preferred stimulation events in connection with each of the 20 objects (10 per container) the participant was scheduled to collect, transport to, and deposit at the right destination within a session. Sessions lasted until the participant had collected, transported, and put away all 20 objects or a 30-min limit had elapsed, whichever came first (see the “Technology System” section).

Measures

The measures were activity responses completed correctly and session duration. During baseline, an activity response was completed correctly if the participant (following the research assistant’s instruction to take an object and transport it to a destination) reached the desk with the two containers, took an object from one of the containers, transported the object to the matching destination, and deposited/put away the object there independently. During the intervention, an activity response was completed correctly if the participant produced the performance sequence described above using the technology system’s support (see the “Technology System” section and Fig. 1). The first measure (i.e., activity responses completed correctly) was recorded by the research assistants throughout the study. The second measure (i.e., session duration) was recorded by the research assistants during baseline, and by the smartphone during the intervention. The smartphone recorded the time elapsed from the delivery of the first instruction to take an object at the desk with the containers to the delivery of the last stimulation event at one of the destinations.

Interrater agreement was assessed (a) in more than 25% of the sessions of each participant on the first measure and (b) in all baseline sessions on the second measure. Agreement was checked through the involvement of a reliability observer in data recording. The percentage of agreement on the first measure (computed for the single sessions by dividing the number of responses on which research assistant and reliability observer had the same “correct” or “incorrect” score by the total number of responses and multiplying by 100%) was within the 90–100% range, with means exceeding 98% for all participants. The percentage of interrater agreement on the second measure (computed by dividing the number of sessions whose reported durations differed less than 1 min by the total number of sessions and multiplying by 100%) was 100%.

Data Analyses

The participants’ data for activity responses completed correctly and session duration were reported in graphic form. In order to simplify the graphic display, the data were summarized into blocks of sessions. Accordingly, each data point appearing in the graphs represents a mean session frequency or a mean session duration computed over a block of sessions. The Kolmogorov–Smirnov test (Siegel & Castellan, 1988) was the statistical tool selected for analyzing the differences between the baseline and intervention frequencies of activity responses completed correctly for those participants who had some levels of data overlap between the two phases.

Results

The seven panels of Fig. 2 report the participants’ mean frequency of activity responses completed correctly and mean session duration over blocks of sessions. Specifically, the black diamonds represent the mean frequency of activity responses completed correctly per session over blocks of three sessions during the baseline and the intervention phase. The empty circles represent the mean session duration for the same blocks of sessions. Baseline and intervention blocks including two sessions (i.e., blocks appearing at the end of the baseline or the intervention phase) are marked with an arrow. The practice sessions occurring at the start of the intervention phase are not reported in the figure.

Fig. 2
figure 2

The seven panels report the participants’ mean frequency of activity responses completed correctly and mean session duration over blocks of sessions. The black diamonds represent the mean frequency of activity responses completed correctly per session over blocks of three sessions during the baseline and the intervention phase. The empty circles represent the mean session duration for the same blocks of sessions. Baseline and intervention blocks including two sessions (i.e., blocks which can appear at the end of the baseline or the intervention phase) are marked with an arrow

During baseline, the frequency of activity responses completed correctly was zero or close to zero for all participants. All baseline sessions were interrupted following four consecutive responses with guidance from the research assistant. The participants’ mean session duration ranged between about 6 and 7 min. During the intervention phase (i.e., with the support of the technology system), the mean frequency of activity responses completed correctly increased for all participants. In fact, the mean frequency ranged from close to 19 (Jane, Adan, and Michah) to close to 20 (Xavier). Occasional lower frequencies of correct activity responses were connected to sessions, which were interrupted because the 30-min time limit had elapsed, or to inaccurate response performance (e.g., failure to take or put away an object). The mean session duration during the intervention phase varied between approximately 16 min (Nylah) and 29 min (Michah).

The baseline and the intervention data for activity responses completed correctly showed no overlaps for any of the participants. Absence of overlaps was seen as clear evidence of the definite difference between the two phases and thus of the effectiveness of the technology system in supporting the participants’ activity performance. Given this evidence, the application of the Kolmogorov–Smirnov test was deemed unnecessary.

Discussion

The data indicate that all participants were successful in using the technology system employed during the intervention. These data support previous findings in the area suggesting that technology-aided programs can be effectively used for promoting functional activity engagement and mobility with participants with intellectual disability and visual impairment (Lancioni et al., 2019, 2021b; Lancioni, Singh, et al., 2017). The same data add to previous evidence in that the technology system used in this study was relatively simple compared to those used before while supporting activities that required differential use of objects (i.e., the two sets of objects available in each session were to be transported to different destinations rather than to a same destination). In light of the above, several considerations may be made.

First, the relevance of technology-aided support for the participants of this study (as well as the participants of previous studies in this area) was emphasized by the great contrast between the highly positive data of the intervention period and the zero or near zero baseline performance. Indeed, the baseline performance suggests that the participants were unable to manage functional occupational engagement and mobility. The intervention performance shows the possibility of enabling the participants to reach constructive occupation as well as orientation and mobility with no need for staff supervision. This achievement can be viewed as meaningful in terms of participants’ behavioral and cognitive growth, social image, and physical exercise (Bertelli et al., 2016; Bouzas et al., 2019; Kocman & Weber, 2018; Warburton & Bredin, 2019).

Second, the results obtained can be considered relevant also in view of the fact that the activities were not simply limited to transporting and putting away objects but involved a differentiated use of the objects, that is, the participants were enabled to transport the objects of one container to one destination and the objects of the other container to a different destination. The ability of using objects differentially may be viewed as a critical qualification for functional engagement (Lancioni et al., 2019; Wehmeyer et al., 2020). It may also be argued that the technology system used in this study could be expanded (i.e., with extra sensors and mini speakers) to include more than two categories of objects to be used differentially (Desideri et al., 2021; Desmond et al., 2018).

Third, the effectiveness of the intervention phase (with the support of the technology system) can easily be ascribed to the conditions implemented during that phase: (a) the calls from the mini speakers of the desk with the containers and of the destinations that were to be reached, (b) the instruction to take or deposit/put away an object, and (c) the praise and preferred stimulation events. The calls were apparently instrumental in helping the participants orient successfully and find the containers and destinations and may also have served to prompt/encourage their performance engagement (Lancioni et al., 2019; Lancioni, O’Reilly, et al., 2017). The instructions may have been essential to ensure that the participants were always aware of what to do with the objects (Lancioni et al., 2021b; Lancioni, Singh, et al., 2017). Praise and the preferred stimulation events available for the activity responses may have fostered the participants’ motivation to carry out those responses throughout the sessions and possibly their satisfaction with the sessions (Catania, 2013; Kazdin, 2012).

Fourth, the technology system used in this study (a) represents a relatively simple and practical tool compared to the systems used by Lancioni, Singh, et al. (2017), Lancioni et al. (2019)), which were specifically built for the purpose of the study or included clusters of smartphones, sensors, mini speakers, and light sources, and (b) has the advantage of supporting differential use of objects compared to that used by Lancioni et al. (2021b). The cost of the present technology system may be estimated at about US $650 (i.e., approximately US $175 for the Samsung smartphone, US $250 for the five Philips Hue sensors, US $75 for the three mini speakers, and US $150 for the Philips Hue Bridge, the Philips Hue smart bulb, and the 4G Long-Term Evolution Wi-Fi router). This cost cannot be considered irrelevant. Yet, the system offers valuable occupation and mobility opportunities and is fairly easy to operate for personnel in charge of the sessions and friendly for the participants (Boot et al., 2018; Borg, 2019; de Witte et al., 2018; Scherer, 2019). A main obstacle in accessing and using such a technology system concerns the fact that it is not a ready-made (off-the-shelf) tool but rather a tool that needs to be arranged with the aforementioned commercial components and programmed for the intervention purpose.

Limitations and Future Research

Three limitations of the study should be noted. The first limitation concerns the relatively small number of participants involved in the study. This limitation, which makes it difficult to draw conclusions about the overall potential and usability of the technology system reported, needs to be addressed via direct and systematic replication studies. The results of these studies will help determine the reliability of such system and the possibilities of improving it and extending its use across individuals (Kazdin, 2011; Locey, 2020; Travers et al., 2016).

The second limitation concerns the lack of assessment of (a) the participants’ satisfaction with the technology system and sessions and (b) the staff’s perception of the applicability and impact of such system and sessions. To assess participants’ satisfaction, one could observe their behavior during the sessions and determine whether they have expressions of positive mood (e.g., smiles) during their engagement (Dillon & Carr, 2007; Lancioni et al., 2020; Parsons et al., 2012). To assess staff’s perception, one could show them video segments of the intervention sessions and ask them to rate those segments and the system being used in terms of efficacy, friendliness, and applicability (Lancioni, O’Reilly, et al., 2017; Plackett et al., 2017; Worthen & Luiselli, 2019).

The third limitation is the lack of generalization and maintenance assessment. With regard to this point, some basic assumptions could be made before any formal assessment is carried out. First, successful generalization results across settings and intervention agents might be expected. In fact, the system’s support responsible for the participants’ performance (i.e., spatial orientation cues, instructions, and praise and preferred stimulation events) would not change irrespective of the setting in which the system is used and of the personnel implementing the intervention sessions (Haegele & Park, 2016; Kazdin, 2012). Second, performance maintenance might also be expected quite realistically if preferred (motivating) stimulation continues to be available contingent on the participants’ activity responses (Kazdin, 2012; Storey & Haymes, 2016).

In conclusion, the results suggest that the technology system evaluated in this study can help people with intellectual and visual disabilities achieve independent occupational engagement involving differential use of objects and mobility. While the results are encouraging, conclusions about the system’s reliability and usability cannot be drawn until new research has addressed the limitations of this study. Future research may also investigate whether it is possible to further develop the present system so as to improve its functioning and adapt its use to more complex activity situations and/or participants with different characteristics.