1 Introduction

1.1 Well-Being in SEN Schools and Interdisciplinary Research

Autism is a lifelong neurodevelopmental condition that affects how a person perceives, communicates and interacts with the world. This is characterised by significant and lasting differences (compared to typical development) in social communications and interaction, restricted and repetitive patterns of behaviour, interests or activities and sensory perception and responses [1]. Current data suggest that as many as 1 in 54 children in the United States of America are on the autism spectrum [2] while other studies suggest a figure of between 1 in 68 to 1 in 100 in the general population [3].

The well-being and emotional regulation of autistic people has been central in supporting meaningful educational experiences [4]. In comparison to their typically-developing peers, autistic children experience greater mental health problems, such as anxiety, depression, anger, and possess lower self-concept and these can impact education and other factors [5]. There have been a range of studies that have implemented technology to assist autistic children in this regard [6,7,8]. However, these studies have often focused on deficit-based models and do not often include users in their design. Other strategies have also been used to support the well-being of autistic pupils. For example, recent work has highlighted that focusing on autistic intense or ‘special’ interests in the classroom may be linked to improved well-being [9]. While the well-being of autistic groups is not well understood or defined, increasing awareness of its impact on students’ academic performance and their adult outcomes is well acknowledged [10]. With greater attention applied to the well-being of autistic pupils, improvements in areas of education and adulthood are often observed [11]. Therefore, considering well-being from an early age and in a setting where this can be supported (i.e. education) represents an important are of enquiry.

This project was established to approach issues in this field that can lack interdisciplinary perspectives and can be limited by siloed working [12]. As a result, to investigate the potential of a social robot in a school setting for autistic children, we built a team consisting of education, robotics and architecture. Thus bringing key interdisciplinary expertise together to explore the role of social robots in a UK-based SEN school which hosts about 145 autistic children.

A note on terminology. The authors take very seriously the use terminology in this field as it relates to person first and/or identity first language and framing. We acknowledge that this can sometimes represent a contentious issue. In this article, we choose to use ‘autistic children’ instead of ‘children with autism‘. The authors consulted, and sought input from, an autistic researcher who helped to guide us in our adoption and use of language. Our position draws as well on the work of [13, 14], and as such we we adopt language that that is currently considered inclusive (i.e. ‘autistic’ and in places, ‘autism’).

2 Related Literature and Derived Hypotheses

Our project is an interdisciplinary investigation of the following overarching aim: How can a social robot integrate into the ecosystem of a special educational needs (SEN) school, and effectively support the development and well-being of the children, in complement to the day-to-day work of the school staff? Extensive prior art related to social robots in the school environment can be found [15]; we focus our analysis of literature to research looking specifically at (1) robots for/with autistic children; (2) the impact of (social) robots on the school ecosystem.

2.1 Robots and Autistic Children

Research in the field of child–robot interaction has already shown that, compared to other technologies, the use of robots can bring unique opportunities for autistic children (and neurotypical populations), including having a powerful impact on their behaviour and development [16, 17]. This is understood to come from the embodied nature of robots [18,19,20] in addition to the unique characteristics that robotic technology has for autistic populations; for example, simplified interactions, safe, predictable and reliable behaviours [21].

Previous research on robotics and autism has predominately focused on using robots to support the acquisition of social skills (like joint attention or mutual gazing) [e.g.  22, 23, 24, 25, 26]. In these previous studies, the robot interaction design is typically a top-down process: experts (typically SEN teachers or autism specialists) suggest age- and development-appropriate activities, implemented on the robots by engineers, and deployed in schools. Our focus was instead on understanding, from the children themselves, their ambitions for using a social robot in their school, without preconceptions or bias towards designing a robot to specifically ‘teach’ social skills.

Despite the potential of social robots in classrooms, research on robots in special educational needs (SEN) settings has highlighted several on-going challenges that can generate barriers to successful adoption. For example, [27] found that the main reasons why special needs schools do not normally use robots in their classrooms include (1) price or availability; (2) difficulty of use; (3) the limited range of activities offered; (4) limited interactions on offer; and (5) the inability to use different robots with the same software. They note that these findings are further compounded by a lack of “involving end users in the design and development of new systems [...] using a user-centred design approach for all the components, including methods of interaction, learning activities and the most suitable type of robots” (p. 59). As such, the lack of user-centred design (at the very least, including school children in deciding the types of interaction they might want or enjoy via a robot in their school) means that uptake, deployment and successful integration is difficult to achieve.

Note that, while the co-design process is a key aspect of this project and is critical to understand our design choices, it is not the main focus of this article. While we briefly outline our approach to participatory design in Sect. 3.1, this article primarily concerns itself with the setup and outcomes of the main three-weeks study, and its eventual impact on the childrens’ well-being. A detailed account of our co-design methodology is provided in [28].

In addition to the barriers to uptake, an increasing body of research has started exploring the possible emerging risks and ethical considerations that could arise and thus should be addressed in the design process of robotic interventions for autistic children [29]. Specifically in this work, we have adopted the ethical framework set out in the UNICEF report on artificial intelligence and children Policy guidance on AI for children [30]. This report lays out nine guidelines that should be considered when designing and building AI systems aimed at children. We discuss in greater details how we applied each of these guidelines to research on social robotics with vulnerable children in [31].

2.2 Impact of Social Robots on School Ecosystems

Robots have now been deployed and studied in the school environment for a significant period of time, initially as non-social tools that support (predominantly STEM) teaching [32], and more recently, as social agents embedded in the classroom environment [15].

While non-social robots were already reported to “play a positive role”, “develop creative thinking”, “increase motivation, engagement and attitude towards learning” [32], the role and impact of social robots in education is specifically investigated by Belpaeme et al. in their 2018 survey [15]. They identify two main roles for social robots in educational settings (tutors and peer learners), and show that, when focused on a restricted set of well-defined tasks, studies in child–robot interactions for education typically find learning outcomes similar to what is achieved by humans. In their conclusions, they acknowledge the open challenges linked to the need for beyond-state-of-art “fluent and contingent” interactions with the children; but they also underline the potential for social robots to “deliver a learning experience tailored to the learner, supporting and challenging students in ways unavailable in current resource-limited educational environments” while freeing up time for teachers to focus on what they do best: “providing a comprehensive, empathic, and rewarding educational experience”.

Specifically looking at long-term deployments in schools [33], pioneering work by Kanda [34, 35], who deployed in 2004 and 2007 their Robovie robot in Japanese primary schools for two months each, evidenced the difficulty of sustaining the children’ interest and engagement beyond the first two weeks. Similar difficulties were encountered in much more recent studies like [36] (with a steep decrease of engagement after the first three weeks). A five-months deployment of a small humanoid robot in a toddler environment by Tanaka et al. [37] did however evidence long-term engagement of children with the robot, albeit through teleoperated toy-like interactions (touching, dancing) with limited social richness. The researchers highlighted that robots can sustain attention of children and found that after 45 days of immersion in a childcare center over a 5-month period, long-term bonding and socialization occurred between toddlers and a social robot. In addition, they also found that “rather than losing interest, the interaction between children and the robot improved over time [and] children exhibited a variety of social and care-taking behaviors toward the robot and progressively treated it more as a peer than as a toy”.

To date, however, the research on social robot in education primarily focuses on explicit, tailored learning sessions with robots, with less spotlight on the broader impact of the robot’s presence on the school ecosystem at large. We aim to investigate this gap (in the specific context of a special needs school). Similiar to Mondada et al.’s approach to looking at the impact of robots on the home ecosystem [38], our research focuses on the impact of a social robot on the school ecosystem, a dynamic fabric comprised of students, teachers, other member of staff, parents, as well as the physical environment of the school. As such, this work feeds into the recent efforts by Charisi et al. [39] to frame the ethical principles that should guide the design of child–robot interactions in the future.

Importantly, most prior research focuses on one-to-one interactions [15] and neglects the social dimension of learning which happen during unstructured groups activities, for instance during the recess or in corridors (Vygotskian social constructivism)—with few noteworthy exceptions like the research on group interactions conducted with the KeepOn robot [22] or the research conducted by Björling and colleagues on participatory design of robots with teens [40]. Following this line of work, our social robot is consequently located in the school corridors, and it interacts autonomously with the children in a non-planned, mostly unstructured way, rather than pre-arranged, in-classroom interactions. This leads to a diversity of interaction patterns, ranging from large groups (\(6+\) children together) who purposefully decide to ‘visit’ the robot and interact for 10–15 min at a time, to short, one-to-one impromptu interactions with a child on their way to e.g. the toilets. To date, only limited literature explores the impact of such a range of types of interaction with a social robot on the dynamics of a school ecosystem; our research provides here some initial insights.

2.3 Aim and Research Questions

Based on this prior art, and the identified limitations, we set out to specifically answer two questions: (1) What are the social & spatial underpinnings of a successful integration of a social robot in a SEN school ecosystem? (in contrast to controlled one-to-one child–robot interventions) (2) Can a social robot deliver a net gain for the well-being of pupils? (in contrast to the traditional focus on quantitative improvement of select cognitive metrics). In particular, to what extent a thorough co-design process supports this goal?

The two research questions themselves were guided and informed by a participatory co-design methodology that focused on gathering autistic children’s views and perspectives of what a social robot might do in their school [28]. We also ensured teachers’ perspectives were included, helping to support the integration of a social robot into their school ecosystem. The co-design approach allowed us to identify the potential roles and behavioural characteristics for the social robot in addition to understanding which locations within the school were preferable. The key, and therein innovations of this work, lie in this participatory co-design methodology as the social robot was entirely designed and developed based by working in collaboration with autistic children and their teachers; this included what the robot should do, where and the types of interactions.

2.3.1 Hypotheses

From our research questions, we derive two hypotheses that drive our investigation: H1: A co-design approach (with children and teachers) supports the creation of a robot that successfully integrates into a school ecosystem; H2: The co-design process leads to the robot having a positive impact on the well-being of the children.

While our work is indeed driven by these hypotheses, we also acknowledge that the level of generality that would be required to firmly support or reject either of these can not be reached with a single study. We explicitly discuss this limitation in the Methodology section (Sect. 3.1). We nevertheless aim to test these hypotheses along several dimensions of the integration and adoption of the robot within the school. For H1, we assess:

  1. (a)

    to what extent the co-design process influenced the interaction design, including the choice of activities and the physical location of the robot;

  2. (b)

    whether the robot successfully integrates into the school ecosystem, by recording the robot’s usage patterns over a meaningful period of time (3 weeks);

  3. (c)

    whether the robot maintains a sustained level of interactions once the initial novelty fades, with a meaningful group of children. This can be measured from the quantitative recordings of daily interactions;

  4. (d)

    whether the pupils perceive and interact with the robot as a social agent, and not merely as a ‘tablet on wheels’. This is analysed using both quantitative data based on in-situ observations, and qualitative data from the children’s and teachers’ questionnaires;

For H2 (about the impact on well-being), we look into:

  1. (a)

    quantitative pre-/post-mood self-reports by the children, recorded everytime they interact with the robot;

  2. (b)

    behavioural data logs of student emotional well-being compiled by the school;

  3. (c)

    the reported acceptance of robot in the school by the teachers and pupils, from the questionnaire data;

  4. (d)

    qualitative feedback from teachers and pupils. In particular, the feedback on the impact of the robot on particular children provided by the school staff reveal important insights.

3 Methodology and Material

3.1 Methodology Overview

Fig. 1
figure 1

Our complete research process and journey. Stages 1 to 3 are the co-design phases, and are presented in detail in [28]. The present paper focuses on Stage 4 and the outcomes of the in-school study

We follow a mixed-method experimental protocol, with a primary focus on building qualitative insights on the design and adoption of a social robot in a special education needs school. As such, we do not make strong claim on how generalisable our results are. Instead, we aim at forming an in-depth understanding of the social and technical underpinnings of a robot-supported intervention on the well-being of autistic children.

Our methodology has two main phases (Fig. 1): (1) the iterative and participatory design of the robot behaviours and context of use; (2) a three-weeks in-situ deployment of the final robot in the school, and the observations of its use and impact on the complex and dynamic context of the school.

Fig. 2
figure 2

Mind-map of the robot’s behaviours based on the results of the focus groups

The research methods employed during the first phase comprise of: (1) a half-day workshop at the school, with two focus groups with the target population, as well as unstructured interactions with other pupils; (2) a one-day workshop among the research team, with an external invited academic, expert in child–robot interactions and responsible AI. This workshop was focused on mind-mapping the outcomes of the children focus groups (Fig. 2); (3) a two hour focus group with the school’s teachers, where the results of the children’s focus group were discussed and built upon; (4) a half-day workshop with an autistic academic, to better account for the autistic community perspective on this line of research and accordingly refine the project framing and interaction design. The detailed methodology and findings of our co-design process are detailed and presented elsewhere in [28].

As previously discussed, one of the key outcomes of this first co-design phase is the focus and framing of our research on the well-being of autistic pupils in their school, rather than robot-supported cognitive development. As a result, we have designed an experimental protocol that allows for loosely structured interactions: no scheduled interaction time slots, robot located in a common space, and available to any passing children, eliciting both one-to-one and group interactions. In addition, the robot is designed to be autonomous (no ‘wizard-of-Oz’ or teleoperation), with interactions that are however always led by the children (i.e. the robot is mainly reacting to the children’s interactions).

The second phase consists in a 3-week-long robot deployment in the school. No dedicated interaction time was scheduled for the pupils, as we wanted to see whether and how robot usage patterns would emerge from within the school ecosystem. As such, the robot was used and observed in a naturalistic context. We argue that this indeed helped formed a better understanding of the actual impact of the robot on a school dynamics compared to e.g. structured and scheduled face-to-face child–robot interactions.

The robot was positioned in a part of the school where only the secondary pupils could engage. This meant that the pupils in this study were aged between 12 and 16 and all pupils had a formal autism diagnosis.

The methods used during this phase comprise of observations made by the researcher, while observing the robot from a distance, children’s self-reported mood before and after interacting with the robot, quantitative measurements of the interactions as experienced by the robot (robot’s logs), post-hoc questionnaires administered to the children, and additional feedback provided by the teachers, school staff, and parents, both qualitative (unstructured interviews, unprompted feedback) and quantitative (behavioural incidents reports). Details of each of the measures are provided hereafter, in Sect. 3.3.2.

Before starting this study and working with the pupils and teachers, we gained ethical approval from the University of the West of England ethical board, under reference ACE.20.10.014. This was granted before any interactions or data collection started. As part of the ethical review process we carefully (and completely) considered, and designed protocols, to ensure the safety of the pupils, both physical and emotional/psychological. We worked in collaboration with the school leadership team to ensure that we located our practices around those in the school. All pupils and their parents/carers provided their consent. Teachers also provided their consent before engaging in data collection.

3.2 Robot Capabilities and Interactions

We used a Softbank Pepper humanoid robot (pictured in Fig. 4). Pepper is a 1.2 m-tall anthropomorphic robot, designed to be safe and approachable [41]. While the robot is mobile, its displacements were mostly limited to turning in place (to face or visually follow children), and the robot remained otherwise in its initial location (see Sect. 3.3.1). The robot’s arms were used for lexical gestures (i.e. gesture to support and accompany the verbal interactions), for choreographed dance sequences, as well as, in one particular activity, for physical interaction (hugging). No object manipulation or deictic gestures were performed.

The robot is also equipped with a tablet, used by the children to initiate an interaction, and answer the robot’s prompts (see below).

3.2.1 Interaction Design

Fig. 3
figure 3

Screenshots of the Pepper’s tablet interface

Choice of interaction modalities

We decided early on in the design process to avoid speech recognition, as it is notoriously difficult to achieve robust child speech recognition on a robot like Pepper, especially in a lively school environment with group interactions [42]. In addition, some of the children are non-verbal.

However, an entirely non-verbal robot would have significantly limited the range of possible activities. After the first focus group, where it was found that the robot’s voice was easy to understand for the children, we eventually decided to use an hybrid approach, with the robot talking to the children, and the children responding by clicking on large icons displayed on Pepper’s tablet (Fig. 3).

Foster the ascription of social agency to the robot and to avoid it being only used as an (expensive) ‘tablet on wheels’, we ensured that: (1) the tablet would be used exclusively to answer the robot’s prompts (e.g. no tablet-focused activities like tablet-based games); (2) the icons would be displayed synchronously with the robot voice, encouraging the children to pay attention to what the robot would say, instead of ‘skipping’ the robot by clicking an icon before the robot has finished speaking; (3) the activities were partially randomised: only a maximum of 2 or 3 random (yet related to the child’s mood) activities would be offered at a given time; the presentation order was random; the content of the activities themselves was partially random (e.g. random jokes), thus always generating slightly different situations.

While these simple strategies would not be sufficient for the child to ascribe complex mental state to the robot (like ‘the robot likes that activity better’ or ‘the robot does not want me to play that game’), they would avoid the system to be perceived as fully static and predictable, supporting the ascription of agency, and shifting the view of the robot by the children from a design stance to an intentional stance [43].

Robot Autonomy and Children–Led Interactions

In order to keep the children in control of the interaction (and thus fully maintain their agency), we decided to let the children (1) initiate the interactions themselves (following the UNICEF recommandations); (2) let them choose what they wish to do with the robot.

However, to ensure the robot is not perceived as an entirely passive system controlled by the children, the robot does automatically start addressing the child/children as soon as they are detected to be engaging (with prompts like ‘Hi! Good to see you’ or ‘Nice to see you! How are you?’).Footnote 1

Engagement detection was not performed automatically by the robot. We initially implemented an automatic engagement state machine that worked robustly in lab conditions. The system however proved unreliable when deployed (due to the number of children, and how dynamic they were). Accordingly, starting on Day 2 of the study, the number of children interacting with the robot at any given time was manually entered by the researcher observing the interaction, providing us with a much more reliable detection of engagement, including group engagement (see details below). Apart from the detection of how many children were interacting with the robot at a given time, the robot was fully autonomous.

Structure of the Interaction

Once detected as engaging, the robot greets the child (or group of children). If the child is alone, their initial ‘mood’ is recorded by showing on the tablet a mood selection screen (Fig. 3a). The child can also skip that step by pressing the blue arrow in the bottom-right corner. If a group of children are detected, the ‘mood’ selection screen is skipped entirely.

The child (or group of children) is then presented with a list of possible activities, whose presentation order is randomised (as explained above). The children also have the possibility to see more activities if the 2 or 3 randomly chosen activities do not satisfy them. The child taps on the desired activity, which then starts. Activities last between 30 s and 3 min.

At the end of each activity, the robot asks the child if they want to continue. If yes, they are taken back to the activity selection screen. Otherwise, the mood selection screen is displayed to record their emotional state at the end of the interaction.

3.2.2 Activities

Based on the outcomes of the co-design process, we implemented a set of eight activities on the robot (visible on Fig. 3b):

  • Calm dance a tai-shi inspired dance accompanied by calm music;

  • Calm music a random calm music out of five atmospheric tunes, lasting between 1’ 30” and 2’ 45”;

  • Relaxing sound: a randomly chosen continuous calm sound (sea side, birds, rain, cricket chirping, village) plays until stopped by the child or until the end of the sound file;

  • Cuddle the robot slowly hugs the child by opening its arm, and closing them in a loose hug for a few seconds. The arms are configured with a low stiffness;

  • Listening (labelled as ‘Chat’ in Fig. 3b) in this activity, the robot prompts the child to speak (for instance: “What do you want to tell me? I’m listening!”), and then simply wait for the child to stop the interaction by pressing the tablet. The robot does not attempt to process the child’s speech or to respond;

  • Story the story activity is adapted with authorisation from the LuniiFootnote 2 interactive story telling system. A total of 48 stories can be generated, by asking the child who should be the hero of the story (a boy or a girl), where the story takes place, who does the hero meet, and one special object. The decision tree and generated stories come from the Lunii system, while the voice is generated by the robot, using naoqi’s animated speech API. The stories last about 4 min each;

  • Fun dance a more energetic dance, randomly chosen from the ‘macarena’, a ‘disco’ dance, and a ‘saxophone player’ dance;

  • Jokes the robot tells two or three jokes randomly picked from a list of 50 primary-school-age jokes. The jokes are non-interactive, the robot simply pausing before the joke’s resolution.

3.3 Study Protocol and Measures

3.3.1 Setup

Figures 4 and 5 show the general setup of the study. The robot was placed in a communal area and was available to interact with from 8:30 (15 min before the normal start of the school day) until 15:15 (i.e. end of school day) each day, with the exception of about 30 min at approximately 12:30 when the researcher would have a break and the robot would be put on charge in a room inaccessible to pupils.

As previously discussed, no dedicated interaction time was scheduled for the pupils, as we wanted to see whether and how robot usage patterns would emerge from within the school ecosystem.

Fig. 4
figure 4

Photo of the physical location of the robot, here depicted while interacting with one child. In the foreground, the researcher observing the interaction, and recording on the tablet how many children are engaged, as well as observations on the interaction

Fig. 5
figure 5

Physical setup of the study at the Mendip school. The robot is placed in the middle of the Secondary corridor, in a 4 m\(\times \)4 m space. Pupils pass-by the robot, and are free to interact with the robot at any time. The researcher is sitting nearby, observing the interactions from outside the main interaction zone

3.3.2 Measures

We laid out in Sect. 2.3 how we intend to test our hypotheses. Accordingly, this section presents all the required quantitative and qualitative measures that we recorded during the study.

In-Situ Measures

Three main types of in-situ measures are recorded during the study: quantitative interaction measures (number of children engaged with the robot at a given time; log of the performed activities), children’s self-reported mood and feelings, and qualitative observations by the researchers.

Quantitative Interaction Measures Our main quantitative interaction measures consist in detailed log of the interactions with the robot: every time a child (or group of children) starts an interaction, the exact activities and time spend on each of them is logged (anonymously, as our protocol did not allow for the tracking of individual pupils).

While we attempted to use Pepper’s human detection features to automatically track behavioural data (people’s distance to the robot, trajectories, gaze direction), the poor reliability of this data (as explained in the previous section on automatic engagement detection) led to excluding it from our analysis.

Children’s Self-reported Mood The children were invited by the robot to self-report their mood and feels before and after interacting, through a screen displayed on the robot’s tablet, and pictured on Fig. 3a. The four, colour-coded categories of feelings were suggested during the teachers’ focus group to match the emotion self-reporting charts already used in the school, and familiar to the pupils.

It must be noted that these four categories do not strictly follow traditional valence/arousal emotional scales, the ‘yellow’ mood including for instance constructs with both positive (Excited) and negative (Frustrated) valence.

Qualitative Observations The researcher’s tablet also featured buttons to quickly record different categories of timestamped audio notes (children’s questions, unexpected events, general observations). A total of about 280 notes were recorded over the 13 days of the study (\(M=21.8\) per day). Notes were transcribed verbatim, and coded using thematic analysis. From this the following categories emerged:

  1. 1.

    observations relating to a child’s Mood modulation/reflection while interacting with the robot;

  2. 2.

    Unstructured, playful interaction with or around the robot;

  3. 3.

    observation evidencing Social ascriptions onto the robot (e.g. treating the robot as a social agent);

  4. 4.

    Group interactions (with other students and/or staff);

  5. 5.

    ‘Hidden’ interactions (e.g. watching the robot from a distance that would not trigger the robot’s interactive behaviours);

  6. 6.

    notes of General comments about the robot (e.g. things they liked/disliked about Pepper, interpretations of the robot’s behaviour);

  7. 7.

    observations of Sensory interactions (e.g. holding Pepper’s hand, moving Pepper, stroking Pepper’s head, hugs);

  8. 8.

    Questions asked by the children about the robot; and finally

  9. 9.

    Meta-observation about the study, including events impacting the data collection (e.g. media days, testing/demonstrating things, robot errors, unresponsive interface).

Where appropriate, notes would be coded with multiple categories (e.g. two children reflecting on their mood with the robot would be coded as categories 1 and 4).

Questionnaires

Post-hoc questionnaires were administered to the children by the researchers to assess three main constructs: Perception of the robot, Role ascription and Reasons for interaction.

The construct Perception of the robot was measured using 5-points Likert scales (adapted for use with the children, using sad to happy emojis, based on consultation with teachers to ensure the children could articulate to suit their communication needs). The 8 questions were adapted from [44]: (1) How much do you like robots in general (not just Pepper)?; (2) How much do you like our Pepper?; (3) Do you want Pepper to stay longer at school?; (4) Do you think Pepper was useful to you?; (5) Do you think Pepper is useful for the school?; (6) I think Pepper is... very boring to very entertaining; (7) I think Pepper is... very mean to very friendly; (8) Do you think Pepper could become your friend?.

Role ascription was measured similarly to [45]: after the prompt Interacting with Pepper is like..., children could circle as many answers as wished amongst Playing with a friend, Playing with a teacher, Playing with a toy, Playing with a pet. They were also free to write down additional answers.

The assessment of the children’s Reasons for interaction is based on a behavioural scheme employed by the school to analyse the motives of observed children behaviours. In collaboration with the school, we designed questions to assess the four families of behaviours: Sensory behaviours (e.g.“I liked touching Pepper”, “I liked holding Pepper’s hands”, “I liked listening to Pepper”); Escape behaviours (e.g.“I went to see Pepper because my work was hard”, “I went to see Pepper because the teacher asked me to do something I didn’t want to”, “When I am bored, I like to play with Pepper”); Attention behaviours (e.g.“I knew Pepper would play with me, listen to me”, “I knew my teacher / teaching assistant would come and get me if I was with Pepper”, “I knew my friends would play with me if Pepper was with me”); Tangible behaviours (e.g.“If I completed my work, I could spend time with Pepper”, “If I completed my work, I could spend time with Pepper and my friends”).

Children were asked to circle sentences they agreed with, amongst 21 ‘speech bubbles’ (see Fig. 6). We purposefully used speech bubbles, uttered by a genderless child emoji: by asking the children to simply agree with a neutral, third-party character’s statements, instead of themselves verbalising (as thus, owning) their own statements, we mitigate the normative pressure that could have otherwise biased some of the answers.

In addition to administering questionnaires to the children, we also captured post-hoc views of the school staff. We sent out questionnaires to all the teachers and staff who had the chance to interact with Pepper and who had pupils in the secondary corridor who were part of the study. The questionnaire was distributed the week after the study ended and all questionnaires were returned within 3 days. The questionnaires were designed to better understand the views and perspectives of the teachers with a total of 12, mostly open-ended questions. This included for instance their views on: “the impact of Pepper on their pupils, either at an individual level or on a group level”; if they “interacted with Pepper?”; whether or not “they observed any impact on the school itself? (e.g. positive and/or negative impact on the school routines or on the children’ routines)”; “If it were possible, would you like for Pepper to stay in your school?”; “Would you change anything about Pepper?”; “Would you say that Pepper has had an impact on your teaching practice or routine? If so, how did you use or integrate Pepper in your teaching and why?”; and “Would you recommend other colleagues/schools to use a robot like Pepper?”.

Fig. 6
figure 6

Examples of post-hoc questions asked to the children. Children were ask to circle the speech bubbles they agreed with. Question 1 is an example of Sensory behaviour motive; Q2 of Attention behaviour; Q3 of Tangible behaviour; Q4 of Escape behaviour

4 Results

We present hereafter our main results, in the order introduced in the Hypothesis testing (Sect. 2.3) and Measures (Sect. 3.3.2) sections.

4.1 In-Situ Quantitative Measures

4.1.1 Robot Usage

The study lasted 13 days school days. The first day involved initial robot setup and introduction, and therefore behavioural data recorded during that day was discarded. Therefore, results reported hereafter only cover the following 12 days.

Figure 7 shows the total daily time the robot interacts with children. While the robot is actively used more than 100 min a day in the first few days, it quickly diminishes to between 40 min and 60 min a day. This interaction profile was expected and corresponds to the typical novelty effect generated by the robot when first introduced in the school.

Fig. 7
figure 7

Daily interaction duration (in min) and percentage of interaction time vs total robot presence time. After Day 7 (23 June), the percentage of interaction time stabilises between 10% and 20% of the total time at school

After Day 7, the interaction time appears to stabilise to about 45–50 min of interaction per day, e.g. between 10 and 20% of the time spent in the school with the robot in the corridor.

In terms of number of interactions, children interacted \(M=25.4, stddev=8.5\) times a day over the whole 13 days, \(M=30.7, stddev=7.6\) during the first 7 days, and \(M=19.2, stddev=4.2\) during the last 6 days.

Based on this data, we consider to be past novelty starting on Day 8, and consequently our subsequent analyses and discussion about the actual impact of the robot on the school ecosystem are based on the data collected over the last 6 days of the intervention only.

Fig. 8
figure 8

Daily interaction time, split by children groups size: interactions are mostly one-to-one, especially after the novelty period

Figure 8 provides insights on the group dynamics. The novelty effect phase (until Day 8) is an initial exploration phase, and we observe that the children mostly interacted with the robot in groups. This dynamic changes in the second half of the study, where most of the interaction are one-to-one or, less often, two-to-one.

Fig. 9
figure 9

Number of children interacting with the robot over 10min-long periods, averaged over the full study. Shaded area represents the 95% confidence interval

Finally, Fig. 9 shows the average number of children interacting with the robot over a day, when split into 10 min periods.

4.1.2 Activities

Figure 10 shows the activities performed with the robot after the initial 7-day novelty phase. During that period, an average of \(M=46.3\) (std=19.2) activities were performed per day (compared to \(M=59.3\) (std=9.0) during the novelty phase). Over half of these activities were fun dances, while the other activities were performed a similar number of times each, about 5 times per day.

Fig. 10
figure 10

Distribution of daily activities performed with the robot, after the novelty period. Activities calm dances, calm music and relaxing sounds are grouped under calm activities

4.2 Children Self-reported Mood and Feelings

Fig. 11
figure 11

Sankey diagram representing changes of self-reported mood before and after interacting with the robot. The scale of each mood state reflects the proportion of answers entered by the children

At the start and end of their interactions with the robot, the children were asked to self-report their mood on a ‘mood board’ (Fig. 3a). This was only recorded for one-to-one interactions (we did not record the mood of a group), and the children could skip it if they wished (in order to reduce random responses if the children were not sure or did not want to report their feelings).

We recorded 105 interactions (out of a total of 325) where a child reported both pre- and post-interaction mood, enabling in-the-moment assessment of their change of mood. Figure 11 represents these changes of mood before and after interacting with the robot.

In 14 of these interactions (13.3% of interactions with self-report), the child’s mood improved over the interaction, to be ‘Happy’ (green color) at the end. In a meaningful amount of cases, however, (9 interactions, 8.6%), children self-reported as being ‘happy’ at the start of the interaction, but ‘angry’ at the end. Qualitative observations (reported in next section) showed that these situations were usually due to the robot malfunctioning or can be attributed to one student who would select the angry response, but verbally confirmed without prompting to the on-site researcher that this was an accident.

4.3 Qualitative In-Situ Observations

As explained in Sect. 3.3.2, one researcher was observing and annotating the interactions during the whole study.

283 notes were recorded over the 13 days of the study, coded following the methodology presented in Sect. 3.3.2. The complete list of transcribed and coded notes is available online: https://bit.ly/3nVUX5y. Table 1 summarizes the distribution.

In this section, notes relating to the hypotheses will be prioritised. In particular this section notes how the robot was used by pupils and staff (H1b, c, the extent to which the robot was perceived as a social entity (H1d), and the impact of the robot on pupils’ well-being (H2).

Table 1 Summary of the in-situ observations, after coding

4.3.1 Mood Modulation/Reflection

There were several observations of children using the robot as a means to reflect or modulate their mood.

One pupil, Student L, often interacted with the robot in an initially agitated state and would specifically select calming sounds or music.

There were instances where pupils selected an emotion that did not match their actual state. For example Student M choosing angry because he liked the way the robot responded, but verbally confirming to the researcher that he felt happy. Another instance was Student L choosing yellow to express excitement, but the robot responded negatively e.g. “Oh you’re not feeling so good?”. Student L then verbally responded “no no delete I’m excited again”. In another instance pupils “approached the robot [and one of them] clicked happy on the mood board despite being very clearly angry and stating explicitly that he was feeling angry. There was another student at the same time who responded [...] on behalf of the first student as angry at the final emotion check.”. These cases suggest that even when the robot’s response is not reflective of the children’s actual mood, they are still taking time to reflect on their, and sometimes other’s, emotions because of the robot’s presence. For some children, Pepper’s presence offered a way to discuss and express their thoughts without having to interact with a person.

There was an occasion where a child went to the robot and talked to it about his feelings. A member of staff watching from a distance remarked that this pupil would not normally talk about his emotions to staff. In a similar instance, a child approached the robot and said “I’m actually getting a little bit angry” while interacting with the mood board he selected just listen and when a second child approached he said “go away I’m having private chat with pepper”.

On other occasions a member of staff prompted a child to talk to Pepper, and he then “told Pepper about his day and it was never a great day then proceeded with couple of other activities but left the interaction seemingly a lot calmer than he initially started it.”

4.3.2 Unstructured, Playful

Other children tended to interact with the robot in a more playful manner, and treated it as a source of entertainment. The majority of these interactions involved the ‘fun dance’ activity which was the most popular of the activities (see Fig. 10).

Not all playful interactions with the robot directly involved the programmed activities. For example, a group of girls interacted with the robot then “decorated it with a rubber snake to make it look like Pepper had hair.” Similarly another student began “play boxing with Pepper, but specifically behind it in effort so as not to engage with the robot”. On another occasion children in a group were trying to “encourage others to join in and daring each other to hug the robot.”

4.3.3 Social Ascriptions

The social status of the robot varied. Some students would regularly interact with Pepper and say “hello” and “goodbye” much like any other social interaction. Some would ask Pepper “how are you?”.

Some students ascribed proto mental states to Pepper, for example saying “she likes me” if the robot looked towards them.

There was a conversation between staff and older pupils concerning these greetings, with the observation that “they don’t say goodbye to Alexa and that is different to not saying goodbye to Pepper”. In this instance there was a clear distinction between the social presence of Pepper compared to other technologies.

For pupils, the robot was only partially regarded as a social entity as seen in the above examples with pupils disclosing to Pepper, but not to other people. This could have been because Pepper did not respond or judge, though this is open to interpretation.

For other pupils, the robot afforded no social status whatsoever, Pepper may as well not have existed for all the attention they gave it.

4.3.4 Group Interactions

Many of the interactions with Pepper were as part of a group (see Fig. 8 where pupils were accompanied by either a member of staff or their peers.

Most of the interactions with staff occurred early on in the study, as staff familiarised themselves and their students with the robot. The larger groups were observed in the first half of the study. The largest of these groups were when a class came to visit the robot. The teacher accompanying the class had “changed plan do a walk and see the robot” and commented that “the class had calmed far quicker than other methods... it was good to have something else for them to focus on”.

In smaller groups of two or three the robot often acted as a focal point for the children’s interactions with each other. Examples included discussing what activity they wanted to do with the robot.

One girl who had initially been unsure of the robot interacted with it in the company of friends and by the end of one interaction had become a lot more comfortable with the robot and spontaneously describing Pepper as “cool” and “beautiful”. Having the support of peers helped to facilitate her initial interactions with the robot.

In one instance the presence of others halted an interaction of one boy “dancing alongside Pepper until a group of older children passed by and he stopped.”

4.3.5 ‘Hidden’ Interactions

There were some ‘interactions’ with the robot that did not meet the criteria for triggering the robot’s behaviour. In these cases it was often because pupils were passing by in the corridor and only verbally greeting Pepper. For some pupils this became part of routine every time they passed by. There were also occasions where pupils would watch the robot from a distance, as they were reluctant to directly engage with the robot.

4.3.6 General Comments About the Robot

One of the most frequent comments about the robot, particularly at the start of the study, was that Pepper was “creepy”. This comment was frequently accompanied by exclamations directed towards the robot asking “why is it looking at me?”. One pupil “elaborated saying that it was [because] the robot doesn’t blink” that they found the robot creepy.

Opinions varied with regards to the quality of some of Pepper’s behaviours. For example, “upon hearing the jokes the teens interacting with it decided the robot wasn’t very good.” In contrast, mere minutes later another “young teen, upon hearing the jokes ... declared [the robot] a legend.”

Though there were many instances of children having positive interactions, there were some who had strong negative feelings towards Pepper. One pupil avoided the corridor entirely. The other expressed to the on-site researcher how much he disliked the robot and wanted it to be moved on an almost daily basis. This pupil spent more time in the classroom than usual to avoid interacting with the robot. His teacher perceived this positively as the pupil was completing more work than usual. There can be indirectly positive impacts of the robot, but this view is not always shared by the pupils. We further discuss this result in the Discussion section.

Table 2 Descriptive statistics for the eight children questions related to the perception of the robot
Fig. 12
figure 12

Perception of the robot; responses are on 5-points Likert scales (labelled from not at all to very much except for Q6: very boring to very entertaining and Q7: very mean to very friendly)

4.3.7 Sensory Interactions

The physical aspect of the robot also presented opportunities for Sensory interactions with the robot. One pupil would regularly and deliberately place “his head into the robots hands asking for his head massage”. This pupil would also “insist on being very close to pepper and to the point of getting headbutted repeatedly”. One girl “was initially quite scared I think at the start of the week maybe even earlier today, but by the end of this interaction was kind of spontaneously saying things like ... “[it] is beautiful” and petting the robot on the head saying that [it] likes to be stroked on the head and other things giving it lots of cuddles being generally quite affectionate towards the robot”.

One comment from a member of staff noted that a pupil “holding the hand of the robot ... normally they have no physical touch.”

4.3.8 Questions

The majority of questions directed at and about Pepper concerned how the robot worked. They frequently centred around how the robot was controlled, and the connection between the robot and the tablet. In these instances the on-site researcher would explain that the tablet ran code that controlled Pepper, and that they (the researcher) were not using the tablet to directly control the robot.

4.4 Questionnaires and Other Post-hoc Measures

4.4.1 Children Questionnaires

After removing data from children for which no data sharing consent was obtained, and two incomplete questionnaires, \(n=30\) questionnaires were collected and analysed.

Table 2 and Fig. 12 reports the questionnaire results for the perception of the robot, from 1 (low/negative) to 5 (high/positive).

Generally, the robot was perceived positively. We can specifically observe that, compared to their general perception of robots, the Pepper robot deployed in the school was well liked. It was generally considered entertaining and friendly.

While the children were uncertain whether or not the robot could be useful to themselves in particular (18 out of 30 thought that it would, 7 that it would not), they were much more confident that it would be useful to the school in general (27 out of 30 positive opinions).

Finally, the students were split regarding whether or not the robot “could become their friend”, 21 thinking that it may, and 8 thinking that it would not.

Figure 13 presents the results of the role ascription question. While the robot was only seldom perceived as a teacher, it was roughly equally often seen as a peer, a pet or a toy.

Table 3 and Fig. 14 presents the results related to the underlying motives of the children. These statements each belong to one of the four behaviours categories identified by the school (presented in Sect. 3.3.2). Because each behaviour does not have the same number of available statements to circle, we normalise the number of responses for each behaviour by the number of statement available for this behaviour.

We found that the first reason for children to interact with the robot was a need for a sensory interaction (for instance, holding hands or stroking the robot), followed by a need for attention. The three most-selected statements related to attention were “I knew my teacher/TA would come and get me if I was with Pepper” (selected 8 times), “I go and see Pepper when I want someone to listen to me” (6 times) and “I knew my friends would play with me if Pepper was with me” (6 times).

Fig. 13
figure 13

Perception of the robot’s social role. Children could choose multiple options

Table 3 Number of children-selected statements explaining their reasons for interacting with the robot, grouped by category
Fig. 14
figure 14

Motives leading to interaction

4.4.2 Teacher’s Questionnaires

After the study ended we surveyed the teachers, teaching assistants and leaders in the school. Responses were received from teachers (\(n=2\)), teaching assistants (\(n=1\)), and school leaders (\(n=3\)). The data revealed insights to their views and perspectives. These were mixed, but shed light on what they felt Pepper could offer in their setting, based on 3-week of activity.

We first consider the teachers’ feedback. When asked about the impact of Pepper on their pupils, responses included: “Some pupils completely engaged with it [...] some pupils actively didn’t like it being in the corridor” and “Initially my students were very engaged in seeing Pepper”. Asked about the impact of Pepper in the school, they suggested that: “The corridor was quieter as certain pupils kept away from it” and “50/50, for some children Pepper caused anxiety and for others joy!”. Leading on this, one of the teachers felt Pepper had an impact on their teaching practice, saying that: “Only a couple of times, seeing Pepper was used as an incentive”. However, they felt that Pepper aroused curiosity in some of their pupils and also enjoyed saying “good morning” to Pepper. The only thing the teachers would change about Pepper would have been “more personalised interaction”.

The sole teaching assistant that responded felt that Pepper was popular, stating: “Some pupils seemed to not want to go to lessons but spend time with pepper”. They also observed/reported some benefits to including technology like Pepper: “Pupils interacting with each other, that would not normally do so”, but also warned of some possible issues: “Some pupils found it scary. The older pupils didn’t want anything to do with it”.

Responses from the leaders of the school suggest a range of impacts on the pupils. One said: “I have seen an increased amount of groups working around Pepper, but these are students who previously would not of interacted with each other. This has been pleasing to see. Individual pupils who are also on a sensory curriculum also appear to have benefited from having a sensory experience from Pepper e.g. holding hands, cuddling, dancing”, while another suggested that: “Individuals have been able to use pepper as a calming activity when arriving at school”. Another mentioned they had observed both “positive and negative interactions with Pepper”. In relation to the overall school, they reported that: “[the] secondary corridor has seemed calmer, more positive” with another suggesting: “initially the excitement meant some pupils used Pepper as an excuse not to be in class [...] later it [Pepper] was able to be used as a motivator to stay in class and be used at reward time”. When asked about what they would change about Pepper (and what Pepper did), they reported: “refine the activities Pepper offers so they are specific to the emotion, e.g. Angry—breathing exercises, counting to 10, visualisations, Low mood—music, jokes etc...perhaps linked to zones of regulation and the strategies that students use themselves—e.g. personalised to each pupil by visual or voice recognition” and “Potentially the ability to differentiate between the feelings within zones”. Asked about the role of technology like Pepper, feedback included: “It’s important to think of Pepper as a teaching aid or resource external to the classroom. It has accessibility where pupils can use it, I still do not believe it is a class room resource” and “Ability to help regulate pupils”. When the Leaders responded to how they personally felt about robots in their school/class, they reported feeling sceptical at first, but “now think there’s lots of potential to develop their ability to support students” and “if refined to link directly to the strategies taught by staff could be a useful tool”.

Taken together, and when asked about recommending Pepper (“Would you recommend other colleagues/schools to use a robot like Pepper?”), all the respondents reported recommend using Pepper to other colleagues and schools (\(n=6\)). In addition, they all stated that Pepper/social robots could be used in schools, like theirs, to support their pupils. However, one of the School Leaders commented that: “I do believe it is a niche field, that isn’t necessarily something all schools need. However from what I have seen, their use in SEND/ASD is valuable”, also that social robots: “can’t be in isolation it [Pepper] needs to be part of the wider school strategies and supports”. Noticeably, all of the respondents, apart from the Teaching Assistant, reported interacting and spending time with Pepper. They all reported “missing Pepper” and one of the school leaders reported that: “Our pupils on a sensory curriculum were gaining additional support through an aid we didn’t need to be with. This had immense benefits for them and their school day. From my own teaching, on reflection I can say key pupils who struggle in the afternoon were generally calmer if they had seen Pepper at lunch time”.

4.5 Behavioural Data from School

Fig. 15
figure 15

Behavioural incidents reported by the teachers, on the week they occurred. Students J and L are presented separately as, together, they were responsible for about half of the reported behavioural incidents. However the evolution of their behaviour over the study does not significantly deviate from the rest of the group

Within the school they have mechanisms to report and document the pupils behaviour. In doing so, behaviours noted in school are written by class teachers or teaching assistants into an online system that is use to track and monitor individuals, classes and cohorts. Behaviours are separated into ‘incidents’ (negative behaviours) and ‘information’ (changes to behaviour support plans to help make positive changes). Data are collected in real-time so the school are able to track and monitor across the school day, peak times etc, as opposed to ‘when a report is written’.

During the study and time that Pepper was in the school, data related to incidents were continued to be captured as they would have been any other week(s) described above. Figure 15 shows the ‘incident’ data for two specific children, as well as the whole secondary corridor (so all pupils who were able to engage with Pepper in the secondary part of the school). These results highlight a clear reduction in recorded incidents during the time Pepper was positioned in their corridors. Recorded incidents did increase the week after Pepper left the school and, for one child (L), returned higher than his baseline the week before Pepper arrived.

5 Discussion

We next discuss the findings in-line with the original hypotheses. The hypotheses were outlined as:

  • H1 A co-design approach (with children and teachers) supports the creation of a robot that successfully integrates into a school ecosystem;

  • H2 The co-design process leads to the robot having a positive impact on the well-being of the children.

Within each hypotheses we gathered a range of data which we discuss next, following the structure outlined in Sect. 2.3.

5.1 H1: Co-design Leads to Successful Integration into School

Evidence and data support this hypothesis. We captured data from child–robot interactions based on co-designed interactions, the robot usage patterns, sustained level of interactions, in-situ observations, and post-study pupil and teacher questionnaires.

5.1.1 Interaction Design and Location of the Robot

As data in Fig. 10 highlight, the pupils engaged successfully with all the planned/programmed child–robot activities. The Listening activity was perceived as limited, as the robot would never answer. Children, when asking for instance “how are you?” would not get an answer. A chatbot-like interaction should be investigated in future work.

Other potential activities were suggested by the teaching staff, including Mackaton (a simplified sign language used by e.g. speech therapists in the school). We had one particular instance where a non-verbal primary-school age child indeed tried to teach the robot the sign for ‘more’, and adding support for using and learning expressive gestures would certainly be an interesting development.

The school has a variety of different architectural spaces, from formal classrooms to informal social spaces; the differing characteristics of which affect the emotional states of individuals and play various roles for social relationships. The space within which the social robot occupies is therefore an important factor to consider in the successful deployment of these technologies. In the focus groups, the children expressed a clear preference for the corridor space as a location. At this particular school, the corridor is primarily used for circulation but is also often a place used for emotional regulation by the children. When the classroom becomes a space in which children are overstimulated, for example angry or frustrated, the children may leave the classroom and use the corridor as a space in which to self-regulate (often to calm down and chill out). The corridor is also a space for informal social interaction outside of the formal classroom settings as children pass by each other between lessons or during breaks. The use of the robot in the corridor continued the existing ‘culture’ of behaviours in this non-formal teaching space. The results stated above reveal that the robot was often used to facilitate children to shift from e.g. an angry zone to a happier mental space. The relative quietude of the corridor also enabled children to reveal their emotions and feelings to the robot (in the absence of fellow humans). Teachers also used the robot in this space to facilitate informal social interactions between children (including children who might otherwise not come into contact during formal classroom activities). There was no attempt by teachers to undertake formal teaching activities in the corridor with the robot. The robot maintained a presence in the school corridor space that a human would not typically undertake for such long durations; this enhanced durational time frame offers future potential for enhancing and enriching a space within the school ecosystem beyond that which is currently afforded by human interactions in the school.

5.1.2 Usage Patterns and Sustained Interactions

We also found that the robot was successfully integrated into the school ecosystem through the robot’s usage patterns. By working with the children from the outset, and co-designing what Pepper did, we saw an initial (and expected) interest in Pepper (week 1 and 2), but also sustained activity across the final week. So, while the interest/activities reduced by nearly half over the first 10 days, as expected from the vanishing novelty effect [46], they did stabilise to a meaningful level of daily interactions (more than 19 unique interactions per day in average, either one-to-one or group interactions). The intra-day usage pattern (Fig. 7) also evidence that the robot was used throughout the day, following the overall structure of the daily school routine (recess, lunch breaks, in-class periods).

5.1.3 Ascription of Agency

We also evidenced Pepper being engaged with as a social agent. Evidence from the pupil and teacher questionnaires, directly after the study concluded, suggested that observed behaviours included social ascription’s, group interactions, sensory interactions, and asking questions. So more than merely a ‘tablet on wheels’, the robot elicited a range of interaction types, and ones that facilitated group activities while others involved clear child–robot engagement. The teachers reported their pupils “engaged with Pepper” and “aroused curiosity”. Data from the pupils revealed that they felt Pepper was a “friend” more than any of the other options (pet, toy, teacher; Fig. 13), once again suggesting that they did not see Pepper as a tablet on wheels. There was a greater connection for some of the pupils than this. Indeed, when looking at the children motives for interaction (Fig. 14), sensory interactions are the often cited reason for interaction: The embodied, physical nature of Pepper was valuable and important to these pupils.

5.2 H2: Co-design Leads to Positive Impact on the Well-Being

Evidence and data support this hypothesis as well. Data from the post-study data revealed mostly positive insight to Pepper and the well-being of the pupils. It was certainly the case that not all pupils and teachers were positive, however. Pupils reported either really liking Pepper or disliking Pepper. Reports from the teachers suggest that “some pupils completely engaged with it [while] some pupils actively didn’t like it being in the corridor” and at least one pupil was observed ‘hiding’ from Pepper by spending more time in the their class and verbally telling the researchers he didn’t want Pepper in the school. By carefully and sensitively working with the pupil and teacher we proceeded with the study; conversely Pepper’s presence meant the pupil engaged with classroom activities with greater attention. However, and despite the mixed opinions, some of our data lead us to conclude that taken together this hypothesis was supported.

5.2.1 Self-reported Mood Changes

The Sankey diagram (Fig. 11) suggests that the reporting of their mood before and after interacting with Pepper either remained happy or transitioned more from frustrated, sad or angry to happy. Although these data are less conclusive to support a claim that Pepper managed their well-being. What the data started to provide us with, however, are insights to the nature of interactions pupils had with Pepper. Data on the specific activities performed with the robot (Fig. 10) revealed that fun dances and calm activities were the most popular activities engaging with Pepper, at least via the interface (touch screen). Cuddles were also popular, and taken together indicate that pupils engaged with mostly stimulating and relaxing activities; all these were included due to the outcomes of the co-design sessions with the pupils.

5.2.2 Impact on Behaviour

The social behaviour of the children in the corridor was often rowdy, for example play-fighting including grappling and wrestling. At times, children included Pepper in these play fights, for example grabbing Pepper around the neck, shoving or grabbing the robot’s arms. Mostly, this was harmless playful behaviour and the robot was not damaged. However, at times, the researcher or a passing teacher had to intervene for the safety of the robot (to stop it being damaged) or children (i.e. the robot falling and hurting a child). These behaviours evidently raise issues of health and safety when leaving robots unguarded in such contexts. School children being physically boisterous in certain school spaces would require careful consideration in terms of human supervision and intervention of a robot in future studies. The inclusion of Pepper in their physical play suggests the acceptance of robots by the children as a social agent. The robot’s physical presence as a humanoid gives affordance to its inclusion in play-fighting (which would not be the case if it were perceived as merely a tablet). Furthermore, the children do not include adults in this play-fighting, which tentatively suggests the children perceive the robot as a peer.

Figure 15 presents the results of the staff-reported behavioural incidents in the secondary corridor of the school, where the robot was located. We found that the robot’s presence tended to support the reduction in behaviour incidents. This suggests some connection between the co-design approach and providing a resources that follows the input from users (pupils) and beneficiaries (teachers). This, in turn, led to a significant decrease of reported behavioural incidents – a change observed over the whole three weeks of the study, with a return to higher levels of behavioural incidents after the robot left the school.

This finding is similar to that of [47] who suggest the main message from their co-design approach with autism professionals and autistic children was that “personalisation to the needs of the individual child at hand” (p. 3091) is paramount. We echo this, and by locating this personalisation through co-design sessions we able to go further than the work of [47], and articulate their views into practice and evaluate them in-situ (i.e. the school).

5.2.3 Acceptance of the Robot and Qualitative Feedback

The majority of children who encountered Pepper responded with positive attitudes towards the robot. It is worth contextualizing that any form of ‘change’ is often perceived negatively by the autistic children at this school, for example more disruptive behaviour at the start/end of days and at start/end of week. Even in this context of a dislike of change, the introduction of the robot was favourable for most students. The increased levels of interaction during the first 6 days reflecting the popularity of the intervention.

Results from the post-study children questionnaire (Table 2; Fig. 12) paint a positive picture of the robot’s acceptance by the children: the pupils overwhelmingly rated Pepper as being “useful for the school” in addition to finding Pepper to being “entertaining”. Despite this, data also suggested that the pupils were less positive about “Pepper becoming their friend” or that Pepper was “useful to them”. They finally reported that they preferred “our Pepper” over robots generally. This insight perhaps reveals that due to “our Pepper” being designed to suit their needs and designed with them, might support the finding that they felt more positive towards “our Pepper” than robots per se.

Despite the overwhelmingly positive response from students, it is important to note that a small group of students held negative attitudes towards the robot. Child J expressed concern and worry about the robot’s presence (although Child J held negative views on many aspects of his daily school life) on repeated occasions. His teacher’s reported that the robot presence in the corridor prompted child J to stay in the room, which ironically helped with his more formal learning. Another child had a generalized (but very strong) fear of digital technology not working correctly and the robot fitted into this collection of technologies. It was clear from the data that a small minority of children had highly negative perceptions of the robot (or robots in general) and their future presence needs very careful management by the school to successfully manage the integration of robots into schools’ ecosystems.

The assistant head teacher identified how the robot was creating interaction opportunities for groups of children to interact together, noting that even children who would not usually interact in groups, would take part. This effect seems to be mostly attributable to the initial novelty and apparent complexity of the robot, and the data revealed that the vast majority of interactions in the later stages of the study were one-to-one (Fig. 8). This highlighted that group interaction was seen as important by the head teacher (seeing this as a positive) but also that over time these interactions transfer to one-to-one. This could be interesting for two reasons. Firstly, as it could suggest that individual pupils grew in confidence to use/interface with Pepper as time went on. It might have been that the group interactions (or gatherings) helped to enable less confident pupils to use Pepper with their peers and then feel more confident towards the end of the three weeks to interact one-on-one. Second, this finding indicates that on some level (albeit only the initial two weeks) that a robot like Pepper can encourage collaborative skills and prolonging their attention span while mediating and encouraging social interaction and learning appropriate physical interactions. All of which are established in this field already; therefore our observations of social interactions of our pupils support previous findings [47, 48] and helps to build greater understanding of interaction patterns over a sustained period of time; something not well reported in robot-autism research.

6 Conclusions

6.1 A Social Robot in a SEN School: What for?

We presented in this paper the design and results of a three-weeks long deployment of a humanoid social robot in a school for autistic children. The robot was operating autonomously in a school corridor, engaging with the pupils in a range of activities co-designed with the children and the teachers. Over the course of these three weeks, the robot performed 330 interactions (most of the time one-to-one interactions), resulting in more than 16 h of continuous interaction time.

We recorded and analysed a range of different metrics, both quantitative and qualitative, including the detailed log of the activities performed with the robot, the number of children interacting with the robot at a given time, the self-reported mood and feelings of the children before and after interacting with the robot, children and teacher feedback via questionnaires, and finally, a large set of observations (more than 280 notes), that were coded post-hoc.

This large amount of data was meant to answer two main questions: can we (co-)design a social robot that ends up being ‘adopted’ by the pupils and the school staff, and integrates nicely in the broader school ecosystem? and: does it have a positive impact on the well-being of the pupils?

Our data and the qualitative feedback from the children and teachers support both of these hypotheses. The robot did integrate within the school ecosystem; it maintained a high level of usage; it was perceived as a social agent rather than a mechanistic tool; the children liked it; they thought the robot was useful the them and their school; it did have a positive impact on the well-being of several children.

Of course, the picture needs to be nuanced: some children did not like the robot, and at time, expressed this dislike in explicit ways, including a handful of cases with physical or verbal violence. While we immediately suggested to stop the study, the school staff did ask us not to: beyond the fact that some pupils were well identified by the school staff for showing patterns of rough behaviours, teachers also saw the negative reactions to the robot as opportunities for them to work with the students on their emotional regulation. We did not anticipate that teachers would exploit negative feelings toward the robot in such a way. While it did cause at that time a level of un-easiness amongst the research team , it was also a strong testimony of how the end-users—and the context of use they bring with them—are ultimately the ones deciding and shaping the use of the technology, as part of a mutual shaping [49] process, where the technology and the society influence each other, in a co-development process. We further reflect on the broader ethical framework of our study in [31].

Our study does bring some elements of answer to the broader question: “A social robot in a SEN school: what for?”

First, unlike most previous research on robotics and autism where robots were used for cognitive training, the starting point of our study design was that social robots could also contribute to the wellbeing of children. Based on insights from our participatory approach, we opted for unstructured, child–led interactions taking place outside of the classroom. This choice was well-received by the children and the teachers: a social robot can effectively contribute to the well-being of the children.

Second, the physical and social presence of the robot is important. The children engaged with numerous sensory interactions with the robot, and referred to it as a social agent that could provide them with psycho-social support. Those capabilities are permitted (and elicited) by embodied nature of the robot, and would not be possible with alternative technologies like tablets.

6.2 Limitations and Future Work

While we provide an account of a study involving the deployment of Pepper the robot in a school for autistic pupils, it is important to highlight several limitations that exist within our study.

One of the limitations of the study is that the spatial configuration of the school was already fixed during the research. It would have been beneficial if the children could have contributed suggestions and insights towards how the school space itself might be configured to facilitate and support the integration of the robot. This points towards the need for future research into the design of social-robot-enhanced architectural spaces from users’ perspectives. It also raises interesting and troublesome design challenges that situate the robot as a social agent with a need for school designers to consider social robots as one of the users (along with other human stakeholders such as teachers, students, cleaners etc) and the possible ‘right’ for robots to have their own space within the school.

We also acknowledge that what we found is mostly contextualised in the school we worked with and their practices. Other schools would most likely require a bespoke interface building to support their pupils. However, we do suggest that what we uncovered from the co-design sessions and the interface we designed (with the pupils and teachers) could be transferred and re-tested in a similar setting (SEN and autism) with similar results expected.

We acknowledge that having a researcher present to help control Pepper, solve possible technical problems, and record data, might have influenced the interactions the pupils had. Despite the researcher not being central to the area where Pepper was based, they were close enough for the pupils to clearly see them. This may have had a bearing on what they did and when they did it. Future research might consider a way to reduce this presence to capture even more naturalistic data.

Future work should also consider personalisation to create bespoke experiences; endow the robot with the ability to be taught by the children new behaviours (eg: one child wanted the robot to touch his head, he tried without success to show that to Pepper). We suggest that building up a personal interaction history between a child and the robot would lead to a continuously refined interaction experience, always a bit different, that would as well help sustain the interest over longer period of time. We also recommend that future work in this field continues to engage end users and do so in a meaningful and complete way. Simply testing solutions and ideas with users is not enough. We located new insights to child–robot interactions by working with children in meaningful ways to co-design and lead the design of interactions. This is vital if we are to (1) produce more focused and beneficial uses of robots in the future and (2) understand users needs from the ground up and to ensure they are included in decisions that impact how technology like this will behave. Finally, future work should liaise closer with teachers and professionals to better locate their needs. We did this though a focus group, but could have gone further. For example, one teacher suggested that we used Pepper in their speech and language therapy sessions (using the interactive story function), but we were not able to test this or build this into our project. Researchers need to find the time to work more closely and in a sustained manner with these individuals.

7 Notes on Results Replication

The scripts used to control the robot, as well as all the interface elements (robot tablet interface, researcher tablet interface, verbal prompts) are available online: github.com/severin-lemaignan/robots4sen-supervisor. This source code repository also includes detailed explanations about the software architecture of the system.

Robot logs, researcher notes, questionnaires and questionnaire results are available in this repository: github.com/severin-lemaignan/robots4sen-data-analysis. The repository also contains two iPython notebooks that allows the reproduction of all our data analysis results.