Next Article in Journal
Outage Probability Performance Prediction for Mobile Cooperative Communication Networks Based on Artificial Neural Network
Next Article in Special Issue
Online Signature Verification Based on a Single Template via Elastic Curve Matching
Previous Article in Journal
SensorTalk: An IoT Device Failure Detection and Calibration Mechanism for Smart Farming
Previous Article in Special Issue
Superpixel-Based Temporally Aligned Representation for Video-Based Person Re-Identification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Sensor-Based Technology for Social Information Processing in Autism: A Review

by
Andrea E. Kowallik
1,2,3 and
Stefan R. Schweinberger
1,2,3,4,5,*
1
Early Support and Counselling Center Jena, Herbert Feuchte Stiftungsverbund, 07743 Jena, Germany
2
Social Potential in Autism Research Unit, Friedrich Schiller University, 07743 Jena, Germany
3
Department of General Psychology and Cognitive Neuroscience, Friedrich Schiller University Jena, Am Steiger 3/Haus 1, 07743 Jena, Germany
4
Michael Stifel Center Jena for Data-Driven and Simulation Science, Friedrich Schiller University, 07743 Jena, Germany
5
Swiss Center for Affective Science, University of Geneva, 1202 Geneva, Switzerland
*
Author to whom correspondence should be addressed.
Submission received: 11 October 2019 / Revised: 29 October 2019 / Accepted: 30 October 2019 / Published: 4 November 2019
(This article belongs to the Special Issue Biometric Systems)

Abstract

:
The prevalence of autism spectrum disorders (ASD) has increased strongly over the past decades, and so has the demand for adequate behavioral assessment and support for persons affected by ASD. Here we provide a review on original research that used sensor technology for an objective assessment of social behavior, either with the aim to assist the assessment of autism or with the aim to use this technology for intervention and support of people with autism. Considering rapid technological progress, we focus (1) on studies published within the last 10 years (2009–2019), (2) on contact- and irritation-free sensor technology that does not constrain natural movement and interaction, and (3) on sensory input from the face, the voice, or body movements. We conclude that sensor technology has already demonstrated its great potential for improving both behavioral assessment and interventions in autism spectrum disorders. We also discuss selected examples for recent theoretical questions related to the understanding of psychological changes and potentials in autism. In addition to its applied potential, we argue that sensor technology—when implemented by appropriate interdisciplinary teams—may even contribute to such theoretical issues in understanding autism.

1. Introduction

Throughout the last decades, the number of people diagnosed with an Autism Spectrum Disorder (ASD) increased dramatically [1,2] and so did the need for high-quality diagnostic protocols and therapies. With the ongoing progress in computer sciences and hardware, a lot of creative ideas emerged on how to use sensor data to identify and observe autistic markers, support diagnostic procedures and enhance specific therapies to improve individuals’ outcomes.
ASD is a behaviorally defined group of neurodevelopmental disorders that are specified by impaired reciprocal social communication and restricted, repetitive patterns of behavior or activities (DSM-5), [3]. The symptoms are usually apparent from early childhood and tend to persist throughout life [4]. Common social impairments include a lack of social attention as evident in abnormal eye gaze or eye contact [5] and social reciprocity such as in reduced sharing of emotions in facial [6] or vocal behavior [7]. Further, only a minority of the affected people report having mutual friendships [8]. Related to the restricted and repetitive behaviors, stereotyped motor movements and speech are the stand-out features in many people with ASD [9]. Other symptoms are insisting on sameness and routines [10], special interests and hyper- or hyporeactivity to sensory input from various modalities [11]. The exact profile and severity of symptoms in people with ASD as well as their personal strengths and coping capabilities vary to a great degree, and so does their need for support.
Reasons for the increased prevalence over the last decades include a more formalized diagnostic approach and heightened awareness. The current ‘gold standard’ for a diagnosis of ASD consists of an assessment of current behavior, a biographical anamnesis, and a parental report, all collected and evaluated by a trained multi-professional team [12]. Although the screening and diagnostic methods for ASD improved throughout the last years, many affected people, especially women [13] and high functioning people, still receive a late diagnosis. Since early interventions have been shown to be most effective for improving adaptive behavior, as well as IQ and language skills [14], there is continued demand for methods promoting early assessment in order to avoid follow-up problems. In this context, progress in automatic and sensor-assisted identification of ASD-specific behavioral patterns could make an important contribution to an earlier and less biased diagnosis.
Even beyond assessment, advances in digital technology are highly relevant for autism, and in more than just one way. First, there is some evidence that many autistic people show behavioral tendencies to interact with technology and to potentially prefer such interactions to interactions with humans. It is thought that autistic traits are related to systemizing, the drive to analyze how systems work, as well as to predict, control and construct systems [15]. In this context, high information technology (IT) employment rates are often used as a proxy for higher rates of strong systemizers in a population. Intriguingly, recent research from the Netherlands reported that the prevalence of childhood autism diagnoses, but not of two control neurodevelopmental diagnoses (i.e., ADHD and dyspraxia), was substantially higher in Eindhoven, a classical (IT) region, when compared to two control regions (Utrecht and Haarlem) that had been selected for high demographic and socio-economic similarity in criteria other than the proportion of IT-related jobs [16]. Second, and qualifying any simplistic interpretation of this correlative (but not necessarily causal) relationship, there is evidence that technology can potentially provide powerful social support for children with autism. For instance, children with ASD often perform better with a social robot than a human partner (e.g., in terms of enhanced levels of social behavior towards robots), tend to perceive interactions with robots as positive [17], and subsequently show reduced levels of repetitive or stereotyped actions. For a recent review, see Pennisi et al. [18].
Scientific interest in the utilization of sensor technology to gain an understanding of people with ASD has increased considerably in the recent past. Some fields of research focus on different neurobiological assessments and try to identify autism-specific signals or ‘biomarkers’ to better understand the neurobiological underpinnings of the disorder. Good overviews covering methods including electroencephalography (EEG), magnetoencephalography (MEG), and functional magnetic resonance imaging (fMRI) are provided by Billeci et al. [19] and by Marco et al. [20]. Other research focused on an autonomic activity such as heart rate variability (HRV) or skin conductance responses (SCR). These can be studied with Wearable devices, typically in the context of emotional monitoring in ASD as seen in a review by Taj-Eldin, et al. [21]. Applications in VR environments [22,23,24] have also been reviewed as promising methods to train and practice social skills.
The aim of this review is to provide an overview of the current state of research using sensor-based technology in the context of ASD. We focus on sensor technology that is applicable without constraining natural movement, and on sensory input from the face, the voice, or body movements. Accordingly, this review does not consider evidence from wearable technology, VR, or psychophysiological and neurophysiological recordings. Note also that while we provide details of our procedure for identifying relevant original findings to enhance reproducibility, this paper represents a thematic review in which we pre-selected for contents as described below, and in which we occasionally discuss additional relevant findings that were not formally identified by this literature search. For instance, we may refer to some key findings regarding psychological theories of human social or emotional communication where relevant, even when the findings were not obtained with individuals with autism.

2. Literature Search

We performed parallel literature searches on Web of Science, PsychInfo, PubMed, and IEEE Xplore. Considering the rapid advance of technology throughout the last decades, we focussed on the last 10 years (2009–2019) to give an overview of recent developments in this field. Searches were performed on 29 May 2019. Our search terms were (autis* OR ASD OR Asperger) AND (sensor OR sensors) AND (fac* OR voice* OR body motion* OR person* OR regoc* OR identi* OR emotio* OR diagnos*). Furthermore, the considered language was restricted to English. Overall, the search resulted in 386 articles (Web of Science N = 185, IEEE Explore N = 109, PubMed N = 52, PsychInfo N = 40). Note that we only included publications with original data from an ASD or high-risk group, such that new algorithms on preexisting data sets were not included. We also included the reference lists of identified articles with respect to additional screening for relevant publications. After removing duplicates, this selection process resulted in a total of 36 articles. These are discussed below, and those articles that focused on assessment or intervention are additionally summarized in Tables 1 and 2, respectively.

3. Supporting Assessment: Identification of ASD Related Features

3.1. Facial Information

3.1.1. Facial Movements

For facial movements, researchers tended to focus on emotional facial information, following the idea that impaired social communication in ASD can be framed as deficits in emotional communication regarding both perception and expression [25], and the finding that people with ASD may show reduced or idiosyncratic emotional expressions [6,26]. Against this context, it may be useful to keep in mind that ASD-specific impairments in face perception are not restricted to emotional expressions, but also affect other aspects such as facial identity [27]. Similarly, expression in social communication may be affected in subtle ways that go beyond emotional expressions [28]. The original articles on sensor-based assessment in terms of identification of autism spectrum disorders (ASD)-related features, regarding facial movements and other forms, represented in this review are listed in Table 1.
Samad et al. [29] compared facial expressions from an ASD and a typically developing (TD) group, with 8 participants each. Computing facial curvatures from 3D point cloud data, they found equally intense but more asymmetrical facial expressions in the ASD group. Another study [30] used a single webcam mounted on a TV screen to record toddlers’ spontaneous facial expressions when confronted with emotional cartoons. Comparing descriptive data from small groups with five children each, they reported that the lower face seemed to be more important in distinguishing between the ASD and TD groups. Leo et al. [31] presented a new processing pipeline on 2D video data that aimed at assessing facial expressions in ASD children specifically. They estimated the production skills for individual children based on verbal instructions to express these emotions, separately for individual face parts and emotions. The performance of 17 boys with ASD was variable, and some boys exhibited differential production scores for different categories of emotions. Note that this finding is in broad agreement with theories that emphasize category-specific mechanisms, and with componential approaches to emotion [32,33]. Samad et al. [34] used a storytelling avatar, also with the aim of eliciting spontaneous emotions to differentiate between ASD and TD groups with 10 participants each. Comparing 3D data on the level of facial action units (AUs) they found overall lower AU activations in the ASD group and lower correlations between the AUs. The deviant activation in AUs 6, 12, and 15 (cheek raiser, lip corner puller, and lip corner depressor, respectively), were found to be promising markers for ASD. As a more serious screening approach, the App ‘Autism & Beyond’ [35] recorded toddlers’ facial responses to certain stimuli with the front camera and classified them into positive, neutral, or negative. Using a large database of 1756 children, the associated facial expressions, along with eye gaze and parental reports, with Autism Spectrum risk status. Comparisons between high- and low-risk groups revealed that high risk for ASD was associated with higher frequencies of neutral facial expressions, and with lower frequencies of positive expressions.

3.1.2. Eye Gaze

One major domain in which autistic people are frequently described to behave unusually relates to oculomotor behavior, including low levels of eye contact during communication, and low levels of directional signaling via eye gaze. In fact, a current target article challenges the common belief that autistic people lack social interest in others, and suggests that this interpretation could be erroneously elicited by unusual behavior such as low levels of eye contact [36]. While this may be a useful context to keep in mind when reading this section, there is agreement that the assessment of eye gaze can help to identify behavioral patterns that are relevant for autism [37].
Chawarska and Shic [38] compared toddlers with ASD and TD of different age groups (2 and 4 years old) with an eye-tracking system while they watched neutral faces. Although all groups spent equal time looking at the screen, a restricted scanning pattern in the ASD group was found, with relative neglect of the mouth area. Additionally, the older ASD group spent less time looking at inner facial features in general, even compared to the younger ASD group. According to the authors, this indicates a different scanning pattern of children with ASD that emerges throughout early childhood and suggests less looking at the mouth as the best early predictor for ASD. Liu et al. [39] used a machine-learning framework on eye-tracking data of 29 ASD vs. 29 TD children (mean age 7.90 years) looking at images of faces embedded in a learning task with repeated presentations. Their proposed framework was able to identify ASD children both from age- (mean age 7.86 years) and IQ-matched (mean age 5.74 years) control groups with high classification accuracy (up to 88.5%). Król and Król [40] used eye-tracking to study the effect of including temporal information into spatial eye-tracking of face stimuli, thus creating scan paths for 21 ASD and 23 TD individuals (mean age around 16 years). They found a difference in face-scanning not only in spatial properties but also in temporal aspects of eye gaze, even within short exposures (about 2 s) to a facial image. Classification of group membership based on a machine-learning algorithm on spatial and temporal data led to better accuracy (55.5%) than classification based on spatial data alone (53.9%), although overall accuracy was rather low. Note that discrepancies between classification accuracies in this study and the study by Liu et al. [39] need to be seen in the context of specific conditions of each study, and could potentially be attributed to the facial stimuli and trial numbers used, the experimental task instructions, methodological differences in data analysis, or differences in the samples tested.
In a more general study on the visual scanning of natural scene images, Wang et al. [41] compared an ASD and a TD group (N = 20 and 19, respectively) at different levels of perception. The ASD group was found to fixate more towards the center of an image (pixel level), less on objects in general (object level) and less on certain objects (e.g., faces or objects indicated by social gaze), but more on manipulatable objects (semantic level).
Overall, facial data suggest that facial movements, either in response to emotional stimuli or as an imitation of seen facial expressions, comprise relevant features that may be markers for autism. Recordings of spontaneous emotional responses could be of particular benefit when assessing small children or nonverbal people, although the importance of age-appropriate task instructions should be considered. Across several studies, eye gaze data reveal different scanning patterns for people with ASD, particularly when viewing faces. However, we believe that further systematic research into fixation patterns and scan paths for more complex natural scenes could enrich our current understanding with additional insights. Similarly, the observation of reduced eye contact/mutual gaze in ASD, reviewed in Jaiswal et al. [50], points at a technologically challenging but theoretically relevant field of future investigations using sensor technology.

3.2. Voices

A scientific metaphor that has become somewhat popular describes the voice as an “auditory face” [55], emphasizing the fact that voices, just like faces, provide rich information about a person ́s emotions, but also about identity, gender, socioeconomic or regional background, or age (for review, see [56]). In the context of autism, and like for faces, researchers focused on deficits in vocal emotional communication and pointed out that these deficits tend to affect multiple modalities including voices, faces, and body movement [57]. At the same time, deficits in vocal communication may also affect other aspects such as vocal identity perception [58,59], and vocal expression of autistic people in communication could be affected beyond emotional expressions.
One path for detecting auditory markers for autism in voices is linked to the symptom of repetitive behaviors or ‘vocal stereotypies’. One study conducted a subspace analysis from acoustic data of autistic children and reported a good detection of vocalized non-word sounds [42]. Subsequently, Min and Fetzner [43] also used subspace learning for vocal stereotypies and trained dictionaries to differentiate between vocal stimming (a nonverbal vocalization often observed in autism) and other noises. Using a small sample of 4 children with ASD who lacked verbal communication (age not reported), the authors could e.g., detect vocal stimming and predict perceived frustration reasonably well, although the study was regarded as preliminary. For verbal children, other potential vocal markers could include prosody. Marchi et al. [44] created an evaluation database in three languages with ASD and TD children’s emotionally toned voices which were analyzed using the COMPARE feature set. Groups comprised between 7 and 11 individuals per language, and children were between 5 and 11 years old. Comparisons between groups and emotions showed relatively poor classification performance ASD children’s’ voices, particularly for ‘Anger’ for both the Swedish and English dataset, and for ‘Afraid’ for Hebrew-speaking children with ASD. Although using a pre-existing dataset, a study on automatic voice perception showed that an algorithm successfully classified up to 61.1% of voice samples of children actually diagnosed with and without autism [60]. Ringeval et al. [45] assessed verbal prosody in ASD children, children with pervasive developmental disorder (PDD), children with specific language impairment (SLI), and TD children (aged around 9–10 years, with 10–13 children per clinical group). Specifically, these authors recorded performance during an imitation task for sentences with different intonations (e.g., rising, falling). The rising intonation condition was reported to best discriminate between groups, and the authors interpreted their findings as indicating a pronounced pragmatic impairment to prosodic intonation in ASD.
When compared with facial data, the use of sensors for automatic analysis of markers for autism in voices clearly is still at an early and preliminary stage. The studies reviewed above that use original data typically rely on small data sets from few people. Although initial findings suggest that the systematic search for vocal markers of autistic traits could be highly promising, more research is clearly warranted at this stage. When considering that similar impairments in facial and vocal emotional processing could be present in ASD [57], multisensory assessment of emotional behavior in future studies could be particularly promising.

3.3. Body Movement

The identification of ASD-associated movement patterns has been the subject of intense research, primarily focusing data from accelerometer sensors [42,61]. As the display of stereotypical body movement patterns is a core symptom of ASD, it has been a target feature in computer vision. Gonçalves et al. [46] used a simple gesture recognition algorithm on 3D visual data automatically detecting hand-flapping movements. Validation of these data with 2D video data suggested that automatic ‘hand flapping’ detection delivers valuable information for monitoring autistic children, as in the case of special needs schools. Jazouli et al. [47] also used the same sensor but based their analysis on a $P Point-Cloud Recogniser to automatically detect body rocking, hand flapping, fingers flapping, hand on the face and hands behind back with an overall mean accuracy of 94%.
Rynkiewicz et al. [47] studied the role of non-verbal communication in the setting of an ADOS-2 assessment for children aged 5–10. They used a 3D sensor for automatic gesture analysis of the upper body, while the boys (N = 17) and girls (N = 16) with high-functioning ASD performed two assessment-related tasks. Females had a higher gesture index than boys although they had less verbal communication skills and a more impaired ability to read mental states from faces. The authors suggested that the vivid use of gestures in girls (generally less common in our current understanding of the autistic phenotype) may contribute to possible under-diagnosis of autism in females. This might further lead to the more general use of automatic gesture analysis that is currently performed by professional human raters in the course of these assessments. A study by Anzulewicz et al. [48] used the touch and inertial sensors of a tablet to assess the specific movement patterns of children with autism (N = 37) and an age- and gender-matched TD group (N = 45) while playing simple games. Apart from interesting findings including the use of greater force, larger and more distal gestures and faster screen taps in the ASD group, they also tested different machine learning algorithms to classify between groups, with promising results.
In summary, the use of 3D data has been preferred by researchers investigating body movements. While it seems to be possible to detect certain stereotypical movements, there is still a lack of large-scale studies. However, even coarser movement indices (including, for instance, gesture indices or general movement patterns) may also provide meaningful information for the identification and differentiation of autistic behavioral markers.

3.4. Multimodal Information

Multimodal approaches to automatic analysis of behavior in ASD are still infrequent. Samad et al. [49] used facial motion, eye-gaze and hand movement analysis of adolescents with (N = 8) and without ASD (N = 8) on tasks involving 3D expressive faces. They found a reduced synchronization of facial expression and visual engagement with the stimuli in the ASD group, as well as poorer correlations in eye-gaze and hand movements. Jaiswal et al. [50] designed an automatic approach based on 3D data of facial features and head movements that detected ASD (vs. TD and vs. comorbid ASD and ADHD, with N between 11 and 22 per group) in adults with high accuracy in classification performance. Their recorded data consisted of participants reading, listening to and answering to questions taken from the ‘Strange Stories’ task [62] often used in ASD-assessments.
Overall, the integration of multimodal information, unfortunately, is not yet common in this field, despite the fact that there are reasons to expect that multimodal assessments will provide ample additional information (for instance, on synchronization or complementarity of signals from different channels). Thus, it can be expected that future use of multimodal assessments has the potential to substantially improve both the identification of markers for autistic traits and classification results.

3.5. Cognition and Social Behavior

A substantial body of research has suggested that autism is characterized by changes in cognitive information processing. For instance, autistic people may have difficulties in cognitively simulating the observed actions of communication partners – a process that has been related to the so-called human mirror neuron system, and that has been inferred via neurophysiological recordings [63]. Via experiments with social behavioral assessments, autism has also been related to a deficit in forming a theory of mind about other people [64]. Researchers in this field have long focused on inferences about “core” cognitive deficits from observational data, even though results can be very inconsistent across situations or studies [65,66]. We also welcome an increasing awareness of the danger that researchers wrongly infer putative core deficits by misinterpreting observational data [36]. Of course, inferences about core cognitive processes most typically need to be made from observational information. It is this context in which we foresee that sensor data, beyond their immediate value for an individual study, might eventually contribute to resolving theoretical controversies. One of these is the degree to which we should frame cognitive changes in autism in terms of general “core” deficits, or rather in terms of domain-specific deficits that emerge in a concrete perceptual and interactional context.
The automatic analysis of playing behavior could be a promising screening tool to identify stages of development, developmental delays and specific forms of deficits. In a preliminary study with only a single completely analyzed data set of a TD child [51], play data from smart toys with embedded acceleration sensors, as well as simultaneous video and audio recordings, were automatically characterized as certain forms of play behavior (exploratory, relational, functional), roughly indicating a certain stage of development. While this approach seems interesting, current limitations include the verbal prompting nature of their setting, which only allows children with good verbal skills to be assessed.
Joint attention is a basic communicational skill of sharing attention between communicational agents towards an object and has long been considered to be a precursor of a theory of mind [67] which may be reduced in people with ASD. In a study investigating a robot-assisted joint attention task [52] in children with and without ASD (N = 16 each), they captured participants’ orientation using 3D sensors over a timespan. In addition to findings of significantly less joint attention in the ASD group’s interaction with the robot, the ASD group also showed decreased micro-stability in the trunk area that was tentatively interpreted as a consequence of increased cognitive cost. In another study [53], attentional and orienting responses to name calls were studied in toddlers, aged 16–31 months, with ASD (N = 22) and TD (N = 82). Achieving a high intra-class-correlation with human raters, automated coding offered a reliable method to detect the differential social behavior of toddlers with ASD, who responded to name calls less often and with longer latency. Petric et al. [54] designed and further developed a robot-assisted subset of the ADOS assessment. They included certain social tasks (name-calling, joint attention) to assess information on eye gaze, gestures, and vocal utterances. The agreement of the robot’s classifier and clinicians’ judgments were evaluated as promising, but the results should be regarded with caution given the very small sample of children (ASD N = 3, TD N = 1).
Overall, while researchers have begun to assess joint attention in ASD, sensor-based assessment of other functional domains of cognition and social behavior still awaits systematic research. In particular, research on a few putative “core” areas of social cognition, including on observational tasks that probe theory of mind in ASD, is currently lacking.

4. Supporting Interventions

While accurate early diagnosis and an assessment of specific impairments are crucial, they also are prerequisites that inform environmental adjustment, intervention, and training approaches which ultimately can be valuable for the individual person with ASD following diagnosis. Technically-assisted training often has the benefit of being readily available, problem-specific, cost-effective, and widely accepted by affected children. Additionally, smart responses of training systems that give reliable, immediate feedback and appraisal can be highly beneficial for fast learning results. Note that although we identified many publications on interventions aiming at autism as a target condition, many of these reported conceptual or technological contributions and few of them presented original data from people with ASD that qualified them for inclusion in this review (cf. Section 2). The original articles presented in this review, regarding sensor-based supporting interventions for ASD, are listed in Table 2.

4.1. Emotion Expression and Recognition

Emotion expression, especially from the face, is considered highly relevant in autism research. A game called FaceMaze [68], in combination with automatic online expression recognition of the user was specifically developed to improve facial expression production. In this game, children played a maze game while posing ‘happy’ or ‘angry’ facial expressions to overcome obstacles in the game. In a pre-post-rating with naive human raters, quality ratings for both trained expressions (happy and angry expression) in ASD children (N = 17, aged 6–18) increased in the post-test while ratings for an untrained emotion (surprise) did not change. Another smaller study created a robot-child-interaction and tested it with three children with ASD that were to imitate a robot’s facial expression [69]. The robot correctly recognized the children’s’ imitated expressions through an embedded camera in half of the cases. In these cases, it was able to give immediate positive feedback. It also correctly did not respond in about one-third of the trials where there was no imitation by the participants. Piana, et al. [70] designed a serious game with online 3D data acquisition that trained children with ASD (N = 10) in several sessions to recognize and express emotional body-movements. Both emotion expression (mean accuracy gain = 21%) and recognition (mean accuracy gain = 28%) increased throughout the sessions. Interestingly, performance in the T.E.C. (Test of Emotion Comprehension, assessing emotional understanding more generally), increased as well (mean gain = 14%).

4.2. Social Skills

Robins et al. [71] created an interactive robot (KASPAR) with force sensitive resistor sensors. They later planned to use KASPAR for robot-assisted play to teach touch, joint attention and body awareness [72], although conclusive data on experimental results from interactions between individuals with autism and KASPAR may still be in the pipeline. Learning social skills also presupposes attention to potential social cues and social engagement. Costa et al. [73] reported preliminary research on using the LEGO mindstorm robots with adolescents with ASD (N = 2), in an attempt to increase openness and induce communication since the participants actively had to provide verbal commands or instructional acts. They reported that the two participants behaved differently, one being indifferent, and one being increasingly interested in the interaction. Wong and Zhong [74] used a robotic platform (polar bear) to teach children with ASD (N = 8) social skills. They found, that within five sessions an increase in turn-taking, joint attention and eye contact was observable, resulting in overall 90% achievement of individually defined goals.
Greeting is a basic element of communication. In a greeting game with 3D body movement as well as voice acquisition [75], a participant would play an avatar with his own face, learning to greet (vocalization, eye contact and waving) and get immediate appraisal upon success. A single case study suggested that this intervention can be effective at teaching greeting behavior. As a more complex pilot intervention, Mower et al. [76] created the embodied conversational agent ‘Rachel’ that acted as an emotional coach in guiding children through emotional problem-solving tasks. Of their two participants with ASD, audio and video data were acquired for post hoc analysis and tentatively suggested that the interface could elicit interactive behavior.
Overall, there is some evidence that sensor technology can improve social skills in people with autism, and the use of sophisticated robotic platforms can be regarded as particularly promising. As limitations, it needs to be noted that all studies that met the criteria to be included in this review only tested very few participants, and that there typically were no real-world follow-up tests reported. As a result, a systematic quantitative assessment of treatment effects and effect sizes, as well as a comparison with more conventional interventions (e.g., social competence training) will require substantial cross-disciplinary research. Moreover, most studies were driven by a combination of theoretically interesting and technically advanced approaches, and from the perspective of typical development. Designing more user-centered and irritation-free approaches could promote both usability and motivation for people with autism to engage in technology-driven interventions.

5. Monitoring

Monitoring a child’s emotional state or behavioral changes can be crucial for the outcomes of a learning environment. As discussed above, emotional expressions from people with ASD may differ in several respects from those of TD people. As a result, there is a higher risk that caregivers or interaction partners overlook or misinterpret the emotional state of people with autism.
Del Coco et al. [77] created a humanoid and tablet-assisted therapy setup that was trained to monitor behavioral change in children with ASD via a video processing module. Besides creating a visual output of behavioral cues, they computed a score for affective engagement (happiness related features) from visual cues such as facial AUs, head pose and gaze that provides the practitioner with a behavioral trend along with the treatment. Dawood et al. [78] used facial expressions, eye gaze and head movements to identify five discrete emotional states of young adults with ASD in learning situations (e.g., anxiety, engagement, uncertainty). Their resulting model yielded a high validity in identifying emotional states of participants with high-functional ASD. At the same time, a lower validity was found for TD participants, suggesting differential facial expressions of certain emotional states in ASD. For monitoring social interactions, Winoto et al. [79] created a machine-learning-based social interaction coding of 3D data around a target user. Kolakowska et al. [80] approached automatic progress recognition with different tablet games. Over a 6-month time window, they were able to identify movement patterns in their study group of children with ASD (N = 40), that not only related to development in fine motor skills but also other fields like communication and socio-emotional skills. Overall, these initial studies suggest that sensor-based monitoring of emotional and behavioral changes may support caregivers in optimizing learning outcomes.

6. Discussion

The studies discussed above demonstrate substantial research activities towards using sensor-based technology in the context of autism overall, with attention to multiple aspects including diagnosis/classification and intervention. At the same time, it appears that much current research is largely driven by fast technological progress in terms of innovative engineering and data analysis methods. It remains a significant challenge to reconcile these developments with the specific testing of psychological or neuroscientific theories regarding functional changes and potentials in autism. Similarly, systematic studies with theory-driven protocols and larger samples are required to evaluate in more detail both the diagnostic and interventional potential of sensor-based technology. For the ultimate goal of evaluating its practical relevance, quantitative assessments of diagnostic sensitivities and specificities, or of treatment effect sizes, will be as important as will be comparative studies with more traditional approaches to diagnosis and intervention.
One of many examples of how sensor technology has the potential to go beyond application, and to contribute to current neurocognitive theories of communication is related to the theory of a tight link between perception and motor action in communication. This link now has been firmly established in speech communication [81], but there are reasons to believe that perception and action are also closely linked in nonverbal emotional and social communication. For instance, listening to laughter normally activates premotor and primary motor cortex [82], and may involuntarily elicit orofacial responses in a perceiver in parallel. In turn, there also is initial evidence that voluntary motor imitation can actually facilitate facial emotion recognition, particularly in people with high levels of autistic traits [83] who are thought to engage less in spontaneous imitation. A consistent theoretical account for such findings is that imitation, and covert sensorimotor simulation of others’ actions, may be based in part on the so-called mirror neuron system. This system consists of neurons that fire not only when a person performs an action, but also when s/he observes the same action in another individual. However, the human mirror neuron system is thought be specifically impaired in autism [63], and a subset of promising intervention approaches for autism using neurofeedback [84] are based on this theory. However, it should be noted that the underlying theory remains disputed [85].
Findings such as those by Lewis and Dunn [83] may be taken to suggest that interventions that promote facial imitation of emotions in autistic people should also support their abilities for emotion recognition and bidirectional communication. However, it is technically challenging to objectively quantify the degree of facial imitation, and in fact, a limitation of the study by Lewis and Dunn was that these authors failed to quantify imitation beyond simply asking participants to rate their own degree of imitation. Other studies measured facial imitation more objectively but typically did so by measuring the facial muscle response for selected target action units with electromyography (EMG, e.g., [86,87]). Although this can provide an objective measure of facial imitation, the fact that the method uses recording electrodes attached to facial muscles has many drawbacks. For instance, one concern is that this technology could draw the participants’ attention to their own facial behavior, which in turn could influence facial action. We believe that contact- and irritation-free assessment of imitation as provided by modern sensor and real-time facial emotion recognition technologies is the method of choice to promote better understanding not only the role of spontaneous facial imitation in emotion recognition in normal communication, but also to determine the potential role of impaired links between perception and action for communication difficulties in people with autism.
While the research discussed in this review appears to underline a great potential for the use of sensor technology, in particular in the context of autism, it is equally clear that many current tests of assessments or interventions will benefit in validity from a clear conceptual framework of autism spectrum disorders in the developmental perspective. At present, and honoring findings of large individual variability within both people with ASD and TD, results that were obtained with only a few participants (not always well described, and sometimes obtained in the absence of a TD group) or with experimental groups that are not comparable with respect to their basic characteristics (e.g., age, gender, IQ) need to be interpreted with caution in order to avoid biased or overgeneralized interpretation of individual study findings.
Other potential obstacles relate to sophisticated developments (and costs) of some of the systems used, which make them unlikely to become available in greater quantities. Moreover, even readily available systems may get discontinued or run out of support, such as in the case of Microsoft’s Kinect in 2017, and this provides great challenges for large-sample research in autism which often takes years to complete. Research aiming at training and modeling behavior of people with ASD also will increasingly need to consider usability, to the extent that the relevant systems are to be used by individuals with ASD, their parents, caregivers, and therapists.
Finally, compared to the typical approach of developing sensor-based technology with neurotypical individuals before applying it to people with autism, a more promising strategy may be one in which technology design originates from a user-centered perspective, with autistic people as users actively involved in the process. Such an approach has been forcefully advocated by Rajendran [88], who argues that this may both enhance our understanding of autism and promote better inclusivity of people with autism in an increasingly digital world. At the same time, such technologies ultimately can be useful for people without autism as well. This is because autism is seen as a unique window into social communication and social learning more generally.

7. Conclusions

Technical advancements and the ongoing developments in sensor technology and data science promise to unlock huge potentials for the diagnosis and understanding of autism, and for supporting affected people with training or intervention programs that can be tailored to their specific needs. At the same time, living up to these potentials calls for a concerted and interdisciplinary effort in which computer scientists, engineers, psychologists, and neuroscientists jointly collaborate in large-scale research projects that can uncover, in a quantitative manner, the efficiency of these approaches. In our view, this will be the route not only for establishing routine contributions to evidence-based diagnosis and interventions in autism [89] but also to ensure that more people with autism can genuinely benefit from tailor-made technology.

Funding

Previous research by SRS on related topics has been funded by a grant from the Bundesministerium für Bildung und Forschung (BMBF), in a project on an irritation-free and emotion-sensitive training system (IRESTRA; Grant Reference: 16SV7210), and another BMBF project on the psychological measurement of anxiety in human-robot interaction (3DimIR, Grant Reference 03ZZ0459B).

Acknowledgments

AEK and SRS would like to thank the Herbert Feuchte Stiftungsverbund for supporting the research in the Social Potentials in Autism Research Unit (www.autismus.uni-jena.de). SRS would like to thank the Swiss Center for Affective Sciences at the University of Geneva, Switzerland, for hosting a sabbatical leave in summer 2019 during which this paper was written.

Conflicts of Interest

The authors declare no conflict of interest. In particular, funding bodies had no role in the planning, collection, or interpretation of evidence reviewed in this paper; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Matson, J.L.; Kozlowski, A.M. The increasing prevalence of autism spectrum disorders. Res. Autism Spectr. Disord. 2011, 5, 418–425. [Google Scholar] [CrossRef]
  2. Weintraub, K. The prevalence puzzle: Autism counts. Nature 2011, 479, 22–24. [Google Scholar] [CrossRef] [PubMed]
  3. American Psychiatric Association. Diagnostic and Statistical Manual of Mental Disorders (DSM-5®); American Psychiatric Pub.: Washington, DC, USA, 2013. [Google Scholar]
  4. Seltzer, M.M.; Shattuck, P.; Abbeduto, L.; Greenberg, J.S. Trajectory of development in adolescents and adults with autism. Ment. Retard. Dev. Disabil. Res. Rev. 2004, 10, 234–247. [Google Scholar] [CrossRef] [PubMed]
  5. Dawson, G.; Toth, K.; Abbott, R.; Osterling, J.; Munson, J.; Estes, A.; Liaw, J. Early Social Attention Impairments in Autism: Social Orienting, Joint Attention, and Attention to Distress. Dev. Psychol. 2004, 40, 271–283. [Google Scholar] [CrossRef] [PubMed]
  6. Brewer, R.; Biotti, F.; Catmur, C.; Press, C.; Happé, F.; Cook, R.; Bird, G. Can neurotypical individuals read autistic facial expressions? Atypical production of emotional facial expressions in autism spectrum disorders. Autism Res. 2016, 9, 262–271. [Google Scholar] [CrossRef] [PubMed]
  7. Green, H.; Tobin, Y. Prosodic analysis is difficult… but worth it: A study in high functioning autism. Int. J. Speech Lang. Pathol. 2009, 11, 308–315. [Google Scholar] [CrossRef]
  8. Orsmond, G.I.; Krauss, M.W.; Seltzer, M.M. Peer relationships and social and recreational activities among adolescents and adults with autism. J. Autism Dev. Disord. 2004, 34, 245–256. [Google Scholar] [CrossRef]
  9. Leekam, S.R.; Prior, M.R.; Uljarevic, M. Restricted and repetitive behaviors in autism spectrum disorders: A review of research in the last decade. Psychol. Bull. 2011, 137, 562–593. [Google Scholar] [CrossRef]
  10. Szatmari, P.; Georgiades, S.; Bryson, S.; Zwaigenbaum, L.; Roberts, W.; Mahoney, W.; Goldberg, J.; Tuff, L. Investigating the structure of the restricted, repetitive behaviours and interests domain of autism. J. Child Psychol. Psychiatry 2006, 47, 582–590. [Google Scholar] [CrossRef]
  11. Leekam, S.R.; Nieto, C.; Libby, S.J.; Wing, L.; Gould, J. Describing the sensory abnormalities of children and adults with autism. J. Autism Dev. Disord. 2007, 37, 894–910. [Google Scholar] [CrossRef]
  12. Falkmer, T.; Anderson, K.; Falkmer, M.; Horlin, C. Diagnostic procedures in autism spectrum disorders: A systematic literature review. Eur. Child Adolesc. Psychiatry 2013, 22, 329–340. [Google Scholar] [CrossRef]
  13. Green, R.M.; Travers, A.M.; Howe, Y.; McDougle, C.J. Women and Autism Spectrum Disorder: Diagnosis and Implications for Treatment of Adolescents and Adults. Curr. Psychiatry Rep. 2019, 21, 22. [Google Scholar] [CrossRef] [PubMed]
  14. Oono, I.P.; Honey, E.J.; McConachie, H. Parent-mediated early intervention for young children with autism spectrum disorders (ASD). Evid.-Based Child Heal. A Cochrane Rev. J. 2013, 8, 2380–2479. [Google Scholar] [CrossRef]
  15. Baron-Cohen, S.; Richler, J.; Bisarya, D.; Gurunathan, N.; Wheelwright, S. The systemizing quotient: an investigation of adults with Asperger syndrome or high-functioning autism, and normal sex differences. Philos. Trans. R. Soc. B Boil. Sci. 2003, 358, 361–374. [Google Scholar] [CrossRef] [PubMed]
  16. Roelfsema, M.T.; Hoekstra, R.A.; Allison, C.; Wheelwright, S.; Brayne, C.; Matthews, F.E.; Baron-Cohen, S. Are autism spectrum conditions more prevalent in an information-technology region? A school-based study of three regions in the Netherlands. J. Autism Dev. Disord. 2012, 42, 734–739. [Google Scholar] [CrossRef]
  17. Dautenhahn, K. Socially intelligent robots: Dimensions of human–robot interaction. Philos. Trans. R. Soc. B Boil. Sci. 2007, 362, 679–704. [Google Scholar] [CrossRef]
  18. Pennisi, P.; Tonacci, A.; Tartarisco, G.; Billeci, L.; Ruta, L.; Gangemi, S.; Pioggia, G. Autism and social robotics: A systematic review. Autism Res. 2016, 9, 165–183. [Google Scholar] [CrossRef]
  19. Billeci, L.; Sicca, F.; Maharatna, K.; Apicella, F.; Narzisi, A.; Campatelli, G.; Calderoni, S.; Pioggia, G.; Muratori, F. On the Application of Quantitative EEG for Characterizing Autistic Brain: A Systematic Review. Front. Hum. Neurosci. 2013, 7, 442. [Google Scholar] [CrossRef] [Green Version]
  20. Marco, E.J.; Hinkley, L.B.N.; Hill, S.S.; Nagarajan, S.S.; Hinkley, L.B.N. Sensory processing in autism: A review of neurophysiologic findings. Pediatr. Res. 2011, 69, 48R–54R. [Google Scholar] [CrossRef]
  21. Taj-Eldin, M.; Ryan, C.; O’Flynn, B.; Galvin, P. A Review of Wearable Solutions for Physiological and Emotional Monitoring for Use by People with Autism Spectrum Disorder and Their Caregivers. Sensors 2018, 18, 4271. [Google Scholar] [CrossRef]
  22. Parsons, S.; Mitchell, P. The potential of virtual reality in social skills training for people with autistic spectrum disorders. J. Intellect. Disabil. Res. 2002, 46, 430–443. [Google Scholar] [CrossRef] [PubMed]
  23. Bellani, M.; Fornasari, L.; Chittaro, L.; Brambilla, P. Virtual reality in autism: State of the art. Epidemiol. Psychiatr. Sci. 2011, 20, 235–238. [Google Scholar] [CrossRef]
  24. Pan, X.; Hamilton, A.F.D.C. Why and how to use virtual reality to study human social interaction: The challenges of exploring a new research landscape. Br. J. Psychol. 2018, 109, 395–417. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Bölte, S.; Hubl, D.; Feineis-Matthews, S.; Prvulovic, D.; Dierks, T.; Poustka, F. Facial affect recognition training in autism: Can we animate the fusiform gyrus? Behav. Neurosci. 2006, 120, 211–216. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Kasari, C.; Sigman, M.; Mundy, P.; Yirmiya, N. Affective sharing in the context of joint attention interactions of normal, autistic, and mentally retarded children. J. Autism Dev. Disord. 1990, 20, 87–100. [Google Scholar] [CrossRef] [PubMed]
  27. Weigelt, S.; Koldewyn, K.; Kanwisher, N. Face identity recognition in autism spectrum disorders: A review of behavioral studies. Neurosci. Biobehav. Rev. 2012, 36, 1060–1084. [Google Scholar] [CrossRef]
  28. Sheppard, E.; Pillai, D.; Wong, G.T.-L.; Ropar, D.; Mitchell, P. How Easy is it to Read the Minds of People with Autism Spectrum Disorder? J. Autism Dev. Disord. 2016, 46, 1247–1254. [Google Scholar] [CrossRef]
  29. Samad, M.D.; Bobzien, J.L.; Harrington, J.W.; Iftekharuddin, K.M. [INVITED] Non-intrusive optical imaging of face to probe physiological traits in Autism Spectrum Disorder. Opt. Laser Technol. 2016, 77, 221–228. [Google Scholar] [CrossRef]
  30. Del Coco, M.; Leo, M.; Carcagni, P.; Spagnolo, P.; Luigi Mazzeo, P.; Bernava, M.; Marino, F.; Pioggia, G.; Distante, C. A Computer Vision based Approach for Understanding Emotional Involvements in Children with Autism Spectrum Disorders. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 1401–1407. [Google Scholar]
  31. Leo, M.; Carcagnì, P.; Distante, C.; Spagnolo, P.; Mazzeo, P.; Rosato, A.; Petrocchi, S.; Pellegrino, C.; Levante, A.; De Lumè, F. Computational Assessment of Facial Expression Production in ASD Children. Sensors 2018, 18, 3993. [Google Scholar] [CrossRef]
  32. Phillips, M.L.; Young, A.W.; Senior, C.; Brammer, M.; Andrew, C.; Calder, A.J.; Bullmore, E.T.; Perrett, D.I.; Rowland, D.; Williams, S.C.R.; et al. A specific neural substrate for perceiving facial expressions of disgust. Nature 1997, 389, 495–498. [Google Scholar] [CrossRef]
  33. Sander, D.; Grandjean, D.; Scherer, K.R. An Appraisal-Driven Componential Approach to the Emotional Brain. Emot. Rev. 2018, 10, 219–231. [Google Scholar] [CrossRef] [Green Version]
  34. Samad, M.D.; Diawara, N.; Bobzien, J.L.; Taylor, C.M.; Harrington, J.W.; Iftekharuddin, K.M. A pilot study to identify autism related traits in spontaneous facial actions using computer vision. Res. Autism Spectr. Disord. 2019, 65, 14–24. [Google Scholar] [CrossRef]
  35. Egger, H.L.; Dawson, G.; Hashemi, J.; Carpenter, K.L.; Sapiro, G. 23.1 Autism and Beyond: Lessons from an Iphone Study of Young Children. J. Am. Acad. Child Adolesc. Psychiatry 2018, 57, S33–S34. [Google Scholar] [CrossRef]
  36. Jaswal, V.K.; Akhtar, N. Being versus appearing socially uninterested: Challenging assumptions about social motivation in autism. Behav. Brain Sci. 2019, 42, e82. [Google Scholar] [CrossRef]
  37. Tanaka, J.W.; Sung, A. The “Eye Avoidance” Hypothesis of Autism Face Processing. J. Autism Dev. Disord. 2016, 46, 1538–1552. [Google Scholar] [CrossRef]
  38. Chawarska, K.; Shic, F. Looking but not seeing: Atypical visual scanning and recognition of faces in 2 and 4-year-old children with autism spectrum disorder. J. Autism Dev. Disord. 2009, 39, 1663–1672. [Google Scholar] [CrossRef]
  39. Liu, W.; Li, M.; Yi, L. Identifying children with autism spectrum disorder based on their face processing abnormality: A machine learning framework. Autism Res. 2016, 9, 888–898. [Google Scholar] [CrossRef]
  40. Król, M.; Król, M.E. A Novel Eye Movement Data Transformation Technique that Preserves Temporal Information: A Demonstration in a Face Processing Task. Sensors 2019, 19, 2377. [Google Scholar] [CrossRef]
  41. Wang, S.; Jiang, M.; Duchesne, X.M.; Laugeson, E.A.; Kennedy, D.P.; Adolphs, R.; Zhao, Q. Atypical Visual Saliency in Autism Spectrum Disorder Quantified through Model-Based Eye Tracking. Neuron 2015, 88, 604–616. [Google Scholar] [CrossRef] [Green Version]
  42. Min, C.-H.; Tewfik, A.H. Novel pattern detection in children with autism spectrum disorder using iterative subspace identification. In Proceedings of the 2010 IEEE International Conference on Acoustics, Speech and Signal Processing, Dallas, TX, USA, 14–19 March 2010; pp. 2266–2269. [Google Scholar]
  43. Min, C.-H.; Fetzner, J. Vocal Stereotypy Detection: An Initial Step to Understanding Emotions of Children with Autism Spectrum Disorder. In Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 18–21 July 2018; pp. 3306–3309. [Google Scholar]
  44. Marchi, E.; Schuller, B.; Baron-Cohen, S.; Lassalle, A.; O’Reilly, H.; Pigat, D.; Golan, O.; Friedenson, S.; Tal, S.; Bolte, S. Voice Emotion Games: Language and Emotion in the Voice of Children with Autism Spectrum Conditio. In Proceedings of the 3rd International Workshop on Intelligent Digital Games for Empowerment and Inclusion (IDGEI 2015) as part of the 20th ACM International Conference on Intelligent User Interfaces, IUI 2015, Atlanta, GA, USA, 29 March–1 April 2015; p. 9. [Google Scholar]
  45. Ringeval, F.; DeMouy, J.; Szaszak, G.; Chetouani, M.; Robel, L.; Xavier, J.; Cohen, D.; Plaza, M. Automatic Intonation Recognition for the Prosodic Assessment of Language-Impaired Children. IEEE Trans. Audio Speech Lang. Process. 2010, 19, 1328–1342. [Google Scholar] [CrossRef] [Green Version]
  46. Gonçalves, N.; Costa, S.; Rodrigues, J.; Soares, F. Detection of stereotyped hand flapping movements in Autistic children using the Kinect sensor: A case study. In Proceedings of the 2014 IEEE international conference on autonomous robot systems and competitions (ICARSC), Espinho, Portugal, 14–15 May 2014; pp. 212–216. [Google Scholar]
  47. Jazouli, M.; Majda, A.; Merad, D.; Aalouane, R.; Zarghili, A. Automatic detection of stereotyped movements in autistic children using the Kinect sensor. Int. J. Biomed. Eng. Technol. 2019, 29, 201–220. [Google Scholar] [CrossRef]
  48. Anzulewicz, A.; Sobota, K.; Delafield-Butt, J.T. Toward the Autism Motor Signature: Gesture patterns during smart tablet gameplay identify children with autism. Sci. Rep. 2016, 6, 31107. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  49. Samad, M.D.; Diawara, N.; Bobzien, J.L.; Harrington, J.W.; Witherow, M.A.; Iftekharuddin, K.M. A Feasibility Study of Autism Behavioral Markers in Spontaneous Facial, Visual, and Hand Movement Response Data. IEEE Trans. Neural Syst. Rehabil. Eng. 2017, 26, 353–361. [Google Scholar] [CrossRef] [PubMed]
  50. Jaiswal, S.; Valstar, M.F.; Gillott, A.; Daley, D. Automatic detection of ADHD and ASD from expressive behaviour in RGBD data. In Proceedings of the 2017 12th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2017), Washington, DC, USA, 30 May–3 June 2017; pp. 762–769. [Google Scholar]
  51. Westeyn, T.L.; Abowd, G.D.; Starner, T.E.; Johnson, J.M.; Presti, P.W.; Weaver, K.A. Monitoring children’s developmental progress using augmented toys and activity recognition. Pers. Ubiquitous Comput. 2012, 16, 169–191. [Google Scholar] [CrossRef]
  52. Anzalone, S.M.; Tilmont, E.; Boucenna, S.; Xavier, J.; Jouen, A.-L.; Bodeau, N.; Maharatna, K.; Chetouani, M.; Cohen, D.; Group, M.S. How children with autism spectrum disorder behave and explore the 4-dimensional (spatial 3D+ time) environment during a joint attention induction task with a robot. Res. Autism Spectr. Disord. 2014, 8, 814–826. [Google Scholar] [CrossRef]
  53. Campbell, K.; Carpenter, K.L.; Hashemi, J.; Espinosa, S.; Marsan, S.; Borg, J.S.; Chang, Z.; Qiu, Q.; Vermeer, S.; Adler, E. Computer vision analysis captures atypical attention in toddlers with autism. Autism 2019, 23, 619–628. [Google Scholar] [CrossRef]
  54. Petric, F.; Hrvatinić, K.; Babić, A.; Malovan, L.; Miklić, D.; Kovačić, Z.; Cepanec, M.; Stošić, J.; Šimleša, S. Four tasks of a robot-assisted autism spectrum disorder diagnostic protocol: First clinical tests. In Proceedings of the IEEE Global Humanitarian Technology Conference (GHTC 2014), San Jose, CA, USA, 10–13 October 2014; pp. 510–517. [Google Scholar]
  55. Belin, P.; Fecteau, S.; Bedard, C. Thinking the voice: neural correlates of voice perception. Trends Cogn. Sci. 2004, 8, 129–135. [Google Scholar] [CrossRef]
  56. Schweinberger, S.R.; Kawahara, H.; Simpson, A.P.; Skuk, V.G.; Zäske, R. Speaker perception. Wiley Interdiscip. Rev. Cogn. Sci. 2014, 5, 15–25. [Google Scholar] [CrossRef]
  57. Philip, R.C.M.; Whalley, H.C.; Stanfield, A.C.; Sprengelmeyer, R.; Santos, I.M.; Young, A.W.; Atkinson, A.P.; Calder, A.J.; Johnstone, E.C.; Lawrie, S.M.; et al. Deficits in facial, body movement and vocal emotional processing in autism spectrum disorders. Psychol. Med. 2010, 40, 1919–1929. [Google Scholar] [CrossRef] [Green Version]
  58. Schelinski, S.; Roswandowitz, C.; von Kriegstein, K. Voice identity processing in autism spectrum disorder. Autism Res. 2017, 10, 155–168. [Google Scholar] [CrossRef]
  59. Skuk, V.G.; Palermo, R.; Broemer, L.; Schweinberger, S.R. Autistic traits are linked to individual differences in familiar voice identification. J. Autism Dev. Disord. 2019, 49, 2747–2767. [Google Scholar] [CrossRef] [PubMed]
  60. Fruhholz, S.; Marchi, E.; Schuller, B. The Effect of Narrow-Band Transmission on Recognition of Paralinguistic Information from Human Vocalizations. IEEE Access 2016, 4, 6059–6072. [Google Scholar] [CrossRef]
  61. Gilchrist, K.H.; Hegarty-Craver, M.; Christian, R.B.; Grego, S.; Kies, A.C.; Wheeler, A.C. Automated detection of repetitive motor behaviors as an outcome measurement in intellectual and developmental disabilities. J. Autism Dev. Disord. 2018, 48, 1458–1466. [Google Scholar] [CrossRef] [PubMed]
  62. Happé, F.G.E. An advanced test of theory of mind: Understanding of story characters’ thoughts and feelings by able autistic, mentally handicapped, and normal children and adults. J. Autism Dev. Disord. 1994, 24, 129–154. [Google Scholar] [CrossRef]
  63. Oberman, L.M.; Hubbard, E.M.; Mccleery, J.P.; Altschuler, E.L.; Ramachandran, V.S.; Pineda, J.A. EEG evidence for mirror neuron dysfunction in autism spectrum disorders. Cogn. Brain Res. 2005, 24, 190–198. [Google Scholar] [CrossRef]
  64. Schneider, D.; Slaughter, V.P.; Bayliss, A.P.; Dux, P.E. A temporally sustained implicit theory of mind deficit in autism spectrum disorders. Cognition 2013, 129, 410–417. [Google Scholar] [CrossRef]
  65. Low, J.; Apperly, I.A.; Butterfill, S.A.; Rakoczy, H. Cognitive Architecture of Belief Reasoning in Children and Adults: A Primer on the Two-Systems Account. Child Dev. Perspect. 2016, 10, 184–189. [Google Scholar] [CrossRef] [Green Version]
  66. Kulke, L.; Von Duhn, B.; Schneider, D.; Rakoczy, H. Is Implicit Theory of Mind a Real and Robust Phenomenon? Results from a Systematic Replication Study. Psychol. Sci. 2018, 29, 888–900. [Google Scholar] [CrossRef]
  67. Premack, D.; Woodruff, G. Does the chimpanzee have a theory of mind? Behav. Brain Sci. 1978, 1, 515–526. [Google Scholar] [CrossRef] [Green Version]
  68. Gordon, I.; Pierce, M.D.; Bartlett, M.S.; Tanaka, J.W. Training Facial Expression Production in Children on the Autism Spectrum. J. Autism Dev. Disord. 2014, 44, 2486–2498. [Google Scholar] [CrossRef]
  69. Leo, M.; Del Coco, M.; Carcagni, P.; Distante, C.; Bernava, M.; Pioggia, G.; Palestra, G. Automatic emotion recognition in robot-children interaction for ASD treatment. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Santiago, Chile, 11–18 December 2015; pp. 145–153. [Google Scholar]
  70. Piana, S.; Malagoli, C.; Usai, M.C.; Camurri, A. Effects of Computerized Emotional Training on Children with High Functioning Autism. IEEE Trans. Affect. Comput. 2019, 1, 1. [Google Scholar] [CrossRef]
  71. Robins, B.; Amirabdollahian, F.; Ji, Z.; Dautenhahn, K. Tactile interaction with a humanoid robot for children with autism: A case study analysis involving user requirements and results of an initial implementation. In Proceedings of the 19th International Symposium in Robot and Human Interactive Communication, Viareggio, Italy, 13–15 September 2010; pp. 704–711. [Google Scholar]
  72. Mengoni, S.E.; Irvine, K.; Thakur, D.; Barton, G.; Dautenhahn, K.; Guldberg, K.; Robins, B.; Wellsted, D.; Sharma, S. Feasibility study of a randomised controlled trial to investigate the effectiveness of using a humanoid robot to improve the social skills of children with autism spectrum disorder (Kaspar RCT): A study protocol. BMJ Open 2017, 7, e017376. [Google Scholar] [CrossRef] [PubMed]
  73. Costa, S.; Resende, J.; Soares, F.O.; Ferreira, M.J.; Santos, C.P.; Moreira, F. Applications of simple robots to encourage social receptiveness of adolescents with autism. In Proceedings of the 2009 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Minneapolis, MN, USA, 3–6 September 2009; pp. 5072–5075. [Google Scholar]
  74. Wong, H.; Zhong, Z. Assessment of robot training for social cognitive learning. In Proceedings of the 2016 16th International Conference on Control, Automation and Systems (ICCAS), Gyeongju, South Korea, 16–19 October 2016; pp. 893–898. [Google Scholar]
  75. Uzuegbunam, N.; Wong, W.-H.; Cheung, S.-C.S.; Ruble, L. In MEBook: Kinect-based self-modeling intervention for children with autism. In Proceedings of the 2015 IEEE International Conference on Multimedia and Expo (ICME), Turin, Italy, 29 June–3 July 2015; pp. 1–6. [Google Scholar]
  76. Mower, E.; Black, M.P.; Flores, E.; Williams, M.; Narayanan, S. Rachel: Design of an emotionally targeted interactive agent for children with autism. In Proceedings of the 2011 IEEE International Conference on Multimedia and Expo, Barcelona, Spain, 11–15 July 2011; pp. 1–6. [Google Scholar]
  77. Del Coco, M.; Leo, M.; Carcagnì, P.; Fama, F.; Spadaro, L.; Ruta, L.; Pioggia, G.; Distante, C. Study of mechanisms of social interaction stimulation in autism spectrum disorder by assisted humanoid robot. IEEE Trans. Cogn. Dev. Syst. 2017, 10, 993–1004. [Google Scholar] [CrossRef]
  78. Dawood, A.; Turner, S.; Perepa, P. Affective Computational Model to Extract Natural Affective States of Students with Asperger Syndrome (AS) in Computer-Based Learning Environment. IEEE Access 2018, 6, 67026–67034. [Google Scholar] [CrossRef]
  79. Winoto, P.; Chen, C.G.; Tang, T.Y. The development of a Kinect-based online socio-meter for users with social and communication skill impairments: A computational sensing approach. In Proceedings of the 2016 IEEE International Conference on Knowledge Engineering and Applications (ICKEA), Singapore, Singapore, 28–30 September 2016; pp. 139–143. [Google Scholar]
  80. Kołakowska, A.; Landowska, A.; Anzulewicz, A.; Sobota, K. Automatic recognition of therapy progress among children with autism. Sci. Rep. 2017, 7, 13863. [Google Scholar] [CrossRef] [Green Version]
  81. Pickering, M.J.; Garrod, S. An integrated theory of language production and comprehension. Behav. Brain Sci. 2013, 36, 329–347. [Google Scholar] [CrossRef] [Green Version]
  82. Warren, J.E.; Sauter, D.A.; Eisner, F.; Wiland, J.; Dresner, M.A.; Wise, R.J.S.; Rosen, S.; Scott, S.K. Positive Emotions Preferentially Engage an Auditory–Motor “Mirror” System. J. Neurosci. 2006, 26, 13067–13075. [Google Scholar] [CrossRef]
  83. Lewis, M.B.; Dunn, E. Instructions to mimic improve facial emotion recognition in people with sub-clinical autism traits. Q. J. Exp. Psychol. 2017, 70, 1–14. [Google Scholar] [CrossRef]
  84. Pineda, J.A.; Carrasco, K.; Datko, M.; Pillen, S.; Schalles, M. Neurofeedback training produces normalization in behavioural and electrophysiological measures of high-functioning autism. Philos. Trans. R. Soc. B Boil. Sci. 2014, 369, 20130183. [Google Scholar] [CrossRef] [Green Version]
  85. Caramazza, A.; Anzellotti, S.; Strnad, L.; Lingnau, A. Embodied Cognition and Mirror Neurons: A Critical Assessment. Annu. Rev. Neurosci. 2014, 37, 1–15. [Google Scholar] [CrossRef] [Green Version]
  86. Dimberg, U.; Thunberg, M.; Elmehed, K. Unconscious facial reactions to emotional facial expressions. Psychol. Sci. 2000, 11, 86–89. [Google Scholar] [CrossRef] [PubMed]
  87. Korb, S.; With, S.; Niedenthal, P.; Kaiser, S.; Grandjean, D. The Perception and Mimicry of Facial Movements Predict Judgments of Smile Authenticity. PLoS ONE 2014, 9, e99194. [Google Scholar] [CrossRef] [PubMed]
  88. Rajendran, G. Virtual environments and autism: A developmental psychopathological approach. J. Comput. Assist. Learn. 2013, 29, 334–347. [Google Scholar] [CrossRef]
  89. Wong, C.; Odom, S.L.; Hume, K.A.; Cox, A.W.; Fettig, A.; Kucharczyk, S.; Brock, M.E.; Plavnick, J.B.; Fleury, V.P.; Schultz, T.R. Evidence-Based Practices for Children, Youth, and Young Adults with Autism Spectrum Disorder: A Comprehensive Review. J. Autism Dev. Disord. 2015, 45, 1951–1966. [Google Scholar] [CrossRef]
Table 1. Original Articles on Sensor-Based Assessment in Terms of Identification of autism spectrum disorders (ASD)-related features.
Table 1. Original Articles on Sensor-Based Assessment in Terms of Identification of autism spectrum disorders (ASD)-related features.
Domain of BehaviorReferenceSolution NameSensor/ParametersSample (Mean Age)Setting and StimuliResults
Facial MovementsSamad et al., 2016 [29]Sony EVID70 Color Camera, 3dMD2D, 3D Imaging ASD: 8 (13 y), TD: 8 (16 y)Emotion Recognition Task: 12 3D FacesASD: Intense, Asymmetrical Facial Expressions with Lack of Differential Facial Muscle Actions
Facial MovementsDel Coco et al., 2017 [30]Webcam2D Imaging ASD: 5 (5 y), TD: 5 (age-, Gender-Matched)Watching 9 Emotion Eliciting Videos Taken from Famous CartoonsASD Descriptively Exhibited Less Facial Behavioral Complexity; Lower Face Seemed More Significant Than Upper Face to Distinguish Between TD-ASD
Facial MovementsLeo et al., 2018 [31]Off-The-Shelf Camera2D Imaging ASD: 17 (9 y)Emotion Production Task: 4 Expressions (Happiness, Sadness, Fear, Anger)Descriptive Scores for Upper and Lower Face Parts
Facial MovementsEgger et al., 2018 [35]iPhone2D Imaging (Front Camera)High Risk: 555 (3 y), Low Risk: 1199 (3 y)Watching Video ClipsMore Neutral and Less Positive Emotional Reactions in High-Risk Group
Facial MovementsSamad et al., 2019 [34]PrimeSense3D Imaging ASD: 10 (14 y), TD: 10 (13 y)Watching Story Content Narrated by The Animated AvatarASD: Overall Lower FAU Activation, Higher Activations of FAU 15, Limited Activation of FAUs In Response to Encountering Negative Emotional States; Concurrent Activations of Several FAU Pairs Are Found to Be Absent
Eye GazeChawarska and Shic, 2009 [38]iView X™ RED Eye-TrackingASD: 44, TD: 30Watching Color Images of Affectively Neutral Female FacesASD: Attended to Visual Scenes Containing Faces to A Similar Extent; ASD Look Less at Inner Facial Features, Older ASD Spent Less Time with Inner Features Than Younger, ASD Spent Less Time Looking at the Mouth
Eye GazeLiu et al., 2016 [39]Tobii T60 Eye-TrackingASD: 29 (8 y), TD-Age: 29 (8 y), TD-Ability: 29 (6 y)Face Memorization and Recognition Task Classification Accuracy Predicting ASD: 88.51% (p < 001)
Eye GazeKról and Król, 2019 [40]SMI RED250Eye-TrackingASD: 21 (16 y), TD: 23 (16 y)Face Perception Tasks with 60 Color Photographs of Faces (FACES Database)Prediction ASD vs. TD Group Membership: Based on Only “Spatial” Information (M = 53.9%) Was Significantly Smaller Than That of the “Spatial + Temporal” Model (M = 55.5%)
Eye GazeWang et al., 2015 [41]Tobii X300Eye-TrackingASD: 20 (31 y), TD:19 (32 y) Free Viewing Task with 700 Natural Scene Images (OSIE Dataset)ASD: Higher Saliency Weights for Low Level Properties, Lower Weights for Object and Semantic-Based Properties
VoiceMin and Tewfik, 2010 [42]MicrophoneAudio, AccelerometerASD: 4 No Context Given22/24 Vocal Stimming Events Detected by Classifier
VoiceMin and Fetzner, 2018 [43]MicrophoneAudio, 2D ImagingASD: 4 No Context GivenTrained Dictionaries Detect Vocal Stimming with Sensitivity: 73–93%
VoiceMarchi et al., 2015 [44] Zoom H1 Handy Recorder (Hebrew, English), Zoom H4 With RØDE NTG-2 Microphone (Swedish)AudioASD: 7 (Hebrew), 11 (Swedish), 9 (English); TD:10 (Hebrew), 9 (Swedish), 9 (English)Repeating Sentences From 9 Emotional StoriesASD: Generally Perform Poorer, In English And Swedish Angry Was Poorly Performed, In Hebrew Afraid Was Poorly Performed
VoiceRingeval et al., 2010 [45]Logitech USB Desktop MicrophoneAudioAD: 12 (10 y), PDD-NOS: 10 (10 y), SLI: 13 (10 y), TD: 73 (10 y)Reading 26 Sentences with Certain Prosodic Dependencies (Descending, Falling, Floating, Rising)AD: Intonation for Falling Floating and Especially Rising Was Worse Compared to TD
Body MovementGonçalves et al., 2014 [46] Microsoft Kinect Color Depth Sensors, IR Emitter, MicrophoneASD: 5 (9 y) playing session with robotGood Detection of Hand Flapping with DTW, But Susceptible to Noise
Body MovementJazouli et al., 2019 [47] Microsoft Kinect V1Color Sensor, IR Depth Sensors, IR Emitter, MicrophoneASD: 5 (5–10 y), TD: 5 (Training Data)No Context Given94% Overall Recognition Rate For Stereotyped Gesture Recognition ($P Algorithm)
Body MovementRynkiewicz et al., 2016 [47]Microsoft KinectColor Sensor, IR Depth Sensors, IR EmitterASD: 33 (5–10 y) ADOS-2-Tasks (Cartoon Task, Demonstration Task)ASD: Females Present Better Non-Verbal Skills (Gestures), Although Communication Skills Were Lower
Body MovementAnzulewicz et al., 2016 [48] iPadTouch Sensor; AccelerometerASD: 37 (4 y); TD: 45 (5 y)Playing 2 Serious Games (Sharing, Coloring)Differences in Pressure Going into the Device as Well as Differences in Gesture Kinematics and Form
MultimodalSamad et al., 2017 [49]Sony EVI-D70, Mirametrix S2 2D Imaging, Eye TrackerASD: 8 (13 y); TD: 8 (16 y)Emotion Recognition and Manipulation Task Of 3D Faces ASD: Have Uncontrolled Manifestation of FAU 12, Spontaneous Facial Responses Are Not Synchronized with Their Visual Engagement with Facial Expressions, Poor Correlation in Dynamic Eye-Hand- Movements
MultimodalJaiswal et al., 2017 [50]Microsoft Kinect V2Color Sensor, IR Depth Sensors, IR Emitter, MicrophoneASD: 22; ADHD: 4, ASD + ADHD: 11, TD: 18 Read and Listen to A Set Of 12 Short Stories (From ‘Strange Stories’ Task), Accompanied By 2–3 QuestionsClassifier sensitivity for Control vs. Clinical Condition: 96.4%; ASD+ADHD vs. ASD: 93,9%
Social BehaviorWesteyn et al., 2012 [51]BlueSense, Video CamerasTouch Sensors, Motion Sensor, Microphone, 2D ImagingHigh-Risk: 1, TD: 10 Adults; 12 ChildrenPlaying with Smart Toys with Scripted Play PromptsRetrieval Score of 59% for a Single Child Using Models Constructed from Adult Play Data
Social BehaviorAnzalone et al., 2014 [52]Microsoft Kinect; NaoColor Sensor, IR Depth Sensors, IR Emitter, MicrophoneASD: 16v (9 y); TD: 16 (8 y)Nao Prompts JA by Gazing/By Gazing and Pointing/By Gazing, Pointing and Vocalizing at PicturesASD: Trunk Position Showed Less Stability in 4D Compared to TD Controls, Gazing Exploration Showed Less Accuracy
Social BehaviorCampbell et al., 2019 [53]iPad2D Imaging Sensor (Front Camera)ASD: 22 (2 y); TD: 82 (2 y)Reaction to Name-Calling While Watching VideosASD: Classifying by Atypical Orientation: Sensitivity: 96%, Specificity: 38%
Social BehaviorPetric et al., 2014 [54]Nao RobotCameras, Microphones, Ultrasound Range Sensors, Tactile Sensors, Force Sensitive Resistors, Accelerometers ASD: 3 (5–8 y), TD: 1 (6 y)ADOS-Tasks (Name Calling, JA, Play Request, Imitation)Descriptively High Correspondence of Human Rater and Algorithm
Note: Age in sample description typically refers to mean age of participants per group, with the exception of a few studies which report either age ranges, individual age of single cases, or did not specify exact age.
Table 2. Original Articles on Sensor-Based Supporting Interventions in ASD.
Table 2. Original Articles on Sensor-Based Supporting Interventions in ASD.
Domain of BehaviorReferenceSolution NameSensor/ParametersSample (Mean Age)Setting and StimuliResults
Emotion Gordon et al., 2014 [68]Webcam2D Imaging SensorASD: 30 (11 y), TD: 23 (11 y)Playing Computer Game (FaceMaze)ASD: Increase in Happy and Angry Expression Performance
Emotion Leo et al., 2015 [69]Camera, Robokind™ R25 Robot2D Imaging SensorASD: 3Imitate Expression from Robot (Happiness, Sadness, Anger, And Fear)31/60 Interactions Recognized; 19/60 No Imitation
Emotion Piana et al., 2019 [70] Microsoft Kinect V2Color Sensor, IR depth Sensors, IR EmitterASD: 10 (10 y)10 Sessions Body Emotion Expression and Recognition TaskIncreased Accuracy in Expression and Recognition After Training Sessions In The Trained Group, Transfer Effect On Facial Expression Recognition
Social SkillsRobins et al., 2010 [71]KASPAR, Video CamerasTouch Sensor, 2D ImagingASD: 3Unconstrained Interaction with The RobotInteraction Evaluation Through Sensor Activation, Differential Interaction and Maintenance of Interaction
Social SkillsCosta et al., 2009 [73]LEGO Mindstorms TM, Video CamerasTouch Sensor, Sound SensorASD: 2 (17, 19 y)4- 5 Sessions of Feedbacked Interaction with RobotInteraction Evaluation Through Sensor Activation, Differential Interaction and Maintenance of Interaction
Social SkillsWong and Zhong, 2016 [74]CuDDler Robot, Video CameraMicrophone, Contact Microphone, Tactile and Posture SensorsASD: 8 (5 y)5 Sessions of ABA (Didactic Teaching Followed by Role Modeling by Either A Robot (RT) Or Human (CT)Robot Training Significantly Facilitated Verbal and Gestural Communicative Skills, Increased Eye Contact Duration
Social SkillsUzuegbunam et al., 2015 [75]Microsoft KinectColor Sensor, IR Depth Sensors, IR Emitter, MicrophoneASD: 3 (7–12 y)Greeting Game with Participants Face, Reacting to Participant, Getting AppraisalAll 3 showed Increased Social Greeting Behavior Throughout and After Intervention
Social SkillsMower et al., 2011 [76]HDR-SR12 High Definition Handycam CamcordersMicrophones, 2D Video SensorsASD: 2 (6, 12 y)4 Sessions with Embodied Conversational Agent RACHEL Going Through Emotional ScenariosTool for Eliciting Interactive Behavior
Note: Age in sample description, where specified in a study, either refer to mean age of participants per group, to age ranges, or to individual ages in small samples of single cases.

Share and Cite

MDPI and ACS Style

Kowallik, A.E.; Schweinberger, S.R. Sensor-Based Technology for Social Information Processing in Autism: A Review. Sensors 2019, 19, 4787. https://0-doi-org.brum.beds.ac.uk/10.3390/s19214787

AMA Style

Kowallik AE, Schweinberger SR. Sensor-Based Technology for Social Information Processing in Autism: A Review. Sensors. 2019; 19(21):4787. https://0-doi-org.brum.beds.ac.uk/10.3390/s19214787

Chicago/Turabian Style

Kowallik, Andrea E., and Stefan R. Schweinberger. 2019. "Sensor-Based Technology for Social Information Processing in Autism: A Review" Sensors 19, no. 21: 4787. https://0-doi-org.brum.beds.ac.uk/10.3390/s19214787

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop