Download as pdf or txt
Download as pdf or txt
You are on page 1of 10

Hindawi

Computational Intelligence and Neuroscience


Volume 2022, Article ID 8709145, 10 pages
https://doi.org/10.1155/2022/8709145

Research Article
Classification and Detection of Autism Spectrum Disorder
Based on Deep Learning Algorithms

Fawaz Waselallah Alsaade and Mohammed Saeed Alzahrani


College of Computer Science and Information Technology, King Faisal University, P.O. Box 4000, Al-Ahsa, Saudi Arabia

Correspondence should be addressed to Fawaz Waselallah Alsaade; falsaade@kfu.edu.sa

Received 1 January 2022; Revised 14 January 2022; Accepted 7 February 2022; Published 28 February 2022

Academic Editor: Deepika Koundal

Copyright © 2022 Fawaz Waselallah Alsaade and Mohammed Saeed Alzahrani. This is an open access article distributed under the
Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided
the original work is properly cited.
Autism spectrum disorder (ASD) is a type of mental illness that can be detected by using social media data and biomedical images.
Autism spectrum disorder (ASD) is a neurological disease correlated with brain growth that later impacts the physical impression
of the face. Children with ASD have dissimilar facial landmarks, which set them noticeably apart from typically developed (TD)
children. Novelty of the proposed research is to design a system that is based on autism spectrum disorder detection on social
media and face recognition. To identify such landmarks, deep learning techniques may be used, but they require a precise
technology for extracting and producing the proper patterns of the face features. This study assists communities and psychiatrists
in experimentally detecting autism based on facial features, by using an uncomplicated web application based on a deep learning
system, that is, a convolutional neural network with transfer learning and the flask framework. Xception, Visual Geometry Group
Network (VGG19), and NASNETMobile are the pretrained models that were used for the classification task. The dataset that was
used to test these models was collected from the Kaggle platform and consisted of 2,940 face images. Standard evaluation metrics
such as accuracy, specificity, and sensitivity were used to evaluate the results of the three deep learning models. The Xception
model achieved the highest accuracy result of 91%, followed by VGG19 (80%) and NASNETMobile (78%).

1. Introduction skills. As these disorders also affect how a person perceives


and socializes with others, children may suddenly become
Autism spectrum disorders (ASD) refer to a group of introverted or aggressive in the first five years of life as they
complex neurodevelopmental disorders of the brain such as experience difficulties in interacting and communicating
autism, childhood disintegrative disorders, and Asperger’s with society. While ASD appears in childhood, it tends to
syndrome, which, as the term “spectrum” implies, have a persist into adolescence and adulthood [7].
wide range of symptoms and levels of severity [1]. These Advanced information technology that uses artificial
disorders are currently included in the International Sta- intelligence (AI) models has helped to diagnose ASD early
tistical Classification of Diseases and Related Health Prob- through facial pattern recognition. Yolcu et al. [8] used the
lems under Mental and Behavioral Disorders, in the category convolutional neural network (CNN) algorithm to train data
of Pervasive Developmental Disorders [2]. The earliest for extracting components of human facial expressions and
symptoms of ASD often appear within the first year of life proposed the use of such algorithm to detect facial ex-
[3–6] and may include lack of eye contact, lack of response to pressions in many neurological disorders [9]. In 2018, Haque
name calling, and indifference to caregivers. A small number and Valles [10], using deep learning approaches, updated the
of children appear to develop normally in the first year, and Facial Expression Recognition 2013 dataset to recognize
then show signs of autism between 18 and 24 months of age facial expressions of autistic children. Rudovic et al. [11]
[5], including limited and repetitive patterns of behavior, a presented the CultureNet deep learning model that was used
narrow range of interests and activities, and weak language to identify 30 videos.
2 Computational Intelligence and Neuroscience

Several studies have discovered important features of (iv) The developing system has been validated and ex-
autism in numerous ways to detect autism, such as feature amined using various methods.
extraction [12], eye tracking [13], facial recognition, medical
The rest of this paper is organized as follows. Section 2
image analysis [9], and voice recognition [14]. However,
describes materials and methods of pretrained deep learning
facial recognition plays a more significant role in detecting
models. Section 3 provides experiments. Section 4 and 5
autism than the person’s emotional state. Facial recognition
provide results and discussion. The conclusion of papers is
is a common way to identify persons and to prove that they
presented in Section 6.
are normal or abnormal. It involves mining pertinent in-
formation to disclose behavior patterns [15, 16]. Duda et al.
[17] introduced a new method of producing samples of
2. Materials and Methods
differentials between autism and attention deficit hyperac- This study proposes a deep learning model based on transfer
tivity disorder (ADHD) and using such differentials to learning, namely, Xception, NASNETMobile, and VGG1 9 to
recognize autism. Sixty five samples of differentials on facial detect autism using facial features of autistic and normal
expressions of social responsiveness were collected for the children. Facial features can be used to determine if a child has
datasets. Deshpande et al. [18] designed metrics for studying autism or is normal. The models extracted significant facial
brain activities to identify autism. AI and soft computing features from the images. One of the advantages offered by deep
approaches have also been applied to detect autism. Dif- learning algorithms is the ability to extract very small details
ferent studies have been conducted on detecting autism, but from an image, which a person cannot notice with the naked
few of them have focused on brain neuroimaging. Parikh eye. Figure 1 shows the framework of our study, from the data
et al. [19] used machine learning approaches to develop a acquisition to the data preprocessing and loading, to the model
system for extracting the characteristics of autism. They preparation and training, and to the model performance test.
gathered their dataset from 851 persons whom they classified
as having and not having ASD. Thabtah and Peebles [20]
used rule-based machine learning (RBML) to detect ASD 2.1. Dataset. This study analyzed facial images of autistic
traits. Al Banna et al. [21] developed a smart system for children and normal children obtained from Kaggle plat-
monitoring ASD patients during the coronavirus disease form, which is publicly accessible online [34]. The dataset
2019 (COVID-19) pandemic. consisted of 2,940 face images, half of which were of autistic
Many real-life applications of AI and machine algorithms children and the other half were of nonautistic children. This
have been designed to help solve social problems. AI has been dataset was collected through Internet sources such as
used in all health applications to help doctors control diseases websites and Facebook pages that are interested in autism.
such as autism. Artificial neural networks have been partic- Table 1 shows the distribution of the split dataset samples.
ularly focused on as a way to extract features of ASD patients The splitting of the input data is presented in Figure 2.
that can be used to discriminate between persons with and
without ASD. Among the techniques that have been used or 2.2. Preprocessing. The purpose of the data preprocessing
proposed to detect autism in children are deep learning was to clean and crop the images. Because the data were
techniques, particularly, CNNs and recurrent neural networks collected from Internet resources by Piosenka [34], they had
(RNNs) [22, 23], and the bidirectional long short-term to be preprocessed before they could be used to train the
memory (BLSTM) model [24]. Lately, more studies have been deep learning model. The dataset creator automatically
conducted to diagnose ASD using machine learning ap- cropped the face from the original image. Then, the dataset
proaches such as [24, 25], brain imaging [26–28], analysis of was split into 2,540 images for training, 100 for validation,
data on physical biomarkers [29–33], assessment of the be- and 300 for testing. To scaling, the normalization method
havior of persons with autism, and assessment of clinical data was applied; the dataset was rescaling the parameters of all
using the machine learning approach [33]. the images from the pixel values [0, 255] to [0, 1].
Our study demonstrated the use of a well-trained clas-
sification model (based on transfer learning) to detect autism
from an image of a child. With the advent of high-speci- 2.3. Convolutional Neural Network Models. AI has been
fication mobile devices, this model can readily provide a remarkably developed to assist humans in their daily life, for
diagnostic test of putative autistic traits by taking an image example, through medical applications, which are based on a
with cameras. The main contributions of our research are as branch of AI called “computer vision.” Hence, the CNN
follows: algorithm has contributed to the detection of diseases and to
behavioral and psychological analysis.
(i) Three pretrained deep learning algorithms were
applied for ASD detection: NASNETMobile,
2.3.1. Basic Components of the CNN Model. The convolu-
Xception, and VGG19
tional neural network (CNN) is one of the most famous deep
(ii) The Xception model showed the best performance learning algorithms. It takes the input image and assigns
of the three pretrained deep learning algorithms importance to learnable weights and biases in order to
(iii) A system was designed to help health officials to recognize the class of the image. The neuron can be said to be
detect ASD through eye and face identification a simulation of the communication pattern of the neurons of
Computational Intelligence and Neuroscience 3

Load training and Model Fine


Data acquisition Prepare Model
Test Dataset Tuning

Confusion Matrix
Prepare the
and Calculate Prediction Model Fitting
Testing dataset
Results

Figure 1: The framework of our study.

Table 1: Splitting of the dataset for training, testing, and validation.


Total face images Training set (%) Validation set (%) Testing set (%)
2,940 2,540 100 300

3500
3000 2940
2540
2500
Numbers of images

2000

1500

1000

500 300
100
0

-500
Testing set (%) Validation set (%) Training Total face Total face images
images
Datasets
Figure 2: Splitting of the dataset.

the human brain through the interconnection and com- max pooling
munication between cells. In this section, we will explain the
20 30
basic components of the CNN model: the input layer,
112 37
convolutional layer, activating function, pooling layer, fully 12 20 30 0
connected layer, and output prediction.
8 12 2 0
34 70 37 4 average pooling
2.3.2. Convolutional Layer with a Pooling Layer. The input 112 100 25 12
of the convolutional layer is an image as a matrix of pixel 13 8
values. The objective of the convolutional layer is to reduce the 79 20
images into a form that can be easily processed without losing
Figure 3: The convolution layer and max pooling and average
their important features that will help to detect autism. The
pooling.
first layer of the CNN model is responsible for extracting low-
level features such as edges and color. The build of the CNN
model allowed us to add more layers to it to enable it to extract processes. The slide window of the kernel extracts the features
the high-level features that will help it understand images. Due from the input image and converts the image into a matrix,
to the large number of parameters outputted from the con- after which the number of the parameters is mathematically
volution layer, which may significantly prolong the arithmetic reduced through max pooling and average pooling.
operations of the matrices, the number of weights was reduced
by using one of the following two techniques: max pooling or
average pooling. Max pooling is based on the maximum values 2.3.3. Fully Connected Layer and Activation Function.
in each window in the stride, while average pooling is based on The fully connected (FC) layer is a nonlinear combination of
the mean value of each window in the stride. In this study, the the high-level features that received the input from the
model was based on max pooling. Figure 3 shows the con- hidden layers and are represented as outputs. In the FC layer,
volutional layer and the max pooling and average pooling the input image is represented as a column vector. The
4 Computational Intelligence and Neuroscience

training of the model has two paths: the forward neural


network and backpropagation. The forward neural network
feeds form flattened output layer. In the backpropagation,
the neural network minimizes the loss errors and learns
more features by applying the number of the training it-
erations. Most deep learning models show high performance
while increasing the number of the hidden layers and the
training iterations, which allows the neural network to
extract the low-level features deeply. The softmax classifier
receives the parameters from the FC layer and calculates the
properties to predict the output, as shown in Figure 4. A
Softmax output of 0 means the image belongs to class 0 and a Fully-Connected Layer Activation function
Softmax output of 1 means the image belongs to class 1 In Figure 4: The fully connected layer and the activation function.
this study, class 0 is the autism class and class 1 is the normal
class.
input volume size, max pooling layers were employed as
assigners in VGG19. In the VGG19 structure, two FC layers
2.4. Deep Learning Models. This paper is based on three were adopted with 4,096 neurons to associate each layer with
pretrained models for autism detection using facial feature others in the model. The structure of VGG19 is presented in
images: NASNetMobile, VGG19, and Xception. Figure 6.
(i) Input layer: the function of this layer is to take and
2.4.1. Xception Model. The Xception model was trained on accept an image input with a size of 224 × 224 × 3.
the ImageNet dataset [2, 3] for the image recognition and For the ImageNet race, the inventors of the model
classification task. Xception is a deep CNN that provides picked out the center 224 × 224 × 3 patch in each
new inception layers. The inception layers are built from image to retain the input size of the image
depthwise convolution layers and are followed by a point- dependable.
wise convolution layer. Transfer learning has two concepts: (ii) Convolutional layers: VGG’s convolutional layers
feature extraction and fine-tuning. In this study, the feature leverage a minimal receptive field, i.e., 3 × 3, the
extraction method used the pretraining model, which was minimum probable size that still releases up/down
trained on the standard dataset to extract the feature from and left/right. The convolution stride is fixed at 1
the new dataset and to remove the top layers of the model. pixel to conserve the spatial resolution after the
The new top layers were added to the model for custom convolution (the stride is the number of pixel shifts
classification based on the number of classes. Fine-tuning over the input matrix).
has been used to adapt the generic features to a given class to
(iii) Hidden layers: all the hidden layers in the VGG
avoid overwriting. Figure 5 shows the network used in the
network use ReLU. VGG does not frequently in-
Xception model architecture for extracting the image fea-
fluence Local Response Normalization (LRN), as it
tures for the dataset. The Xception architecture used the
increases the memory utilization and the training
features maps, followed by a global max pooling layer and
time. Furthermore, it does not enhance the overall
two dense layers, 128 and 64, respectively, with a rule ac-
accuracy of the model.
tivation function. Then, the output of the dense layers was
passed to the flatten layer, which took the input as a feature (iv) Fully connected layers: of the three FC layers, the
map and the output as a vector. Batch normalization was first two had 4,096 channels each, and the third had
used to enhance the output by avoiding overfitting. Keras 1,000 channels.
supported the early-stopping method, which stopped the
training when the validation loss of the model did not 3. Experiments
improve. In this model, the RMSprop optimizer was used to
reduce the error learning rate or loss during the training of The results of the deep learning models are presented in this
the parameters of the CNN model. The last layer for the section, and the significant results of developing system are
output prediction used the Softmax function. declared.

3.1. Experimental Setup. The experiment was executed on


2.4.2. Visual Geometry Group Network (VGG) Model.
different libraries of python and hardware devices for devel-
VGG19 stands for the visual geometry group network
oping an intelligent autism detection system (ADS). Table 2
(VGG19), a deep artificial neural network model with a 19
shows the main requirements for the design of the ADS.
artificial multilayers process. Primarily, VGG19 is based on
the CNN technique, is commonly implemented on the
ImageNet dataset, and is valuable because of its straight- 3.2. Evaluation Metrics. This study uses different types of
forwardness, as 3 × 3 convolutional layers are attached to its performance evaluation metrics for the three pretrained
upper side to upsurge with the gravity level. To reduce the models such as accuracy, sensitivity, and specificity and a
Computational Intelligence and Neuroscience 5

Xception Model Dense layer


Input Face Image Average-
Pooling Layer
Softmax
Function

Autism

Normal

Figure 5: Xception model architecture in our study.

224×224×64
224×224×3 112×112×128
Input Face Image 56×56×256
28×28×512
14×14×512
7×7×512
1×1×4096 1×1×1000

Figure 6: VGG19 model.

Table 2: Experimental setup. TP + TN


Accuracy � × 100%,
Hardware Software/libraries FP + FN + TP + TN
Processor core I7 TensorFlow library TN
Keras library Specificity � × 100%, (1)
Panda seaborn TN + FN
8 GB RAM
Matplotlib TP
Numpy Sensitivity � × 100%,
TP + FP
Note. The model parameters were set with 100 epochs and with a batch size
of 32. To avoid overfitting and to optimize the training time, the early stop where TP is the True Positive, FP is the False Positive, TN is
strategy was used, whereby the training was stopped after 28 epochs. The the True Negative, and FN is the False Negative. Specificity is
parameters of the deep learning model are shown in Table 3.
the capacity of the model to correctly identify the normal
children, and sensitivity is the capacity of the model to
Table 3: Parameters used in the pretrained deep learning models.
correctly identify autistic children.

Parameter name Value


4. Results
Global max pooling layer size 3∗3
Dense layer 128, 64 This section presents the testing results of the experiments
Batch size 32 conducted to detect ASD. Table 4 summarizes the testing
Number of epochs 100 results of the used deep learning models.
Output classification layer Softmax
In these experiments, three different pretrained deep
Optimizer ADAM
Activation function Rule
learning models, namely, Xception, VGG19, and NAS-
NetMobile were implemented to detect ASD. Each model
was trained and tested to extract the traits that categorize
confusion matrix. A confusion matrix is a type of measure of children as autistic and as normal based on their facial
classification performance that represents a table of the true features. Figure 7 shows the confusion metrics of the
and false values of the testing results. In the confusion matrix three deep learning models. They show that the Xception
of the Xception model, the True Positives were 132 autistic model had the highest testing accuracy, 91%, of the three
children out of 150 autistic children, the False Negatives models, and the NASNetMobile model had the lowest
were 18 children classified as autistic children, the True performance level of 78%. Although the dataset was
Negatives were 141 correctly classified children out of the collected from Internet sources by the data generator,
150 normal children, and the False Positives were 9 children. which clearly showed the difference in ages and the
The equations for these metrics are as follows: quality of the photography, the Xception model showed
6 Computational Intelligence and Neuroscience

Table 4: Testing results of the pretrained deep learning models.


Model name Specificity Sensitivity Accuracy
Xception 0.94 0.88 0.91
VGG19 0.83 0.78 0.80
NASNETMobile 0.75 0.82 0.78

Confusion Matrix Confusion Matrix


120
120

Autism 127 23 Autism 121 29 100


100
True label

True label
80 80

60 60
Normal 36 114 Normal 35 115
40
40

Autism Normal Autism Normal


Predicted label Predicted label
(a) (b)
Confusion Matrix

120

Autism 129 21 100


True label

80

60
Normal 40 110
40

Autism Normal
Predicted label
(c)

Figure 7: Confusion matrices for the. (a) VGG19 model, (b) NASNetmobile model, and (c) Xception model.

the highest accuracy, with only a small percentage of axis representing the score percentage and the x-axis indi-
errors. cating the number of epochs. Its accuracy in the training
The performance of the VGG19 model for the training stage was 70–90% and in the validation stagewas 65–78%. It
and validation of the data for ASD detection is presented in had good results in the training phase but poorer results in
Figure 8(a), with the y-axis representing the score percentage the validation phase than another deep learning model. Its
and the x-axis indicating the number of epochs. In the entropy training and validation losses are presented in
training process, the accuracy of the VGG19 model rose Figure 9(b). Its training loss was 4.0, and its validation loss
from 56% to 85% after 25 epochs, and in the validation was 3.2. Its performance plot shows overfitting in the
process, its accuracy was 82%. Its training and validation training process, which explains why its accuracy percentage
losses are presented in Figure 8(b). Its training loss was 3.5, was lower.
and its validation loss was 2.5. The Xception model achieved 91% accuracy for the
Figure 9 shows the performance of the NASNetMobile testing data and 100% accuracy for the training data.
model in detecting ASD. The graphic plot shows that the Figure 10(a) shows its accuracy performance. Its accuracy in
model did not have good results. Its training and validation the training process was 100% and in the validation process
accuracy percentages are shown in Figure 9(a), with the y- was from 70% to 91%. It was observed that the Xception
Computational Intelligence and Neuroscience 7

Accuracy over 100 Epochs Loss over 100 Epochs


0.85 3.5

0.80 3.0

0.75 2.5

0.70 2.0

0.65 1.5

0.60 1.0

0.5
0.55
0 5 10 15 20 25 0 5 10 15 20 25
Train_accuracy Train_loss
valid_accuracy valid_loss
(a) (b)

Figure 8: Overall performance plot for the VGG19 model. (a) Model performance. (b) Model loss.

Accuracy over 100 Epochs Loss over 100 Epochs

0.90 4.0

0.85 3.5
3.0
0.80
2.5
0.75
2.0
0.70
1.5
0.65
1.0
0.60
0.5
0 5 10 15 20 0 5 10 15 20
Train_accuracy Train_loss
valid_accuracy valid_loss
(a) (b)

Figure 9: Overall performance plot for the NASNetMobile model. (a) Model performance. (b) Model loss.

model is the appropriate deep learning model for detecting and NASNetMobile were considered for use in diagnosing
ASD. Its validation loss is shown in Figure 10(b) as 2.0. autism. The empirical results of these models were pre-
sented, and it was noted that the Xception model attained
5. Results and Discussion the highest accuracy of 91%.
The results of the comparative prediction analysis of the
People with autism face difficulties and challenges in un- Xception model and the existing system are presented in
derstanding the world around them and in understanding Table 5. Musser [35] used the VGGFace model that attained
their thoughts, feelings, and needs. The world surrounding a 85% accuracy. Tamilarasi et al. [36, 37] used deep learning
person with autism seems to him like a horror movie, and he ResNet50 to detect autism but with small data, that is, only
finds some sounds and lights and even smells and tastes of 49 images. Its accuracy was 89%. Jahanara et al. [38] in-
foods frightening and sometimes painful. Thus, when a troduced the VGG19 model to identify autism through facial
sudden change occurs in their world, they are terrified that images. The VGG19 model achieved 84% accuracy in the
no one else can understand. validation accuracy process.
Diagnosing autism is very important to save the lives of In this study, three deep learning algorithms were used
many children. Developing an intelligence system based on to detect autism. We observed that the Xception model
AI can help identify autism early. In this study, three ad- showed significant accuracy, 91%, compared with existing
vanced deep learning models, namely, Xception, VGG19, systems. Figure 11 shows the results of the comparison of
8 Computational Intelligence and Neuroscience

1.00 4.0
3.5
0.95
3.0
0.90
2.5
0.85 2.0

0.80 1.5
1.0
0.75
0.5
0.70 0.0
0 5 10 15 20 25 0 5 10 15 20 25
Train_accuracy Train_loss
valid_accuracy valid_loss
(a) (b)

Figure 10: Plot of the training and validation accuracy of the Xception model. (a) Accuracy performance. (b) Model loss.

Table 5: Comparison of the results of our system and existing models.


Author Datasets Model Accuracy (%)
Han et al. [35] Same as our dataset VGGFace 85.0
Haque et al. [10] The dataset included 19 TD and 20ASD children ResNet50 89.2
Beary et al. [38] Same as our dataset VGG19 84.0
Developing model MobileNet 91.0

Accuracy
100
98
96
94
92 91
90 89
88
%

86 85
84
84
82
80
78
76
Musser et al,2020 Tamilarasi et al.,2020 Jahanara et al.,2020 proposed system
Results
Figure 11: Comparison results of the developing system with existing systems.

our developing system with various existing systems. due to which researchers and academics have intensified
Overall, our developing system outperformed all the existing their efforts to uncover the causes of autism and to detect it
systems. early in order to give autistic people behavioral development
treatment programs that should help them integrate into
6. Conclusions society and leave the isolation of the autistic world.
This paper evaluated the performance of three deep
Interest in child autism has risen due to the advances in learning models in detecting ASD through facial features:
global health know-how and capacities. Moreover, the NASNETMobile, Xception, and VGG19. Each model was
number of autistic children has increased in recent years, trained on a publicly available dataset on the Internet, and
Computational Intelligence and Neuroscience 9

the best result for classification accuracy was achieved by the convolutional neural network,” Multimedia Tools and Ap-
Xception model (91%). The results of the model classifica- plications, vol. 78, no. 22, pp. 31581–31603, 2019.
tion showed us the possibility of using such models based on [10] M. I. U. Haque and D. Valles, “A facial expression recognition
deep learning and computer vision as automatic tools for approach using DCNN for autistic children to identify
specialists and families to accurately and more quickly di- emotions,” in Proceedings of the 2018 IEEE 9th Annual In-
formation Technology, Electronics and Mobile Communication
agnose autism. Computer techniques contribute to the
Conference (IEMCON), pp. 546–551, Vancouver, Canada, 1–3
successful conduct of complex behavioral and psychological November 2018.
analyses for autism diagnosis that require a longer time and [11] O. Rudovic, Y. Utsumi, J. Lee et al., “CultureNet: a deep
great effort. learning approach for engagement intensity estimation from
face images of children with autism,” in Proceedings of the
Data Availability 2018 IEEE/RSJ International Conference on Intelligent Robots
and Systems (IROS), pp. 339–346ISSN 2153-0866, Madrid,
The dataset used to support the findings of the study are Spain, 1–5 October 2018.
available at https://www.kaggle.com/cihan063/autism- [12] M. S. Satu, F. Farida Sathi, M. S. Arifen, M. Hanif Ali, and
image-data (accessed on 10 December 2021). M. A. Moni, “Early detection of autism by extracting features:
a case study in Bangladesh,” in Proceedings of the 2019 In-
Conflicts of Interest ternational Conference on Robotics, Electrical and Signal
Processing Techniques (ICREST), pp. 400–405, Dhaka, Ban-
The authors declare that they have no conflicts of interest. gladesh, 10–12 January 2019.
[13] Q. Guillon, N. Hadjikhani, S. Baduel, and B. Rogé, “Visual
Acknowledgments social attention in autism spectrum disorder: insights from
eye tracking studies,” Neuroscience & Biobehavioral Reviews,
The authors extend their appreciation to the Deanship of vol. 42, pp. 279–297, 2014.
Scientific Research at King Faisal University for funding this [14] T. Akter, M. S. Satu, L. Barua, F. F. Sathi, and M. H. Ali,
research work through project no. NA00075. “Statistical analysis of the activation area of fusiform gyrus of
human brain to explore autism,” International Journal of
Computer Science and Information Security, vol. 15, pp. 331–
References 337, 2017.
[1] C. S. Paula, S. H. Ribeiro, E. Fombonne, and [15] S. Schelinski, K. Borowiak, and K. von Kriegstein, “Temporal
M. T. Mercadante, “Brief report: prevalence of pervasive voice areas exist in autism spectrum disorder but are dys-
developmental disorder in Brazil: a pilot study,” Journal of functional for voice identity recognition,” Social Cognitive and
Autism and Developmental Disorders, vol. 41, no. 12, Affective Neuroscience, vol. 11, no. 11, pp. 1812–1822, 2016.
pp. 1738–1742, 2011. [16] X. Jiang and Y.-F. Chen, “Facial image processing,” in Applied
[2] L. C. Nunes, P. R. Pinheiro, M. C. D. Pinheiro et al., A Hybrid Pattern Recognition, H. Bunke, A. Kandel, and M. Last, Eds.,
Model to Guide the Consultation of Children with Autism Springer, Berlin/Heidelberg, Germany, pp. 29–48, 2008.
Spectrum Disorder, A. Visvizi and M. D. Lytras, Eds., Springer [17] M. Duda, R. Ma, N. Haber, and D. P. Wall, “Use of ma-
International Publishing, View at: Google Scholar, chine learning for behavioral distinction of autism and
pp. 419–431. ADHD,” Translational Psychiatry, vol. 6, no. 2, Article ID
[3] Apa–American Psychiatric Association, “Diagnostic and e732, 2016.
statistical manual of mental disorders (DSM –5),” 2020, [18] G. Deshpande, L. E. Libero, K. R. Sreenivasan,
https://www.psychiatry.org/psychiatrists/practice/dsm. H. D. Deshpande, and R. K. Kana, “Identification of neural
[4] R. Carette, F. Cilia, G. Dequen, J. Bosche, J.-L. Guerin, and connectivity signatures of autism using machine learning,”
L. Vandromme, “Automatic autism spectrum disorder de- Frontiers in Human Neuroscience, vol. 7, p. 670, 2013.
tection thanks to eye-tracking and neural network-based [19] M. N. Parikh, H. Li, and L. He, “Enhancing diagnosis of
approach,” in Proceedings of the International Conference on autism with optimized machine learning models and personal
IoT Technologies for Healthcare, pp. 75–81. 4, Springer, An- characteristic data,” Frontiers in Computational Neuroscience,
gers, France, 24–25 October 2017. 2019.
[5] L. Kanner, “Autistic disturbances of affective contact,” Nerv. [20] F. Thabtah and D. Peebles, “A new machine learning model
Child, vol. 2, pp. 217–250, 1943. based on induction of rules for autism detection,” Health
[6] E. Fombonne, “Epidemiology of pervasive developmental Informatics Journal, vol. 26, no. 1, pp. 264–286, 2020.
disorders,” Pediatric Research, vol. 65, no. 6, pp. 591–598, [21] M. H. Al Banna, T. Ghosh, K. A. Taher, M. S. Kaiser, and
2009. M. Mahmud, “A monitoring system for patients of autism
[7] “International statistical classification of diseases and related spectrum disorder using artificial intelligence,” in Proceedings
health problems (ICD),” https://www.who.int/standards/ of the International Conference on Brain Informatics,
classifications/classification-of-diseases. pp. 251–262, Springer, Padua, Italy, July 2020.
[8] G. Yolcu, I. Oztel, S. Kazan et al., “Deep learning-based facial [22] M. Li, D. Tang, J. Zeng et al., “An automated assessment
expression recognition for monitoring neurological disor- framework for atypical prosody and stereotyped idio-
ders,” in Proceedings of the 2017 IEEE International Confer- syncratic phrases related to autism spectrum disorder,”
ence on Bioinformatics and Biomedicine (BIBM), Computer Speech & Language, vol. 56, pp. 80–94, 2019.
pp. 1652–1657, Kansas City, MO, USA, 13–16 November [23] F. B. Pokorny, B. W. Schuller, P. B. Marschik et al., “Earlier
2017. identification of children with autism spectrum disorder: an
[9] G. Yolcu, I. Oztel, S. Kazan et al., “Facial expression recog- automatic vocalisation-based approach,” in Proceedings of the
nition for monitoring neurological disorders based on Annual Conference of the International Sp, August 2017.
10 Computational Intelligence and Neuroscience

[24] F. Eyben, K. R. Scherer, B. W. Schuller et al., “The Geneva


minimalistic acoustic parameter set (GeMAPS) for voice
research and affective computing,” IEEE Transactions on
Affective Computing, vol. 7, no. 2, pp. 190–202, 2016.
[25] A. Jack, “Neuroimaging in neurodevelopmental disorders:
focus on resting-state fMRI analysis of intrinsic functional
brain connectivity,” Current Opinion in Neurology, vol. 31,
no. 2, pp. 140–148, 2018.
[26] C. H. Y. Fu and S. G. Costafreda, “Neuroimaging-based
biomarkers in psychiatry: clinical opportunities of a paradigm
shift,” Canadian Journal of Psychiatry, vol. 58, no. 9,
pp. 499–508, 2013.
[27] S. J. Moon, J. Hwang, R. Kana, J. Torous, and J. W. Kim,
“Accuracy of machine learning algorithms for the diag-
nosis of autism spectrum disorder: systematic review and
meta-analysis of brain magnetic resonance imaging
studies,” JMIR Mental Health, vol. 6, no. 12, Article ID
e14108, 2019.
[28] S. Sarabadani, L. C. Schudlo, A. A. Samadani, and A. Kushski,
“Physiological detection of affective states in children with
autism spectrum disorder,” IEEE Transactions on Affective
Computing, vol. 11, no. 4, pp. 588–600, 2020.
[29] W. Liu, M. Li, and L. Yi, “Identifying children with autism
spectrum disorder based on their face processing abnormality:
a machine learning framework,” Autism Research, vol. 9, no. 8,
pp. 888–898, 2016.
[30] M. Alcañiz Raya, I. A. Chicchi Giglioli, J. Marı́n-Morales et al.,
“Application of supervised machine learning for behavioral
biomarkers of autism spectrum disorder based on electro-
dermal activity and virtual reality,” Frontiers in Human
Neuroscience, vol. 14, p. 90, 2020.
[31] J. Hashemi, G. Dawson, K. L. H. Carpenter et al., “Computer
vision analysis for quantification of autism risk behaviors,”
IEEE Trans. Affect. Comput.vol. 3045, pp. 1–12, 2018.
[32] A. V. Dahiya, C. McDonnell, E. DeLucia, and A. Scarpa, “A
systematic review of remote telehealth assessments for early
signs of autism spectrum disorder: video and mobile ap-
plications,” Practice Innovations, vol. 5, no. 2, pp. 150–164,
2020.
[33] S. R. Shahamiri and F. Thabtah, “Autism AI: a new autism
screening system based on artificial intelligence,” Cognitive
Computation, vol. 12, no. 4, pp. 766–777, 2020.
[34] G. Piosenka, “Detect autism from a facial image,” 10 De-
cember 2021, https://www.kaggle.com/cihan063/autism-
image-data accessed on.
[35] https://towardsdatascience.com/detecting-autism-spectrum-
disorder-in-children-with-computer-vision-8abd7fc9b40arep
lace.
[36] C. Garcia, J. Ostermann, and T. Cootes, “Facial image pro-
cessing,” Eurasip J. Image Video Process, vol. 2007, pp. 1-2, 2008.
[37] F. C. Tamilarasi and J. Shanmugam, “Convolutional Neural
Network based Autism Classification,” in 2020 5th Interna-
tional Conference on Communication and Electronics Systems
(ICCES), pp. 1208–1212, IEEE, (2020, June).
[38] S. Jahanara and S. Padmanabhan, “Detecting autism from
facial image,” 2021.

You might also like