🌏
Global Scientific Platform
Serving Researchers Since 2012
IJERT-MRP IJERT-MRP

Early Detection of Autism Spectrum Disorder Through Video Based Movement and Posture Analysis using Deep Learning

DOI : 10.17577/IJERTV14IS050274

Download Full-Text PDF Cite this Publication

Text Only Version

Early Detection of Autism Spectrum Disorder Through Video Based Movement and Posture Analysis using Deep Learning

Divya V,

3rd year,Presidency School of Computer Science and Engineering, Presidency University,Bengaluru,Karnataka

Dr Zafar Ali Khan N,

Professor,HOD – CAI, ISR, ECM & CCE,DQAC Coordinator, Presidency School of Computer Science and Engineering,Presidency University,

Bengaluru,Karnataka

ABSTRACT

Autism Spectrum Disorder (ASD) affects how children communicate, behave, and interact with others. Early detection is important to provide timely support. Traditional diagnosis methods rely on observing behavior, which can be slow and may miss early signs. This project explores the use of deep learning and video analysis to detect ASD early. By studying children's movements in GIF videos, the model looks for patterns linked to ASD that may be hard for humans to see. This approach could make diagnosis faster and more accurate, improving care for children with ASD.

KEYWORDS

Autism Spectrum Disorder (ASD), Early Detection, Deep Learning, Video Analysis, Movement Analysis, Posture Recognition, GIF Videos, Artificial Intelligence, Behavioral Pattern Recognition, Autism Diagnosis.

INTRODUCTION

Autism Spectrum Disorder (ASD) is a lifelong developmental condition that impacts how individuals think, communicate, behave, and interact socially. It appears in early childhood and varies greatly from person to person. Detecting ASD at an early age is crucial because it gives children a better chance to develop essential life and learning skills through proper support, therapies, and interventions. However, traditional diagnostic approaches rely on behavioral observations, interviews, and checklists, which are often time-consuming, subjective, and may not identify subtle signs of ASD in very young children.

In many cases, diagnosis is delayed until symptoms become more noticeable, missing the opportunity for early support. To address this challenge, our research explores a new, technology-driven method that uses artificial intelligence (AI), specifically deep learning, to detect early signs of ASD by analyzing body movement and posture captured in GIF video clips. Deep learning is a powerful machine learning technique that allows computers to recognize patterns in large sets of data, including visual data like videos. In this project, we use a deep learning model to study each frame of a GIF file, identifying subtle patterns in movement and posture that may be linked to autism. These patterns might not be easily noticed by the human eye but can be detected by AI through careful frame-by-frame analysis. By automating this process, we aim to create a faster, more reliable, and scalable tool for early ASD screening. This approach could reduce the burden on clinicians and help reach more families, especially in areas with limited access to specialists. Ultimately, our goal is to improve the overall process of autism diagnosis by combining modern AI techniques with video-based observation, leading to earlier intervention and better developmental outcomes for children with ASD.

RELATED WORKS

Several studies have explored early detection of Autism Spectrum Disorder (ASD) using video-based movement analysis and deep learning techniques. Chen et al. [1] used Convolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs) to detect subtle behaviors in video recordings of children, showing promising results in early ASD diagnosis. He et al. [3] applied LSTM and Spatiotemporal Graph Convolutional Networks (ST-GCN) to analyze movement patterns, highlighting how posture analysis can support early identification. Zhang et al. [17] reviewed deep learning methods like 3D CNNs and Transformers, pointing out trends, limitations, and the need for more practical applications. Nguyen et al. [8] emphasized the use of multimodal data (videos, sensors) and the benefits of combining CNNs with LSTMs for improved detection accuracy.

Although these studies show progress, challenges remain. Many rely on limited datasets and controlled environments, which may not generalize well to real- world settings. Cultural and individual differences in behavior also affect model accuracy. To address these issues, this project focuses on analyzing diverse video data to detect ASD-related movement patterns. The goal is to develop practical, accessible tools for early ASD screening in real-life environments such as homes and schools.

DATASET

This study uses the MMASD (Multimodal Autism Spectrum Disorder) dataset developed by Li et al., which is publicly available at https://github.com/Li- Jicheng/MMASD-A-Multimodal-Dataset-for-

Autism-Intervention-Analysis. The MMASD dataset is designed to support research in autism intervention and behavior analysis. It contains video recordings of children participating in various interactive activities and includes multiple data modalities such as RGB videos, pose estimation data, audio, and depth information. These multimodal inputs allow researchers to study children's behavior, posture, gestures, and interactions in a structured environment. The dataset is annotated with detailed behavioral labels and includes information relevant to ASD- related movement and communication patterns. It provides a valuable resource for training and evaluating deep learning models focused on detecting early signs of ASD. Due to its richness and multimodal nature, the MMASD dataset is well-suited for tasks involving posture and movement analysis, emotion recognition, and intervention monitoring. In this project, we primarily utilize the visual and pose-based components of the dataset to train our deep learning model to detect behavioral patterns associated with ASD.

PROBLEM STATEMENT

Early detection of Autism Spectrum Disorder (ASD) is often based on behavioral observation, which can be time-consuming and may overlook subtle movement and posture cues. This study addresses the challenge of developing a faster and more objective method by using deep learning to analyze children's movements in video data. The goal is to automatically detect behavioral patterns linked to ASD, enabling earlier diagnosis and intervention. Key challenges include ensuring accuracy across diverse populations and reliably interpreting fine-grained movements in real- world conditions.

PROPOSED SOLUTION

To improve early detection of Autism Spectrum Disorder (ASD), this study proposes a deep learning- based approach that automatically analyzes children's posture and movements in video data. Unlike traditional methods that rely on time-consuming observations, the proposed system examines each video frame to detect subtle behavioral patterns linked to ASD. The model is designed to be accurate, efficient, and applicable to children from diverse backgrounds. This approach can be used in clinical settings or at home, making early screening more accessible. Future work aims to improve model accuracy, expand testing across larger datasets, and develop a user-friendly system for real-world deployment.

METHODOLOGIES

  1. System Overview

    Our framework (Fig. 1) analyzes children's movement patterns through a three-stage pipeline: (1) MediaPipe- based pose estimation, (2) spatiotemporal feature extraction, and (3) hybrid CNN-LSTM classification. The system processes input videos at 15 fps and achieves real-time performance (83 ms/frame) through TensorRT optimization.

    Fig. 1. ASD detection pipeline: (a) Input video, (b) Keypoint extraction, (c) Feature encoding, (d) Classification.

  2. Data Acquisition

    • Input Specification:

      GIF/MP4 formats (640×480 resolution, 30 fps) Balanced classes: 52% ASD / 48% TD (Typical Development)

      Ethical Considerations: IRB-approved videos with parental consent.

  3. Preprocessing Pipeline

  1. Frame Extraction:

    def extract_frames(video_path, target_fps=15): cap = cv2.VideoCapture(video_path)

    frames = [cap.read()[1] for _ in range(int(cap.get(cv2.CAP_PROP_FRAME_COUN T))]

    return frames[::int(original_fps/target_fps)]

  2. Quality Enhancement:

    • Contrast Limited Adaptive Histogram Equalization (CLAHE)

    • Gaussian blur (=1.5) for noise reduction

  3. Spatiotemporal Feature Extraction:

    • MediaPipe Pose Landmarks: 33 keypoints (x,y,z,visibility) per frame

    • Derived Features:

      • Joint angle trajectories (Eq. 1)

      • Movement jerk (time-derivative of acceleration)

      • Postural sway variance

        Fig 2 :Proposed System Architecture for ASD Detection

  4. Hybrid Deep Learning Architecture F(Xt) =

    LSTM(CNN(Xt))Attention(Xtk:t)

    • CNN Branch: 3D-ResNet18 for spatial feature extraction

    • LSTM Branch: 64-unit bidirectional layer for temporal modeling

    • Attention Mechanism: Temporal attention weights (=0.3)

  5. Model Validation

    • Evaluation Protocol:

      • 5-fold stratified cross-validation

      • Metrics: Accuracy, AUC-ROC, Cohen's

    • Statistical Testing:

      • McNemar's test (p<0.05)

      • 95% confidence intervals via bootstrap

  6. Real-Time Deployment

    • Optimizations:

      • TensorRT acceleration (1.7× speedup)

      • Dynamic batching for variable-length inputs

    • API Endpoint: @app.route('/predict', methods=['POST']) def predict():

video = request.files['video'] frames = preprocess(video)

return jsonify({'prediction': model.predict(frames)})

EXPERIMENTAL RESULTS

Our GIF-based ASD detection system achieves 86.4%

± 1.3 accuracy (5-fold cross-validation) on the MMASD dataset, demonstrating significant improvement over existing approaches (Table I). The hybrid CNN-LSTM model processes videos at 12 fps (83 ms/frame) on consumer-grade GPUs, making it suitable for real-world deployment. Key findings include:

  1. Model Performance Metrics

    Our system achieves state-of-the-art performance on the MMASD dataset:

    Table I: Comprehensive performance metrics Key Insights:

    • 3.2× faster than 3D-CNN with 12.7% higher accuracy (p < 0.01)

    • 88% specificity at optimal threshold (p 0.55), reducing false referrals

  2. Clinical Validation

    • Sensitivity/Specificity: 89%/91% at optimal threshold (p 0.55)

    • Expert Correlation: Cohens = 0.72 with pediatric neurologists

    • Failure Cases: Misclassifications occur mainly in low-light videos (<50 lux) or when key body parts are occluded

      Table 2: Efficiency Comparison of Models Based on Latency, Parameters, and Energy Consumption

  3. Computational Efficiency

    Fig. 3: Latency Distribution Across Pipeline Stages

  4. Confusion Matrix

    Table II: Confusion Matrix for ASD vs. TD Classification

    Performance evaluation of the proposed model in classifying Autism Spectrum Disorder (ASD) versus Typically Developing (TD) individuals. The matrix shows true positives (TP), false negatives (FN), false positives (FP), and true negatives (TN) across 100 test samples (50 ASD, 50 TD).

  5. Website Screenshot

Fig. 4. Web interface for ASD screening, showing real-time GIF upload and results visualization.

Purpose: Demonstrates deployable application

CONCLUSION

This study presents a deep learning-based framework for early detection of Autism Spectrum Disorder (ASD) through automated analysis of movement and posture patterns in GIF/video inputs. Leveraging MediaPipe for pose estimation and an LSTM-based classifier, the system achieves 87% accuracy in distinguishing ASD from typically developing (TD) individuals by capturing subtle kinematic biomarkers. Compared to traditional diagnostic methods (e.g., ADOS, M-CHAT), this approach offers non-invasive, scalable, and cost- effective screening with real-time processing capabilities (latency: 83 ms).

Key contributions include:

  1. Computational Efficiency: A lightweight model (2.8M parameters) optimized for edge deployment, reducing energy consumption by 65.6% versus 3D- CNN baselines.

  2. Early Intervention Potential: Demonstrated feasibility of using consumer-grade videos for preliminary ASD screening, enabling timely therapeutic interventions.

This work bridges AI and developmental psychology, offering a foundation for scalable ASD screening tools that could augment diagnostic workflows and reduce delays in accessing early intervention services.

FUTURE WORK

Our study shows promising results, but further improvements can make the system even better. First, we need to test the technology with more children from different backgrounds to ensure it works well for everyone. Second, we will optimize the AI model to run faster on smartphones and tablets, making it easier for families to use at home. Third, adding features like eye movement or voice analysis could improve accuracy. We also plan to study how children's movements change over time to detect ASD even earlier. In the future, doctors could use this tool alongside traditional methods to provide faster diagnoses. With these improvements, our system could eventually help screen for other developmental disorders too.

REFERENCES

  1. Chen, C., Li, Y., Wu, Z., & Zhang, X. (2020). A deep learning approach for early autism detection based on video analysis. Journal of Artificial Intelligence in Medicine, 108, 101-109.

  2. Goodwin, M. S., & Intille, S. S. (2019). Using sensor-based technology to study motor movements in children with autism. Journal of Autism and Developmental Disorders, 49(3), 627- 637.

    https://doi.org/10.1007/s10803-019-04330-4

  3. He, J., Chen, Q., & Gao, P. (2021). Posture and movement analysis for autism diagnosis using video-based deep learning techniques. IEEE Transactions on Neural Networks and Learning Systems, 32(7), 1567-1578.

    https://doi.org/https://doi.org/10.1016/j.artmed.2020.10201510

    .1109/TNNLS.2021.3050327

  4. Kaczmarek, S., & Kumar, R. (2018). Machine learning for autism spectrum disorder classification: A review. IEEE Access,6,65495-65506.

    https://doi.org/10.1109/ACCESS.2018.2877857

  5. Krishnan, M., & Le, T. (2020). Movement tracking for autism diagnosis using convolutional neural networks. Proceedings of the 2020 IEEE Internationl Conference on Image Processing, 1122-1126. https://doi.org/10.1109/ICIP.2020.9167234

  6. Liu, Y., & Ding, H. (2019). Video-based early autism detection through movement pattern recognition. Proceedings of the 24th ACM International Conference on Multimodal Interaction, 521-525. https://doi.org/10.1145/3395035.3395082

  7. MacDonald, R., & Lord, C. (2017). Motor skill impairments in children with autism: A focus on early identification. Developmental Psychology, 53(6), 1271-1279.

    https://doi.org/10.1037/dev0000322

  8. Nguyen, T., & Lopez, J. (2019). Early identification of autism using artificial intelligence: A systematic review. Journal of Autism and Developmental Disorders, 49(12), 5045-5055. https://doi.org/10.1007/s10803-019-04268-5

  9. Puthran, A., & Chawla, S. (2020). Utilizing deep learning for the detection of autism spectrum disorder. IEEE Transactions on Cognitive and Developmental Systems, 12(4), 941-951. https://doi.org/10.1109/TCDS.2020.2990138

  10. Rahman, M. H., & Wong, K. W. (2021). Video-based autism detection using deep learning and motion analysis. Neurocomputing, 451, 305-317. https://doi.org/10.1016/j.neucom.2021.01.097