Multimodal AI Robotic System for ASD Therapy

Started: January 2025 | Status: In Progress

Python AI Nao Linux virtualenv NLP
Project image 1

Project Overview

A social robotic system using a NAO robot for emotion education in children with autism spectrum disorder (ASD). The system integrates a NAO robot with a dual-interface architecture: one interface displays emotional stimuli for the child, while the other is a Graphical User Interface (GUI) for a human facilitator. This design balances the robot's autonomous operation with essential human oversight.

The therapeutic approach is based on a five-activity framework that incrementally increases the complexity of interactions. A session target four core emotions—happy, sad, surprised, and angry—through multimodal exchanges that include verbal, facial, bodily, and contextual emotional cues.

system overview
Figure1. System architecture

The system leverages several advanced technologies to facilitate these activities. ChatGPT/Whisper is used for adaptive conversation, while DeepFace is employed for facial emotion analysis. MediaPipe is used for real-time body pose recognition by monitoring key anatomical landmarks on the child. A critical component is the bespoke GUI for the facilitator, which offers granular control to initiate, repeat, or omit any part of the session, enabling tailored pacing for children with ASD who often benefit from repeated exposure to concepts.

GUi
Figure2. GUI for the facilitator

Work in Progress

  • Start & stop recording button
  • Smoother transition between games
  • Story Interaction

Next Steps

  • More complex body movements
  • Longer session