Develop real-time application that translates American Sign Language (ASL) into spoken and written language, enabling seamless communication between deaf and hearing individuals.
The original approach was inspired by SignFormer-GCN
The approach taken leans heavier on GCN influence by adding two more GCN paths which process the right/left hand coordinates extracted by mediapipe
How2Sign
Farida Shittu & Shane O'Donnell
Clone repo locally or in colab
git clone --recurse-submodules https://github.com/sodonne6/EEP55C34_Advanced_AI.git
cd \inside\repo\rootClone Submodules
git submodule update --init --recursiveThis repository contains both the runtime demo code and the training framework used during development.
- See
src/README.mdfor environment setup and instructions on running the application end-to-end. - See
SLT/external/README.mdfor training-related notes, including the original SignFormer GCN reference repo, required override files, and the training launch workflow.
Recommended starting points:
- Running the system:
src/README.md - Training a new model:
SLT/external/README.md
[1] S. H. Arib, R. Akter, S. Rahman, and S. Rahman, "SignFormer-GCN: Continuous sign language translation using spatio-temporal graph convolutional networks," PLoS ONE, vol. 20, no. 2, p. e0316298, Feb. 2025, doi: 10.1371/journal.pone.0316298.
[2] A. Duarte et al., “How2Sign: a large-scale multimodal dataset for continuous American sign language,” arXiv.org, Aug. 18, 2020. https://arxiv.org/abs/2008.08143

