ECFOR is an end-to-end web-based communication framework designed to bridge the communication gap between hearing/speaking individuals and people with speech or hearing impairments.
The system enables bi-directional translation between:
- Speech → Sign Language
- Sign Language → Text & Speech
By integrating computer vision, deep learning, and speech processing, ECFOR provides an accessible, real-time, and scalable communication solution.
Individuals with speech or hearing impairments often face significant communication barriers in everyday interactions. Most existing tools are limited to one-directional translation, lack real-time performance, or fail to generalize well across different users and environments.
ECFOR addresses these challenges by offering a unified, real-time, two-way communication platform.
- 🔊 Speech to Sign Language translation
- 🤟 Sign Language (video) to text and speech
- 🎥 Real-time video capture and processing
- 🧠 Deep learning–based gesture recognition
- 🌐 Web-based interface for accessibility
- 🔐 User authentication and session management
- 📁 Modular and extensible system design
The system consists of three core modules:
- Captures speech input via microphone
- Converts speech to text using speech recognition
- Maps recognized text to corresponding sign language gestures
- Uses the WLASL dataset for sign representations
- Captures real-time video input from a camera
- Processes frames using OpenCV
- Recognizes hand gestures using a MobileNet-based deep learning model
- Converts recognized signs into text
- Optionally converts text into synthesized speech
- Built using Django
- Manages routing, backend logic, and user authentication
- Integrates ML modules with the frontend
- Handles media files and real-time interaction
- Python
- Django
- OpenCV
- MobileNet
- WLASL Dataset
- NumPy
- TensorFlow / PyTorch (model dependent)
- Speech-to-Text APIs
- Text-to-Speech engines
- HTML
- CSS
- JavaScript
- Django Templates
Ecfor/
├── Authentication/ # User authentication logic
├── Communication/ # Core communication modules
├── Ecfor/ # Django project configuration
├── speech_to_txt/ # Speech-to-text processing
├── static/ # Static files (CSS, JavaScript)
├── templates/ # HTML templates
├── manage.py # Django entry point
├── merged_drum.avi # Sample sign language video
├── output.mp3 # Generated speech output
├── text_to_speech.txt # Intermediate text output
└── README.md # Project documentation
- Python 3.8 or higher
- pip
- Virtual environment (recommended)
- Webcam and microphone (for full functionality)
# Clone the repository
git clone https://github.com/Hasan-Saju/Ecfor.git
cd Ecfor
# Create and activate virtual environment
python -m venv venv
source venv/bin/activate # Windows: venv\Scripts\activate
# Install dependencies
pip install -r requirements.txt
# Apply database migrations
python manage.py migrate
# Run the development server
python manage.py runserver- WLASL (Word-Level American Sign Language Dataset)
- Used for training and evaluating the sign language recognition model
- Supports word-level gesture classification
- Enables real-time communication for individuals with speech or hearing impairments
- Reduces reliance on human interpreters
- Improves accessibility and social inclusion
- Applicable to education, healthcare, and public service domains
- WLASL Dataset contributors
- OpenCV and deep learning open-source communities