d-kavinraja/MouthMap
MouthMap is a deep learning-based lip reading system that converts silent video sequences into sentences. This research-focused project leverages computer vision and sequence modeling to assist in speech-impaired communication and human-computer interaction.
No commits in the last 6 months.
Stars
4
Forks
4
Language
Jupyter Notebook
License
Apache-2.0
Category
Last pushed
Aug 23, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/d-kavinraja/MouthMap"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
astorfi/lip-reading-deeplearning
:unlock: Lip Reading - Cross Audio-Visual Recognition using 3D Architectures
deepconvolution/LipNet
Automated Lip reading from real-time videos in tensorflow in python
articulateinstruments/DeepLabCut-for-Speech-Production
Trained deep neural-net models for estimating articulatory keypoints from midsagittal ultrasound...
ZakirCodeArchitect/Sonic-Lipsync-AI
A Google Colab-based Gradio app for generating lip-synced videos using the Sonic model. It...
Cl0ud-9/Lip-Sync-Video-Generator
An AI-powered pipeline that transforms text into realistic lip-synced talking face videos using...