deepconvolution/LipNet
Automated Lip reading from real-time videos in tensorflow in python
This project helps decipher spoken words by analyzing mouth movements in video, even in noisy environments. It takes a video of a person speaking and outputs the predicted word or phrase they uttered. This is particularly useful for individuals with hearing impairments, those trying to understand speech amidst background noise, or intelligence agencies for covert operations.
164 stars. No commits in the last 6 months.
Use this if you need to understand spoken content from video recordings where audio is unclear or unavailable, such as in noisy settings or for assistive technology for the deaf.
Not ideal if you require real-time, interactive lip-reading beyond sentence-level prediction or if you don't have video input of a speaker's mouth.
Stars
164
Forks
48
Language
Jupyter Notebook
License
—
Category
Last pushed
Mar 20, 2018
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/deepconvolution/LipNet"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
astorfi/lip-reading-deeplearning
:unlock: Lip Reading - Cross Audio-Visual Recognition using 3D Architectures
articulateinstruments/DeepLabCut-for-Speech-Production
Trained deep neural-net models for estimating articulatory keypoints from midsagittal ultrasound...
MrfoxAK/Evaluate-Lip-reading-using-Deep-Learning-Techniques.
This paper explores Silent Sound Technology, focusing on its potential to enhance communication...
BenedettoSimone/Lipnet-ITA
LipReadingITA: Keras implementation of the method described in the paper 'LipNet: End-to-End...
Cl0ud-9/Lip-Sync-Video-Generator
An AI-powered pipeline that transforms text into realistic lip-synced talking face videos using...