deepconvolution/LipNet

Automated Lip reading from real-time videos in tensorflow in python

40
/ 100
Emerging

This project helps decipher spoken words by analyzing mouth movements in video, even in noisy environments. It takes a video of a person speaking and outputs the predicted word or phrase they uttered. This is particularly useful for individuals with hearing impairments, those trying to understand speech amidst background noise, or intelligence agencies for covert operations.

164 stars. No commits in the last 6 months.

Use this if you need to understand spoken content from video recordings where audio is unclear or unavailable, such as in noisy settings or for assistive technology for the deaf.

Not ideal if you require real-time, interactive lip-reading beyond sentence-level prediction or if you don't have video input of a speaker's mouth.

hearing-assistance speech-transcription surveillance accessibility video-analysis
No License Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 10 / 25
Maturity 8 / 25
Community 22 / 25

How are scores calculated?

Stars

164

Forks

48

Language

Jupyter Notebook

License

Last pushed

Mar 20, 2018

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/deepconvolution/LipNet"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.