VQA-Team/Visual-Question-Answering
The project is an Android application aimed to help the visually impaired by giving them the ability to take a picture, ask questions about it and the application will provide them with the answers using machine learning techniques and tools.
This Android application helps visually impaired individuals understand their surroundings better. Users can take a picture and verbally ask questions about its content. The app processes the image and question to provide spoken answers, acting as a personal visual assistant.
No commits in the last 6 months.
Use this if you are visually impaired and need a convenient way to get spoken descriptions and answers about objects and scenes in your immediate environment.
Not ideal if you need detailed, nuanced descriptions or if the environment involves highly complex or abstract visual information.
Stars
7
Forks
4
Language
Jupyter Notebook
License
—
Category
Last pushed
May 28, 2022
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/VQA-Team/Visual-Question-Answering"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
open-mmlab/mmpretrain
OpenMMLab Pre-training Toolbox and Benchmark
facebookresearch/mmf
A modular framework for vision & language multimodal research from Facebook AI Research (FAIR)
adambielski/siamese-triplet
Siamese and triplet networks with online pair/triplet mining in PyTorch
HuaizhengZhang/Awsome-Deep-Learning-for-Video-Analysis
Papers, code and datasets about deep learning and multi-modal learning for video analysis
KaiyangZhou/pytorch-vsumm-reinforce
Unsupervised video summarization with deep reinforcement learning (AAAI'18)