kyegomez/VisionLLaMA

Implementation of VisionLLaMA from the paper: "VisionLLaMA: A Unified LLaMA Interface for Vision Tasks" in PyTorch and Zeta

22
/ 100
Experimental

This project helps machine learning engineers and researchers explore and implement advanced vision models. It takes image data as input and processes it through a LLaMA-like architecture to produce outputs for various computer vision tasks, such as image classification. This is primarily used by AI/ML practitioners focused on cutting-edge model development.

No commits in the last 6 months.

Use this if you are an AI/ML engineer or researcher working with PyTorch and Zeta, and you want to experiment with a unified LLaMA interface for vision tasks as described in the VisionLLaMA paper.

Not ideal if you are looking for a plug-and-play solution for common computer vision problems without deep involvement in model architecture or framework-level development.

Computer Vision Deep Learning Research Model Architecture Image Classification AI Development
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 6 / 25
Maturity 16 / 25
Community 0 / 25

How are scores calculated?

Stars

16

Forks

Language

Python

License

MIT

Last pushed

Nov 11, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/kyegomez/VisionLLaMA"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.