kyegomez/VisionLLaMA
Implementation of VisionLLaMA from the paper: "VisionLLaMA: A Unified LLaMA Interface for Vision Tasks" in PyTorch and Zeta
This project helps machine learning engineers and researchers explore and implement advanced vision models. It takes image data as input and processes it through a LLaMA-like architecture to produce outputs for various computer vision tasks, such as image classification. This is primarily used by AI/ML practitioners focused on cutting-edge model development.
No commits in the last 6 months.
Use this if you are an AI/ML engineer or researcher working with PyTorch and Zeta, and you want to experiment with a unified LLaMA interface for vision tasks as described in the VisionLLaMA paper.
Not ideal if you are looking for a plug-and-play solution for common computer vision problems without deep involvement in model architecture or framework-level development.
Stars
16
Forks
—
Language
Python
License
MIT
Category
Last pushed
Nov 11, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/kyegomez/VisionLLaMA"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
open-mmlab/mmpretrain
OpenMMLab Pre-training Toolbox and Benchmark
facebookresearch/mmf
A modular framework for vision & language multimodal research from Facebook AI Research (FAIR)
HuaizhengZhang/Awsome-Deep-Learning-for-Video-Analysis
Papers, code and datasets about deep learning and multi-modal learning for video analysis
KaiyangZhou/pytorch-vsumm-reinforce
Unsupervised video summarization with deep reinforcement learning (AAAI'18)
adambielski/siamese-triplet
Siamese and triplet networks with online pair/triplet mining in PyTorch