Hodasia/Awesome-Vision-Language-Finetune
Awesome List of Vision Language Prompt Papers
This is a curated list of research papers focused on making large vision-language AI models, like CLIP, more effective for specific tasks. It helps AI researchers and machine learning engineers find efficient ways to adapt these powerful models to new image and text datasets. The list takes academic papers on 'prompting,' 'finetuning,' and 'adapters' as input and organizes them chronologically for easy review.
No commits in the last 6 months.
Use this if you are an AI researcher or machine learning engineer exploring methods to efficiently adapt pre-trained vision-language models for custom datasets or specialized tasks.
Not ideal if you are looking for a ready-to-use tool or code for deploying vision-language models without needing to understand the underlying research and adaptation techniques.
Stars
47
Forks
1
Language
—
License
—
Category
Last pushed
Nov 09, 2023
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/Hodasia/Awesome-Vision-Language-Finetune"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
open-mmlab/mmpretrain
OpenMMLab Pre-training Toolbox and Benchmark
facebookresearch/mmf
A modular framework for vision & language multimodal research from Facebook AI Research (FAIR)
HuaizhengZhang/Awsome-Deep-Learning-for-Video-Analysis
Papers, code and datasets about deep learning and multi-modal learning for video analysis
KaiyangZhou/pytorch-vsumm-reinforce
Unsupervised video summarization with deep reinforcement learning (AAAI'18)
adambielski/siamese-triplet
Siamese and triplet networks with online pair/triplet mining in PyTorch