microsoft/AdaMix

This is the implementation of the paper AdaMix: Mixture-of-Adaptations for Parameter-efficient Model Tuning (https://arxiv.org/abs/2205.12410).

37
/ 100
Emerging

This project helps machine learning engineers improve the performance of large language models like BERT and RoBERTa on specific natural language processing tasks. It takes an existing pre-trained model and fine-tunes it using a 'mixture-of-adaptations' approach, resulting in a specialized model that performs better on your chosen task while being more parameter-efficient than traditional fine-tuning.

138 stars. No commits in the last 6 months.

Use this if you are a machine learning engineer working on fine-tuning large language models for tasks such as text classification, natural language inference, or semantic similarity, and want to achieve better performance with fewer trainable parameters.

Not ideal if you are looking for a general-purpose, out-of-the-box solution for end-users or if you are not comfortable working with deep learning model training and evaluation scripts.

natural-language-processing large-language-models model-fine-tuning text-classification machine-learning-engineering
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 11 / 25

How are scores calculated?

Stars

138

Forks

11

Language

Python

License

Apache-2.0

Last pushed

Aug 14, 2023

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/microsoft/AdaMix"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.