amazon-science/ContraCLM
[ACL 2023] Code for ContraCLM: Contrastive Learning For Causal Language Model
This project helps machine learning engineers pre-train causal language models like GPT2 and CodeGen for better performance on various NLP and programming tasks. It takes in large natural language or programming language datasets and outputs a more robust, discriminative language model. Data scientists and ML researchers who work with large language models would use this to enhance their models beyond simple text generation.
No commits in the last 6 months.
Use this if you are a machine learning engineer or researcher looking to improve the representation quality and discriminative power of your decoder-only language models for tasks beyond just text generation.
Not ideal if you are looking for a pre-trained model to use directly in an application, as this is a framework for pre-training models.
Stars
35
Forks
2
Language
Python
License
Apache-2.0
Category
Last pushed
Dec 20, 2023
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/amazon-science/ContraCLM"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
filipnaudot/llmSHAP
llmSHAP: a multi-threaded explainability framework using Shapley values for LLM-based outputs.
microsoft/automated-brain-explanations
Generating and validating natural-language explanations for the brain.
CAS-SIAT-XinHai/CPsyCoun
[ACL 2024] CPsyCoun: A Report-based Multi-turn Dialogue Reconstruction and Evaluation Framework...
wesg52/universal-neurons
Universal Neurons in GPT2 Language Models
ICTMCG/LLM-for-misinformation-research
Paper list of misinformation research using (multi-modal) large language models, i.e., (M)LLMs.