wyt2000/InverseCoder
[AAAI 2025] The official code of the paper "InverseCoder: Unleashing the Power of Instruction-Tuned Code LLMs with Inverse-Instruct"(https://arxiv.org/abs/2407.05700).
This project helps AI engineers and machine learning researchers enhance the capabilities of large language models for code generation. It takes existing code snippets and automatically generates high-quality programming instructions for them. The output is a refined dataset that makes code LLMs better at understanding and responding to natural language prompts for coding tasks.
No commits in the last 6 months.
Use this if you are a developer or researcher looking to create more robust and accurate code generation LLMs by automatically expanding and improving their training data.
Not ideal if you are looking for a tool to directly write code for your projects; this is a toolkit for training code generation models, not for direct code production.
Stars
14
Forks
1
Language
Python
License
—
Category
Last pushed
Jul 10, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/wyt2000/InverseCoder"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
jncraton/languagemodels
Explore large language models in 512MB of RAM
microsoft/unilm
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
haizelabs/verdict
Inference-time scaling for LLMs-as-a-judge.
albertan017/LLM4Decompile
Reverse Engineering: Decompiling Binary Code with Large Language Models
bytedance/Sa2VA
Official Repo For Pixel-LLM Codebase