PhoebusSi/Alpaca-CoT
We unified the interfaces of instruction-tuning data (e.g., CoT data), multiple LLMs and parameter-efficient methods (e.g., lora, p-tuning) together for easy use. We welcome open-source enthusiasts to initiate any meaningful PR on this repo and integrate as many LLM related technologies as possible. 我们打造了方便研究人员上手和使用大模型等微调平台,我们欢迎开源爱好者发起任何有意义的pr!
This platform helps researchers quickly customize large language models for specific tasks. You provide instruction-tuning datasets (like Alpaca-CoT) and a base large language model, then select parameter-efficient training methods. The output is a fine-tuned model tailored to your specialized instructions. This is designed for AI researchers or data scientists who want to experiment with or deploy custom LLMs.
2,801 stars. No commits in the last 6 months.
Use this if you need to adapt an existing large language model to perform better on a very specific set of instructions or data, without having to rebuild the model from scratch.
Not ideal if you're looking for an off-the-shelf solution for general language tasks or if you don't have access to instruction-tuning datasets.
Stars
2,801
Forks
251
Language
Jupyter Notebook
License
Apache-2.0
Category
Last pushed
Dec 12, 2023
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/PhoebusSi/Alpaca-CoT"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
shibing624/MedicalGPT
MedicalGPT: Training Your Own Medical GPT Model with ChatGPT Training Pipeline....
lyogavin/airllm
AirLLM 70B inference with single 4GB GPU
GradientHQ/parallax
Parallax is a distributed model serving framework that lets you build your own AI cluster anywhere
CrazyBoyM/llama3-Chinese-chat
Llama3、Llama3.1 中文后训练版仓库 - 微调、魔改版本有趣权重 & 训练、推理、评测、部署教程视频 & 文档。
CLUEbenchmark/CLUE
中文语言理解测评基准 Chinese Language Understanding Evaluation Benchmark: datasets, baselines, pre-trained...