A-baoYang/instruction-finetune-datasets
Collect and maintain high quality instruction finetune datasets in different domain and languages. 搜集並維護高品質各專業領域及語言的指令微調資料集
This project offers a collection of high-quality, pre-structured datasets designed to help you build specialized AI assistants or chatbots. You provide these datasets to a language model, and it learns to respond to specific instructions or questions in a particular domain. It's ideal for domain experts, educators, or business owners who want to create AI tools that understand and generate content relevant to their field, even if they have limited technical resources.
No commits in the last 6 months.
Use this if you need ready-to-use instruction datasets to train a large language model to perform specific tasks, generate code, or answer questions in various languages and professional domains like healthcare or general knowledge.
Not ideal if you're looking for raw, unstructured text data or if you need datasets for general, non-instruction-based language model training.
Stars
20
Forks
—
Language
—
License
CC0-1.0
Category
Last pushed
Apr 06, 2023
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/A-baoYang/instruction-finetune-datasets"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
MantisAI/sieves
Plug-and-play document AI with zero-shot models.
xiaoya-li/Instruction-Tuning-Survey
Project for the paper entitled `Instruction Tuning for Large Language Models: A Survey`
rafaelpierre/bullet
bullet: A Zero-Shot / Few-Shot Learning, LLM Based, text classification framework
TencentARC-QQ/TagGPT
TagGPT: Large Language Models are Zero-shot Multimodal Taggers
amazon-science/adaptive-in-context-learning
AdaICL: Which Examples to Annotate of In-Context Learning? Towards Effective and Efficient Selection