airllm and Chinese-LLaMA-Alpaca

These are complements: AirLLM provides memory-efficient inference techniques that could optimize the deployment of Chinese-LLaMA-Alpaca models on resource-constrained hardware, while Chinese-LLaMA-Alpaca provides Chinese-adapted model weights and training procedures that AirLLM's quantization and offloading methods could enhance.

airllm
67
Established
Chinese-LLaMA-Alpaca
48
Emerging
Maintenance 10/25
Adoption 12/25
Maturity 25/25
Community 20/25
Maintenance 2/25
Adoption 10/25
Maturity 16/25
Community 20/25
Stars: 13,828
Forks: 1,368
Downloads:
Commits (30d): 0
Language: Jupyter Notebook
License: Apache-2.0
Stars: 18,970
Forks: 1,868
Downloads:
Commits (30d): 0
Language: Python
License: Apache-2.0
No risk flags
Stale 6m No Package No Dependents

About airllm

lyogavin/airllm

AirLLM 70B inference with single 4GB GPU

This project helps AI developers and researchers run powerful Large Language Models (LLMs) on hardware with limited GPU memory. It takes a large LLM like Llama3.1 405B and allows it to generate text on a single 8GB GPU. This means you can deploy sophisticated AI capabilities without needing expensive, high-end graphics cards, making advanced LLMs more accessible.

AI model deployment LLM inference edge AI resource optimization text generation

About Chinese-LLaMA-Alpaca

ymcui/Chinese-LLaMA-Alpaca

中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)

This project helps developers integrate large language models (LLMs) with enhanced Chinese language capabilities into their applications. It provides the foundational Chinese LLaMA models for text completion and the instruction-tuned Chinese Alpaca models for understanding and responding to commands. Developers can input Chinese text or instructions and receive contextually relevant Chinese text generation or answers, making it suitable for building AI products tailored for Chinese speakers.

natural-language-processing chinese-language-ai ai-application-development chatbot-creation text-generation

Scores updated daily from GitHub, PyPI, and npm data. How scores work