marella/ctransformers
Python bindings for the Transformer models implemented in C/C++ using GGML library.
This tool helps AI developers efficiently run large language models (LLMs) like LLaMA and Falcon on their own machines, even without powerful GPUs. It takes a pre-trained LLM and allows you to generate text or integrate it into other AI frameworks. Developers building AI applications or prototypes can use this to make their models accessible on standard hardware.
1,882 stars. No commits in the last 6 months.
Use this if you are a developer who needs to run transformer-based large language models efficiently on consumer-grade hardware, including CPUs or entry-level GPUs.
Not ideal if you are a non-developer seeking a ready-to-use application, as this project requires programming knowledge to implement.
Stars
1,882
Forks
143
Language
C
License
MIT
Category
Last pushed
Jan 28, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/marella/ctransformers"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
lucidrains/x-transformers
A concise but complete full-attention transformer with a set of promising experimental features...
kanishkamisra/minicons
Utility for behavioral and representational analyses of Language Models
lucidrains/simple-hierarchical-transformer
Experiments around a simple idea for inducing multiple hierarchical predictive model within a GPT
lucidrains/dreamer4
Implementation of Danijar's latest iteration for his Dreamer line of work
Nicolepcx/Transformers-in-Action
This is the corresponding code for the book Transformers in Action