GusLovesMath/Llama3_MacSilicon
Repository for running LLMs efficiently on Mac silicon (M1, M2, M3). Features Jupyter notebook for Meta-Llama-3 setup using MLX framework, with install guide & perf tips. Aims to optimize LLM performance on Mac silicon for devs & researchers.
This project helps you run the powerful Meta-Llama-3 language model directly on your Apple Mac with M-series chips. You input text prompts, and it generates human-like responses, from simple answers to solving math problems. It's designed for researchers and developers who want to experiment with or integrate large language models efficiently on their personal Mac hardware.
No commits in the last 6 months.
Use this if you are a developer or researcher with a Mac M-series computer and want to run the Llama 3 large language model locally for experimentation or application development without needing cloud services.
Not ideal if you don't have a Mac with an M1, M2, or M3 chip, or if you need to deploy large-scale, high-throughput AI applications in a production environment.
Stars
11
Forks
1
Language
Jupyter Notebook
License
—
Category
Last pushed
May 04, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/GusLovesMath/Llama3_MacSilicon"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
OpenNMT/CTranslate2
Fast inference engine for Transformer models
mechramc/Orion
Local AI runtime for training & running small LLMs directly on Apple Neural Engine (ANE). No...
Pomilon/LEMA
LEMA (Layer-wise Efficient Memory Abstraction): A hardware-aware framework for fine-tuning LLMs...
dilbersha/llm-inference-benchmarking-3080
A production-grade telemetry-aware suite for benchmarking LLM inference performance on NVIDIA RTX 3080.
Yuan-ManX/infera
Infera — A High-Performance Inference Engine for Large Language Models.