optimum and optimum-graphcore
Optimum-graphcore is a specialized hardware backend plugin for Optimum that enables transformer training specifically on Graphcore IPUs, making them complementary tools designed to be used together rather than alternatives.
About optimum
huggingface/optimum
🚀 Accelerate inference and training of 🤗 Transformers, Diffusers, TIMM and Sentence Transformers with easy to use hardware optimization tools
This tool helps machine learning engineers and researchers accelerate the performance of their large language, image, and sentence models. It takes your existing AI models built with popular frameworks like Hugging Face Transformers or Diffusers and optimizes them for faster execution and training on specialized hardware. The output is a more efficient model that runs quicker and uses fewer resources.
About optimum-graphcore
huggingface/optimum-graphcore
Blazing fast training of 🤗 Transformers on Graphcore IPUs
This project helps machine learning engineers or researchers accelerate the training and fine-tuning of large language models and other AI models. It provides tools to efficiently run popular Hugging Face Transformers models on Graphcore Intelligence Processing Units (IPUs), which are specialized AI processors. You input your existing Transformer models and datasets, and it outputs faster-trained or fine-tuned models ready for deployment.
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work