All Transformer Models
7,795 models ranked by quality score · Page 16 of 78
| # | Model | Score | Tier |
|---|---|---|---|
| 1501 |
parvbhullar/superpilot
LLMs based multi-model framework for building AI apps. |
|
Emerging |
| 1502 |
deep-symbolic-mathematics/Multimodal-Symbolic-Regression
[ICLR 2024 Spotlight] SNIP on Symbolic Regression: Deep Symbolic Regression... |
|
Emerging |
| 1503 |
jaco-bro/MLX.zig
MLX.zig: Phi-4, Llama 3.2, and Whisper in Zig |
|
Emerging |
| 1504 |
Infini-AI-Lab/vortex_torch
Vortex: A Flexible and Efficient Sparse Attention Framework |
|
Emerging |
| 1505 |
InhwanBae/LMTrajectory
Official Code for "Can Language Beat Numerical Regression? Language-Based... |
|
Emerging |
| 1506 |
daniel-furman/sft-demos
Lightweight demos for finetuning LLMs. Powered by 🤗 transformers and... |
|
Emerging |
| 1507 |
zjukg/KoPA
[Paper][ACM MM 2024] Making Large Language Models Perform Better in... |
|
Emerging |
| 1508 |
Longyichen/Alpaca-family-library
Summarize all open source Large Languages Models and low-cost replication... |
|
Emerging |
| 1509 |
hao-ai-lab/Consistency_LLM
[ICML 2024] CLLMs: Consistency Large Language Models |
|
Emerging |
| 1510 |
AIoT-MLSys-Lab/Efficient-LLMs-Survey
[TMLR 2024] Efficient Large Language Models: A Survey |
|
Emerging |
| 1511 |
miranthajayatilake/nanoQA
Question-answering on your own data with Large Language Models (LLMs) |
|
Emerging |
| 1512 |
ivanfioravanti/wine_variety_classification
Examples on how to use various LLM providers with a Wine Classification problem |
|
Emerging |
| 1513 |
otadk/nuxt-edge-ai
Nuxt module for local-first AI apps with server-side WASM inference via... |
|
Emerging |
| 1514 |
EagleW/Stage-wise-Fine-tuning
Code for Stage-wise Fine-tuning for Graph-to-Text Generation |
|
Emerging |
| 1515 |
dbmdz/berts
DBMDZ BERT, DistilBERT, ELECTRA, GPT-2 and ConvBERT models |
|
Emerging |
| 1516 |
rohit901/VANE-Bench
[NAACL'25] Contains code and documentation for our VANE-Bench paper. |
|
Emerging |
| 1517 |
dohlee/chromoformer
The official code implementation for Chromoformer in PyTorch. (Lee et al.,... |
|
Emerging |
| 1518 |
samestrin/llm-newsletter-generator
llm-newsletter-generator transforms a valid RSS feed into a "Newsletter"... |
|
Emerging |
| 1519 |
WENGSYX/LMTuner
LMTuner: Make the LLM Better for Everyone |
|
Emerging |
| 1520 |
kyegomez/qformer
Implementation of Qformer from BLIP2 in Zeta Lego blocks. |
|
Emerging |
| 1521 |
amin-tehrani/ollama-colab
Serve Ollama LLMs on Google Colab (free plan) using Ngrok |
|
Emerging |
| 1522 |
cocktailpeanut/dalai
The simplest way to run LLaMA on your local machine |
|
Emerging |
| 1523 |
RightNow-AI/TIDE
Dynamic per-token early exit for LLM inference. Skip layers tokens don't need |
|
Emerging |
| 1524 |
Kagamma/llama-pas
Free Pascal bindings for llama.cpp |
|
Emerging |
| 1525 |
jie-jw-wu/human-eval-comm
HumanEvalComm: Evaluating Communication Skill of Code LLM and LLM Agent |
|
Emerging |
| 1526 |
pmichel31415/are-16-heads-really-better-than-1
Code for the paper "Are Sixteen Heads Really Better than One?" |
|
Emerging |
| 1527 |
ma2za/telegram-llm-bot
Telegram LLM bot backed by OpenAI, Whisper, Beam, LLaMA, Weaviate, MinIO and MongoDB |
|
Emerging |
| 1528 |
IvanBongiorni/maximal
A TensorFlow-compatible Python library that provides models and layers to... |
|
Emerging |
| 1529 |
cmhungsteve/Awesome-Transformer-Attention
An ultimately comprehensive paper list of Vision Transformer/Attention,... |
|
Emerging |
| 1530 |
chenhan97/TimeLlama
The official repo of TimeLlama, an instruction-finetuned Llama2 series that... |
|
Emerging |
| 1531 |
hasanirtiza/PedesFormer-Transformer-Networks-For-Pedestrian-Detection
Transformer Networks for Pedestrian Detection |
|
Emerging |
| 1532 |
AnkitNayak-eth/llmBench
llmBench is a high-depth benchmarking tool designed to measure the raw... |
|
Emerging |
| 1533 |
di37/finetuning-quantize-evaluate
Fine-Tune, Quantize, Evaluate: The Complete Guide — LLMs, VLMs, and Embedding Models |
|
Emerging |
| 1534 |
takara-ai/go-attention
A full attention mechanism and transformer in pure go. |
|
Emerging |
| 1535 |
botisan-ai/sentence-transformers.js
Run sentence-transformers (SBERT) compatible models in Node.js or browser. |
|
Emerging |
| 1536 |
rust-dd/iTransformer
An iTransformer implementation in Rust |
|
Emerging |
| 1537 |
pyladiesams/eval-llm-based-apps-jan2025
Create an evaluation framework for your LLM based app. Incorporate it into... |
|
Emerging |
| 1538 |
albrateanu/ModalFormer
[2025] ModalFormer: Multimodal Transformer for Low-Light Image Enhancement |
|
Emerging |
| 1539 |
AmpereComputingAI/llama.cpp
Ampere optimized llama.cpp |
|
Emerging |
| 1540 |
mbzuai-oryx/Awesome-LLM-Post-training
Awesome Reasoning LLM Tutorial/Survey/Guide |
|
Emerging |
| 1541 |
datawhalechina/diy-llm
🎓 系统性大语言模型构建课程|🛠️ 覆盖预训练数据工程、Tokenizer、Transformer、MoE、GPU 编程... |
|
Emerging |
| 1542 |
rosinality/halite
Acceleration framework for Human Alignment Learning |
|
Emerging |
| 1543 |
iflytek/VLE
VLE: Vision-Language Encoder (VLE: 视觉-语言多模态预训练模型) |
|
Emerging |
| 1544 |
bwittmann/transoar
A 3D medical Detection Transformer library. Papers accepted @ MIDL22 & MELBA23/02. |
|
Emerging |
| 1545 |
biswassanket/DocSegTr
A Bottom-Up Instance Segmentation Strategy for segmenting document instances... |
|
Emerging |
| 1546 |
lenguajenatural-ai/autotransformers
A Python package for automatically training and comparing language models. |
|
Emerging |
| 1547 |
viddexa/moderators
One package to moderate them all |
|
Emerging |
| 1548 |
osainz59/Ask2Transformers
A Framework for Textual Entailment based Zero Shot text classification |
|
Emerging |
| 1549 |
EvilFreelancer/impruver
A set of scripts and configurations for pretraining of Large Language Models (LLM) |
|
Emerging |
| 1550 |
Sandipan99/IndMask
IndMask: Inductive Explanation for Multivariate Time Series Black-box Model |
|
Emerging |
| 1551 |
Nkluge-correa/TeenyTinyLlama
A pair of tiny foundational models trained in Brazilian Portuguese.🦙🦙 |
|
Emerging |
| 1552 |
yizhangele/llm-guided-mod-optimization
Implementation for “Hierarchical Optimization via LLM-Guided Objective... |
|
Emerging |
| 1553 |
epfml/llm-optimizer-benchmark
Benchmarking Optimizers for LLM Pretraining |
|
Emerging |
| 1554 |
DirtyHarryLYL/Transformer-in-Vision
Recent Transformer-based CV and related works. |
|
Emerging |
| 1555 |
Kirill-Kravtsov/drophead-pytorch
An implementation of drophead regularization for pytorch transformers |
|
Emerging |
| 1556 |
dcaffo98/transpormer
TranSPormer: a transformer for the Travelling Salesman Problem |
|
Emerging |
| 1557 |
TrevTron/indiedroid-nova-llm
Running Llama 3.1 8B and other LLMs on RK3588 NPU - benchmarks and setup guides |
|
Emerging |
| 1558 |
kolinko/effort
An implementation of bucketMul LLM inference |
|
Emerging |
| 1559 |
NiuTrans/LMT
Building a inclusive, scalable, and high-performance multilingual translation model |
|
Emerging |
| 1560 |
jlin816/dynalang
Code for "Learning to Model the World with Language." ICML 2024 Oral. |
|
Emerging |
| 1561 |
ymoslem/Adaptive-MT-LLM-Fine-tuning
Fine-tuning Open-Source LLMs for Adaptive Machine Translation |
|
Emerging |
| 1562 |
yueyu1030/AttrPrompt
[NeurIPS 2023] This is the code for the paper `Large Language Model as... |
|
Emerging |
| 1563 |
mikemayuare/apetokenizer
Tokenizer for chemnical SMILES and SELFIES for use in transformers models. |
|
Emerging |
| 1564 |
shufangxun/LLaVA-MoD
[ICLR 2025] LLaVA-MoD: Making LLaVA Tiny via MoE-Knowledge Distillation |
|
Emerging |
| 1565 |
OFA-Sys/OFASys
OFASys: A Multi-Modal Multi-Task Learning System for Building Generalist Models |
|
Emerging |
| 1566 |
awneesht/KVShuttle
Benchmark & decision framework for KV cache transfer compression in... |
|
Emerging |
| 1567 |
HillZhang1999/ICD
Code & Data for our Paper "Alleviating Hallucinations of Large Language... |
|
Emerging |
| 1568 |
ZongXR/8th-National-AI-Training-Competition
第八届全国职工职业技能大赛人工智能训练师赛项 |
|
Emerging |
| 1569 |
OFA-Sys/ExpertLLaMA
An opensource ChatBot built with ExpertPrompting which achieves 96% of... |
|
Emerging |
| 1570 |
LostBeard/SpawnDev.BlazorJS.TransformersJS
Use Transformers.js from Blazor WebAssembly to run pretrained models with... |
|
Emerging |
| 1571 |
katanaml/table-query-model
Table Query with ML |
|
Emerging |
| 1572 |
GiovanniGatti/socratic-llm
Training pipeline for fine tuning Phi-3-mini-instruct to follow the Socratic method |
|
Emerging |
| 1573 |
wenge-research/YAYI
雅意大模型:为客户打造安全可靠的专属大模型,基于大规模中英文多领域指令数据训练的 LlaMA 2 & BLOOM... |
|
Emerging |
| 1574 |
Curated-Awesome-Lists/awesome-llms-fine-tuning
Explore a comprehensive collection of resources, tutorials, papers, tools,... |
|
Emerging |
| 1575 |
JinhaoLee/WCA
[ICML 2024] Visual-Text Cross Alignment: Refining the Similarity Score in... |
|
Emerging |
| 1576 |
minosvasilias/godot-dodo
Finetuning large language models for GDScript generation. |
|
Emerging |
| 1577 |
InnovatorLM/Innovator-VL
Fully Open-source Multimodal Language Models for Science Discovery |
|
Emerging |
| 1578 |
OnlyTerp/kvtc
First open-source KVTC implementation (NVIDIA, ICLR 2026) -- 8-32x KV cache... |
|
Emerging |
| 1579 |
iVishalr/GPT
A minimal and efficient Pytorch implementation of OpenAI's GPT (Generative... |
|
Emerging |
| 1580 |
ManasVardhan/bench-my-llm
🏎️ Dead-simple LLM benchmarking CLI - latency, cost, and quality metrics |
|
Emerging |
| 1581 |
VikingOwl91/vessel
A lightweight, local-first web UI for managing Ollama models. |
|
Emerging |
| 1582 |
icon-lab/SLATER
Official implementation of the paper: Unsupervised MRI Reconstruction via... |
|
Emerging |
| 1583 |
arrmansa/Basic-UI-for-GPT-J-6B-with-low-vram
A repository to run gpt-j-6b on low vram machines (4.2 gb minimum vram for... |
|
Emerging |
| 1584 |
stylellm/stylellm_models
StyleLLM文风大模型:基于大语言模型的文本风格迁移项目。Text style transfer base on Large Language... |
|
Emerging |
| 1585 |
sotiraslab/AgileFormer
This the repo for the paper tiltled "AgileFormer: Spatially Agile... |
|
Emerging |
| 1586 |
JosefAlbers/VL-JEPA
VL-JEPA (Vision-Language Joint Embedding Predictive Architecture) in MLX |
|
Emerging |
| 1587 |
kyaiooiayk/Awesome-LLM-Large-Language-Models-Notes
What can I do with a LLM model? |
|
Emerging |
| 1588 |
efeslab/Nanoflow
A throughput-oriented high-performance serving framework for LLMs |
|
Emerging |
| 1589 |
SqueezeAILab/LLM2LLM
[ACL 2024] LLM2LLM: Boosting LLMs with Novel Iterative Data Enhancement |
|
Emerging |
| 1590 |
eqimp/hogwild_llm
Official PyTorch implementation for Hogwild! Inference: Parallel LLM... |
|
Emerging |
| 1591 |
zhanshijinwat/Steel-LLM
Train a 1B LLM with 1T tokens from scratch by personal |
|
Emerging |
| 1592 |
kyegomez/CNNGPT
This CNN-based language model leverages causal and dilated convolutions,... |
|
Emerging |
| 1593 |
anthonyfoust/ai-stack-homelab
Complete AI automation stack optimized for Mac Mini M4, but can work in... |
|
Emerging |
| 1594 |
Gurumurthy30/Stackformer
Modular PyTorch transformer library for building, training, and... |
|
Emerging |
| 1595 |
itsnamgyu/block-transformer
Block Transformer: Global-to-Local Language Modeling for Fast Inference... |
|
Emerging |
| 1596 |
Sakeeb91/text2sql-agent
Self-correcting AI agent for natural language to SQL using HuggingFace... |
|
Emerging |
| 1597 |
WhereIsAI/BiLLM
Tool for converting LLMs from uni-directional to bi-directional by removing... |
|
Emerging |
| 1598 |
tomekkorbak/pretraining-with-human-feedback
Code accompanying the paper Pretraining Language Models with Human Preferences |
|
Emerging |
| 1599 |
sayakpaul/probing-vits
Probing the representations of Vision Transformers. |
|
Emerging |
| 1600 |
ccdv-ai/convert_checkpoint_to_lsg
Efficient Attention for Long Sequence Processing |
|
Emerging |