AdityaNG/kan-gpt
The PyTorch implementation of Generative Pre-trained Transformers (GPTs) using Kolmogorov-Arnold Networks (KANs) for language modeling
This project offers a Python implementation for building language models, specifically Generative Pre-trained Transformers (GPTs), using a different type of neural network called Kolmogorov-Arnold Networks (KANs). It takes in text data and can generate new, coherent text based on the patterns it learned. This tool is designed for researchers and machine learning engineers who are exploring novel neural network architectures for natural language processing.
725 stars. No commits in the last 6 months. Available on PyPI.
Use this if you are a machine learning researcher or engineer interested in experimenting with KANs for language generation and want to compare their performance against traditional MLP-based GPT models.
Not ideal if you are looking for a ready-to-use, production-grade language model for immediate deployment or if you are not comfortable with advanced machine learning development.
Stars
725
Forks
55
Language
Python
License
MIT
Category
Last pushed
Nov 25, 2024
Commits (30d)
0
Dependencies
12
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/AdityaNG/kan-gpt"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related models
tabularis-ai/be_great
A novel approach for synthesizing tabular data using pretrained large language models
EleutherAI/gpt-neox
An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron...
shibing624/textgen
TextGen: Implementation of Text Generation models, include LLaMA, BLOOM, GPT2, BART, T5, SongNet...
ai-forever/ru-gpts
Russian GPT3 models.
zemlyansky/gpt-tfjs
GPT in TensorFlow.js