milmor/GPT
Implementation of Generative Pretrained Transformer Model in Tensorflow / Keras
This is an implementation for developers who want to experiment with or build upon a Generative Pretrained Transformer (GPT) model using TensorFlow. You can feed it a dataset like OpenWebText to train a new language model, or use a pre-trained GPT-Mini to generate new text based on a given prompt. This is for machine learning engineers or researchers looking to understand and apply foundational large language models.
No commits in the last 6 months.
Use this if you are a machine learning engineer or researcher wanting a clear, functional TensorFlow implementation of a GPT model for training or text generation experiments.
Not ideal if you are looking for a ready-to-use application or a high-level API for natural language processing without diving into model implementation details.
Stars
34
Forks
11
Language
Python
License
MIT
Category
Last pushed
Jun 08, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/milmor/GPT"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
Nixtla/nixtla
TimeGPT-1: production ready pre-trained Time Series Foundation Model for forecasting and...
andrewdalpino/NoPE-GPT
A GPT-style small language model (SLM) with no positional embeddings (NoPE).
sigdelsanjog/gptmed
pip install gptmed
akanyaani/gpt-2-tensorflow2.0
OpenAI GPT2 pre-training and sequence prediction implementation in Tensorflow 2.0
samkamau81/FinGPT_
FinGPT is an AI language model designed to understand and generate financial content. Built upon...