DorsaRoh/transformer-from-scratch

Complete transformer from scratch, using only numpy

25
/ 100
Experimental

This project helps anyone working with language models to understand how a Transformer neural network processes sequential information. It takes an array of real numbers, representing pieces of information like words or sounds, and transforms them through layers to output a probability distribution of what comes next. This is for machine learning researchers or practitioners interested in the foundational mechanics of large language models.

No commits in the last 6 months.

Use this if you need to deeply understand the mathematical operations behind Transformer models for natural language processing or sequence prediction, without relying on high-level libraries.

Not ideal if you're looking for a pre-built, production-ready language model or a tool to quickly apply existing Transformer architectures to your data.

natural-language-processing sequence-prediction deep-learning-research language-modeling neural-network-architecture
No License Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 7 / 25
Maturity 8 / 25
Community 10 / 25

How are scores calculated?

Stars

41

Forks

4

Language

Jupyter Notebook

License

Last pushed

Aug 27, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/generative-ai/DorsaRoh/transformer-from-scratch"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.