Minimalist GPT-like transformer implementation for educational purposes
Top 99.5% on sourcepulse
This project provides an absolute minimalistic implementation of a GPT-like transformer using only NumPy, targeting developers and researchers who want to understand the core mechanics of transformer models without the complexity of large frameworks. It offers a learning tool for building and training transformer architectures from scratch.
How It Works
Tensorli implements a GPT-like transformer using a custom Tensorli
object that mimics PyTorch's tensor functionality, built entirely on NumPy. It includes automatic differentiation and essential neural network components like Linearli
, Embeddingli
, MultiheadAttentionli
, and LayerNorm
, along with the Adamli
optimizer. This approach prioritizes clarity and educational value over performance or scalability.
Quick Start & Requirements
conda env create -f environment.yml
or mamba env create -f environment.yml
.conda activate tensorli
.export PYTHONPATH=$PWD
.pytest
.Highlighted Details
Maintenance & Community
The project appears to be a personal learning project with no explicit mention of contributors, sponsorships, or community channels.
Licensing & Compatibility
The README does not specify a license. Compatibility for commercial or closed-source use is not addressed.
Limitations & Caveats
This library is not optimized and is not intended for production use or large-scale applications. It is purely a learning tool. Dropout and additional experimental architectures are planned but not yet implemented.
1 year ago
1 day