Discover and explore top open-source AI tools and projects—updated daily.
joennlaeMinimalist GPT-like transformer implementation for educational purposes
Top 99.1% on SourcePulse
This project provides an absolute minimalistic implementation of a GPT-like transformer using only NumPy, targeting developers and researchers who want to understand the core mechanics of transformer models without the complexity of large frameworks. It offers a learning tool for building and training transformer architectures from scratch.
How It Works
Tensorli implements a GPT-like transformer using a custom Tensorli object that mimics PyTorch's tensor functionality, built entirely on NumPy. It includes automatic differentiation and essential neural network components like Linearli, Embeddingli, MultiheadAttentionli, and LayerNorm, along with the Adamli optimizer. This approach prioritizes clarity and educational value over performance or scalability.
Quick Start & Requirements
conda env create -f environment.yml or mamba env create -f environment.yml.conda activate tensorli.export PYTHONPATH=$PWD.pytest.Highlighted Details
Maintenance & Community
The project appears to be a personal learning project with no explicit mention of contributors, sponsorships, or community channels.
Licensing & Compatibility
The README does not specify a license. Compatibility for commercial or closed-source use is not addressed.
Limitations & Caveats
This library is not optimized and is not intended for production use or large-scale applications. It is purely a learning tool. Dropout and additional experimental architectures are planned but not yet implemented.
1 year ago
Inactive
wandb
microsoft
huggingface
huggingface
facebookresearch
NVIDIA