Coding workshop for understanding LLM implementation and usage
Top 38.0% on sourcepulse
This repository provides materials for a 4-hour coding workshop focused on understanding and implementing Large Language Models (LLMs) from scratch. It targets coders interested in the foundational building blocks, architecture, and practical application of LLMs, enabling them to build, pretrain, and finetune their own models.
How It Works
The workshop guides participants through coding a GPT-like LLM using PyTorch. It covers essential components: data input pipelines (tokenization, DataLoaders), core architectural elements, and the pretraining process. Subsequently, it demonstrates loading pretrained weights and finetuning LLMs using the LitGPT library, offering a practical bridge from foundational concepts to real-world usage.
Quick Start & Requirements
setup
folder.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The pretraining section uses a small text sample, meaning the self-built LLM will only generate basic sentences. The workshop focuses on understanding the core mechanics rather than achieving state-of-the-art performance.
6 months ago
1 day