Educational resource for building a Storyteller AI LLM
Top 1.0% on sourcepulse
This repository outlines a comprehensive, end-to-end course for building a Storyteller AI Large Language Model (LLM) from scratch. Aimed at individuals seeking a deep understanding of AI and LLMs, it guides users through creating, refining, and illustrating stories with an AI, culminating in a ChatGPT-like web application.
How It Works
The course progresses from fundamental concepts like Bigram Language Models and micrograd for backpropagation to advanced topics including attention mechanisms, Transformer architectures (GPT-2), tokenization (minBPE), optimization techniques (AdamW), and speed enhancements via device utilization, mixed precision, and distributed training. It covers dataset handling, inference optimizations (kv-cache, quantization), various finetuning methods (SFT, RLHF), and deployment strategies.
Quick Start & Requirements
This is a course syllabus and does not contain runnable code. The development is ongoing by Eureka Labs.
Highlighted Details
Maintenance & Community
This repository is currently archived as the course is under development by Eureka Labs. Further details on community or roadmap are not yet available.
Licensing & Compatibility
The repository is archived and does not specify a license.
Limitations & Caveats
The course content is under development and not yet available. The repository is archived, indicating the project is not actively maintained in its current state.
1 year ago
Inactive