PyTorch pre-training for LaMDA research paper
Top 65.6% on sourcepulse
This repository provides an open-source PyTorch implementation of Google's LaMDA architecture, focusing on a 2B parameter model suitable for researchers and developers interested in replicating or extending large language models for dialog applications. It aims to incorporate Reinforcement Learning from Human Feedback (RLHF) similar to ChatGPT.
How It Works
The implementation follows a GPT-like decoder-only architecture, utilizing T5's relative positional bias in attention and Gated GELU activation in the feed-forward layers. It employs a Sentencepiece byte-pair encoded tokenizer for efficient text processing. The model is designed for autoregressive generation with Top-k sampling.
Quick Start & Requirements
pip install
is planned but not yet available.Highlighted Details
Maintenance & Community
The project is authored by Enrico Shippole, with updates available on Twitter and LinkedIn. Key TODO items include Sentencepiece tokenizer training/integration, detailed documentation, finetuning scripts, and a PyPI installer.
Licensing & Compatibility
The repository does not explicitly state a license. Compatibility for commercial use or closed-source linking is not specified.
Limitations & Caveats
The project is under active development with several planned features (Sentencepiece, finetuning, pip installer) yet to be implemented. Official documentation is also pending. The absence of an explicit license may pose restrictions on commercial use.
1 year ago
Inactive