Workshop for local LLM application development
Top 73.3% on sourcepulse
This repository provides a hands-on workshop and example code for developing applications with local Large Language Models (LLMs). It targets developers and researchers interested in building Retrieval Augmented Generation (RAG) chatbots that can query custom documents, with a focus on open-source tools and local deployment. The primary benefit is enabling users to create private, document-aware AI assistants without relying on external cloud services.
How It Works
The application utilizes Streamlit for the user interface, LlamaIndex for document indexing and retrieval, and Ollama for serving local LLMs. This stack allows for RAG by indexing documents into a vector store and then retrieving relevant chunks to augment LLM prompts. The approach emphasizes using open-source components and local LLMs, making it accessible for users without powerful GPUs or those prioritizing data privacy.
Quick Start & Requirements
python -m venv venv
source venv/bin/activate
pip install -r requirements.txt
streamlit run app.py
ollama pull zephyr
).OLLAMA_HOST
environment variable if needed.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
1 year ago
1 day