VS Code extension for local AI code completion
Top 22.3% on sourcepulse
Llama Coder offers a self-hosted, privacy-focused alternative to GitHub Copilot for VS Code users. It leverages Ollama and Code Llama models to provide local AI-powered code completion, aiming for performance comparable to commercial solutions while running on user hardware.
How It Works
The extension integrates with Ollama, a framework for running large language models locally. Users can select from various Code Llama model sizes and quantization levels, with larger, more quantized models generally offering better performance. The system is designed for flexibility, allowing users to run inference on their local machine or offload it to a dedicated server via a configurable Ollama endpoint.
Quick Start & Requirements
Highlighted Details
stable-code:3b-code-q4_0
, codellama:7b-code-q4_K_M
).Maintenance & Community
Recent updates include features like pausing completions, bearer token support for remote inference, and improved Jupyter notebook support. The project is actively maintained with frequent releases.
Licensing & Compatibility
The README does not explicitly state a license. Compatibility for commercial use or closed-source linking is not specified.
Limitations & Caveats
Performance is highly dependent on user hardware, with Apple Silicon or high-end NVIDIA GPUs recommended. Some models may perform slowly on older NVIDIA cards or macOS. The project does not specify which models are supported beyond Code Llama and DeepSeek.
1 year ago
1 week