Self-hosted code completion engine (deprecated)
Top 13.0% on sourcepulse
TurboPilot was an open-source, self-hosted code completion engine designed to run large language models locally on CPU, targeting developers seeking an alternative to cloud-based AI coding assistants. It aimed to provide efficient, private code suggestions by leveraging quantized models and the llama.cpp library.
How It Works
TurboPilot utilizes the llama.cpp library to run quantized versions of large language models, such as Salesforce Codegen, WizardCoder, and Starcoder, on consumer hardware. This approach allows for local inference, reducing reliance on external servers and enhancing privacy. The project supports various model formats and quantization levels, enabling users with limited RAM (as low as 4GB) to run capable models, while also offering GPU offloading for enhanced performance.
Quick Start & Requirements
./turbopilot -m starcoder -f ./models/santacoder-q4_0.bin
or docker run --rm -it -v ./models:/models -e THREADS=6 -e MODEL_TYPE=starcoder -e MODEL="/models/santacoder-q4_0.bin" -p 18080:18080 ghcr.io/ravenscroftj/turbopilot:latest
.Highlighted Details
vscode-fauxpilot
plugin.Maintenance & Community
TurboPilot is deprecated and archived as of September 30, 2023. The author recommends exploring more mature alternatives.
Licensing & Compatibility
The project's licensing is not explicitly stated in the README, but it relies on GGML and llama.cpp, which are typically under permissive licenses. Compatibility for commercial use or closed-source linking would require verification of the specific model licenses and the project's own licensing.
Limitations & Caveats
The project is explicitly marked as deprecated and archived. It was considered a proof-of-concept with potentially slow autocompletion and only supports one GPU device at a time.
1 year ago
1 day