Discover and explore top open-source AI tools and projects—updated daily.
nomic-aiPython bindings for local LLM inference (deprecated)
Top 36.8% on SourcePulse
This repository provides Python bindings for llama.cpp and gpt4all, enabling local execution of large language models. It targets developers and researchers seeking to integrate LLM capabilities into Python applications without relying on cloud APIs. The primary benefit is local, private, and cost-effective LLM inference.
How It Works
The project offers Python bindings that interface directly with the llama.cpp C++ library. This allows users to load and run various quantized LLM architectures (like LLaMA, GPT-J, and GPT-Neo) efficiently on local hardware, leveraging CPU or GPU acceleration through llama.cpp's optimized inference engine.
Quick Start & Requirements
pip install gpt4allllama.cpp compilation.Highlighted Details
llama.cpp for efficient CPU and GPU inference.Maintenance & Community
This repository is archived. Future development and support are handled by the main nomic-ai/gpt4all repository.
Licensing & Compatibility
The project is released under the MIT License, permitting commercial use and integration into closed-source applications.
Limitations & Caveats
This repository is archived and no longer actively maintained. The pygpt4all PyPI package may diverge from backend model updates. Users are directed to use the gpt4all package for current bindings.
2 years ago
Inactive
Atome-FE
ModelTC
OpenBMB
nomic-ai