Inference code for CodeLlama models
Top 2.9% on sourcepulse
Code Llama provides state-of-the-art, open-source large language models specifically for code-related tasks. It offers foundation models, Python-specialized versions, and instruction-following variants across various parameter sizes (7B to 70B), targeting developers, researchers, and businesses seeking advanced code generation, infilling, and large context handling capabilities.
How It Works
Code Llama is built by fine-tuning the Llama 2 architecture on a large corpus of code. It supports long input contexts (up to 100k tokens) and features infilling capabilities for 7B and 13B models, allowing code completion based on surrounding context. The instruction-following models are fine-tuned with specific prompt formatting for better conversational and task-oriented performance.
Quick Start & Requirements
pip install -e .
within a conda environment with PyTorch/CUDA.wget
and md5sum
for downloading weights. PyTorch with CUDA support is essential.Highlighted Details
Maintenance & Community
Developed by Meta AI. Issues can be reported on the GitHub repository. Feedback on risky content is available via a dedicated Facebook link.
Licensing & Compatibility
Models and weights are licensed for both research and commercial use, with an accompanying Acceptable Use Policy.
Limitations & Caveats
The models are trained on sequences of 16k tokens, with improvements noted on inputs up to 100k tokens. Output generated by Code Llama may be subject to third-party licenses. The README notes that Code Llama is a new technology with potential risks, and testing cannot cover all scenarios.
11 months ago
Inactive