Secure code sandbox for LLM-generated code execution and evaluation
Top 62.0% on sourcepulse
SandboxFusion provides a secure, containerized environment for executing and evaluating code generated by Large Language Models (LLMs). It is designed for researchers and developers working with LLM-based code generation, offering support for numerous programming languages and popular code evaluation benchmarks.
How It Works
The system utilizes Docker containers to isolate code execution, ensuring security and reproducibility. It supports a wide array of languages including Python, C++, Java, Go, Node.js, and even CUDA for GPU acceleration. SandboxFusion also integrates with various code evaluation datasets like HumanEval, MultiPL-E, and MBPP, facilitating robust benchmarking of LLM-generated code.
Quick Start & Requirements
conda
and poetry
.conda
, poetry
. For Docker, a base image is provided, with instructions to customize the server image.conda
and poetry
.Highlighted Details
Maintenance & Community
The project lists several contributors from Bytedance. Further community engagement details (e.g., Discord, Slack) are not specified in the README.
Licensing & Compatibility
Licensed under the Apache License, Version 2.0. This license is permissive and generally compatible with commercial use and closed-source linking.
Limitations & Caveats
The README does not detail specific limitations, unsupported platforms, or known issues. The setup for specific language runtimes requires manual execution of provided shell scripts.
1 month ago
1+ week