Discover and explore top open-source AI tools and projects—updated daily.
amruth-snAgentic reverse engineer for binaries
Top 52.9% on SourcePulse
The Kong project addresses the significant challenge of reverse engineering stripped binaries by automating the recovery of crucial context such as function names, type information, and symbols. It is designed for reverse engineers and security researchers, offering a substantial benefit by accelerating the analysis of obfuscated code through advanced LLM orchestration and a novel agentic deobfuscation pipeline.
How It Works
Kong employs a sophisticated five-phase pipeline orchestrated by a supervisor: triage, analysis, cleanup, synthesis, and export. It constructs rich context windows from Ghidra's program database, incorporating decompilation, cross-references, and data flow, before querying Large Language Models (LLMs). Functions are analyzed in a bottom-up order based on the call graph, ensuring that callers benefit from the already-resolved context of their callees. A unique agentic deobfuscation pipeline is integrated to identify and remove various obfuscation techniques. The synthesis phase then unifies naming conventions across the binary and synthesizes struct definitions, with all recovered information exported back into Ghidra's program database.
Quick Start & Requirements
uv pip install kong-re. Alternatively, clone from source and use uv sync.kong setup wizard for initial configuration.kong analyze ./path/to/stripped_binary.Highlighted Details
Maintenance & Community
The project is actively maintained by amruth-sn. Community engagement is encouraged through GitHub Issues. The author is also reachable via X (formerly Twitter) and LinkedIn for further discussion.
Licensing & Compatibility
Kong is licensed under the Apache License 2.0. This license is compatible with Ghidra's licensing and explicitly permits commercial use.
Limitations & Caveats
Confidence levels for architecture support vary, with lower confidence noted for Rust and Go binaries on ARM, MIPS, and PowerPC architectures. The size of the binary, LLM costs, and time to completion scale positively with the number of functions, while analysis confidence scales negatively.
1 week ago
Inactive
LaurieWired