Discover and explore top open-source AI tools and projects—updated daily.
teamchongVector compression and fast search for web and edge
New!
Top 87.1% on SourcePulse
This repository provides an experimental WebAssembly (WASM) build of TurboQuant, enabling efficient vector compression and fast dot product searches directly on compressed data within browsers and Node.js environments. It targets developers needing to reduce the memory footprint of embedding vectors for applications like real-time search, image similarity, or LLM KV cache compression, offering a significant 6x size reduction without requiring a training step.
How It Works
The project leverages the TurboQuant algorithm, employing polar decomposition and QJL rotation, implemented in Zig and compiled to WASM with relaxed SIMD instructions for CPU execution. It offers a dual-substrate approach: the core turboquant-wasm npm package utilizes WASM for general vector operations, while an optional WebGPU compute shader path accelerates dotBatch operations by processing compressed vectors directly on the GPU. This design provides a single, dependency-light package with transparent fallback mechanisms.
Quick Start & Requirements
npm install turboquant-wasmHighlighted Details
dotBatch method transparently utilizes WebGPU compute shaders when available, falling back to WASM SIMD.Maintenance & Community
No specific details regarding maintainers, community channels (e.g., Discord, Slack), or project roadmap were found in the provided README text.
Licensing & Compatibility
Limitations & Caveats
This is an experimental build, and the compression ratio (~4.5 bpd) is less aggressive than methods like PQ/OPQ (1-2 bpd). Query speed for dot operations is slower than PQ/OPQ due to per-pair decompression, though dotBatch offers acceleration. Support relies on modern browser/Node.js runtimes with relaxed SIMD capabilities.
2 weeks ago
Inactive
tensorchord
timescale
marqo-ai
pgvector
facebookresearch