Python library for Intel NPU acceleration (now end-of-life)
Top 50.8% on sourcepulse
This Python library aimed to accelerate AI computations on Intel Neural Processing Units (NPUs), targeting developers working with Intel Core Ultra processors. It provided low-level access to NPU hardware for high-speed matrix operations and model inference, with the goal of boosting application efficiency.
How It Works
The library leverages Intel's NPU architecture, which includes dedicated Neural Compute Engines for AI operations like matrix multiplication and convolution, and Streaming Hybrid Architecture Vector Engines for general computing. It utilizes compiler technology to optimize AI workloads by tiling compute and data flow, maximizing utilization of on-chip SRAM and minimizing DRAM transfers for performance and power efficiency.
Quick Start & Requirements
pip install intel-npu-acceleration-library
Highlighted Details
torch.compile
for NPU optimization (Windows torch.compile
not supported; use explicit intel_npu_acceleration_library.compile
).Maintenance & Community
This project is no longer under active management by Intel and has been archived. Intel has ceased development, maintenance, bug fixes, and contributions. The project is available for reference, and users are encouraged to fork it for independent development. Intel recommends adopting OpenVINO™ and OpenVINO™ GenAI for NPU acceleration.
Licensing & Compatibility
The license is not explicitly stated in the provided README text. Compatibility for commercial use or closed-source linking would require clarification of the licensing terms.
Limitations & Caveats
The project is officially announced as End-of-Life and will not receive further updates or maintenance from Intel. macOS is not supported. Windows torch.compile
is not supported. Users are directed to OpenVINO™ and OpenVINO™ GenAI for current NPU acceleration solutions.
3 months ago
1 day