BigDL  by intel

AI scaling library for Spark/Flink/Ray, from laptop to cloud

created 8 years ago
2,680 stars

Top 18.0% on sourcepulse

GitHubView on GitHub
Project Summary

BigDL is a unified distributed AI framework designed to scale data analytics and AI applications from laptops to clusters. It offers specialized libraries for large language models (LLMs), big data AI pipelines, transparent program acceleration, deep learning on Spark, time series analysis, recommendation systems, and hardware-secured AI.

How It Works

BigDL leverages Apache Spark and Ray for distributed execution, enabling users to scale single-node Python or Scala/Java programs. Its core strength lies in providing high-level APIs that abstract away distributed computing complexities, allowing seamless integration with popular deep learning frameworks like TensorFlow, PyTorch, and Keras. The Nano library further enhances performance by transparently applying CPU optimizations.

Quick Start & Requirements

  • Installation: pip install bigdl (recommended via conda environment).
  • Prerequisites: Python, Spark/Ray (for distributed modes). Specific libraries may have additional requirements.
  • Documentation: BigDL Docs

Highlighted Details

  • LLM Support: Optimized for Intel CPUs and GPUs, though the LLM library is deprecated in favor of ipex-llm.
  • Orca: Scales TensorFlow, PyTorch, and OpenVINO programs on Spark/Ray clusters, supporting distributed data processing and model training.
  • Nano: Transparently accelerates TensorFlow and PyTorch programs on CPUs with optimizations like BF16, INT8 quantization, and JIT compilation, offering up to 10x speedup.
  • PPML: Provides hardware-secured AI execution using Intel SGX/TDX for enhanced data privacy.

Maintenance & Community

  • Development Focus: Future development for LLMs is directed to the ipex-llm project.
  • Community: Support via Mail List, User Group, and GitHub Issues.

Licensing & Compatibility

  • License: Apache 2.0.
  • Compatibility: Generally compatible with commercial and closed-source applications.

Limitations & Caveats

  • The bigdl-llm library is deprecated and users should migrate to ipex-llm. Performance optimizations are primarily targeted at Intel hardware.
Health Check
Last commit

1 month ago

Responsiveness

1 day

Pull Requests (30d)
0
Issues (30d)
0
Star History
10 stars in the last 90 days

Explore Similar Projects

Starred by Stas Bekman Stas Bekman(Author of Machine Learning Engineering Open Book; Research Engineer at Snowflake) and Zhiqiang Xie Zhiqiang Xie(Author of SGLang).

veScale by volcengine

0.1%
839
PyTorch-native framework for LLM training
created 1 year ago
updated 3 weeks ago
Starred by Jeff Hammerbacher Jeff Hammerbacher(Cofounder of Cloudera) and Stas Bekman Stas Bekman(Author of Machine Learning Engineering Open Book; Research Engineer at Snowflake).

InternEvo by InternLM

1.0%
402
Lightweight training framework for model pre-training
created 1 year ago
updated 1 week ago
Starred by Jeff Hammerbacher Jeff Hammerbacher(Cofounder of Cloudera), Stas Bekman Stas Bekman(Author of Machine Learning Engineering Open Book; Research Engineer at Snowflake), and
2 more.

gpustack by gpustack

1.6%
3k
GPU cluster manager for AI model deployment
created 1 year ago
updated 3 days ago
Starred by Patrick von Platen Patrick von Platen(Core Contributor to Hugging Face Transformers and Diffusers), Michael Han Michael Han(Cofounder of Unsloth), and
1 more.

ktransformers by kvcache-ai

0.4%
15k
Framework for LLM inference optimization experimentation
created 1 year ago
updated 3 days ago
Starred by Chip Huyen Chip Huyen(Author of AI Engineering, Designing Machine Learning Systems), Omar Sanseviero Omar Sanseviero(DevRel at Google DeepMind), and
5 more.

TensorRT-LLM by NVIDIA

0.6%
11k
LLM inference optimization SDK for NVIDIA GPUs
created 1 year ago
updated 1 day ago
Starred by George Hotz George Hotz(Author of tinygrad; Founder of the tiny corp, comma.ai), Anton Bukov Anton Bukov(Cofounder of 1inch Network), and
16 more.

tinygrad by tinygrad

0.1%
30k
Minimalist deep learning framework for education and exploration
created 4 years ago
updated 1 day ago
Starred by Andrej Karpathy Andrej Karpathy(Founder of Eureka Labs; Formerly at Tesla, OpenAI; Author of CS 231n), Nat Friedman Nat Friedman(Former CEO of GitHub), and
32 more.

llama.cpp by ggml-org

0.4%
84k
C/C++ library for local LLM inference
created 2 years ago
updated 23 hours ago
Feedback? Help us improve.