Omega-AI  by dromara

Java DL framework for model training/inference, supporting multi-GPU

Created 6 years ago
491 stars

Top 62.9% on SourcePulse

GitHubView on GitHub
Project Summary

Omega-AI is a deep learning framework built in Java, designed to simplify the creation, training, and inference of neural networks for Java developers. It supports automatic differentiation, multi-threading, and GPU acceleration via CUDA and cuDNN, enabling rapid development of AI models.

How It Works

Omega-AI provides a comprehensive set of layers, optimizers, and loss functions, allowing users to build various neural network architectures. Its core advantage lies in its Java-native implementation, aiming to lower the barrier to entry for AI development within the Java ecosystem. The framework emphasizes performance through GPU acceleration and optimized CUDA kernels for operations like matrix multiplication and convolution.

Quick Start & Requirements

Highlighted Details

  • Supports a wide range of models including BP, CNN, RNN, VGG16, ResNet, YOLO, Transformer, GPT, LLaMA, and Diffusion models.
  • Features automatic differentiation and multi-threaded CPU/GPU computation.
  • Includes demos for various tasks like image classification (MNIST, CIFAR-10), object detection (YOLOv1, v3, v7), text generation (RNN, GPT), and image generation (GAN, Diffusion).
  • Provides implementations for common layers (Convolution, Pooling, Fully Connected, RNN, LSTM, Transformer blocks) and activation/normalization functions (ReLU, Leaky ReLU, BN, LN).

Maintenance & Community

  • Active development with regular updates and new model implementations.
  • Community discussion via QQ group: 119593195.
  • Contact: 465973119@qq.com.

Licensing & Compatibility

  • The specific license is not explicitly stated in the README, but the project is hosted on Gitee and GitHub, suggesting a permissive open-source license. Compatibility for commercial use would require verification of the license.

Limitations & Caveats

  • The project is primarily Java-focused, which may limit its adoption by users accustomed to Python-based deep learning ecosystems.
  • GPU support relies on JCUDa and specific CUDA/cuDNN versions, requiring careful environment setup.
  • Some demos and model implementations might be in early stages or require specific dataset formats.
Health Check
Last Commit

2 days ago

Responsiveness

1 week

Pull Requests (30d)
0
Issues (30d)
0
Star History
6 stars in the last 30 days

Explore Similar Projects

Starred by Peter Norvig Peter Norvig(Author of "Artificial Intelligence: A Modern Approach"; Research Director at Google), Alexey Milovidov Alexey Milovidov(Cofounder of Clickhouse), and
29 more.

llm.c by karpathy

0.2%
28k
LLM training in pure C/CUDA, no PyTorch needed
Created 1 year ago
Updated 2 months ago
Feedback? Help us improve.