bdh  by pathwaycom

Biologically inspired LLM architecture bridging neuroscience and deep learning

Created 5 months ago
3,377 stars

Top 14.1% on SourcePulse

GitHubView on GitHub
1 Expert Loves This Project
Project Summary

Baby Dragon Hatchling (BDH) presents a novel, biologically inspired large language model architecture that bridges deep learning principles with neuroscience foundations. Developed by Pathway, it offers a theoretical and practical framework for understanding emergent reasoning and generalization in AI systems. BDH targets researchers and engineers seeking interpretable, brain-like AI models that match Transformer performance without sacrificing transparency.

How It Works

BDH employs a scale-free, locally interacting network of neurons, mimicking biological connectivity and dynamics. Its core approach utilizes excitatory/inhibitory neuron particles and Hebbian working memory based on synaptic plasticity, promoting monosemanticity. A GPU-friendly state-space formulation enables efficient implementation, yielding sparse, positive, and interpretable activations. This architecture formalizes a bridge between neural computation and machine language understanding, demonstrating how macro-level reasoning emerges from micro-level neuron dynamics guided by graph theory and local computation.

Quick Start & Requirements

  • Installation: pip install -r requirements.txt
  • Training: python train.py
  • Prerequisites: Python environment (via requirements.txt). GPU acceleration is implied by the architecture's design.
  • Resources: Links to discussions are available on Hugging Face Papers, Alphaxiv, and EmergentMind. A SuperDataScience podcast and media coverage (Forbes, Semafor, etc.) are also provided.

Highlighted Details

  • Matches GPT-2–scale Transformers in language and translation tasks at equivalent parameter scales (10M–1B).
  • Features a scale-free network topology mimicking biological connectivity.
  • Employs locally interacting neuron particles with excitatory/inhibitory dynamics.
  • Incorporates Hebbian working memory based on synaptic plasticity, displaying monosemanticity.
  • Utilizes a GPU-friendly state-space formulation for efficient implementation.
  • Generates interpretable activations that are sparse and positive.
  • Attention mechanisms emerge naturally from neuron-level interactions.
  • Follows Transformer-like scaling laws, maintaining parameter efficiency.

Maintenance & Community

Several community forks exist, including adamskrodzki/bdh, mosure/burn_dragon_hatchling, severian42/bdh, Git-Faisal/bdh, and GrahLnn/bdh. The project has garnered media attention from outlets like Forbes and Semafor, and discussions are active on platforms like Hugging Face Papers.

Licensing & Compatibility

License information is not specified in the provided README. This absence requires further investigation for commercial use or integration into closed-source projects.

Limitations & Caveats

The README does not explicitly detail limitations, known bugs, or the project's development stage (e.g., alpha/beta). The reference to an arXiv paper dated 2025 suggests a research-oriented focus rather than a production-ready library.

Health Check
Last Commit

1 week ago

Responsiveness

Inactive

Pull Requests (30d)
2
Issues (30d)
0
Star History
26 stars in the last 30 days

Explore Similar Projects

Feedback? Help us improve.