weightnorm  by openai

Example code for weight normalization research paper

created 8 years ago
364 stars

Top 78.4% on sourcepulse

GitHubView on GitHub
Project Summary

This repository provides example implementations of Weight Normalization, a technique to accelerate deep neural network training. It targets researchers and practitioners working with deep learning frameworks like Theano (via Lasagne), TensorFlow, and Keras. The primary benefit is faster convergence and improved training stability.

How It Works

Weight Normalization reparameterizes network weights by separating the direction and magnitude of weight vectors. This decoupling allows the optimizer to focus on learning the optimal direction, leading to faster convergence compared to standard weight parameterization.

Quick Start & Requirements

  • Lasagne/Theano: Requires Lasagne and Theano.
  • TensorFlow: Requires TensorFlow.
  • Keras: Requires Keras.
  • CIFAR-10: The Lasagne examples were used for CIFAR-10 experiments.

Highlighted Details

  • Implements Weight Normalization as described in the Salimans & Kingma (2016) paper.
  • Includes example code for Lasagne (Theano), TensorFlow, and Keras.
  • Code was used for CIFAR-10 experiments in the original paper.

Maintenance & Community

Status: Archive (code is provided as-is, no updates expected).

Licensing & Compatibility

The repository does not explicitly state a license.

Limitations & Caveats

The code is archived and no longer maintained or updated. Compatibility with current versions of the underlying frameworks (Theano, TensorFlow, Keras) is not guaranteed.

Health Check
Last commit

6 years ago

Responsiveness

Inactive

Pull Requests (30d)
0
Issues (30d)
0
Star History
0 stars in the last 90 days

Explore Similar Projects

Feedback? Help us improve.