numpy_ml/neural_nets/README.md
This module implements building-blocks for larger neural network models in the Keras-style. This module does not implement a general autograd system in order emphasize conceptual understanding over flexibility.
Activations. Common activation nonlinearities. Includes:
Losses. Common loss functions. Includes:
Wrappers. Layer wrappers. Includes:
Layers. Common layers / layer-wise operations that can be composed to create larger neural networks. Includes:
Optimizers. Common modifications to stochastic gradient descent. Includes:
Learning Rate Schedulers. Common learning rate decay schedules.
Initializers. Common weight initialization strategies.
Modules. Common multi-layer blocks that appear across many deep networks. Includes:
same-convolution) residual blocks (He et al., 2015)Models. Well-known network architectures. Includes:
vae.py: Bernoulli variational autoencoder (Kingma & Welling, 2014)wgan_gp.py: Wasserstein generative adversarial network with gradient
penalty (Gulrajani et al., 2017;
Goodfellow et al., 2014)w2v.py: word2vec model with CBOW and skip-gram architectures and
training via noise contrastive estimation (Mikolov et al., 2012)Utils. Common helper functions, primarily for dealing with CNNs. Includes:
im2colcol2imconv1Dconv2Ddilatedeconv2Dminibatch