User Tools

Site Tools


ml:nn_architectures

Neural Network Architectures

Overviews

Feedforward Networks

Connections

  • ReZero Similar to residual connections, but with a trainable parameter that controls the strength of the nonlinearity (which is initialized to zero).

Sequence Networks

See also State-Space Models.

Tree Networks

Graph Networks

See also Wu et al 2019 - A Comprehensive Survey on Graph Neural Networks and Graph Neural Networks.

  • Graph convolution networks
  • Graph transformers

Activation Functions

See also the table in Wikipedia's Activation functions.

Comparisons:

Matrices

Various representations of matrices, such as sparse, or low-dimensional ones.

Set and Pooling Networks

Memory Architectures

RNN Cells

See also Wikipedia - Recurrent Neural Networks and Yu et al 2019 - A Review of Recurrent Neural Networks: LSTM Cells and Network Architectures

  • Feedforward network (Elman network)
  • Feedforward network with residual connections (with careful tuning, has been shown to perform as well as LSTMs I believe)
  • LSTM
    • Forget gate
    • Peephole connections
  • GRU (has been shown not to perform as well as the LSTM cell, for example here)
  • Minimal Gated Unit (MGU)

Position Embeddings

Attention Mechanisms

See also the Attention Mechanisms page.

Neurosymbolic Networks

Dynamic Neural Networks

Miscellaneous Architectures

ml/nn_architectures.txt · Last modified: 2025/03/25 07:34 by jmflanig

Donate Powered by PHP Valid HTML5 Valid CSS Driven by DokuWiki