User Tools

Site Tools


ml:nn_architectures

This is an old revision of the document!


Neural Network Architectures

Overviews

Feedforward Networks

Connections

  • ReZero Similar to residual connections, but with a trainable parameter that controls the strength of the nonlinearity (which is initialized to zero).

Sequence Networks

See also State-Space Models.

Tree Networks

Graph Networks

See also Wu et al 2019 - A Comprehensive Survey on Graph Neural Networks and Graph Neural Networks.

  • Graph convolution networks
  • Graph transformers

Activation Functions

See also the table in Wikipedia's Activation functions.

Comparisons:

Set and Pooling Networks

Memory Architectures

RNN Cells

See also Wikipedia - Recurrent Neural Networks and Yu et al 2019 - A Review of Recurrent Neural Networks: LSTM Cells and Network Architectures

  • Feedforward network (Elman network)
  • Feedforward network with residual connections (with careful tuning, has been shown to perform as well as LSTMs I believe)
  • LSTM
    • Forget gate
    • Peephole connections
  • GRU (has been shown not to perform as well as the LSTM cell, for example here)
  • Minimal Gated Unit (MGU)

Position Embeddings

Attention Mechanisms

See also the Attention Mechanisms page.

Neurosymbolic Networks

Dynamic Neural Networks

Miscellaneous Architectures

ml/nn_architectures.1714465659.txt.gz · Last modified: 2024/04/30 08:27 by jmflanig

Donate Powered by PHP Valid HTML5 Valid CSS Driven by DokuWiki