Skip to content

Architectures

Purpose Meaning Major
Application
Computation
Complexity
Limitation Advantage
Linear Combinations FC
Fully-Connected
Poor scalability for large input sizes
Do not capture “intuitive” invariances
Spatial pattens CNN
(Convolutional)
- Require that activations between layers occur only in “local” manner
- Treat hidden layers themselves as spatial images
- Share weights across all spatial locations
Images, Videos High Reduce parameter count
Capture [some] “natural” invariances
Temporal
(Sequences)
RNN
(Recurrent)
Forward-feed, backward-feed, and self-loop is allowed Time Series
GRU Time Series
LSTM Time Series
Transformer Text Generation
IDK ResNet
(Residual Network)
Add operator Time Series
DenseNet Concat operator
U-Net Basis of diffusion models
Segmentation
Super-Resolution
Diffusion Models
PINN
(Physics-Informed)
Lagrangian
Deep Operator
Fourier Neural Operator
Graph Neural Networks

IDK

image-20240309214029828

Note

skip connections of inputs

  • For structured problems
    • Like DenseNet architecture
    • The raw inputs may possess more meaningful information than the linear/non-linear combination
  • For unstructured problems
    • ResNet is better
Last Updated: 2024-12-26 ; Contributors: AhmedThahir, web-flow

Comments