Jump to content

User:Chakazul/AI

From Wikipedia, the free encyclopedia

Neural Network Notations

[edit]

Dimensions

[edit]
Dimension Variable
# Samples
# Layers (exclude input)
# Units in Input Layer
# Units in Hidden Layer
# Units in Output Layer / # Classes

Constants

[edit]
Constant
Learning Rate
Regularization Factor

Matrices

[edit]
Notation Equation Dimensions Layers
Input (given) (global)
Output (given) (global)
Feedforward
Weight (given / calculated)
Bias (given / calculated)
Input
Weighted Input
Activation
Predicted Output
Backpropagation
Loss Function
(CE or MSE)
Cost Function (scalar) (global)
Optimization
Output Error
Hidden Error
Weight Update
(Gradient Descent)
Bias Update
(Gradient Descent)

Details

[edit]

Functions and Partial Derivatives

[edit]

Chain Rule

[edit]

Weight / Bias Update (Gradient Descent)

[edit]

Examples

[edit]

Remarks

[edit]
  • is the matrix of the previous layer, is that of the next layer, otherwise implicitly refer to the current layer
  • is the activation function (e.g. sigmoid, tanh, ReLU)
  • is the element-wise product
  • is the element-wise power
  • is the matrix's sum of elements
  • is the matrix derivative
  • Variations:
    1. All matrices transposed, matrix multiplcations in reverse order (row vectors instead of column vectors)
    2. combined into one parameter matrix
    3. No term in

References

[edit]