Mark As Completed Discussion

The Math You Really Need

Here's the mathematical terms at play:

  • Weights (W) and biases (b): the parameters we learn.
  • Activation function φ: adds non-linearity (e.g., ReLU(x) = max(0,x)).
  • Loss: scalar measuring error, e.g., MSE for regression, cross-entropy for classification.
  • Gradient: vector of partial derivatives that tells us how to tweak parameters to reduce loss.
  • Gradient descent: update rule θ ← θ − η ∇θ L with learning rate η.