Results for "gradient of density"

15 results

Gradient Descent Intermediate

Iterative method that updates parameters in the direction of negative gradient to minimize loss.

Optimization
Stochastic Gradient Descent Intermediate

A gradient method using random minibatches for efficient training on large datasets.

Foundations & Theory
Gradient Clipping Intermediate

Limiting gradient magnitude to prevent exploding gradients.

AI Economics & Strategy
Policy Gradient Intermediate

Optimizing policies directly via gradient ascent on expected reward.

AI Economics & Strategy
Loss Function Intermediate

A function measuring prediction error (and sometimes calibration), guiding gradient-based optimization.

Foundations & Theory
Batch Size Intermediate

Number of samples per gradient update; impacts compute efficiency, generalization, and stability.

Foundations & Theory
ReLU Intermediate

Activation max(0, x); improves gradient flow and training speed in deep nets.

Foundations & Theory
Weight Initialization Intermediate

Methods to set starting weights to preserve signal/gradient scales across layers.

Foundations & Theory
Saddle Point Intermediate

A point where gradient is zero but is neither a max nor min; common in deep nets.

AI Economics & Strategy
Line Search Intermediate

Choosing step size along gradient direction.

Foundations & Theory
Vanishing Gradient Intermediate

Gradients shrink through layers, slowing learning in early layers; mitigated by ReLU, residuals, normalization.

Foundations & Theory
Exploding Gradient Intermediate

Gradients grow too large, causing divergence; mitigated by clipping, normalization, careful init.

Foundations & Theory
Gradient Noise Intermediate

Variability introduced by minibatch sampling during SGD.

AI Economics & Strategy
Gradient Leakage Intermediate

Recovering training data from gradients.

AI Economics & Strategy
Gradient Advanced

Direction of steepest ascent of a function.

Mathematics