Results for "belief divergence"
Posterior Distribution
Advanced
Updated belief after observing data.
Prior Distribution
Advanced
Belief before observing data.
Exploding Gradient
Intermediate
Gradients grow too large, causing divergence; mitigated by clipping, normalization, careful init.
Cross-Entropy
Intermediate
Measures divergence between true and predicted probability distributions.
Warmup
Intermediate
Gradually increasing learning rate at training start to avoid divergence.
KL Divergence
Intermediate
Measures how one probability distribution diverges from another.