Results for "cross-entropy"
Cross-Entropy
IntermediateMeasures divergence between true and predicted probability distributions.
Cross-entropy measures how well a model's predictions match the actual outcomes. Imagine you have a bag of colored balls, and you want to predict how many of each color are in the bag. If your predictions are far off from the actual counts, the cross-entropy will be high, indicating a poor match....
Measures divergence between true and predicted probability distributions.
A measure of randomness or uncertainty in a probability distribution.
Reduction in uncertainty achieved by observing a variable; used in decision trees and active learning.
A robust evaluation technique that trains/evaluates across multiple splits to estimate performance variability.
Attention between different modalities.
Quantifies shared information between random variables.
A scalar measure optimized during training, typically expected loss over data, sometimes with regularization terms.
A function measuring prediction error (and sometimes calibration), guiding gradient-based optimization.
Penalizes confident wrong predictions heavily; standard for classification and language modeling.
Training objective where the model predicts the next token given previous tokens (causal modeling).
Predicts masked tokens in a sequence, enabling bidirectional context; often used for embeddings rather than generation.
Model that compresses input into latent space and reconstructs it.
Assigning labels per pixel (semantic) or per instance (instance segmentation) to map object boundaries.
Assigning category labels to images.
Pixel-wise classification of image regions.
Learning policies from expert demonstrations.
Learning action mapping directly from demonstrations.
Minimizing average loss on training data; can overfit when data is limited or biased.
When a model fits noise/idiosyncrasies of training data and performs poorly on unseen data.
How well a model performs on new data drawn from the same (or similar) distribution as training.
Halting training when validation performance stops improving to reduce overfitting.
When information from evaluation data improperly influences training, inflating reported performance.
Models that process or generate multiple modalities, enabling vision-language tasks, speech, video understanding, etc.
End-to-end process for model training.
Centralized AI expertise group.
A conceptual framework describing error as the sum of systematic error (bias) and sensitivity to data (variance).