Results for "representation learning"
Representation Learning
IntermediateAutomatically learning useful internal features (latent variables) that capture salient structure for downstream tasks.
Representation learning is like teaching a computer to understand the essence of data without needing someone to explain every detail. Imagine trying to recognize different animals in pictures. Instead of manually pointing out features like fur color or size, a representation learning model can a...
Performance drop when moving from simulation to reality.
Directly optimizing control policies.
Reward only given upon task completion.
Control shared between human and agent.
Inferring human goals from behavior.
Automated assistance identifying disease indicators.
AI-assisted review of legal documents.
Predicting protein 3D structure from sequence.
AI selecting next experiments.
AI tacitly coordinating prices.
Rate at which AI capabilities improve.
Research ensuring AI remains safe.
A conceptual framework describing error as the sum of systematic error (bias) and sensitivity to data (variance).
Updating a pretrained model’s weights on task-specific data to improve performance or adapt style/behavior.
Measures a model’s ability to fit random noise; used to bound generalization error.
The learned numeric values of a model adjusted during training to minimize a loss function.
A scalar measure optimized during training, typically expected loss over data, sometimes with regularization terms.
Minimizing average loss on training data; can overfit when data is limited or biased.
When a model fits noise/idiosyncrasies of training data and performs poorly on unseen data.
How well a model performs on new data drawn from the same (or similar) distribution as training.
Separating data into training (fit), validation (tune), and test (final estimate) to avoid leakage and optimism bias.
A robust evaluation technique that trains/evaluates across multiple splits to estimate performance variability.
Scalar summary of ROC; measures ranking ability, not calibration.
A proper scoring rule measuring squared error of predicted probabilities for binary outcomes.
Penalizes confident wrong predictions heavily; standard for classification and language modeling.
Average of squared residuals; common regression objective.
Uses an exponential moving average of gradients to speed convergence and reduce oscillation.
One complete traversal of the training dataset during training.
Halting training when validation performance stops improving to reduce overfitting.
A parameterized function composed of interconnected units organized in layers with nonlinear activations.