Results for "representation learning"
Representation Learning
IntermediateAutomatically learning useful internal features (latent variables) that capture salient structure for downstream tasks.
Representation learning is like teaching a computer to understand the essence of data without needing someone to explain every detail. Imagine trying to recognize different animals in pictures. Instead of manually pointing out features like fur color or size, a representation learning model can a...
A theoretical framework analyzing what classes of functions can be learned, how efficiently, and with what guarantees.
Training with a small labeled dataset plus a larger unlabeled dataset, leveraging assumptions like smoothness/cluster structure.
Achieving task performance by providing a small number of examples inside the prompt without weight updates.
Training across many devices/silos without centralizing raw data; aggregates updates, not data.
Gradually increasing learning rate at training start to avoid divergence.
Built-in assumptions guiding learning efficiency and generalization.
Modifying reward to accelerate learning.
Training one model on multiple tasks simultaneously to improve generalization through shared structure.
Iterative method that updates parameters in the direction of negative gradient to minimize loss.
Reinforcement learning from human feedback: uses preference data to train a reward model and optimize the policy.
Methods like Adam adjusting learning rates dynamically.
Inferring reward function from observed behavior.
Learning by minimizing prediction error.
Humans assist or override autonomous behavior.
Robots learning via exploration and growth.
A subfield of AI where models learn patterns from data to make predictions or decisions, improving with experience rather than explicit rule-coding.
Reusing knowledge from a source task/domain to improve learning on a target task/domain, typically via pretrained models.
A structured collection of examples used to train/evaluate models; quality, bias, and coverage often dominate outcomes.
Designing input features to expose useful structure (e.g., ratios, lags, aggregations), often crucial outside deep learning.
Human or automated process of assigning targets; quality, consistency, and guidelines matter heavily.
System design where humans validate or guide model outputs, especially for high-stakes decisions.
Strategy mapping states to actions.
Systematic error introduced by simplifying assumptions in a learning algorithm.
Combines value estimation (critic) with policy learning (actor).
Balancing learning new behaviors vs exploiting known rewards.
Continuous cycle of observation, reasoning, action, and feedback.
Loss of old knowledge when learning new tasks.
Combining simulation and real-world data.
RL without explicit dynamics model.
Learning without catastrophic forgetting.