Results for "training loss"

54 results

Dropout Intermediate

Randomly zeroing activations during training to reduce co-adaptation and overfitting.

Foundations & Theory
Next-Token Prediction Intermediate

Training objective where the model predicts the next token given previous tokens (causal modeling).

Foundations & Theory
DPO Intermediate

A preference-based training method optimizing policies directly from pairwise comparisons without explicit RL loops.

Optimization
Curriculum Learning Intermediate

Ordering training samples from easier to harder to improve convergence or generalization.

Foundations & Theory
Data Augmentation Intermediate

Expanding training data via transformations (flips, noise, paraphrases) to improve robustness.

Foundations & Theory
Federated Learning Intermediate

Training across many devices/silos without centralizing raw data; aggregates updates, not data.

Foundations & Theory
Reproducibility Intermediate

Ability to replicate results given same code/data; harder in distributed training and nondeterministic ops.

Foundations & Theory
Compute Intermediate

Hardware resources used for training/inference; constrained by memory bandwidth, FLOPs, and parallelism.

Foundations & Theory
Distillation Intermediate

Training a smaller “student” model to mimic a larger “teacher,” often improving efficiency while retaining performance.

Foundations & Theory
Data Poisoning Intermediate

Maliciously inserting or altering training data to implant backdoors or degrade performance.

Foundations & Theory
Automation Bias Intermediate

Tendency to trust automated suggestions even when incorrect; mitigated by UI design, training, and checks.

Foundations & Theory
Variance Term Intermediate

Error due to sensitivity to fluctuations in the training dataset.

AI Economics & Strategy
Learning Rate Schedule Intermediate

Adjusting learning rate over training to improve convergence.

AI Economics & Strategy
Warmup Intermediate

Gradually increasing learning rate at training start to avoid divergence.

AI Economics & Strategy
Causal Mask Intermediate

Prevents attention to future tokens during training/inference.

AI Economics & Strategy
Gradient Leakage Intermediate

Recovering training data from gradients.

AI Economics & Strategy
Model Inversion Intermediate

Inferring sensitive features of training data.

AI Economics & Strategy
Generative Model Advanced

Models that learn to generate samples resembling training data.

Diffusion & Generative Models
Saddle Plateau Intermediate

Flat high-dimensional regions slowing training.

Foundations & Theory
Deceptive Alignment Advanced

Model behaves well during training but not deployment.

AI Safety & Alignment
Exposure Bias Intermediate

Differences between training and inference conditions.

Model Failure Modes
Simulation Advanced

Artificial environment for training/testing agents.

Simulation & Sim-to-Real
Dataset Shift Intermediate

Differences between training and deployed patient populations.

AI in Healthcare
Hybrid Training Advanced

Combining simulation and real-world data.

Simulation & Sim-to-Real