Results for "training mismatch"

AdvertisementAd space — search-top

156 results

Compute Intermediate

Hardware resources used for training/inference; constrained by memory bandwidth, FLOPs, and parallelism.

Foundations & Theory
Data Poisoning Intermediate

Maliciously inserting or altering training data to implant backdoors or degrade performance.

Foundations & Theory
Privacy Attack Intermediate

Attacks that infer whether specific records were in training data, or reconstruct sensitive training examples.

Foundations & Theory
Variance Term Intermediate

Error due to sensitivity to fluctuations in the training dataset.

AI Economics & Strategy
Learning Rate Schedule Intermediate

Adjusting learning rate over training to improve convergence.

AI Economics & Strategy
Scaling Laws Intermediate

Empirical laws linking model size, data, compute to performance.

AI Economics & Strategy
Generative Model Advanced

Models that learn to generate samples resembling training data.

Diffusion & Generative Models
Data Scaling Intermediate

Increasing performance via more data.

AI Economics & Strategy
Chinchilla Scaling Intermediate

Scaling law optimizing compute vs data.

AI Economics & Strategy
Domain Randomization Advanced

Randomizing simulation parameters to improve real-world transfer.

Simulation & Sim-to-Real
Adaptive Optimization Intermediate

Methods like Adam adjusting learning rates dynamically.

Foundations & Theory
AI Hallucination Intermediate

Fabrication of cases or statutes by LLMs.

AI in Law
Meta-Learning Intermediate

Methods that learn training procedures or initializations so models can adapt quickly to new tasks with little data.

Machine Learning
Parameters Intermediate

The learned numeric values of a model adjusted during training to minimize a loss function.

Foundations & Theory
Regularization Intermediate

Techniques that discourage overly complex solutions to improve generalization (reduce overfitting).

Foundations & Theory
Overfitting Intermediate

When a model fits noise/idiosyncrasies of training data and performs poorly on unseen data.

Foundations & Theory
Generalization Intermediate

How well a model performs on new data drawn from the same (or similar) distribution as training.

Foundations & Theory
Cross-Validation Intermediate

A robust evaluation technique that trains/evaluates across multiple splits to estimate performance variability.

Foundations & Theory
Stochastic Gradient Descent Intermediate

A gradient method using random minibatches for efficient training on large datasets.

Foundations & Theory
Learning Rate Intermediate

Controls the size of parameter updates; too high diverges, too low trains slowly or gets stuck.

Foundations & Theory
Vanishing Gradient Intermediate

Gradients shrink through layers, slowing learning in early layers; mitigated by ReLU, residuals, normalization.

Foundations & Theory
Normalization Intermediate

Techniques that stabilize and speed training by normalizing activations; LayerNorm is common in Transformers.

Foundations & Theory
Large Language Model Intermediate

A high-capacity language model trained on massive corpora, exhibiting broad generalization and emergent behaviors.

Large Language Models
Few-Shot Learning Intermediate

Achieving task performance by providing a small number of examples inside the prompt without weight updates.

Foundations & Theory
SFT Intermediate

Fine-tuning on (prompt, response) pairs to align a model with instruction-following behaviors.

Foundations & Theory
Reward Model Intermediate

Model trained to predict human preferences (or utility) for candidate outputs; used in RLHF-style pipelines.

Foundations & Theory
Federated Learning Intermediate

Training across many devices/silos without centralizing raw data; aggregates updates, not data.

Foundations & Theory
Reproducibility Intermediate

Ability to replicate results given same code/data; harder in distributed training and nondeterministic ops.

Foundations & Theory
LoRA Intermediate

PEFT method injecting trainable low-rank matrices into layers, enabling efficient fine-tuning.

Foundations & Theory
Quantization Intermediate

Reducing numeric precision of weights/activations to speed inference and reduce memory with acceptable accuracy loss.

Foundations & Theory

Welcome to AI Glossary

The free, self-building AI dictionary. Help us keep it free—click an ad once in a while!

Search

Type any question or keyword into the search bar at the top.

Browse

Tap a letter in the A–Z bar to browse terms alphabetically, or filter by domain, industry, or difficulty level.

3D WordGraph

Fly around the interactive 3D graph to explore how AI concepts connect. Click any word to read its full definition.