Results for "learning like humans"

AdvertisementAd space — search-top

366 results

Stochastic Approximation Intermediate

Optimization under uncertainty.

Foundations & Theory
Feedback Loop Collapse Intermediate

Model trained on its own outputs degrades quality.

Model Failure Modes
Model-Based RL Advanced

RL using learned or known environment models.

Reinforcement Learning
Dynamics Model Advanced

Predicts next state given current state and action.

Reinforcement Learning
Behavior Cloning Advanced

Learning action mapping directly from demonstrations.

Reinforcement Learning
Fraud Detection Intermediate

Identifying suspicious transactions.

AI Economics & Strategy
Scientific ML Advanced

AI applied to scientific problems.

AI in Science
Meta-Cognition Frontier

Awareness and regulation of internal processes.

AGI & General Intelligence
Domain Shift Intermediate

A mismatch between training and deployment data distributions that can degrade model performance.

MLOps & Infrastructure
Concept Drift Intermediate

The relationship between inputs and outputs changes over time, requiring monitoring and model updates.

Foundations & Theory
Model Intermediate

A parameterized mapping from inputs to outputs; includes architecture + learned parameters.

Foundations & Theory
Stochastic Gradient Descent Intermediate

A gradient method using random minibatches for efficient training on large datasets.

Foundations & Theory
Adam Intermediate

Popular optimizer combining momentum and per-parameter adaptive step sizes via first/second moment estimates.

Optimization
Vanishing Gradient Intermediate

Gradients shrink through layers, slowing learning in early layers; mitigated by ReLU, residuals, normalization.

Foundations & Theory
Normalization Intermediate

Techniques that stabilize and speed training by normalizing activations; LayerNorm is common in Transformers.

Foundations & Theory
System Prompt Intermediate

A high-priority instruction layer setting overarching behavior constraints for a chat model.

Reinforcement Learning
DPO Intermediate

A preference-based training method optimizing policies directly from pairwise comparisons without explicit RL loops.

Optimization
Reward Model Intermediate

Model trained to predict human preferences (or utility) for candidate outputs; used in RLHF-style pipelines.

Foundations & Theory
Bias Intermediate

Systematic differences in model outcomes across groups; arises from data, labels, and deployment context.

Foundations & Theory
SHAP Intermediate

Feature attribution method grounded in cooperative game theory for explaining predictions in tabular settings.

Foundations & Theory
Inter-Annotator Agreement Intermediate

Measure of consistency across labelers; low agreement indicates ambiguous tasks or poor guidelines.

Foundations & Theory
Class Imbalance Intermediate

When some classes are rare, requiring reweighting, resampling, or specialized metrics.

Machine Learning
Data Augmentation Intermediate

Expanding training data via transformations (flips, noise, paraphrases) to improve robustness.

Foundations & Theory
Model Governance Intermediate

Policies and practices for approving, monitoring, auditing, and documenting models in production.

Governance & Ethics
Model Card Intermediate

Standardized documentation describing intended use, performance, limitations, data, and ethical considerations.

Foundations & Theory
Datasheet for Datasets Intermediate

Structured dataset documentation covering collection, composition, recommended uses, biases, and maintenance.

Foundations & Theory
Parameter-Efficient Fine-Tuning Intermediate

Techniques that fine-tune small additional components rather than all weights to reduce compute and storage.

Foundations & Theory
Monitoring Intermediate

Observing model inputs/outputs, latency, cost, and quality over time to catch regressions and drift.

MLOps & Infrastructure
LoRA Intermediate

PEFT method injecting trainable low-rank matrices into layers, enabling efficient fine-tuning.

Foundations & Theory
Adversarial Example Intermediate

Inputs crafted to cause model errors or unsafe behavior, often imperceptible in vision or subtle in text.

Foundations & Theory

Welcome to AI Glossary

The free, self-building AI dictionary. Help us keep it free—click an ad once in a while!

Search

Type any question or keyword into the search bar at the top.

Browse

Tap a letter in the A–Z bar to browse terms alphabetically, or filter by domain, industry, or difficulty level.

3D WordGraph

Fly around the interactive 3D graph to explore how AI concepts connect. Click any word to read its full definition.