Difficulty: Intermediate
Running models locally.
A continuous vector encoding of an item (word, image, user) such that semantic similarity corresponds to geometric closeness.
Capabilities that appear only beyond certain model sizes.
Coordination arising without explicit programming.
Minimizing average loss on training data; can overfit when data is limited or biased.
Protecting data during network transfer and while stored; essential for ML pipelines handling sensitive data.
Models that define an energy landscape rather than explicit probabilities.
A measure of randomness or uncertainty in a probability distribution.
One complete traversal of the training dataset during training.
European regulation classifying AI systems by risk.
System for running consistent evaluations across tasks, versions, prompts, and model settings.
Logging hyperparameters, code versions, data snapshots, and results to reproduce and compare experiments.
Techniques to understand model decisions (global or local), important in high-stakes and regulated settings.
Requirement to provide explanations.
Legal or policy requirement to explain AI decisions.
Credit models with interpretable logic.
Gradients grow too large, causing divergence; mitigated by clipping, normalization, careful init.
Balancing learning new behaviors vs exploiting known rewards.
Differences between training and inference conditions.
The range of functions a model can represent.
Harmonic mean of precision and recall; useful when balancing false positives/negatives matters.
Graphical model expressing factorization of a probability distribution.
Ensuring models comply with lending fairness laws.
Failure to detect present disease.
US approval process for medical AI devices.
A measurable property or attribute used as model input (raw or engineered), such as age, pixel intensity, or token ID.
Designing input features to expose useful structure (e.g., ratios, lags, aggregations), often crucial outside deep learning.
Centralized repository for curated features.
Training across many devices/silos without centralizing raw data; aggregates updates, not data.
Using output to adjust future inputs.