Results for "data → model"
A formal privacy framework ensuring outputs do not reveal much about any single individual’s data contribution.
Separating data into training (fit), validation (tune), and test (final estimate) to avoid leakage and optimism bias.
Shift in model outputs.
Models that learn to generate samples resembling training data.
A parameterized mapping from inputs to outputs; includes architecture + learned parameters.
When a model cannot capture underlying structure, performing poorly on both training and test data.
Competitive advantage from proprietary models/data.
Model that compresses input into latent space and reconstructs it.
A branch of ML using multi-layer neural networks to learn hierarchical representations, often excelling in vision, speech, and language.
Running models locally.
Probability of data given parameters.
Enables external computation or lookup.
The learned numeric values of a model adjusted during training to minimize a loss function.
Learns the score (∇ log p(x)) for generative sampling.
End-to-end process for model training.
Model trained on its own outputs degrades quality.
A subfield of AI where models learn patterns from data to make predictions or decisions, improving with experience rather than explicit rule-coding.
Minimizing average loss on training data; can overfit when data is limited or biased.
Automatically learning useful internal features (latent variables) that capture salient structure for downstream tasks.
Forcing predictable formats for downstream systems; reduces parsing errors and supports validation/guardrails.
Learning from data generated by a different policy.
Updated belief after observing data.
Requirement to preserve relevant data.
Trend reversal when data is aggregated improperly.
Central system to store model versions, metadata, approvals, and deployment state.
RL using learned or known environment models.
Built-in assumptions guiding learning efficiency and generalization.
Train/test environment mismatch.
Learning physical parameters from data.
How well a model performs on new data drawn from the same (or similar) distribution as training.