Results for "data preservation"
Systematic review of model/data processes to ensure performance, fairness, security, and policy compliance.
Attacks that infer whether specific records were in training data, or reconstruct sensitive training examples.
Methods to protect model/data during inference (e.g., trusted execution environments) from operators/attackers.
A narrow minimum often associated with poorer generalization.
Systematic error introduced by simplifying assumptions in a learning algorithm.
Built-in assumptions guiding learning efficiency and generalization.
Estimating parameters by maximizing likelihood of observed data.
Detecting unauthorized model outputs or data leaks.
Neural networks that operate on graph-structured data by propagating information along edges.
Models that define an energy landscape rather than explicit probabilities.
Learns the score (∇ log p(x)) for generative sampling.
Exact likelihood generative models using invertible transforms.
Two-network setup where generator fools a discriminator.
CNNs applied to time series.
Attention between different modalities.
End-to-end process for model training.
Running predictions on large datasets periodically.
Centralized repository for curated features.
Scaling law optimizing compute vs data.
Competitive advantage from proprietary models/data.
Belief before observing data.
Train/test environment mismatch.
Model trained on its own outputs degrades quality.
Startup latency for services.
Storing results to reduce compute.
Software pipeline converting raw sensor data into structured representations.
Learning physical parameters from data.
Models estimating recidivism risk.
Finding mathematical equations from data.
Learning only from current policy’s data.