Results for "privacy attack"
Privacy Attack
IntermediateAttacks that infer whether specific records were in training data, or reconstruct sensitive training examples.
Imagine if someone could tell whether your personal information was used to train a smart assistant just by asking it questions. That’s what a privacy attack does—it tries to find out if specific data was part of the training set. This can be really concerning because it means that private detail...
A formal privacy framework ensuring outputs do not reveal much about any single individual’s data contribution.
Attacks that infer whether specific records were in training data, or reconstruct sensitive training examples.
Compromising AI systems via libraries, models, or datasets.
Recovering training data from gradients.
Privacy risk analysis under GDPR-like laws.
Inferring sensitive features of training data.
Attacks that manipulate model instructions (especially via retrieved content) to override system goals or exfiltrate data.
Artificially created data used to train/test models; helpful for privacy and coverage, risky if unrealistic.
Information that can identify an individual (directly or indirectly); requires careful handling and compliance.
Training across many devices/silos without centralizing raw data; aggregates updates, not data.
A discipline ensuring AI systems are fair, safe, transparent, privacy-preserving, and accountable throughout lifecycle.
Maliciously inserting or altering training data to implant backdoors or degrade performance.
Stepwise reasoning patterns that can improve multi-step tasks; often handled implicitly or summarized for safety/privacy.
US framework for AI risk governance.
Running models locally.
AI predicting crime patterns (highly controversial).
International agreements on AI.