Back to glossary
AI GLOSSARY
Data Exfiltration
Security & Adversarial AI
The unauthorized extraction of sensitive data from an AI system or its associated infrastructure, whether training data, model weights, user inputs, or outputs. Data exfiltration can occur through direct system compromise, model inversion attacks, or by exploiting the model itself to leak information it should not reveal, such as personal data from training sets or confidential system prompts.
See also: data poisoning, adversarial attack, Privacy.