Back to glossary
AI GLOSSARY
Model Extraction
Security & Adversarial AI
An attack where an adversary queries a model repeatedly and uses the outputs to train a replica that approximates the original's behavior, effectively stealing the model's capabilities without access to its weights or training data. Model extraction threatens the intellectual property of AI developers and can also be a precursor to more targeted attacks against the replicated model.
See also: model theft.