{"version":"1.0","type":"rich","provider_name":"gaks.ai AI Glossary","provider_url":"https://gaks.ai/glossary","title":"Evasion Attack — AI Glossary","author_name":"Glenn Katrud Solheim","author_url":"https://gaks.ai","width":600,"height":200,"html":"<div style=\"font-family:sans-serif;border:1px solid #e0e0e0;border-radius:8px;padding:16px;max-width:600px;background:#ffffff;color:#111111;\"><p style=\"margin:0 0 4px;font-size:11px;color:#666;\">AI Glossary — gaks.ai</p><h3 style=\"margin:0 0 8px;font-size:16px;\">Evasion Attack</h3><p style=\"margin:0 0 12px;font-size:14px;line-height:1.6;\">An attack where an adversary crafts inputs at inference time that cause an AI model to misclassify or produce incorrect outputs, without modifying the model itself. Unlike data poisoning, which targets training, evasion attacks target deployment. A classic example is a stop sign with carefully placed stickers that cause an autonomous vehicle's perception system to misidentify it.  See also: adversarial attack, adversarial example, data poisoning.</p><a href=\"https://gaks.ai/glossary/evasion-attack\" style=\"font-size:12px;color:#0077aa;\">Source: gaks.ai/glossary/evasion-attack →</a></div>"}