{"version":"1.0","type":"rich","provider_name":"gaks.ai AI Glossary","provider_url":"https://gaks.ai/glossary","title":"Adversarial Attack — AI Glossary","author_name":"Glenn Katrud Solheim","author_url":"https://gaks.ai","width":600,"height":200,"html":"<div style=\"font-family:sans-serif;border:1px solid #e0e0e0;border-radius:8px;padding:16px;max-width:600px;background:#ffffff;color:#111111;\"><p style=\"margin:0 0 4px;font-size:11px;color:#666;\">AI Glossary — gaks.ai</p><h3 style=\"margin:0 0 8px;font-size:16px;\">Adversarial Attack</h3><p style=\"margin:0 0 12px;font-size:14px;line-height:1.6;\">A deliberate attempt to manipulate an AI system by crafting inputs designed to cause errors, bypass safety measures, or produce unintended outputs. Adversarial attacks exploit weaknesses in how models generalize, where small, carefully designed perturbations, often imperceptible to humans, can cause confident misclassification or harmful outputs.</p><a href=\"https://gaks.ai/glossary/adversarial-attack\" style=\"font-size:12px;color:#0077aa;\">Source: gaks.ai/glossary/adversarial-attack →</a></div>"}