{"version":"1.0","type":"rich","provider_name":"gaks.ai AI Glossary","provider_url":"https://gaks.ai/glossary","title":"Purple Team — AI Glossary","author_name":"Glenn Katrud Solheim","author_url":"https://gaks.ai","width":600,"height":200,"html":"<div style=\"font-family:sans-serif;border:1px solid #e0e0e0;border-radius:8px;padding:16px;max-width:600px;background:#ffffff;color:#111111;\"><p style=\"margin:0 0 4px;font-size:11px;color:#666;\">AI Glossary — gaks.ai</p><h3 style=\"margin:0 0 8px;font-size:16px;\">Purple Team</h3><p style=\"margin:0 0 12px;font-size:14px;line-height:1.6;\">A collaborative security approach that brings together red team attackers and blue team defenders to work jointly, sharing findings, techniques, and insights in real time rather than operating in separate silos. Purple teaming accelerates security improvement by ensuring that offensive findings are immediately translated into defensive actions.</p><a href=\"https://gaks.ai/glossary/purple-team\" style=\"font-size:12px;color:#0077aa;\">Source: gaks.ai/glossary/purple-team →</a></div>"}