Back to glossary

AI GLOSSARY

Content Moderation

Safety, Alignment & Ethics

The practice of reviewing and managing AI-generated or user-generated content to prevent harmful, illegal, or policy-violating material from being produced or distributed. In AI systems, content moderation involves a combination of model-level training, output filtering, and human review, balancing the need to prevent harm against the risk of over-restricting legitimate use.
See also: content filtering, behavioral policy, abuse monitoring.