
Llama Guard
About
The Llama Guard family of LLMs, developed by Meta AI, offers content safety classification capabilities for managing human-AI interactions. These models work by scrutinizing both inputs (prompts) and outputs (responses) to flag potentially unsafe content, utilizing a comprehensive safety risk taxonomy 14. Initially focused on text, the Llama Guard 3 Vision model extended this functionality to multimodal inputs, including image analysis 2. These models are known for their performance, which equals or surpasses current content moderation solutions on renowned benchmarks 1. Moreover, they are instruction-tuned, offering adaptability to various use cases and safety frameworks 14. Llama Guard models, including version 3-8B and its variants, are accessible via Hugging Face 4.