OMNIGUARD: An Efficient Approach for AI Safety Moderation Across Modalities
Abstract
OMNIGUARD detects harmful prompts across languages and modalities by identifying aligned internal representations in large language models, achieving high accuracy and efficiency.
The emerging capabilities of large language models (LLMs) have sparked concerns about their immediate potential for harmful misuse. The core approach to mitigate these concerns is the detection of harmful queries to the model. Current detection approaches are fallible, and are particularly susceptible to attacks that exploit mismatched generalization of model capabilities (e.g., prompts in low-resource languages or prompts provided in non-text modalities such as image and audio). To tackle this challenge, we propose OMNIGUARD, an approach for detecting harmful prompts across languages and modalities. Our approach (i) identifies internal representations of an LLM/MLLM that are aligned across languages or modalities and then (ii) uses them to build a language-agnostic or modality-agnostic classifier for detecting harmful prompts. OMNIGUARD improves harmful prompt classification accuracy by 11.57\% over the strongest baseline in a multilingual setting, by 20.44\% for image-based prompts, and sets a new SOTA for audio-based prompts. By repurposing embeddings computed during generation, OMNIGUARD is also very efficient (approx 120 times faster than the next fastest baseline). Code and data are available at: https://github.com/vsahil/OmniGuard.
Community
We build a new AI safety moderation model, OmniGuard, that can detect harmful prompts across multiple languages and multiple modalities, all using one approach. It achieves SOTA results for detecting harmful prompts in 3 modalities: (multilingual) text, images, and audios.
OmniGuard operates by finding:
- internal representations of a model (LLM or MLLM) that are universally shared across languages or different modalities, and
- building a classifier using these representations
Using the internal representations for safety classification bypasses the need for a separate Guard model while making OmniGuard ~120X faster than the next fastest baseline Guard model.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- MrGuard: A Multilingual Reasoning Guardrail for Universal LLM Safety (2025)
- DREAM: Disentangling Risks to Enhance Safety Alignment in Multimodal Large Language Models (2025)
- Multilingual Collaborative Defense for Large Language Models (2025)
- PolyGuard: A Multilingual Safety Moderation Tool for 17 Languages (2025)
- X-Guard: Multilingual Guard Agent for Content Moderation (2025)
- Robustifying Vision-Language Models via Dynamic Token Reweighting (2025)
- Logic Jailbreak: Efficiently Unlocking LLM Safety Restrictions Through Formal Logical Expression (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper