Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

JailGuard: A Universal Detection Framework for LLM Prompt-based Attacks #2

Open
ramimac opened this issue Jun 19, 2024 · 0 comments
Open

Comments

@ramimac
Copy link
Collaborator

ramimac commented Jun 19, 2024

https://arxiv.org/pdf/2312.10766

we propose JailGuard, a universal detection framework for jailbreaking and hijacking
attacks across LLMs and MLLMs. JailGuard operates on the principle that attacks are inherently less robust
than benign ones, regardless of method or modality. Specifically, JailGuard mutates untrusted inputs to
generate variants and leverages discrepancy of the variants’ responses on the model to distinguish attack
samples from benign samples

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant