Security
Rebuff
Catch sneaky attempts to trick your AI chatbot before they cause damage
Using Rebuff is like having a security guard at the door of your AI who pats down every visitor for hidden weapons, remembers everyone who's tried to sneak something in before, and gets better at spotting troublemakers every day.
Rebuff is a free security tool that protects AI applications from 'prompt injection' attacks β when bad actors try to trick your AI into ignoring its rules, leaking private data, or doing things it shouldn't. Think of it as a bouncer that inspects every message before it reaches your AI assistant. It learns from each attack it sees, so it gets smarter and harder to fool over time. If you're building anything that lets users chat with an AI, Rebuff helps make sure those users can't hijack it.
Best for
How well does it fit you?
Rough fit scores (1β10) for different kinds of people. Tap a row to highlight it.
Great at
Not ideal for
See it in action
Real prompts you could paste into the product β pick a persona tab below.
Use case
Protecting a customer support chatbot from manipulation
Try this prompt
Integrate Rebuff to check every incoming user message before it reaches our GPT-4 support agent, flagging anything that looks like an attempt to override system instructions.
Performance, trust, value, improving fast, here to stay
Score shape
We check this tool every day. The SovereignScoreβ’ and its five dimensions update automatically when our pipeline detects meaningful changes across benchmarks, pricing, GitHub activity, trust signals, and longevity data. Below is a transparent log of the most recent applied adjustments.
No automated score adjustments have been published for this tool yet. When our scoring engine approves a change, it will appear here with the reasoning we used.
Self-hardening prompt injection detector with pluggable vector memory.
No published updates for this tool yet.
Same category β with a plain-English note on how they differ when we have comparison copy stored.
A security guard for your AI β blocks prompt attacks, jailbreaks, and harmful responses before they cause damage
Rebuff is a free, open-source tool you set up yourself to catch prompt injection attacks, while Lakera is a paid commercial service that offers broader protection (including filtering AI responses for harmful content) with less hands-on work.
Make sure your AI actually gives you the answer you asked for β every single time.
Guardrails AI checks what your AI sends out to make sure the answers follow your rules, while Rebuff watches what users send in to block people trying to trick or hijack your AI.
Vendors can verify ownership and request corrections to how we describe or score your product.
Email claims deskExports and email alerts when ratings change β for teams evaluating many tools.
For builders who want the same update feed in their own apps β see /api/changelog.