The IBM Granite Guardian 3.0 2B and 8B models are designed to detect risks in prompts and/or responses.
3,117 Pulls Updated 4 weeks ago
Updated 4 weeks ago
4 weeks ago
edeec72aa741 · 5.6GB
Readme
Granite guardian models
The IBM Granite Guardian 3.0 2B and 8B models are designed to detect risks in prompts and/or responses. They can help with risk detection along many key dimensions catalogued in the IBM AI Risk Atlas. They are trained on unique data comprising human annotations and synthetic data informed by internal red-teaming, and they outperform other open-source models in the same space on standard benchmarks.
Parameter Sizes
The model will produce a single output token, either Yes
or No
. By default, the general-purpose harm
category is used, but other categories can be selected by setting the system prompt.
2B:
ollama run granite3-guardian:2b
>>> /set system profanity
8B:
ollama run granite3-guardian:8b
>>> /set system violence
Supported Uses
Risk detection in prompt text or model response (i.e. as guardrails), such as:
- Harm (
harm
): content considered generally harmful - Social Bias (
social_bias
): prejudice based on identity or characteristics - Jailbreaking (
jailbreak
): deliberate instances of manipulating AI to generate harmful, undesired, or inappropriate content - Violence (
violence
): content promoting physical, mental, or sexual harm - Profanity (
profanity
): use of offensive language or insults - Sexual Content (
sexual_content
): explicit or suggestive material of a sexual nature - Unethical Behavior (
unethical_behavior
): actions that violate moral or legal standards
- Harm (
RAG (retrieval-augmented generation) to assess:
- Context relevance (
relevance
): whether the retrieved context is relevant to the query - Groundedness (
groundedness
): whether the response is accurate and faithful to the provided context - Answer relevance (
answer_relevance
): whether the response directly addresses the user’s query
- Context relevance (
Granite dense models
The Granite dense models are available in 2B and 8B parameter sizes designed to support tool-based use cases and for retrieval augmented generation (RAG), streamlining code generation, translation and bug fixing.
Granite mixture of experts models
The Granite MoE models are available in 1B and 3B parameter sizes designed for low latency usage and to support deployment in on-device applications or situations requiring instantaneous inference.
Learn more
- Developers: IBM Research
- GitHub Repository: ibm-granite/granite-guardian
- Website: Granite Guardian Docs
- Cookbook: Granite Guardian Snack
- Release Date: October 21st, 2024
- License: Apache 2.0.