org.llm4s.agent.guardrails.builtin.LLMFactualityGuardrail
See theLLMFactualityGuardrail companion object
LLM-based factual accuracy validation guardrail.
Uses an LLM to evaluate whether content is factually accurate given a reference context. Useful for RAG applications where you want to ensure the model's response aligns with retrieved documents.
Value parameters
llmClient
The LLM client to use for evaluation
referenceContext
The reference text to fact-check against
threshold
Minimum score to pass (default: 0.7)
Attributes
Example
val context = "Paris is the capital of France. It has a population of 2.1 million."
val guardrail = LLMFactualityGuardrail(client, context, threshold = 0.8)
agent.run(query, tools, outputGuardrails = Seq(guardrail))
Companion
object
Graph
Reset zoom Hide graph Show graph
Supertypes
class Object
trait Matchable
class Any
Show all
Members list
Compose this guardrail with another sequentially.
Compose this guardrail with another sequentially.
The second guardrail runs only if this one passes.
Value parameters
other
The guardrail to run after this one
Attributes
Returns
A composite guardrail that runs both in sequence
Inherited from:
Guardrail
Optional completion options for the judge LLM call. Override to customize temperature, max tokens, etc.
Optional completion options for the judge LLM call. Override to customize temperature, max tokens, etc.
Attributes
Inherited from:
LLMGuardrail
Validate content using the LLM as a judge.
Validate content using the LLM as a judge.
The implementation:
Constructs a prompt with evaluation criteria and content
Calls the LLM to get a score
Parses the score and compares to threshold
Returns success if score >= threshold, error otherwise
Attributes
Definition Classes
Inherited from:
LLMGuardrail
Optional description of what this guardrail validates.
Optional description of what this guardrail validates.
Attributes
Natural language prompt describing the evaluation criteria.
Natural language prompt describing the evaluation criteria.
The prompt should instruct the model to return a score between 0 and 1. The content being evaluated will be provided separately.
Attributes
Example
"Rate if this response is professional in tone. Return only a number between 0 and 1."
The LLM client to use for evaluation. Can be the same client used by the agent or a different one.
The LLM client to use for evaluation. Can be the same client used by the agent or a different one.
Attributes
Name of this guardrail for logging and error messages.
Name of this guardrail for logging and error messages.
Attributes
Minimum score required to pass validation (0.0 to 1.0). Default is 0.7 (70% confidence).
Minimum score required to pass validation (0.0 to 1.0). Default is 0.7 (70% confidence).
Attributes