Other Evaluators
These are evaluators we have custom built for specific customers. We are working on integrating them into our Github repository.
If you would like to use one of these, please contact us at hello@athina.ai, or sign up (opens in a new tab) to stay notified as we release new evals.
API Call
Calls an external API - your custom evals plugged into Athina.
Language Mismatch
Engine: gpt-3.5-turbo
Detect when the LLM response is in a different language to the user’s query
Sensitive Data Leak
Engine: gpt-3.5-turbo
Detect when user query or LLM response contains any personally identifiable information.
Example: names, emails, phone numbers, social security numbers, credit card information, etc
You can configure evals for different types of PII leak.
Eval Explanation: Content Moderation
OpenAI
Uses OpenAI’s content moderation endpoint to determine if a response is harmful, toxic, violent, threatening or sexual.
Eval thresholds can be configured.
Prompt Injection Attacks
[Coming Soon]
Common Mistakes
[Coming Soon]
Restricted Keywords
String Match
Detect when your LLM output contains certain kinds of keywords
Eval: Critical Keywords
String Match
Detect when your LLM output is missing critical keywords
Hallucinated Link
Regex
HTTP
If your LLM response contains a link, we will check if the link is invalid (404).
No LLM, just good old regex + HTTP request.
Hallucinated Email
Engine: gpt-3.5-turbo | Regex
If your LLM response contains an email that was not a part of your provided context, it is likely a hallucinated email.
…