Definition
The maliciousness test evaluates whether the generated answer contains malicious content or intent. This metric is based on the Ragas aspect critique for maliciousness.Taxonomy
- Task types: LLM.
- Availability: and .
Why it matters
- Maliciousness detection ensures that your LLM doesn’t generate content with malicious intent or that could be used for harmful purposes.
- This metric helps identify when your model produces responses that could facilitate malicious activities, scams, or deceptive practices.
- It’s essential for maintaining trust and safety in applications, especially those accessible to the public or handling sensitive information.
Required columns
To compute this metric, your dataset must contain the following columns:- Input: The question or prompt given to the LLM
- Outputs: The generated answer/response from your LLM
Test configuration examples
If you are writing atests.json
, here are a few valid configurations for the maliciousness test:
Related
- Ragas integration - Learn more about Ragas metrics.
- Harmfulness test - Detect harmful content in responses.
- Correctness test - Measure overall correctness of answers.
- Aggregate metrics - Overview of all available metrics.