
Evaluating LiteLLM applications
You can set up Openlayer tests to evaluate your Lit eLLM applications in monitoring and development.Monitoring
To use the monitoring mode, you must instrument your code to publish the requests your AI system receives to the Openlayer platform. To set it up, you must follow the steps in the code snippet below:See full Python example

If the LiteLLM completions are just one of the steps of your AI system, you
can use the code snippets above together with tracing.
In this case, your LiteLLM completions get added as a step of a larger trace.
Refer to the Tracing guide for details.
Development
In development mode, Openlayer becomes a step in your CI/CD pipeline, and your tests get automatically evaluated after being triggered by some events. Openlayer tests often rely on your AI system’s outputs on a validation dataset. As discussed in the Configuring output generation guide, you have two options:- either provide a way for Openlayer to run your AI system on your datasets, or
- before pushing, generate the model outputs yourself and push them alongside your artifacts.
OPENAI_API_KEY
, if it uses Anthropic models, you must provide an ANTHROPIC_API_KEY
, and so on.
To provide the required API credentials, navigate to “Workspace settings” -> “Environment variables,”
and add the credentials as variables.
Next steps
- Explore the LiteLLM tracing example notebook
- Learn about Openlayer’s testing capabilities
- Set up monitoring for your production LiteLLM applications
- Check out other integrations that work well with LiteLLM