
Building multi-agent systems with Google Agent Development Kit? Check out
the Google ADK integration page for comprehensive
tracing of agent conversations, handoffs, and tool usage.
Evaluating Google Gemini models
You can set up Openlayer tests to evaluate your Google Gemini models in monitoring and development.Monitoring
To use the monitoring mode, you must instrument your code to publish the requests your AI system receives to the Openlayer platform. To set it up, you must follow the steps in the code snippet below:Python
See full Python example
If the Google Gemini model call is just one of the steps of your AI system,
you can use the code snippets above together with
tracing. In this case, your Gemini calls get added as a
step of a larger trace. Refer to the Tracing guide for
details.
Development
In development mode, Openlayer becomes a step in your CI/CD pipeline, and your tests get automatically evaluated after being triggered by some events. Openlayer tests often rely on your AI system’s outputs on a validation dataset. As discussed in the Configuring output generation guide, you have two options:- either provide a way for Openlayer to run your AI system on your datasets, or
- before pushing, generate the model outputs yourself and push them alongside your artifacts.
GOOGLE_AI_API_KEY.
If you don’t add the required Google AI API key, you’ll encounter a “Missing API key”
error when Openlayer tries to run your AI system to get its outputs.
Make sure to configure the Gemini API with the API key from the
environment in the script you provide as the
batchCommand in
the openlayer.json:
