Maxim Logo
How toEvaluate Chains

Test your agentic workflows using chains

Test Prompt Chains using datasets to evaluate performance across examples

After testing in the playground, evaluate your Prompt Chains across multiple test cases to ensure consistent performance. You can do the same via test run.

Create a Dataset

Add test cases by creating a Dataset. For this example, we'll use a Dataset of product images to generate descriptions.

Dataset with product images for testing

Build your Prompt Chain

Create a Prompt Chain that processes your test examples. In this case, the chain generates product descriptions, translates them to multiple languages, and formats them to match specific requirements.

Prompt chain for product description generation

Start a test run

Open the test configuration by clicking in the top right corner.

Configure your test

Select your dataset and add Evaluators to measure the quality of outputs.

Test configuration with dataset and evaluator options

Review results

Monitor the test run to analyze the performance of your Prompt Chain across all inputs.

Test run results showing performance metrics

On this page