Maxim requires at least one provider with access to GPT-3.5 and GPT-4 models. We use industry-standard encryption to securely store your API keys.
Running your first test
Learn how to get started with your first test run in Maxim
1. Set up your environment
First, configure your AI model providers:
Go to Settings
→ Models
.
Click on the tab of the provider for which you want to add an API key.
Click on Add New
and fill in the required details.
To learn more about API keys, inviting users, and managing roles, refer to our Workspace and roles guide.
2. Create your first prompt
Start experimenting with AI by crafting and testing prompts:
Navigate to the Prompts
tab under the Evaluate
section and click on Single prompts.
Click Create prompt
or Try sample
to get started.
Write your system prompt and user prompt in the respective fields.
Configure additional settings like model, temperature, and max tokens.
Click Run
to test your prompt and see the AI's response.
Iterate on your prompt based on the results.
When satisfied, click Save
to create a new version of your prompt.
To learn more about prompts, refer to our detailed guide on Single prompts.
3. Prepare your dataset
Organize and manage the data you'll use for testing and evaluation:
Navigate to the Datasets tab under the Library
section.
Click Create New
or Upload CSV
. We also have a sample dataset created for you. Click on View our sample dataset
to get started.
If creating a new dataset, enter a name and description for your dataset.
Add columns to your dataset (e.g., 'input' and 'expected_output').
Add entries to your dataset, filling in the values for each column.
Click Save
to create your dataset.
To learn more about datasets, refer to our detailed guide on Datasets.
4. Create a workflow
Set up an HTTP workflow to test your AI application end-to-end:
Navigate to the Workflows tab under the Evaluate
section.
Click Create Workflow
or Try sample
.
Enter your API endpoint URL in the "URL" field.
Configure any necessary headers or parameters. You can use dynamic variables like {input}
to reference static context easily in any part of your workflow using {}
Click Run
to test your endpoint in the playground.
In the "Output Mapping" section, select the part of the response you want to evaluate (e.g., data.response
).
Click Save
to create your workflow.
To learn more about workflows, refer to our detailed guide on Workflows.
5. Add evaluators
Set up evaluators to assess your prompt or workflow's performance:
Navigate to the Evaluators
tab under the Library
section.
Click Add Evaluator
to browse available evaluators.
Choose an evaluator type (e.g., AI, Programmatic, API, or Human).
Configure the evaluator settings as needed.
Click Save
to add the evaluator to your workspace.
To learn more about evaluators, refer to our detailed guide on Evaluators.
6. Run Your First Test
Execute a test run to evaluate your prompt or workflow:
Navigate to your saved prompt or workflow.
Click Test
in the top right corner.
Select the dataset you created earlier.
Choose the evaluators you want to use for this test run.
Click Trigger Test Run
to start the evaluation process.
If you've added human evaluators, you'll be prompted to enter their email addresses. The test run will evaluate your prompt or workflow across all inputs in your dataset using the selected evaluators.
7. Analyze test results
Review and analyze the results of your test run:
Navigate to the Runs
tab in the left navigation menu.
Find your recent test run and click on it to view details.
Review the overall performance metrics and scores for each evaluator.
Drill down into individual queries to see specific scores and reasoning.
Use these insights to identify areas for improvement in your prompt or workflow.
To learn more about test runs, refer to our detailed guide on Test runs.
Next steps
Now that you've completed your first cycle on the Maxim platform, consider exploring these additional capabilities:
- Prompt comparisons: Evaluate different prompts side-by-side to determine which ones produce the best results for a given task.
- Prompt chains: Create complex, multi-step AI workflows. Learn how to connect prompts, code, and APIs to build powerful, real-world AI systems using our intuitive, no-code editor.
- Context sources: Integrate Retrieval-Augmented Generation (RAG) into your workflows.
- Prompt tools: Enhance your prompts with custom functions and agentic behaviors.
- Observability: Use our stateless SDK to monitor real-time production logs and run periodic quality checks.
By following this guide, you've learned how to set up your environment, create prompts, prepare datasets, set up workflows, add evaluators, run tests, and analyze results. This foundational knowledge will help you leverage Maxim's powerful features to develop and improve your AI applications efficiently.