Experiment
Test multiple models simultaneously, compare prompt templates, evaluate results, and select winning configurations based on performance
Prerequisites
Before starting an experiment, ensure you have:
- Added a dataset to Future AGI
- Created and tested prompts
- Set up evaluation metrics (optional)
Step-by-Step Guide
1. Dataset Selection
Click on your target dataset in the dashboard. If no datasets are visible, follow the Add Dataset guide first.
2. Access Experiment Interface
- Ensure you have created prompts following the Run Prompt guide
- Select Experiment from the top right corner
3. Configure Experiment
Basic Setup
-
Assign a descriptive name to track your experiment
-
Select the column containing prompt responses
-
Create your prompt template using double curly braces to reference dataset columns
Model Configuration
-
Choose your target model(s) from the dropdown menu
-
Enter the required API keys in the popup window
-
Click the + button to add each model to your experiment
Evaluation Setup
- Select evaluation metrics from Added Evaluations if previously configured
- To create new metrics:
- Click + Create Eval
- Follow the Choosing Evals guide
4. Track Results
Viewing Experiments
Access the Experiments dashboard from the top-left corner
The dashboard displays:
- Experiment names
- Status
- Number of models used
- Number of metrics used
- Creation date
Analyzing Results
-
Click on an experiment to view detailed results
-
Review evaluation scores per model and metric
-
Perform additional evaluations using the Evaluate button
Summary Dashboard
-
Click Summary to compare models and metrics
-
Select winning experiments:
- Click Choose winner
- Set selection criteria
- Click Save & Run
Winners are marked with a crown symbol based on your criteria.
Best Practices
- Use descriptive experiment names for easy tracking
- Test multiple models with varying parameters
- Include diverse evaluation metrics
- Document selection criteria for winning experiments
- Review results thoroughly before selecting winners