Run Prompt in Dataset

Learn how to execute prompts against your dataset and generate responses

What it is

Run Prompt lets you add a new column to your dataset that is filled by a model (LLM, Text-to-Speech, Speech-to-Text, or Image Generation). You define a prompt (messages with placeholders that pull from other columns), pick a model and settings, and the system runs the prompt on each row and writes the model output into that column. The result is a dynamic column of responses you can use for evals, comparison, or export.

Use cases

  • Generate answers or text – Use an LLM to answer questions, summarize, or complete text per row (e.g. a column of questions → a column of answers).
  • Produce audio – Use Text-to-Speech to turn a text column into an audio column (e.g. scripts → voice clips).
  • Transcribe audio – Use Speech-to-Text to turn an audio column into a text column for evals or search.
  • Batch test a prompt – Run the same prompt across many rows to see how the model behaves and then run evals on the outputs.
  • Generate images – Use Image Generation to create images from text (or text + image) per row; the new column stores image URLs.
  • Structured output – Use response format (e.g. JSON schema) to get structured fields (object, array) in the new column for downstream use.

How to

Navigate to Run Prompt

Click the “Run Prompt” button (e.g. in the top-right or dataset toolbar) to add a new run-prompt column. This creates a dynamic column that will store the model output for each row. Run Prompt

Assign prompt name

Enter a name for the prompt. This name is used as the new column name in your dataset. Each row will have one cell in this column holding the model response for that row. Run Prompt

Choose model type and model

Select the type of task, then pick the model to use. Models are filtered by type; you need an API key (or custom model) for the chosen provider.

Choose LLM for text generation (chat). Use for Q&A, summarization, or any text-in, text-out task. Select a chat model from the list; ensure the provider has an API key configured. LLM

Tip

Click here to learn how to create custom models.

Choose Text-to-Speech to generate audio from text. The prompt output column will store audio (e.g. URLs). You can configure voice and format for supported TTS models. Text-to-Speech

Tip

Click here to learn how to create custom models.

Choose Speech-to-Text to transcribe audio into text. Use when a column contains audio; the model output will be text in the new column. Speech-to-Text

Tip

Click here to learn how to create custom models.

Choose Image Generation to create images from text (or image + text) prompts. The prompt output column will store image URLs. Select an image-generation model and ensure the provider has an API key configured. Image Generation

Tip

Click here to learn how to create custom models.

Build the prompt (messages)

Define the prompt as a list of messages (system, user, assistant). In the message content, use placeholders that reference dataset columns (e.g. {{column_name}}). At runtime, these are replaced by the cell value for that column in each row. The first message must not be from the assistant.

Configure Prompt with Roles

Define your prompt using roles. You can configure messages with different roles:

  • User Role (Required): The main input message from the user perspective. This role is required for the prompt to work.
  • System Role (Optional): System-level instructions that guide the model’s behavior and set the context.

Using Variables

You can reference dataset columns as variables within your prompt using the {{ }} syntax. Simply wrap the column name in double curly braces:

Basic Example:

System: You are a helpful assistant that summarizes content.

User: Please summarize the following text: {{column_name}}

The variables (column names) will be dynamically replaced with actual values from your dataset when the prompt runs.

JSON Dot Notation

For JSON type columns, you can access nested fields directly using dot notation. This allows you to reference specific keys within structured data without additional processing:

JSON Example:

User: Based on this prompt: {{column_name.key_name}}, generate a response that addresses {{column_name.key_name}}

In this example:

  • {{column_name.key_name}} accesses the key_name field within the column_name JSON column

This feature significantly simplifies complex data handling and speeds up setup when working with structured JSON data in your dataset.

Configure Model Parameters (optional)

Adjust model parameters such as temperature, max tokens, top_p, and other settings to fine-tune the model’s behavior according to your needs.

Configure Tools (optional)

Add tools or functions that the model can use during execution. This enables the model to perform specific actions or access external capabilities.

Configure Concurrency

Set the concurrency level to control how many prompt executions run in parallel. Higher concurrency speeds up processing but may consume more resources.

Run Prompt

Click the “Run” button to execute the prompt across your dataset. The responses will be generated and saved as a new dynamic column in your dataset.

What you can do next

Was this page helpful?

Questions & Discussion