Working with Batches
An AI Task Builder Batch allows you to collect human annotations on your existing data. You provide a dataset, define instructions, and participants evaluate each datapoint according to your instructions.
This guide covers the workflow for creating, configuring, and publishing a Batch.
Workflow overview
Creating a dataset
Create a dataset to hold your data.
Response
Uploading your data
Upload your dataset as a CSV file using presigned URLs.
Step 1: Request a presigned URL
For example:
Step 2: Upload to S3
Use the presigned URL from the response to upload your CSV file directly to S3.
CSV format
Your CSV should contain one row per datapoint. Each column is displayed to participants alongside the instructions.
For advanced options including metadata columns and custom task grouping, see Working with Datasets.
Monitoring dataset status
Poll the dataset endpoint to check when processing is complete.
Wait for the status to change to READY before proceeding.
Dataset status
Creating a batch
Once your dataset is ready, create a batch and attach the dataset.
Task details
The optional task_details object provides context to participants:
All three fields support basic HTML formatting.
Response
Batch status
A batch transitions through the following states:
Creating instructions
Instructions define what participants should do with each datapoint. Each instruction is displayed to participants sequentially alongside the datapoint.
Instruction types
By default, when there are 5 or more options, a dropdown is rendered instead of checkboxes or radio buttons. Set disable_dropdown: true to always use checkboxes/radio buttons. See Instructions for full details on all instruction fields.
Example: Sentiment classification
Example: Free text explanation
Setting up the batch
Once your instructions are created, trigger task generation. Each datapoint in your dataset is paired with all instructions to create a task. Tasks are then organized into task groups — participants complete one task group per submission.
The tasks_per_group parameter controls how many tasks are randomly assigned to each group. If omitted, each task group contains a single task.
Participants complete all tasks within their assigned group in a single submission. No participant will be assigned the same task group twice, even if they complete multiple submissions.
For custom task grouping based on your own criteria, see Working with Datasets.
This triggers task generation. The batch status will change to PROCESSING and then to READY once complete.
Monitoring batch status
Poll the batch endpoint to check when task generation is complete.
Wait for the status to change to READY before creating a study.
The total_task_count reflects the number of datapoints in your dataset.
Publishing a batch
To make your batch available to participants, create a Prolific study that references it.
When creating the study, set data_collection_method to AI_TASK_BUILDER_BATCH and provide your batch ID:
Use annotators_per_task in data_collection_metadata to specify how many participants should annotate each datapoint. The default is 1. After publishing, this value can only be increased.
Then publish the study:
Retrieving responses
After participants have completed their tasks, download the annotated data as a CSV.
This returns your original CSV with additional columns containing participant responses for each instruction.
By using AI Task Builder, you agree to our AI Task Builder Terms.