Your agent can now run studies and evaluations on Prolific

Integrate feedback from real, verified humans faster with the Prolific CLI

Trusted by leading engineering teams

Google
Hugging Face
Ai2
Stanford
Why Prolific

Built for pipelines

One network for every interface
Reach the same trusted human network and combine with 300+ filters whether you're calling via UI, CLI or REST API. A cohort defined in the dashboard resolves identically when called from your pipeline — same hash, same provenance.
Designed for repeat runs
Idempotency keys on study creation, stable cohort hashes across re-recruits, paginated response export, and JSONL output. Filter provenance is preserved in every response record so downstream training data carries its own cohort description.
Latency that fits a training loop
Studies launch in minutes and return first responses in hours. Webhooks on response.submitted and study.completed let a pipeline ingest incrementally rather than block on a full cohort.
What customers say

"I want to remove any barrier between my agent and the results."

Emerging Products Director: Fortune 500 software company

How teams do more with the CLI

Four programmatic workflows across one human intelligence network.

01
Preference collection in the loop
Launch pairwise preferences, Likert ratings or step-level rationale tasks directly from your training pipeline between runs. JSONL export feeds into RLHF, DPO and reward-model workflows.
02
Human-in-the-loop checkpoints
Agents call Prolific mid-run for human review, escalation or disambiguation — filtered down to the right participants for the task. Responses return as structured JSON.
03
Embed human evaluation
Run pre-release human evaluation as a gated step in CI. Specify a cohort by filter, aggregate responses against a threshold, and pass or fail the job on the resulting metric. Reproducible across model versions.
04
On demand cohorts for user simulation
Programmatically recruit targeted participants to role-play end users in multi-turn evaluation. Stable cohort hashes mean the same group can be re-recruited for longitudinal comparison.
Get started

Do more from where you already operate

Start where it matches your stack.

API reference
Studies, cohorts, responses, webhooks, submissions. OpenAPI schema, auth model, rate limits, idempotency semantics.
docs.prolific.com/api
CLI reference
Launch studies, wait on completion, export responses, manage filters from a shell. Scriptable in any pipeline.
docs.prolific.com/cli
03
Events & webhooks
Subscribe to response.submitted, study.completed, and submission review events. Signed payloads, retry policy.
docs.prolific.com/webhooks

Fast-moving AI teams using Prolific

Trusted by AI/ML developers, researchers, and leading organizations across industries.

Prolific's CLI in practice
See how agent can launch tasks to real humans on Prolific - from dataset to results
Watch the demo
prolific
Building breakthrough AI faster
Ai2 reduced human data collection from weeks to hours with Prolific, building state-of-the-art multimodal AI models faster without sacrificing quality.
Read more
Ai2
Unpacking human preference for LLMs - The HUMAINE framework
Prolific's human-centered leaderboard ranks frontier AI models by how real users actually experience them - not just technical benchmarks. Featured at ICLR 2026.
Read the paper
humaine leaderboard
Gemini 3 Pro: Frontier safety framework
The frontier safety framework report for Google’s latest model.
Read more
google ai

Start building automated workflows

Sign up free and get rich human feedback within hours.
FAQ

Questions from engineers integrating Prolific