A research platform built for experiments that mix humans, bots, and AI agents.
Real participants chat alongside scripted bots and LLM-powered agents in the same room — indistinguishable or with visible type badges.
Every participant — human or AI — is assigned a role that determines their capabilities. Communicators chat directly. Mediators observe and orchestrate. Processors assist with AI-powered feedback.
Define exactly how participants interact with AI at every stage. Three modes — Review, Generate, and Real-time Assist — transition automatically based on configurable triggers. Researchers control when AI activates, what it sees, and how participants respond.
Chain different activity types into timed sequences within a single session. Chat, vote, rank, survey, watch media, complete tasks — each with its own timing and transition rules.
Click any card to see details
Upload CSV datasets, annotate with multiple AI models at scale.
Visual drag-and-drop experiment design — no code required.
Describe your experiment in natural language — the AI builds it for you.
Automatic grouping by condition, survey response, or queue order.
OpenAI, Anthropic, and Google models side by side.
Real-time session tracking, alerts, and data export.
13+ trigger types for scripted agent behavior — no code required.
Reusable experiment configs with peer-review workflow.
AI models evolve fast. Your research platform should keep up. Carrier is built so experiments are reproducible across model versions, scalable from pilot studies to large-scale deployments, and cumulative — each new model adds to a growing, comparable body of results.
Every experiment is a saved configuration — same chambers, same roles, same prompts. Re-run with a different model and compare results directly. No ambiguity about what changed.
From a 4-person pilot to hundreds of concurrent sessions. Batch annotation processes thousands of rows. Matching, chatrooms, and data export handle the volume without changing your design.
New model released? Plug it in and re-run. Every experiment adds to a growing body of comparable results — same design, different models, tracked over time.