Pricing Enterprise Docs Agent Skills Learn

MCP

Add Runhuman to your AI coding agent via MCP.

Recommended: For most users, Agent Skills is the simplest way to get started. Use MCP if your agent doesn’t support skills or you prefer MCP configuration.


Installation

One command adds Runhuman to your agent:


Available Tools

Runhuman exposes 7 MCP tools for orchestrating human QA testing:

list_organizations

List all organizations the authenticated user belongs to. Use this to discover your workspace before creating jobs.

No parameters required. Returns organization name, ID, project count, and member count.

list_projects

List projects accessible to the authenticated user, optionally filtered by organization.

ParameterRequiredDescription
organizationIdNoFilter projects by organization ID. If omitted, returns all accessible projects

Returns project name, ID, GitHub repo, and default URL.

create_job

Creates a custom QA test and returns immediately with a job ID. You must follow up with wait to get results.

ParameterRequiredDescription
projectIdYes*Project ID for the test job
organizationIdNoOrganization ID for billing context, or to auto-create a project with githubRepo
urlNoURL to test (publicly accessible)
descriptionNoInstructions for the human tester
templateNoTemplate name to use as base configuration
templateContentNoRaw template content (markdown with YAML frontmatter) as alternative to saved template
outputSchemaNoJSON Schema for structured result extraction. If omitted, only success/explanation returned
resultsTemplateNoMDForm template for free-form text reports (alternative to outputSchema)
targetDurationMinutesNoTime limit in minutes (default: 30, range: 1-60)
allowDurationExtensionNoAllow tester to request more time (default: true)
maxExtensionMinutesNoMax additional minutes, or false for unlimited (default: false)
additionalValidationInstructionsNoCustom instructions for AI validation
deviceClassNo"desktop" or "mobile" (default: “desktop”)
attachmentsNoArray of file attachments for the tester (max 10)
githubRepoNoGitHub repo ("owner/repo") for AI context
githubTokenNoGitHub token for operations without GitHub App installation
prNumbersNoPR numbers to test — triggers AI test plan generation from PR changes (requires githubRepo)
issueNumbersNoIssue numbers to test — triggers AI test plan generation from issues (requires githubRepo)
checkTestabilityNoReject job early if AI determines it’s not testable (default: true when prNumbers/issueNumbers provided)

*projectId is required unless organizationId + githubRepo are provided for auto-creation.

Either url+description, template, or prNumbers/issueNumbers is needed to define what to test.

run_template

Create a job from a pre-configured template. Templates let you reuse test configurations without writing full descriptions every time.

ParameterRequiredDescription
projectIdYesProject ID for the test job
templateYesTemplate name (get from list_templates)
urlNoOverride template’s default URL
descriptionNoAdditional instructions (appended to template)
outputSchemaNoOverride template’s output schema
targetDurationMinutesNoOverride template’s duration (range: 1-60)
deviceClassNoOverride template’s device class
githubRepoNoOverride template’s GitHub repo
attachmentsNoOverride template’s attachments
additionalValidationInstructionsNoExtra validation rules (appended to template)

Use list_templates to see available templates, then reference them by name.

wait

Idiomatic polling — waits for a job to complete and returns results automatically. Polls every 10 seconds until completion, timeout, or failure.

ParameterRequiredDescription
jobIdYesJob ID from create_job or run_template
timeoutSecondsNoMaximum wait time (default: 600, min: 10, max: 3600)

No manual polling needed! Just call wait once and it automatically polls until the job finishes.

Returns: When complete, includes:

  • result: Structured test results matching your schema
  • testerAlias: Tester identification
  • testerResponse: Raw feedback from the human tester
  • costUsd: Exact cost in USD
  • testDurationSeconds: Time spent by tester
  • jobUrl: Link to view results in the dashboard

get_job

Quick status check without waiting. Get current job status instantly without polling.

ParameterRequiredDescription
jobIdYesJob ID to check

Returns different detail levels based on status:

  • Active (pending, preparing, waiting, working): Current state description and next steps
  • Completed: Full results, tester info, cost, and duration
  • Failed (incomplete, abandoned, rejected, error): Failure reason and details

Use this for manual polling control or checking multiple jobs in parallel.

list_templates

List available templates for a project.

ParameterRequiredDescription
projectIdYesProject ID to list templates for
limitNoMax templates to return (default: 50)

Returns template names, default URLs, and descriptions.


Example Prompts

These prompts work well with any AI agent that has Runhuman installed:

Simple page check:

Use Runhuman to verify that example.com loads correctly and shows the main heading.

Login testing:

Use Runhuman to test the login flow on staging.myapp.com. Try email test@example.com with password demo123, then try a wrong password and verify the error message.

Checkout flow:

Use Runhuman to test the checkout on myapp.com. Add a product to cart, fill shipping info, and verify the order total is correct. Give the tester 10 minutes.

Visual issues:

Use Runhuman to check the product page at myapp.com/products/123 for visual issues. Look for broken images, layout problems, or unreadable text.

Mobile testing:

Use Runhuman to test the navigation menu on myapp.com on mobile. Check if it opens and closes correctly and all links work.


What Happens Behind the Scenes

When you ask your agent to use Runhuman:

  1. Agent discovers your workspace with list_organizations and list_projects
  2. Agent checks available templates with list_templates (if applicable)
  3. Agent calls create_job or run_template with your URL/instructions and an output schema it generates
  4. Agent receives a job ID and status message
  5. Agent calls wait with that job ID — this automatically polls until complete!
  6. When complete, agent receives structured results and the tester’s raw response
  7. Agent summarizes the findings for you

The wait tool handles all the polling logic automatically. You just describe what you want tested.


Duration Control

Tell the agent how much time to give the tester:

Use Runhuman to test the signup flow. Allow 10 minutes since this involves email verification.

Use Runhuman to check the homepage. This should be quick, give the tester 3 minutes.

The default is 30 minutes. For simple checks, you can request less time (e.g., 5 minutes).


Writing Good Prompts

Good prompts are specific about what to test:

Instead ofWrite
”Test the app""Test the login flow on myapp.com. Try valid credentials, then invalid password. Verify error messages."
"Check the UI""Check the product page for visual issues. Look for broken images, layout problems, text overflow."
"Test checkout""Test checkout on staging.myapp.com. Add product to cart, fill shipping, verify total before submitting.”

Include:

  1. The URL to test
  2. Specific actions to perform
  3. What success looks like
  4. Time needed for complex flows

Complete Example

Here’s what an agent does when you ask it to test a login page:

You say:

Use Runhuman to test login on myapp.com with email test@example.com and password demo123. Then try wrong password and check the error.

Agent calls create_job:

{
  "projectId": "proj-abc123",
  "url": "https://myapp.com/login",
  "description": "Test login with test@example.com / demo123, then try wrong password",
  "outputSchema": {
    "loginWorks": { "type": "boolean", "description": "Login works with valid credentials" },
    "errorShown": { "type": "boolean", "description": "Error shown for wrong password" },
    "errorMessage": { "type": "string", "description": "Error message text" }
  }
}

Agent calls wait:

{
  "jobId": "job-abc123"
}

The wait tool automatically polls every 10 seconds until complete. No manual polling needed!

After 2-3 minutes, wait returns:

Test Completed

Job ID: job-abc123
Status: completed
Duration: 120s
Cost: $0.216
Tester: Alex

Results:
  loginWorks: true
  errorShown: true
  errorMessage: Invalid username or password

Raw Response:
I tested login with the provided credentials. Login works correctly with test@example.com.
When I tried a wrong password, the error message 'Invalid username or password' appeared.

View in dashboard: https://runhuman.com/dashboard/...

Agent tells you:

Login test complete! Login works with valid credentials. Error message “Invalid username or password” shown for wrong password. Test took 2 minutes, cost $0.22.


Next Steps

TopicLink
Full technical specificationReference
Practical recipesCookbook
Direct API integrationREST API