ES EN
Get Started

Quickstart in 5 minutes

You're going to create your account, connect your first agent, run a test, and see your first report. No credit card required.

Before you start

You'll need two things:

💡 Don't have an AI agent handy? We have a demo site with a fictional banking AI agent ready for you to test the platform end-to-end: artificial-qa-test-bank.vercel.app. Set it up as a Browser Connection and you'll be able to run real test cases against it.

The 6-step flow

1

Create your account

Go to app.artificialqa.com/signup.

Enter your email and a password, or sign up with Google.

We'll send you a verification email. Click the link to confirm the account. Only after verifying your email can you sign in.

🔐 About 2FA. On your first login, the platform will prompt you to set up two-factor authentication (TOTP). It's mandatory for all users. Compatible with Google Authenticator, Authy, 1Password, and any TOTP app.
2

Your organization and project

When you log in for the first time, ArtificialQA automatically creates an organization and a project inside it.

  • The organization groups everything for your company (users, configuration, billing).
  • A project is a logical grouping for the tests of a specific AI agent or use case.

To get started, keep the default project; you can always rename it or create more later.

3

Connect your first agent

Go to the sidebar at Configuration → AI Agents and click + New Connection.

Configure the following fields based on your AI agent type:

If your AI agent is embedded in a webpage

  • Name: a recognizable name (e.g., "Commercial site").
  • Protocol: Browser.
  • Chat URL: the URL where the chat lives (e.g., https://mycompany.com/contact).
  • Execution runs on our cloud workers in headless mode (no visible window).
  • If the chat requires login, configure the Login Steps (selectors + actions to fill in user and password).

If your AI agent has an HTTP/API endpoint (Pro or Enterprise plan)

  • Name: a recognizable name.
  • Protocol: HTTP.
  • Base URL: the agent endpoint.
  • Authentication: Bearer Token, API Key, or Custom Headers as needed.
  • Message Configuration: where to inject the user input and which field to extract the response from.

Save the connection. Once saved, click Test Connection to validate that ArtificialQA can talk to your AI agent. If the test responds OK, you're set.

4

Create your first test cases

There are 4 ways to create test cases in ArtificialQA:

  • Generating them with AI (what we'll do in this Quickstart).
  • Manually one by one in the editor.
  • Importing an Excel or JSON with a downloadable template.
  • Pulling them from the public catalog of ArtificialQA (curated cases by industry).

To get going fast, let's use AI generation. Go to the sidebar at Test Design → AI Generation.

Configure the generation:

  • Industry: pick the industry that best matches your AI agent (Healthcare, Finance, Customer Support, Insurance, etc. — 15 available).
  • Test Type: Simple Q&A or Conversational (multi-turn).
  • Quantity: how many cases to generate (we recommend starting with 5).
  • Language: Spanish or English.
  • Additional Context: describe in natural language what kind of cases you want (e.g., "policy quote requests, complaints about delays, coverage questions").

Further down the form there are two optional advanced sections: User Tone (which tone to simulate the user with) and Risk Categories (focus on vulnerabilities like bias, hallucinations, prompt injection, etc.). For a first try you can leave them empty. More detail in Designing tests.

Click Generate. The AI generates the cases in seconds.

Generated cases land in a review view. For each one you can edit it (adjust input, expected response, asserts) and then decide where it goes: send it to your Test Cases catalog, send it straight to a specific Test Suite, or discard it.

AI Test Generator screen with configuration fields and generation result
The AI Generation form — Industry, Test Type, Quantity, Language, Context Sources, Additional Context, User Tone, and Risk Categories.
5

Run the test against your AI agent

Execution is the first of two final stages: your test cases run against the AI agent and the responses are captured. There's no scoring yet — that comes in step 6.

You need 3 things to run:

  1. A Test Suite with the cases you want to run (Test Design → Test Suites → New Suite, add the cases).
  2. A Test Plan that combines the Test Suite with the Agent Connection (Test Design → Test Plans → New Plan).
  3. Click Run on the created Test Plan.

The execution shows real-time progress (current case, completed, timings). When it finishes, the Run is marked as Completed and appears in the Ready to Evaluate tab on the Evaluations page — ready for the evaluators in the next step.

6

Evaluate the responses and view the report

Evaluation is the second stage: the responses you captured in the previous step are passed to the LLM evaluators for scoring. This is what turns a Run into an Evaluation Report with scores.

Go to Execution → Evaluations, click Evaluate on the run, and pick the evaluators you want to activate (you don't need all 17 — use the ones that make sense for your domain).

When evaluation finishes, click the run to see the Evaluation Report:

  • Overall score and pass rate.
  • Per-evaluator performance.
  • Detail per test case: what happened, which evaluator scored what, textual justification.
  • PDF Report button to export the full report.

What's next?

You closed your first cycle. The next things to look at are:

If something doesn't work