how to start with e2e testing from scratch?

An Octomind tutorial for developers who want start end-to-end testing

octomind app screenshot 10/24

When we're under pressure to ship features faster - testing quality is usually the biggest casualty. Similarly, when applications with a lack of unit and integration tests start to suffer from bugs that should’ve been caught earlier, end-to-end tests are often brought in to catch regressions as they can cover the most critical user flows faster than ramping up TDD efforts.

But without a dedicated testing team, learning, configuring, and integrating automation frameworks such as Playwright also takes developers away from writing code and shipping features.

That’s why we’re building Octomind. We use AI to automate the creation of an end-to-end test suite from scratch, and without any prior automation testing experience.

Let me take you through what getting up to speed with Octomind looks like so you can see if it’s something you’d want to try out for your own websites or apps. 

  1. Getting started
  2. Test creation
  3. Test running
  4. Test coverage
  5. Test status & debugging

1. Getting started

We make it as easy as possible to get started. Just enter a publicly accessible URL for the site you want to test and you’re away.


octomind app sign-in, 10/24
octomind app sign-in, 10/24


But Octomind isn’t limited to public URLs. Only the onboarding flow needs that to create a  ‘known good’ state or ‘baseline’ in testing terminology for future deployments to be tested and verified against. The entire test suite can be run in any internal environment (including development) with a single API call or SDK call.


octomind app test report, 10/24
octomind app test report, 10/24

But this is jumping ahead a bit and I’ll cover this in the Test running section.


Getting back to the sign-up flow, we suggest using the URL as the project name, but totally up to you. Then we’ve got the standard create account and log in flow, and once logged in Octomind’s AI agent (I’ll call it Agent from here on in) checks it can access your site, then gets to work creating test cases.

octomind app sign-up, 10/24
octomind app sign-up, 10/24

2. Test creation

So how does Octomind know what to test?


Our LLMs are continuously trained and evaluated against thousands of different types of sites and applications. With this data set, it can classify each site type and create a unique context for deriving relevant test cases.


AI Agent generating tests for an e-commerce site 10/24
Agent generating tests for an e-commerce site, 10/24


The Agent always starts with detecting cookie banners and required login forms because unless you’re testing the login flow itself, most tests will require one or both of these to complete before every test run.


AI auto-generated cookie banner test, 10/24
auto-generated cookie banner test, 10/24

If it finds a required login form, you’ll be prompted to enter testing credentials - or just leave them blank to test unauthenticated.

credentials input for a login test, 10/24
credentials input for a login test, 10/24


You can also enter them later from the settings panel for use in a future login test.


test account in octomind app account settings, 10/24
test account in octomind app account settings, 10/24


Octomind then analyzes your site to determine the most logical initial set of test cases, with the test stack view showing its progress and the steps for each test case.


test stack view showing Agent progress when generating, 10/24
test stack view showing Agent progress when generating, 10/24


The Agent typically starts with three to four test cases. This isn’t a system limitation, but a practical one - as the best way to scale end-to-end test coverage is to start small, then expand incrementally.

The initial test implementation can take a bit of time (we’re working on it!). As the Agent completes each test case, the number of total and active tests is updated active meaning the test executed and passed successfully.

project overview page with active test cases, 10/24
project overview page with active test cases, 10/24

3. Test running

Once all test cases are implemented, Octomind runs the active tests - creating the first test report.


first test results after sign-up, 10/24
first test results after sign-up, 10/24


To go deeper, the test report details page provides a step-by-step visual representation for each test case.


test run snapshots in test report detail, 10/24
test run snapshots in test report detail, 10/24


Or you can debug a specific test case by inspecting the Playwright Trace Viewer or run the agent-generated code locally for step-debugging.

test run detail, debugging tab, 10/24
test run detail, debugging tab, 10/24


Test runs can be scheduled.

test run scheduler, 10/24
test run scheduler, 10/24



And triggered on-demand in a post-deployment task in CI/CD via our integrations or API.


API key for CI integrations in Octomind app settings, 10/24
API key for CI integrations in Octomind app settings, 10/24

4. Test coverage

This is where things get interesting. You can increase your test coverage simply by asking the Agent to generate more tests.

telling the AI agent to generate sequential tests , 10/24
telling the Agent to generate sequential tests , 10/24


You can also get suggestions for new tests or you create a custom test case by prompting the Agent.

prompting the AI agent to generate a custom test
prompting the Agent to generate a custom test, 10/24



Test cases arranged in a hierarchy, visually showing the execution sequence from parent to child while also providing the Agent with context for the new test cases.

dependency view showing the execution sequence of test cases, 10/24
dependency view showing the execution sequence of test cases, 10/24

5. Test status and debugging

Active indicates the test step design worked as expected, whereas an inactive test (off) means the Agent hit a stumbling block that prevented it from reaching the test verification step. Nothing is wrong with your site - the Agent just needs help refining its test strategy. It's labeled as 'steps need review'.

inactive generated test where steps need review, 10/24
inactive generated test where steps need review, 10/24


There are numerous options for fixing an inactive test.


The agent-generated prompt can be altered to include steps it didn’t think of.

debugging an agent-generated prompt of a test case, 10/24
debugging an agent-generated prompt of a test case, 10/24


The locator (element to interact with) can be adjusted (e.g. if the wrong button was clicked).

hover in snapshot to activate the visual locator picker to change a locator, 10/24
hover in snapshot to activate the visual locator picker to change a locator, 10/24


The assertion and interaction methods can be changed (e.g. if it’s not verifying the state of the correct element).


changing interaction & assertion method with visual locator picker, 10/24
changing interaction & assertion method with visual locator picker, 10/24


You can delete redundant steps.


test step deletion, 10/24
test step deletion, 10/24


Plus you can run and debug the test locally.


test step deletion, 10/24
local test running & debugging, 10/24

Once the test steps have been fixed, click ‘save and run’ to execute the test and if successful, it will change to an active test.

That’s a pretty comprehensive start-to-finish tour of Octomind.  Hopefully you’re ready to give it a try.

And if you have any questions, you can email me at daniel @ octomind.dev or join our Discord community.

maximilian link headshot
Daniel Roedler
Co-founder & CPO
see more blogs
interested in new features?
Subscribe to our monthly product update!
Thank you! We got your submission!
Oops! Something went wrong.