← All articles

Why test automation fails in most organisations

by Rainer Haupt

70 % of all test automation initiatives fail to meet expectations. George Ukkuru has analysed more than 100 initiatives over 15 years and documented the pattern across industries. The tools are rarely the problem. The architecture is missing.

Buy a tool, fail, buy the next one

A global investment platform invests USD 1.2 million in Selenium-based test automation. Sixteen months later, the project is abandoned. Brittle selectors, runaway maintenance, no chance to keep up with changes to the application.

An insurance company expects USD 340,000 in annual savings from automation. The reality: USD 480,000 in annual maintenance cost. Releases slow down by 23 %. Net loss instead of savings.

These cases are not exceptions. The pattern repeats across industries — financial services, pharma, public administration.

Five causes that reinforce each other

Missing test strategy

Most organisations start automating before a test strategy is in place. They automate what takes the most manual effort. Typically GUI tests. The result is an inverted test pyramid: many slow, brittle end-to-end tests, few fast unit and integration tests.

Google quantified this: one team needed over seven days to bring an E2E suite from 5 % to 89 % pass rate. The milestone slipped by a week. The tests ran — but they tested the wrong things at the wrong level.

Tool fixation

Proprietary test tools cost EUR 50,000 to 500,000 per year in licence fees. They promise “testing without programming”. For demos, that works. For 10,000 tests in production, it does not.

Every new tool brings new specialist knowledge. Teams need Tosca experts, UFT specialists, Ranorex insiders — all skills hard to find on the market. Organisations run an average of 7.4 testing tools in parallel; 75 % lose 6–15 hours a week to that sprawl.

Flaky tests destroy trust

At Google, 16 % of all tests are affected by flakiness. 84 % of all transitions from “pass” to “fail” come from unstable tests, not from real defects. Atlassian loses over 150,000 developer-hours per year to reruns.

When tests fail at random, teams ignore the results. Google compares the effect to pilots disabling alarms because of too many false positives. Above 1 % flakiness, the value of a test suite drops sharply.

Test code receives less care than production code

No code review, no modularisation, no versioning. Test debt grows faster than the value. Once the suite reaches a certain size, every change to the system becomes a risk: nobody knows which tests will break and why.

A Fortune 500 manufacturer invested USD 890,000 over 18 months in 1,200 automated test scenarios. Final cost: USD 2.3 million. ROI: minus 180 %. The suite became a maintenance burden, not an accelerator.

The Capgemini World Quality Report 2023-24 confirms the pattern: only 3 % of organisations have fully integrated their regression suites into delivery pipelines. Half the tests run unstably, and nobody knows which failures matter.

QA as a silo

In many organisations, quality assurance is a separate department. Engineering “throws code over the wall” to QA. Late defect detection, long feedback cycles, blame instead of shared responsibility.

The DORA reports show the difference: elite performers deploy 973× more frequently than low performers. They achieve a 3× lower change failure rate. The key: testing is part of development, not a downstream step.

The negative feedback loop

These five causes reinforce each other. Missing strategy leads to wrong tests. Wrong tests become unstable. Unstable tests lose the team’s trust. The team invests less in test quality. Test debt grows. Every change becomes riskier. Pressure rises to evaluate the next tool.

Martin Fowler quotes Justin Searls: “Almost no team writes tests that set clear boundaries, run fast and only fail for useful reasons.” That is not a tool problem. It is an architecture problem.

What architecture does differently

Anyone who treats test automation as software engineering treats test code with the same discipline as production code. Modularity, code review, versioning, CI/CD. No special status for test code, no quality exceptions.

A well-designed test architecture addresses the five causes:

  • Test strategy instead of tool fixation — the right distribution across test levels, not the next tool
  • Reusable building blocks — keywords and libraries that work across projects
  • Stability by design — API-first instead of GUI-only, deterministic tests instead of flakiness
  • Readable tests — scenarios that functional testers and auditors also understand
  • Integrated, not isolated — tests as part of development

A large German social-insurance organisation took this path: from a few manual tests to over 10,000 automated tests. Nightly regression across 20 parallel CI/CD jobs. Zero licence cost. New team members productive in days, not weeks. The tests double as documentation for auditors.

Three questions for your test landscape

  1. Can you quantify what test maintenance costs per month?
  2. Does the team trust the test results, or are failing tests routinely ignored?
  3. Can a new team member understand the tests without asking the author?

If any of those three answers gives pause, the test architecture deserves a closer look.

Sources

Request callback