App Development

AI applied to QA at GooApps®: How we moved from “fast” tests to reliable tests

Introduction: AI in QA is about judgment, not speed

At GooApps®, applying AI to QA has never been about speed. It has always been about judgment and control. As a QA professional, my goal is not to have AI “do the testing for me,” but to use it to improve coverage, detect risks earlier, and standardize test case quality without losing traceability or accountability.

The human problem we aim to solve is clear: test case generation is often costly, repetitive, and highly dependent on context. This leads to inconsistencies across projects and wasted time on low-value tasks. AI can help—but only if it is integrated into a proven, validated, and explainable workflow.

From isolated prompts to a controlled workflow

Over the past year, I have continuously used AI across multiple projects to:

  • Create and expand test cases from functional documentation
  • Detect alternative scenarios, edge cases, and implicit risks
  • Adapt test cases to standardized formats compatible with X-Ray

What started as occasional prompt usage evolved into a fully automated workflow that allows us to:

  • Analyze functional and technical documentation using AI
  • Automatically classify whether the task relates to an App, Backoffice, CRM, WebApp, or API
  • Generate advanced test cases based on product type
  • Transform outputs into CSV format for X-Ray import
  • Automatically create and link test-type tasks in Jira to the parent ticket

AI does not operate as a black box here. It functions as a component within a controlled system.

From improvisation to structured methodology

Initial AI Approach Structured GooApps® Approach
Isolated prompt Replicable automated workflow
Manually generated cases Pre-classification by product type
Inconsistent format X-Ray-compatible standardization
Executor-dependent Traceable and auditable process
Informal review Mandatory human validation
The shift was not technological. It was methodological.

Data, risks, and validation: the QA’s critical role

The primary inputs are functional and technical documentation. This introduces clear risks when applying AI to testing:

  • Misinterpretation of scope
  • False certainty generated by the model
  • Incomplete coverage if documentation is outdated or partial

For this reason, our core principle is simple and non-negotiable: no AI-generated output is considered valid without human review.

Every output goes through:

  • Manual review for coherence and alignment with the ticket
  • Validation against real acceptance criteria
  • Iterative prompt adjustments when ambiguities are detected
  • Monitoring of quality metrics (reduced rework, improved coverage, increased reuse)

AI accelerates the process. Accountability always remains with QA.

Control, explainability, and what happens when AI fails

Control over the workflow always remains with the user. AI proposes, but it does not decide.

We can assert that results are explainable because we know:

  • Which documentation was used
  • Which prompts and rules were applied
  • Which transformations occurred before reaching X-Ray

When the model fails—irrelevant cases, incorrect assumptions, noise—the system is not patched superficially. The workflow is adjusted, prompts are refined, or scope is limited. That learning becomes part of the process and improves future executions.

From chaos to system: designing a QA flow that is explainable, validatable, and improvable

One of the biggest risks of applying AI in QA is remaining at the level of isolated solutions: loose prompts and inconsistent outputs. To avoid this, the focus was on designing a structured workflow, not on “asking AI better questions.”

The workflow always starts from a specific task and automates test creation from the beginning, following GooApps® internal standards. AI does not replace QA judgment; it executes defined steps within a clear process.

Intelligent classification before test generation

Before generating any test case, the system automatically classifies the task type (API, App, CRM, Backoffice, or WebApp) by analyzing documentation and functional context.

This step is critical. It makes no sense to apply the same testing logic to an API and a mobile application.

Classification prevents one of the most common AI-in-QA errors: generating generic test cases that ignore product nature.

API testing: beyond the happy path

When the task is API-related, the workflow generates:

  • Exhaustive functional test cases
  • Negative scenarios and error validations
  • Cases prepared for execution in Postman

The goal is not only to validate correct responses, but also to cover structure, HTTP status codes, invalid inputs, and consistency with functional documentation.

CRM, Backoffice, and WebApps: structure over volume

In these environments, AI works with extended, structured prompts. Test cases follow a consistent order:

  • Visual verification
  • Core functionalities
  • Detailed validations
  • Error handling
  • Security and permissions
  • Responsiveness and usability

AI shifts from being a text generator to becoming an assistant aligned with internal QA standards.

Mobile apps: focus on real-world context

For mobile applications, priority is given to:

  • Permissions and connectivity states
  • Navigation and real user flows
  • Boundary and edge-case behaviors

The objective is not to replicate manual testing, but to expand coverage where time constraints typically limit us.

Repository analysis: when QA enters the codebase

One differentiating step in the workflow is repository code analysis related to the task. AI receives both documentation and relevant source code.

This allows us to:

  • Detect discrepancies between documentation and implementation
  • Identify fragile areas
  • Strengthen negative testing

AI does not replace technical judgment. It reduces exclusive dependency on incomplete documentation.

The final outcome: executable test cases

The final step transforms generated knowledge into executable test cases ready for X-Ray import, in Jira-compatible CSV format.

The cycle is complete: Documentation → analysis → generation → validation → execution

No manual copy-paste.
No later reinterpretation.
Direct traceability between task, test case, and execution.

Key learning from QA

The main learning has not been technical, but methodological: AI in QA only creates value when the process is explainable, validatable, and improvable.

If we do not understand why a test case exists, we should not execute it.

At GooApps®, using AI in QA does not mean losing control. It forces us to define our quality criteria more precisely. And that is what truly elevates testing standards.

Frequently Asked Questions

Can AI replace QA in generating test cases?

No. AI can accelerate creation and expand coverage, but scope interpretation, validation, and accountability remain human responsibilities.

What is the main risk of using AI in QA?

False confidence in coverage. Without human review, AI may generate structurally correct but conceptually incorrect test cases.

Why is product-type classification important before generating tests?

Because each type (API, App, CRM, WebApp) requires different testing logic. Without classification, test cases tend to become generic and ineffective.

What does repository analysis add to AI-driven testing?

It helps detect mismatches between documentation and implementation and enables test case generation based on real technical risks.

What is the key learning when applying AI to QA?

AI only delivers value within a structured, traceable, and reviewed process. Without methodology, it simply generates technical debt.

 

Take the next step

Complete the form and GooApps® will help you find the best solution for your organization. We will contact you very soon!





    Contact


      Privacy policy.