iTnews Asia
  • Home
  • News
  • Data and Analytics

Why AI code quality is the next critical enterprise risk

Why AI code quality is the next critical enterprise risk

While AI is accelerating how software is built, organisations may not fully understand the risks and blind spots.

By Damien Wong on Apr 14, 2026 10:30AM

OpenClaw has quickly become one of the most talked-about tools in software development, with its rapid adoption positioning it as an “always-on” junior developer for engineering teams. It can generate code on demand, accelerate development cycles, and reduce the burden of repetitive tasks.

That momentum reflects a broader shift already underway. In many enterprises today, AI-assisted coding is estimated to account for more than 40 percent of new code, redefining how software is built. But while the pace of creation has changed, the way organisations validate that code has not kept up.

AI-generated code often looks correct, passes standard checks, and integrates cleanly into existing systems. Yet beneath that surface, it can introduce dependencies that do not exist or behave as expected, apply logic that only holds under ideal conditions, or introduce subtle inconsistencies with existing systems and business rules. These issues rarely surface in controlled development environments, where inputs are predictable and testing is scoped. They tend to emerge only in production, where real-world variability, scale, and user behaviour expose gaps that were never accounted for.

These are not obvious defects but structural weaknesses that traditional quality mechanisms were never designed to detect.

Why traditional QA creates a dangerous blind spot

Traditional quality assurance and CI/CD pipelines were built for a world in which code was written by humans, reviewed by humans, and tested under known conditions. In that environment, validating performance and expected output was sufficient to ensure reliability.

But with AI-generated code, the model is put to the test.

Unlike human developers, AI does not reason through intent. It generates output based on patterns and probabilities, often optimising for functionality rather than soundness. This can result in code that is syntactically valid but logically fragile, or functionally correct but misaligned with architectural standards, security principles, or business logic.

- Damien Wong, Senior Vice President, APJ, Tricentis

Recent developments around OpenClaw illustrate this phenomenon clearly. As one of the fastest-growing open-source projects on GitHub, amassing over 100,000 stars within weeks, its rapid adoption has outpaced the safeguards needed to manage it.

In practice, this has already led to real-world issues. In one instance, the agent executed commands that exposed sensitive files while in another, vulnerabilities in third-party “skills” introduced malware risks and unsafe data handling

As AI accelerates development, the volume of code increases, and so does the likelihood that small, hidden defects accumulate. These issues often surface only after deployment, where they are harder to trace, more expensive to fix, and more damaging to business operations and customer trust.

From validation to accountability If traditional QA cannot keep up with how AI-generated code behaves, the answer is not more testing, but smarter, continuous validation embedded across the development lifecycle.

The first step is traceability. Organisations need visibility into how code is created, including whether it originates from human developers or AI systems. This enables teams to identify patterns in defects, refine how AI tools are used, and strengthen guardrails where needed.

The second is moving beyond basic validation to behavioural testing. Traditional unit tests validate expected scenarios, but they rarely expose how code behaves under stress or ambiguity. Techniques such as property-based testing and chaos engineering are better suited to uncover these weaknesses, ensuring that code is not only correct in theory but also resilient in practice.

At the same time, human judgment becomes more important, not less. AI can generate and test code at scale, but it cannot fully account for business context or risk. The most effective model is a layered one, where AI handles execution and scale, while humans remain the decisive layer of accountability.

Finally, validation must extend beyond deployment. As AI systems evolve, so do the outputs they generate. Continuous monitoring and regression testing ensure that code remains reliable over time, not just at release.

What must CIOs do now

This starts with a shift in focus. Not all AI-generated code carries the same risk, so CIOs should prioritise assessing critical systems that directly impact business performance, customer experience, or regulatory compliance, reinforcing strategic control over AI risks.

The next shift is organisational. AI-generated code has blurred traditional boundaries between development, testing, and security. Treating these as separate functions creates gaps in accountability. CIOs need to bring these disciplines together, not through process alone, but through shared ownership of outcome.

Equally important is shifting how testing is perceived. In many organisations, testing is still seen as a bottleneck that slows release cycles. In an AI-driven environment, that mindset becomes a liability. The cost of a single failure caused by unvalidated AI-generated code can far outweigh the perceived gains in speed.

As AI-generated code increasingly underpins core systems, testing and validation must move beyond engineering teams and become a board-level imperative, with clear ownership and accountability at the highest levels of the organisation.

Trust as the differentiator

The rise of tools like OpenClaw is exposing a deeper issue. AI is accelerating how software is built, but not how well organisations understand what they are deploying. In this environment, speed is no longer the competitive edge. Trust is.

But the organisations that succeed will not be those that move fastest without constraint. They will be the ones who can move quickly while maintaining control.

Responsible AI is not proven in governance frameworks alone. It is proven in execution, in systems that run reliably, services that perform consistently, and experiences users can trust.

As AI-generated code becomes embedded in core systems, trust is no longer an abstract principle but an operational requirement. In an AI native world, that trust is what ultimately separates organisations that can scale with confidence from those that cannot.

Damien Wong is Senior Vice President for APJ at Tricentis.

To reach the editorial team on your feedback, story ideas and pitches, contact them here.
© iTnews Asia
Tags:
ai ai trends data and analytics

Related Articles

  • AI readiness could be APAC’s biggest challenge despite rapid adoption
  • Before scaling AI, what must APAC enterprises fix first?
  • Infrastructure gaps, not algorithms, are hindering healthcare innovation
  • A solid data security foundation is fundamental to trusted AI
Share on Twitter Share on Facebook Share on LinkedIn Share on Whatsapp Email A Friend

Most Read Articles

Before scaling AI, what must APAC enterprises fix first?

Before scaling AI, what must APAC enterprises fix first?

AI readiness could be APAC’s biggest challenge despite rapid adoption

AI readiness could be APAC’s biggest challenge despite rapid adoption

Infrastructure gaps, not algorithms, are hindering healthcare innovation

Infrastructure gaps, not algorithms, are hindering healthcare innovation

Can Agentic AI become the next frontier of enterprise transformation?

Can Agentic AI become the next frontier of enterprise transformation?

All rights reserved. This material may not be published, broadcast, rewritten or redistributed in any form without prior authorisation.
Your use of this website constitutes acceptance of Lighthouse Independent Media's Privacy Policy and Terms & Conditions.