Skip to NavigationSkip to Content

23 Apr 2026

Industry:Technology & Communications

5 min reading time

AI Security Testing Cut 2,000 Findings Down to 48 Real Exploits

ai security findings

AI security often fails for a familiar reason: teams inherit thousands of findings but still don’t know which ones represent real risk. Scanning tools generate volume, but they rarely provide the context or proof needed to make confident decisions.

At Chaleit, we approach AI security through a secure-by-design lens, focusing on how systems are built, how components interact, and how they behave when AI introduces autonomous decision-making across those components and whether security assumptions hold up under attack. To support this approach, we developed a context-rich AI security capability that combines local LLM analysis, automated exploitation, and hands-on security engineering.

To show what this looks like in practice, we applied this method to DVWA (Damn Vulnerable Web Application), a widely used security benchmarking environment.

The objective was straightforward: determine whether context-aware analysis and exploitation could replace weeks of manual triage with faster, defensible results.

AI security challenges

Most security programs start with scanning tools that generate volume rather than clarity.

“In a field dominated by AI-based threats, we need to focus on what actually matters, what has context within the organisation and leads to real, material improvement,” explains Dan Haagman, CEO of Chaleit.  

In this benchmark, traditional SAST scanners flagged more than 2,000 potential issues across 319 files of PHP, JavaScript, and HTML. In a real organisation, this typically leads to:

  • 3–4 weeks of manual triage

  • additional time verifying exploitability

  • the majority of findings proving non-actionable

Security teams spend more time managing findings than reducing risk. As AI is introduced, these gaps are amplified, accelerating how quickly weak assumptions are exploited across systems.

Chaleit approach

Chaleit uses a context-aware AI security analysis and exploitation capability designed to identify, validate, and prioritise real risk in complex systems to restructure the workflow around proof rather than assumptions. Here’s what this means in practice:

Context-first analysis

The full codebase was analysed using private, local LLMs with architectural context: understanding data flows, component interactions, business logic, and how decisions are chained, delegated, and executed, particularly where AI components influence outcomes. This reduced 2,000 scanner findings to 249 contextually relevant issues.

Automated exploitation

Our team then attempted real exploitation, specifically targeting the assumptions these systems rely on to behave safely under real-world conditions. Of 55 critical/high-severity issues, 48 were successfully exploited, each with working proof-of-concept code and visual evidence.

Expert validation

Chaleit security engineers reviewed and confirmed the exploits, validated the business impact, and delivered prioritised remediation guidance aligned with material risk.

Results

The outcome was a fundamentally clearer AI security picture:

  • 98% noise reduction — From 2,000 potential issues to 48 confirmed exploits.

  • 0% false positives on critical findings — Every reported critical issue was demonstrably exploitable.

  • 87% exploitation success rate — 48 of 55 critical/high findings exploited automatically.

  • 100% OWASP coverage — All 19 OWASP vulnerability categories identified and tested.

  • 60% time savings — Validation cycles reduced from 5–6 weeks to 2–3 weeks

As Dan Haagman summarises:

“We shouldn’t be creating volumes of findings that sit in isolation. They need context, and they need to lead to real, material security uplift; otherwise, they're not actionable.” 

What this means for security teams

These results translate directly into how security work gets done day-to-day:

  1. Clearer priorities. Teams focus on a short list of proven risks instead of thousands of theoretical findings.

  2. Faster delivery. Security validation no longer stalls releases for weeks. Decisions are made with evidence, not assumptions.

  3. Higher confidence. Every critical issue includes working exploits and visual proof — removing “is this real?” debates.

  4. Better return on effort. Security investment goes into fixing what actually matters, not managing noise.

In our experience, most organisations are not struggling because they lack tools. They struggle because they are building on insecure foundations, and their security work produces volume without clarity.

In practice, this means validating whether systems operate within their intended boundaries, even as AI is layered on top and begins to influence decisions and interactions.

Chaleit’s approach is designed to focus on understanding context, validating risk, and helping teams act on what matters. The result is stronger decisions and security improvements that hold up in the real world.

Explore AI security

Get clear evidence on where your AI systems break and what to fix first.

Talk to an expert

About this article

Industries:

  • Technology & Communications

Service Areas:

  • AI Security

Related Customer Stories

Penetration testing case study

Financial Services

Financial Giant Cuts Remediation Costs by 50 Percent

adversarial simulation

Manufacturing

Partnership in Practice: How Trust and Collaboration Transformed a Manufacturing Client's Security Posture

purple teaming exercise case study

Utilities & Resources

How a Global Energy Firm Enhanced SOC Effectiveness Through Purple Teaming

A woman looking at notes in an office environment.

Law & Business Services

Leading Law Firm Transforms Document Security in Seven Days

Your Cookie Preferences

We use cookies to improve your experience on this website. You may choose which types of cookies to allow and change your preferences at any time. Disabling cookies may impact your experience on this website. By clicking "Accept" you are agreeing to our privacy policy and terms of use.