Key Takeaways

  • SOC analysts reject most AI tooling because of three failure modes: hallucination, context blindness, and black-box decisions they cannot audit or override.
  • The 80/20 approach works: fully automate the routine 80% of alerts while maintaining human-in-the-loop for the critical 20% that requires judgment.
  • LLM-assisted triage combines alert ingestion, enrichment, clustering, and confidence-scored routing to deliver 97% accuracy in automated decisions.
  • Trust is engineered through transparency — every AI decision must include reasoning, evidence references, confidence scores, and a one-click escalation path.
  • The goal is not replacing analysts. It is giving them back 6+ hours per day currently lost to obvious false positives and repetitive triage.

The Trust Problem in Security AI

Ask any SOC analyst what they think of AI-powered security tools and you will get one of two responses: a cynical laugh or a tired sigh. After years of vendors promising "AI-driven" everything — most of which turned out to be glorified if-else logic with a machine learning label — security practitioners have developed a healthy distrust of any tool that claims artificial intelligence.

This skepticism is earned. When a false negative means a missed breach and a false positive means waking someone up at 3 AM, analysts need to trust their tools implicitly. And most AI security products have failed to earn that trust for good reason.

But here is the uncomfortable truth: the alert volume problem is no longer solvable by humans alone. The average enterprise SOC generates 11,000 alerts per day. Analyst teams are not growing. Threats are. Something has to give — and if AI is the answer, it needs to work differently than everything that came before it.

Three Failure Modes of AI in Security Operations

Before we can build AI that analysts trust, we need to understand exactly why they distrust what exists today. Every failed AI security product falls into one or more of these three failure modes:

Failure Mode 1: Hallucination in High-Stakes Decisions

Large language models hallucinate. This is well-documented and well-understood. In a customer service context, a hallucinated response is embarrassing. In a security operations context, a hallucinated IOC can trigger containment of a production server, a false attribution can send an investigation down the wrong path for days, and a hallucinated "all clear" can leave an active threat actor undisturbed in your network.

The problem is not that LLMs hallucinate — it is that most AI security products do not account for hallucination in their architecture. They treat the model's output as ground truth rather than as a hypothesis that requires validation.

Failure Mode 2: Context Blindness

Generic AI models lack the specific context that makes triage decisions correct. They do not know that your finance team legitimately runs PowerShell scripts every Monday at 6 AM. They do not know that the "suspicious" binary is actually your custom monitoring agent. They do not know that the IP address flagged as malicious is your CDN provider's egress.

Context blindness leads to two outcomes: false positives that erode trust over time, and false negatives where legitimate threats are dismissed because the model lacked environment-specific knowledge to recognize them as anomalous.

Failure Mode 3: Black Box Decisions

When an AI system says "this alert is a true positive, severity high" without showing its reasoning, it is asking for blind trust. Security analysts do not give blind trust. They need to validate. They need to understand the reasoning chain so they can identify when the system is wrong — because it will be wrong, and the analyst needs to catch those cases.

A black box that is right 95% of the time sounds impressive until you realize that the 5% it misses could be the breach that ends your company. Analysts need to see the work, not just the answer.

Our Approach: Human-in-the-Loop for the 20%

The fundamental design principle behind trustworthy AI triage is simple: do not try to replace human judgment. Augment it. The goal is not an autonomous system that handles everything — it is an intelligent system that handles the obvious 80% autonomously and presents the complex 20% to humans with full context, enrichment, and a preliminary assessment they can accept, modify, or reject.

This is not a compromise. It is the architecture that actually works in production. Here is why:

Technical Architecture: How LLM-Assisted Triage Works

The triage pipeline consists of four stages, each building on the last. At every stage, the system generates audit-friendly logs that show exactly what happened and why.

Alert Ingestion Enrichment Clustering Routing

Stage 1: Alert Ingestion and Normalization

Alerts arrive from multiple sources — SIEM, EDR, email security, cloud security posture management — in different formats, with different severity scales, and different levels of detail. The ingestion layer normalizes everything into a common schema.

This normalization is not AI-driven. It is deterministic. Every alert gets mapped to a consistent structure with fields for source, timestamp, affected assets, observables (IPs, hashes, domains, users), raw evidence, and source-assigned severity. This structured data is what the AI operates on — not raw, unstructured alert text.

Stage 2: Automated Enrichment

Before the LLM ever sees an alert, it is enriched with contextual data from every available source:

All enrichment is executed via deterministic API calls to connected systems. The AI does not generate enrichment data — it consumes it. This eliminates the hallucination risk at the data-gathering stage entirely.

Stage 3: Intelligent Clustering

Individual alerts rarely tell the full story. A single failed login is noise. Fifty failed logins from the same source IP, followed by a successful login, followed by unusual data access — that is an attack chain. The clustering stage groups related alerts using:

The LLM's role here is specifically the semantic layer — identifying relationships that deterministic rules miss. But every cluster it proposes includes the reasoning chain and the specific evidence that links the alerts together.

Stage 4: Confidence-Scored Routing

This is where the triage decision happens. The LLM receives the normalized, enriched, clustered alert data and produces a structured assessment:

{
  "verdict": "false_positive",
  "confidence": 0.96,
  "reasoning": [
    "Alert source: Sentinel rule 'Suspicious PowerShell Execution'",
    "Asset: FINANCE-SRV-03 (Tier 2, Finance department)",
    "Historical: This exact alert has fired 47 times in 30 days",
    "Previous dispositions: 47/47 resolved as false positive",
    "Pattern: Scheduled task 'Monthly-Report-Gen' runs at this time",
    "Enrichment: No IOC matches, no anomalous network activity"
  ],
  "action": "auto_resolve",
  "escalation_trigger": "Override if confidence < 0.90"
}

The routing logic is straightforward:

Confidence Scoring: When the AI Should Escalate vs Resolve

The confidence score is not a single number pulled from the model's logits. It is a composite score calculated from multiple signals:

This multi-factor confidence scoring means the system naturally becomes more conservative in novel situations and more decisive in well-understood ones. It does not need explicit rules for every scenario — the confidence calculation handles edge cases automatically.

Building Trust Through Transparency

Every automated decision in the system is accompanied by a full audit trail. Analysts can see:

This transparency serves two purposes. First, it lets analysts validate decisions, catch errors, and maintain situational awareness even over automatically-handled alerts. Second, it builds trust over time. When analysts see the system's reasoning and it consistently matches what they would have done — they start trusting it. Not because someone told them to. Because they verified it themselves, repeatedly.

What 97% Accuracy in Auto-Triage Actually Means

When we say our AI triage agents achieve 97% accuracy, here is exactly what that means and what it does not mean:

What it means

What it does not mean

The Result: Analysts Working on What Matters

The end state is not a SOC without analysts. It is a SOC where analysts spend their time on work that actually requires their expertise. Instead of spending 6+ hours per shift on repetitive triage — the same false positive they have closed 200 times, the same benign alert they already know the answer to — they spend that time on:

This is not about reducing headcount. It is about making existing headcount effective at a scale that would be physically impossible without AI assistance. An analyst who can focus on 40 meaningful alerts per day instead of drowning in 400 undifferentiated ones is a better analyst — more engaged, more accurate, and far less likely to burn out and leave.

Getting Started: The Shadow Mode Approach

No organization should deploy AI triage with autonomous action on day one. The path to trust is gradual and verifiable:

  1. Week 1-2: Shadow mode. The AI sees every alert, makes recommendations, but takes no action. Analysts do their normal work and the system's recommendations are compared against actual dispositions.
  2. Week 3-4: Advisory mode. The AI presents its recommendations alongside alerts. Analysts can accept or override. Every override improves the model's environment-specific knowledge.
  3. Week 5-6: Supervised automation. The AI begins auto-resolving alerts above 0.98 confidence. Analysts spot-check 10% of automated decisions daily.
  4. Week 7+: Full operation. The AI handles the 80% autonomously. Confidence thresholds are tuned based on performance data. The analyst team focuses on what they do best.

This approach lets analysts see the system earn their trust incrementally. No leap of faith required. Just data, transparency, and consistent performance over time.

See How AI Triage Works in Practice

Our AI Workflow solutions are built on these exact principles — transparent, confidence-scored, human-in-the-loop where it matters. See how Quandry's automation layer transforms SOC operations without asking your analysts to trust a black box.

See AI Workflow Solutions →