Threat Research Center →Explore

TL;DR

  • The integration of Generative AI (GenAI) and Agentic AI (autonomous agents) has fundamentally altered the threat landscape.

  • To defend against AI cyber attacks like deepfakes or tool chaining, security leaders must combine 1) automation that replaces humans, and 2) integration that puts humans in the loop.

  • We must weaponize AI to defend against AI.

By the Numbers (2025 Reality)

  • Global Impact: The total global cost of cybercrime is estimated to have reached $10.5 Trillion annually by the end of 2025.

  • Widespread Vulnerabilities: 97% of organizations surveyed in IBM’s Cost of a Data Breach Report experienced AI-related security incidents and lacked proper AI controls.

  • Attack Volume: In 2025, approximately 41% of new ransomware families utilized AI-driven code mutation (polymorphism) to evade traditional defenses.

Strategic Preamble: Four Core Considerations for the C-Suite

Before we dismantle the technical threats, every security leader must wrestle with these four paradigm-shifting questions. Your answers will define your organization's resilience in the AI era.

Diagram with quadrants that have the following text: 
1. The Velocity Gap: If an AI agent can execute an attack chain in milliseconds, is our "1-10-60" benchmark (1 minute to detect, 10 to investigate, 60 to remediate) still relevant? Or is it already too slow? 
2. The Identity Crisis: In a world where voice, video, and writing style can be perfectly cloned, how do we prove "who" is on the other end of the connection? "Zero Trust" must now evolve into "Zero Implicit Identity." 
3. The Integrity vs. Confidentiality Shift: Traditional security focuses on preventing data theft. AI introduces the risk of data poisoning. Are we prepared for an attack that doesn't steal our data, but subtly changes it to corrupt our own AI models? 
4. The Autonomy Dilemma: To fight machine-speed attacks, we must authorize defensive AI to take autonomous action (e.g., shutting down a production server). Are we culturally ready to trust the algorithm with the "kill switch"?
  1. The Velocity Gap: If an AI agent can execute an attack chain in milliseconds, is our "1-10-60" benchmark (1 minute to detect, 10 to investigate, 60 to remediate) still relevant? Or is it already too slow?

  2. The Identity Crisis: In a world where voice, video, and writing style can be perfectly cloned, how do we prove "who" is on the other end of the connection? "Zero Trust" must now evolve into "Zero Implicit Identity."

  3. The Integrity vs. Confidentiality Shift: Traditional security focuses on preventing data theft. AI introduces the risk of data poisoning. Are we prepared for an attack that doesn't steal our data, but subtly changes it to corrupt our own AI models?

  4. The Autonomy Dilemma: To fight machine-speed attacks, we must authorize defensive AI to take autonomous action (e.g., shutting down a production server). Are we culturally ready to trust the algorithm with the "kill switch"?

The Rise of AI: Why Incident Response Must Evolve

The New Threat Landscape: From "Scripted" to "Agentic"

We must distinguish between GenAI (creating content) and Agentic AI (executing tasks). While GenAI creates convincing lures, Agentic AI can autonomously navigate our networks, chaining tools together to achieve an objective without human intervention.

Diagram with the following text: 
1. Social engineering- AI-ingested public data used for spear-phishing campaigns.
2. Deepfake & Voice Cloning - Synthetic visuals and voice used to create nearly-undetectable fraud
3. Polymorphic Malware - Rewriting malware code
4. Agentic Attack Chain - Connecting malicious agents to execute data exfiltration

Hyper-Personalized Social Engineering (The GenAI Threat)

The era of "bad grammar" phishing is over. LLMs ingest public data (LinkedIn, X, corporate bios) to generate spear-phishing campaigns that are indistinguishable from legitimate executive communication. By late 2025, roughly 82% of phishing campaigns showed signs of AI-generated content.

The Help Desk Siege: Deepfake & Voice Cloning

This is the fastest-growing vector. Attackers are using Real-Time Voice Cloning (RTVC) to bypass MFA. By capturing just 3-5 seconds of an executive’s voice from a podcast or YouTube video, they can generate a synthetic clone that mimics tone, accent, and cadence.

  • The Attack: The attacker calls the Service Desk posing as a stressed executive who "lost their phone," demanding an immediate MFA reset.

  • The Impact: Deepfake-enabled fraud caused over $200 million in confirmed losses in Q1 2025 alone.

Polymorphic Malware (The "Chameleon" Code)

Attackers are using private, unaligned LLMs to rewrite malware code on the fly. Each time the malware infects a new machine, it changes its file structure and hash while keeping the payload intact. This renders traditional signature-based AV (antivirus) useless.

The Agentic Attack Chain

"Agentic" AI implies a system given a goal (e.g., "Exfiltrate financial data") that autonomously figures out the steps.

  • Tool Chaining: Malicious agents can connect unrelated API endpoints to create a kill chain—for example, using a calendar API to find when an admin is Out of Office, then triggering a password reset flow that requires manual approval, knowing the admin isn't watching.

Executive Takeaway: We are defending against software that can "reason" its way through our defenses 24/7.

The Paradigm Shift: Evolving the Incident Response Team

The traditional "Observe -> Orient -> Decide -> Act" (OODA) loop is too slow. IR teams are evolving and expanding their role beyond log analysis to include deeper expertise in AI‑enabled behaviors and attack patterns.

The Evolution of Response Capabilities

To visualize where your organization stands vs. where it needs to be:

Stage

1. Legacy IR

2. AI-Assisted (Current)

3. Agentic Defense (Target)

Speed

Human Speed (Hours/Days)

Accelerated (Minutes)

Machine Speed (Milliseconds)

Role of AI

None / Basic Rules

Co-Pilot / Summarizer

Autonomous Agent

Human Role

Analyst & Executor

Decision Maker (In-the-Loop)

Supervisor (On-the-Loop)

Balancing Augmentation vs. Automation (The "Safe Kill Switch")

AI systems are capable of assisting human practitioners like an intern that never sleeps. However, it is critical for security teams to differentiate what to automate from what to augment.

Zone 1: Full Automation (The "Machine" Tasks)

Criteria: Tasks that are data-heavy, repeatable, measurable, and have a low cost of error.

Action: Threat enrichment, log parsing, and alert deduplication are prime candidates. These are pattern-driven processes where consistency outperforms creativity.

The "Kill Switch": We can automate the isolation of high-tolerance assets (e.g., employee laptops). If an AI detects Agentic behavior on a laptop, it should have the authority to kill the network connection immediately.

Zone 2: AI Augmentation (The "Human" Tasks)

Criteria: Tasks demanding context, intuition, or ethics.

Action: Incident scoping, attribution, and server shutdowns. Here, AI should augment the human by surfacing indicators, suggesting next steps, or summarizing findings, but the human retains "decision authority."

  • Why: An AI might see a massive data transfer as "exfiltration" and shut it down, not realizing it is a critical, scheduled backup. Context is king.

Graphic with two sides and the following text: 
1. Full Automation - Threat Enrichment, Alert Deduplication, Log Parsing, High-Tolerance Assets
2. AI Augmentation - Incident Scoping, Attribution, Server Shutdowns, Surfacing Indicators

New Roles and Skill Sets

AI Forensics Specialist: Traditional forensics looks at disk images. AI forensics requires analyzing vector databases, prompt logs, and "chain-of-thought" reasoning logs to understand why an AI agent hallucinated or was manipulated.

Adversarial Prompt Engineer: A Red Teamer specialized in "poisoning" RAG databases and testing internal models for prompt injection vulnerabilities.

AI vs. AI: Tactical Defense Strategies

We cannot fight AI with spreadsheets. We must weaponize AI for defense to regain the advantage. Here is how we detect and prevent the top threats.

  1. Countering Phishing: Linguistic & Behavioral Analysis

  • Detection: Deploy email security tools using Natural Language Understanding (NLU) to flag "urgency" and "coercion" patterns.

  • Relationship Mapping: AI maps normal communication strength. If a vendor suddenly emails the CFO with a high-urgency request after 3 years of silence, the AI flags the anomaly regardless of the email content.

2. Countering Deepfakes: Identity Defense

  • Spectral Analysis: Integrate voice analysis software into the VoIP/Call Center stack that looks for "synthetic artifacts"—micro-pauses or frequency gaps that human vocal cords don't produce.

  • The "Challenge Phrase" Protocol: Establish a non-digital "safe word" for executives.

  • Policy: Strict "Out-of-Band" Verification. No sensitive action is ever authorized on the incoming channel. If the CEO calls on Zoom, the verification must happen via a callback to their registered cell phone.

3. Countering Polymorphic Malware: Behavioral Anchoring

  • Abandon Signatures: Stop relying on file hashes. Tune EDR to flag behaviors (e.g., "Process X is attempting to touch ").

  • Memory Scanning: Polymorphic code often unpacks itself only in memory. Frequent memory scanning is critical to catch the payload before it writes to disk.

4. Agentic Deception (Honeypot 2.0)

  • Scenario: An attacker's AI scans our network and finds a "Database Admin" agent (which is actually a trap). The defensive agent engages the attacker, providing fake data and wasting their resources while simultaneously tracing their origin.

Strategic Conclusion

There is currently an asymmetry in the cyber arms race. Attackers are adopting Agentic AI faster than defenders because they are not bound by compliance, governance, or safety regulations.

To close this gap, we must:

  1. Trust Automation: Move high-fidelity containment actions to fully autonomous modes.

  2. Protect the AI: Treat our internal models as critical infrastructure (secure the weights, secure the training data).

  3. Train the People: Upskill the SOC team to understand how LLMs "think" so they can diagnose when they are being tricked.

The goal is to elevate, not replace, the human element. Let the AI fight the bots; let the humans fight the strategy.

Final Thought

The future of Incident Response is about humans.

+ AI vs. Adversarial AI. The organizations that master this symbiosis will do more than just survive the coming wave of automated attacks—they will dictate the terms of the engagement.

Curious about how to defend your network from evolving threats like AI misuse?

Sachin Saurabh
Sachin Saurabh

Staff Engineer - Detection and Response, Aviatrix

I am a cybersecurity professional specializing in threat detection, incident response, and cloud security. My writing focuses on breaking down real-world attacks, security research, and detection engineering into clear, actionable insights for defenders. Through my research blogs, I explore emerging threat techniques and modern approaches to security operations, including the use of AI in defense.

PODCAST

Altitude

Secure The Connections Between Your Clouds and Cloud Workloads

Leverage a security fabric to meet compliance and reduce cost, risk, and complexity.

Cta pattren Image
Cta pattren Image