Large Language Models are Unreliable for Cyber Threat Intelligence

A recent study titled "Large Language Models are Unreliable for Cyber Threat Intelligence" critically examines the application of Large Language Models (LLMs) in automating Cyber Threat Intelligence (CTI) tasks. The research presents an evaluation methodology that assesses LLMs using zero-shot learning, few-shot learning, and fine-tuning approaches, focusing on their consistency and confidence levels. Experiments conducted with three state-of-the-art LLMs on a dataset of 350 CTI reports reveal that LLMs struggle with real-sized reports, exhibiting inconsistent performance and overconfidence. While few-shot learning and fine-tuning offer limited improvements, the findings raise concerns about relying on LLMs for CTI, especially in scenarios lacking labeled datasets where confidence is crucial.  

https://arxiv.org/abs/2503.23175

Comments

Popular posts from this blog

Secure Vibe Coding Guide: Best Practices for Writing Secure Code

KEVIntel: Real-Time Intelligence on Exploited Vulnerabilities

OWASP SAMM Skills Framework Enhances Software Security Roles