Large Language Models are Unreliable for Cyber Threat Intelligence
A recent study titled "Large Language Models are Unreliable for Cyber Threat Intelligence" critically examines the application of Large Language Models (LLMs) in automating Cyber Threat Intelligence (CTI) tasks. The research presents an evaluation methodology that assesses LLMs using zero-shot learning, few-shot learning, and fine-tuning approaches, focusing on their consistency and confidence levels. Experiments conducted with three state-of-the-art LLMs on a dataset of 350 CTI reports reveal that LLMs struggle with real-sized reports, exhibiting inconsistent performance and overconfidence. While few-shot learning and fine-tuning offer limited improvements, the findings raise concerns about relying on LLMs for CTI, especially in scenarios lacking labeled datasets where confidence is crucial.
Comments
Post a Comment