4/15 (Tue.) 16:15 - 17:00 7F 703

Unveiling the Bias in Language Models: A Path to Stability in Security Assessments

Large Language Models (LLMs) have shown great potential in cybersecurity applications. However, to fully harness their value, inherent biases and stability issues in LLM-driven security assessments must be effectively addressed. This talk will focus on these challenges and present our latest research on improving evaluation frameworks.

Our study analyzes how LLMs can be influenced by the order of presented options during the assessment process, leading to biases. We propose ranking strategies and probabilistic weighting techniques that significantly improve scoring accuracy and consistency. Key topics covered in this talk include experimental design and observations on LLM biases, probability-based weighting adjustments, and methodologies for integrating results from multiple ranking permutations. Notably, through validation with the G-EVAL dataset, we demonstrate measurable improvements in model evaluation performance.

Whether you are conducting research on language models or working in cybersecurity technology and decision-making, this talk will provide valuable technical insights and practical takeaways.

Chen, Shu-Yuan
SPEAKER
CyCraft Technology
Data Scientist, Data Science

TOPIC / TRACK
AI Security & Safety Forum

LOCATION
Taipei Nangang Exhibition Center, Hall 2
7F 703

LEVEL
General General sessions explore new cybersecurity knowledge and non-technical topics, ideal for those with limited or no prior cybersecurity knowledge.

SESSION TYPE
Breakout Session

LANGUAGE
Chinese

SUBTOPIC
AI
AI Security
LLM