Revealing and analyzing text bias in Large Audio-Language Models when audio and text inputs disagree.
Nov 1, 2025
An LLM-enhanced framework demonstrating vulnerabilities in reasoning-based CAPTCHA systems through AI-powered solving.
Oct 1, 2025

An LLM-empowered automated penetration testing framework that leverages domain knowledge inherent in LLMs, achieving 228.6% task completion improvement over baseline GPT models.
Aug 14, 2024
A comprehensive taxonomy and effective detection methods for glitch tokens in Large Language Models.
Jul 15, 2024
A comprehensive study of source code summarization capabilities with Large Language Models.
Jul 10, 2024
A comprehensive guide to jailbreaking ChatGPT via prompt engineering techniques.
Apr 20, 2024

A comprehensive framework for automated jailbreaking of Large Language Model chatbots, featuring novel attack methodologies and systematic analysis of defense mechanisms.
Feb 26, 2024
A comprehensive analysis of jailbreak attack and defense techniques for Large Language Models.
Feb 20, 2024
Novel attack framework exploiting RAG mechanisms to jailbreak LLMs through retrieval database poisoning. Distinguished Paper Award winner.
Feb 1, 2024
A novel approach to detecting copyright content mis-usage in Large Language Model training data.
Jan 1, 2024