A Hitchhiker's Guide to Jailbreaking ChatGPT via Prompt Engineering

Apr 20, 2024ยท
Yi Liu
Gelei Deng
Gelei Deng
,
Zhengzi Xu
,
Yuekang Li
,
Yaowen Zheng
,
Ying Zhang
,
Lida Zhao
,
Tianwei Zhang
,
Kailong Wang
ยท 1 min read
Abstract
This work provides a comprehensive guide to jailbreaking ChatGPT through prompt engineering techniques. We systematically categorize jailbreak methods, analyze their effectiveness, and discuss implications for LLM safety and responsible AI deployment.
Type
Publication
4th International Workshop on Software Engineering and AI for Data Quality (SEA4DQ @ ICSE)

This workshop paper provides practical guidance on jailbreaking ChatGPT through prompt engineering, offering insights for both attack and defense research.