Login or signup to connect with paper authors and to register for specific Author Connect sessions (if available).
Cracking the Code: Examining Linguistic Elements in Adversarial Prompt Engineering
Kofi Arhin, Haiyan Jia
In recent times, there has been a surge in the popularity and use of generative artificial intelligence (GAI) systems. While GAI systems can potentially make people and organizations more efficient, malicious actors can exploit existing vulnerabilities in these systems. Prompt engineering - the act of interacting with GAI systems via text to produce content - has been used for adversarial purposes. In this study, we examine how linguistic text features and psychological triggers in prompts impact the success of malicious prompts. Our preliminary results show that less concrete prompts have a positive relation with malicious prompt success, and this is also the case for psychological triggers such as trust and urgency. Meanwhile, psychological triggers such as authority and reward show a negative relationship with adversarial prompt success. To contribute to theory and practice, we intend to elaborate on these findings further using a data-driven, computationally intensive theory-building approach.
AuthorConnect Sessions
No sessions scheduled yet