Zhiyuan Yu
Zhiyuan Yu
Home
News
Publications
Teaching
Contact
Generative AI
Don't Listen To Me: Understanding and Exploring Jailbreak Prompts of Large Language Models
(
USENIX Security 2024
)
This is a systematic study on jailbreak attacks against commercial large language model (LLM) systems. We analyzed existing jailbreak prompts, examined their contributing factors, and conducted user studies to explore human behavioral patterns during jailbreak attempts.
Zhiyuan Yu
,
Xiaogeng Liu
,
Shunning Liang
,
Zach Cameron
,
Chaowei Xiao
,
Ning Zhang
Code
Dataset
Distinguished Paper Award
Artifact Badges - Available, Functional, Results Reproduced
Press
DOI
Website
PDF
AntiFake: Using Adversarial Audio to Prevent Unauthorized Speech Synthesis
(
CCS 2023
)
We introduced the concept of proactive defense to combat unauthorized speech synthesis. Our defense works by adding perturbations to speech samples before releasing them into the public domain (e.g., social media and streaming platforms). While the processed sample still sounds like the victim to humans, when it is used for speech synthesis by the attacker, the resulting synthetic speech would resemble others’ voices rather than the victim’s.
Zhiyuan Yu
,
Shixuan Zhai
,
Ning Zhang
Code
DOI
National Public Radio Report
Winner of the FTC Voice Cloning Challenge
PDF
Cite
×