BriefGPT.xyz
Jul, 2024
大型语言模型能自动破解GPT-4V吗?
Can Large Language Models Automatically Jailbreak GPT-4V?
HTML
PDF
Yuanwei Wu, Yue Huang, Yixin Liu, Xiang Li, Pan Zhou...
TL;DR
本研究解决了GPT-4V在隐私泄露方面的安全隐患,提出了一种名为AutoJailbreak的自动破解技术。通过利用大型语言模型进行优化提示和高效搜索方法,我们的实验结果显示该技术的攻击成功率超过95.3%,显著优于传统方法,凸显了大型语言模型在强化GPT-4V安全性中的作用。
Abstract
GPT-4V
has attracted considerable attention due to its extraordinary capacity for integrating and processing multimodal information. At the same time, its ability of face recognition raises new safety concerns of
Privac
→