BriefGPT.xyz
Nov, 2023
大型语言模型的安全风险分类
A Security Risk Taxonomy for Large Language Models
HTML
PDF
Erik Derner, Kristina Batistič, Jan Zahálka, Robert Babuška
TL;DR
通过针对大型语言模型(LLM)存在的安全风险进行评估,本研究填补了当前研究的空白,并提出了一种基于提示的攻击风险分类方法,以强调LLM在用户-模型通信路径上的安全风险。该分类方法通过具体的攻击示例加以支持,并旨在为安全性强、值得信赖的LLM应用程序的开发提供指导。
Abstract
As
large language models
(LLMs) permeate more and more applications, an assessment of their associated
security risks
becomes increasingly necessary. The potential for exploitation by
→