信息通信技术与政策

信息通信技术与政策

信息通信技术与政策 ›› 2025, Vol. 51 ›› Issue (1): 33-37.doi: 10.12267/j.issn.2096-5931.2025.01.005

专题:网络安全 上一篇    下一篇

面向可信大语言模型智能体的安全挑战与应对机制

Security challenges and response mechanisms for trustworthy large language model agents

张熙1, 李朝卓1, 许诺1, 张力天2   

  1. 1.北京邮电大学网络空间安全学院,北京 100876
    2.北京航空航天大学网络空间安全学院,北京 100191
  • 收稿日期:2024-06-28 出版日期:2025-01-25 发布日期:2025-02-14
  • 通讯作者: 李朝卓, 北京邮电大学网络空间安全学院特聘副研究员,主要从事可信大语言模型、图神经网络和推荐系统等研究工作。
  • 作者简介:
    张熙, 北京邮电大学网络空间安全学院教授,博士生导师,主要从事信息内容安全和人工智能安全等研究工作;
    许诺, 北京邮电大学网络空间安全学院硕士研究生在读,主要从事大模型安全等研究工作;
    张力天, 北京航空航天大学网络空间安全学院博士研究生在读,主要从事大模型安全、大模型多智能体、多模态机器学习等研究工作。

ZHANG Xi1, LI Chaozhuo1, XU Nuo1, ZHANG Litian2   

  1. 1. School of Cyberspace Security, Beijing University of Posts and Telecommunications, Beijing 100876, China
    2. School of Cyber Science and Technology, Beihang University, Beijing 100191, China
  • Received:2024-06-28 Online:2025-01-25 Published:2025-02-14

摘要:

随着大语言模型驱动的智能体在各领域的应用日益深化,潜在的安全隐患逐渐凸显。旨在系统梳理基于大语言模型的智能体面临的信息泄露、模型攻击、幻觉输出、伦理道德风险和法律合规隐患等安全可信问题。通过对这些安全隐患的成因与影响进行深入分析,探讨现有的防护措施和技术手段,提出构建可信大语言模型智能体的建议,为相关研究和实践提供参考。

关键词: 可信大语言模型智能体, 安全, 防御

Abstract:

As the application of large language model-driven agents deepens in various fields, potential security risks are gradually prominent. This paper aims to systematically sort out the security and trustworthiness problems faced by agents based on large language models, including information leakage, model attacks, hallucination outputs, ethical and moral risks, and legal compliance hazards. By conducting an in-depth analysis of the causes and impacts of these security risks, this paper discusses existing protective measures and technical means, and proposes suggestions for building trustworthy large language model agents, providing references for related research and practice.

Key words: trustworthy large language model agent, security, defense

中图分类号: