This study addresses categories of harm surrounding Large Language Models (LLMs) in the field of artificial intelligence. It addresses five categories of harms addressed before, during, and after development of AI applications: pre-development, direct output, Misuse and Malicious Application, and downstream application. By underscoring the need to define risks of the current landscape to ensure accountability, transparency and navigating bias when adapting LLMs for practical applications. It proposes mitigation strategies and future directions for specific domains and a dynamic auditing system guiding responsible development and integration of LLMs in a standardized proposal.
翻译:本研究针对人工智能领域大语言模型(LLMs)相关的危害类别进行了系统分析。研究涵盖了人工智能应用开发前、开发中及开发后五个类别的危害:开发前阶段、直接输出危害、误用与恶意应用、以及下游应用风险。通过强调在将LLMs应用于实际场景时,需明确当前环境中的风险以确保问责制、透明性并规避偏见,本文提出了针对特定领域的缓解策略与未来研究方向,并构建了一个动态审计体系,以标准化方案指导LLMs的负责任开发与集成。