Privacy-Preserving Mechanisms Enable Cheap Verifiable Inference of LLMs
AI 摘要
该论文提出利用隐私保护的LLM推理来实现廉价且可验证的推理,降低验证开销。
主要贡献
- 提出了新的基于隐私保护LLM推理的可验证推理协议
- 提出的协议计算成本低,几乎没有下游影响
- 证明了隐私和可验证性在LLM推理中的联系
方法论
利用现有的隐私保护LLM推理方法,设计新的协议,增加少量计算即可实现推理验证。
原文摘要
As large language models (LLMs) continue to grow in size, fewer users are able to host and run models locally. This has led to increased use of third-party hosting services. However, in this setting, there is a lack of guarantees on the computation performed by the inference provider. For example, a dishonest provider may replace an expensive large model with a cheaper-to-run weaker model and return the results from the weaker model to the user. Existing tools to verify inference typically rely on methods from cryptography such as zero-knowledge proofs (ZKPs), but these add significant computational overhead, and remain infeasible for use for large models. In this work, we develop a new insight -- that given a method for performing private LLM inference, one can obtain forms of verified inference at marginal extra cost. Specifically, we propose two new protocols which leverage privacy-preserving LLM inference in order to provide guarantees over the inference that was carried out. Our approaches are cheap, requiring the addition of a few extra tokens of computation, and have little to no downstream impact. As the fastest privacy-preserving inference methods are typically faster than ZK methods, the proposed protocols also improve verification runtime. Our work provides novel insights into the connections between privacy and verifiability in LLM inference.