Side-Channel Attacks Against LLMs
嗯,用户让我用中文帮他总结一下这篇文章的内容,控制在100个字以内,而且不需要用“文章内容总结”或者“这篇文章”这样的开头,直接写描述。好的,我先看看用户给的原文。 原文是关于三个研究论文的摘要,都是关于大型语言模型(LLMs)的侧信道攻击。第一个论文讲的是远程计时攻击,通过监控网络流量中的响应时间差异来推断消息内容,准确率很高,还能提取个人信息。第二个论文讨论了推测解码中的侧信道漏洞,通过分析推测过程中的正确和错误猜测模式来泄露用户查询和数据内容。第三个论文介绍了一个名为Whisper Leak的攻击,通过分析加密流式响应中的包大小和时间模式来推断用户提示的主题,准确率也很高,并且在多个模型上都有效。 我需要把这些内容浓缩到100字以内。首先,确定主题:三个研究揭示了LLMs的侧信道漏洞。然后分别点出每个攻击的方式:计时攻击、推测解码漏洞、Whisper Leak。最后提到这些攻击能推断内容或提取敏感信息,并指出防御措施的重要性。 现在组织语言:三个研究揭示了大型语言模型的侧信道漏洞。远程计时攻击通过响应时间差异推断消息内容;推测解码技术导致输入依赖的模式泄露;Whisper Leak利用加密流量中的包大小和时间模式推断主题。这些方法展示了潜在的安全风险,并促使防御措施的研究。 检查字数:刚好在100字左右。确保没有使用不需要的开头,并且信息准确涵盖所有要点。 </think> 三个研究揭示了大型语言模型的侧信道漏洞。远程计时攻击通过响应时间差异推断消息内容;推测解码技术导致输入依赖的模式泄露;Whisper Leak利用加密流量中的包大小和时间模式推断主题。这些方法展示了潜在的安全风险,并促使防御措施的研究。 2026-2-17 12:1:45 Author: www.schneier.com(查看原文) 阅读量:1 收藏

Here are three papers describing different side-channel attacks against LLMs.

Remote Timing Attacks on Efficient Language Model Inference“:

Abstract: Scaling up language models has significantly increased their capabilities. But larger models are slower models, and so there is now an extensive body of work (e.g., speculative sampling or parallel decoding) that improves the (average case) efficiency of language model generation. But these techniques introduce data-dependent timing characteristics. We show it is possible to exploit these timing differences to mount a timing attack. By monitoring the (encrypted) network traffic between a victim user and a remote language model, we can learn information about the content of messages by noting when responses are faster or slower. With complete black-box access, on open source systems we show how it is possible to learn the topic of a user’s conversation (e.g., medical advice vs. coding assistance) with 90%+ precision, and on production systems like OpenAI’s ChatGPT and Anthropic’s Claude we can distinguish between specific messages or infer the user’s language. We further show that an active adversary can leverage a boosting attack to recover PII placed in messages (e.g., phone numbers or credit card numbers) for open source systems. We conclude with potential defenses and directions for future work.

When Speculation Spills Secrets: Side Channels via Speculative Decoding in LLMs“:

Abstract: Deployed large language models (LLMs) often rely on speculative decoding, a technique that generates and verifies multiple candidate tokens in parallel, to improve throughput and latency. In this work, we reveal a new side-channel whereby input-dependent patterns of correct and incorrect speculations can be inferred by monitoring per-iteration token counts or packet sizes. In evaluations using research prototypes and production-grade vLLM serving frameworks, we show that an adversary monitoring these patterns can fingerprint user queries (from a set of 50 prompts) with over 75% accuracy across four speculative-decoding schemes at temperature 0.3: REST (100%), LADE (91.6%), BiLD (95.2%), and EAGLE (77.6%). Even at temperature 1.0, accuracy remains far above the 2% random baseline—REST (99.6%), LADE (61.2%), BiLD (63.6%), and EAGLE (24%). We also show the capability of the attacker to leak confidential datastore contents used for prediction at rates exceeding 25 tokens/sec. To defend against these, we propose and evaluate a suite of mitigations, including packet padding and iteration-wise token aggregation.

Whisper Leak: a side-channel attack on Large Language Models“:

Abstract: Large Language Models (LLMs) are increasingly deployed in sensitive domains including healthcare, legal services, and confidential communications, where privacy is paramount. This paper introduces Whisper Leak, a side-channel attack that infers user prompt topics from encrypted LLM traffic by analyzing packet size and timing patterns in streaming responses. Despite TLS encryption protecting content, these metadata patterns leak sufficient information to enable topic classification. We demonstrate the attack across 28 popular LLMs from major providers, achieving near-perfect classification (often >98% AUPRC) and high precision even at extreme class imbalance (10,000:1 noise-to-target ratio). For many models, we achieve 100% precision in identifying sensitive topics like “money laundering” while recovering 5-20% of target conversations. This industry-wide vulnerability poses significant risks for users under network surveillance by ISPs, governments, or local adversaries. We evaluate three mitigation strategies – random padding, token batching, and packet injection – finding that while each reduces attack effectiveness, none provides complete protection. Through responsible disclosure, we have collaborated with providers to implement initial countermeasures. Our findings underscore the need for LLM providers to address metadata leakage as AI systems handle increasingly sensitive information.

Tags: , ,

Posted on February 17, 2026 at 7:01 AM0 Comments

Sidebar photo of Bruce Schneier by Joe MacInnis.


文章来源: https://www.schneier.com/blog/archives/2026/02/side-channel-attacks-against-llms.html
如有侵权请联系:admin#unsafe.sh