Side-Channel Attacks Against LLMs

Feb. 17th, 2026 12:01 pm
[syndicated profile] bruce_schneier_feed

Posted by Bruce Schneier

Here are three papers describing different side-channel attacks against LLMs.

Remote Timing Attacks on Efficient Language Model Inference“:

Abstract: Scaling up language models has significantly increased their capabilities. But larger models are slower models, and so there is now an extensive body of work (e.g., speculative sampling or parallel decoding) that improves the (average case) efficiency of language model generation. But these techniques introduce data-dependent timing characteristics. We show it is possible to exploit these timing differences to mount a timing attack. By monitoring the (encrypted) network traffic between a victim user and a remote language model, we can learn information about the content of messages by noting when responses are faster or slower. With complete black-box access, on open source systems we show how it is possible to learn the topic of a user’s conversation (e.g., medical advice vs. coding assistance) with 90%+ precision, and on production systems like OpenAI’s ChatGPT and Anthropic’s Claude we can distinguish between specific messages or infer the user’s language. We further show that an active adversary can leverage a boosting attack to recover PII placed in messages (e.g., phone numbers or credit card numbers) for open source systems. We conclude with potential defenses and directions for future work.

When Speculation Spills Secrets: Side Channels via Speculative Decoding in LLMs“:

Abstract: Deployed large language models (LLMs) often rely on speculative decoding, a technique that generates and verifies multiple candidate tokens in parallel, to improve throughput and latency. In this work, we reveal a new side-channel whereby input-dependent patterns of correct and incorrect speculations can be inferred by monitoring per-iteration token counts or packet sizes. In evaluations using research prototypes and production-grade vLLM serving frameworks, we show that an adversary monitoring these patterns can fingerprint user queries (from a set of 50 prompts) with over 75% accuracy across four speculative-decoding schemes at temperature 0.3: REST (100%), LADE (91.6%), BiLD (95.2%), and EAGLE (77.6%). Even at temperature 1.0, accuracy remains far above the 2% random baseline—REST (99.6%), LADE (61.2%), BiLD (63.6%), and EAGLE (24%). We also show the capability of the attacker to leak confidential datastore contents used for prediction at rates exceeding 25 tokens/sec. To defend against these, we propose and evaluate a suite of mitigations, including packet padding and iteration-wise token aggregation.

Whisper Leak: a side-channel attack on Large Language Models“:

Abstract: Large Language Models (LLMs) are increasingly deployed in sensitive domains including healthcare, legal services, and confidential communications, where privacy is paramount. This paper introduces Whisper Leak, a side-channel attack that infers user prompt topics from encrypted LLM traffic by analyzing packet size and timing patterns in streaming responses. Despite TLS encryption protecting content, these metadata patterns leak sufficient information to enable topic classification. We demonstrate the attack across 28 popular LLMs from major providers, achieving near-perfect classification (often >98% AUPRC) and high precision even at extreme class imbalance (10,000:1 noise-to-target ratio). For many models, we achieve 100% precision in identifying sensitive topics like “money laundering” while recovering 5-20% of target conversations. This industry-wide vulnerability poses significant risks for users under network surveillance by ISPs, governments, or local adversaries. We evaluate three mitigation strategies – random padding, token batching, and packet injection – finding that while each reduces attack effectiveness, none provides complete protection. Through responsible disclosure, we have collaborated with providers to implement initial countermeasures. Our findings underscore the need for LLM providers to address metadata leakage as AI systems handle increasingly sensitive information.

mific: (Heated rivalry)
[personal profile] mific posting in [community profile] fanart_recs
Fandom: Heated Rivalry
Characters/Pairing/Other Subject: Shane Hollander/Ilya Rozanov
Content Notes/Warnings: none
Medium: digital art
Artist on DW/LJ: n/a
Artist Website/Gallery: luluxa on tumblr, and on AO3 (AO3 ones are often higher-rated)
Why this piece is awesome: Luluxa did Heated Rivalry art! And it's gorgeous - warm skin tones with the boys on vacation somewhere hot - maybe their honeymoon? Just lovely!
Link: something for the Valentine's, backup link here

SGA: on purpose by dedkake

Feb. 17th, 2026 04:21 pm
mific: (John eyeroll Rodney frazzled)
[personal profile] mific posting in [community profile] fancake
Fandom: Stargate Atlantis
Characters/Pairings: John Sheppard/Rodney McKay
Rating: Teen
Length: 2492
Content Notes: no AO3 warnings apply
Creator Links: dedkake on AO3
Themes: Inept in love, Pining, Five things, Friends to lovers

Summary: The thing is, he hadn’t really meant to say it. Not then. Not there. He hadn’t really ever even thought about it before, not in such specific terms. So, it’s as much of a shock to him as it is to anyone else.

or, Rodney's trying so hard and John just doesn't get it.

Reccer's Notes: This is a fun read that makes you want to hit them both upside the head just a little. Rodney keeps telling John how he feels (or trying to), and John keeps missing the point each time, so they're both inept in different ways. Until they aren't!

Fanwork Links: on purpose

Mysteries of the Universe, I guess?

Feb. 16th, 2026 09:22 pm
kalloway: (Lucifer 11 GBF)
[personal profile] kalloway
Weird Question Department- did anyone send me something in the mail around the end of last month? There's a tracking number in the system that scanned once at a sorting center and then has never been seen again. Unfortunately, the system also only has the tracking number, no origin point, and no actual image of the label so we're not even sure whose it is. I can't think of anything unaccounted for! (Not the KS I thought it might be.)

Also mildly convinced there's something going on/I'm supposed to have done by tomorrow but also nothing comes to mind. I have a busy week next week, but this week is pretty open. But it's also Fat Tuesday and the Lunar New Year so perhaps my brain is trying to juggle those in.

Other than that, work last night wasn't ridiculous but I hadn't slept well, so I pretty much deflated early but slept decent. I did get runners checked for both Star Abyss kits so I'm just going to count the entire day as a win anyway, lol. I'll keep working on the Destiny Astray when I have the brain power, and I grabbed a battered KO GM to poke at otherwise.

YEAR OF HORSE

Feb. 17th, 2026 09:46 am
scaramouche: Malaysian dreamwidth sheep (dreamwidth sheep baaa)
[personal profile] scaramouche


We're having a rainy Chinese New Year this time, which is quite unusual, though I vaguely remember we've had that before recently. Maybe the stereotype of a super hot CNY is no longer as typical?

Substitution.

Feb. 16th, 2026 08:35 pm
hannah: (James Wilson - maker unknown)
[personal profile] hannah
In today's dubious triumph over aphasia, I told my client I'd emptied out her utility kit.

Her toolbag. I couldn't remember toolbag and tried to use the next best thing to describe the object in question.

It was a fairly remarkable moment on a number of levels, and I'm pretty sure I'm going to be shaking my head over it for quite some time.

Music Monday: Two Rockin’ Videos

Feb. 16th, 2026 06:43 pm
jesse_the_k: Head inside a box, with words "Thinking inside the box" scrawled on it. (thinking inside the box)
[personal profile] jesse_the_k

The singer and the band are all on roller skates performing Bend Your Knees by Henry Mansfield & Digital Velvet! It’s an NPR Tiny Desk contest entry. Lyrics on bandcamp, video on YouTube or…

Stream it Here )

Thanks to [personal profile] clevermanka for sharing Fabulous, an absolute banger in both fashion and music from MEEK. Not work-safe since the chorus repeats “fucking” 42 times. Video on YouTube with accurate captions and lyrics in the description or …

stream it here )

rachelmanija: (Books: old)
[personal profile] rachelmanija


Five high school friends go on a camping trip and find a mysterious staircase in the woods. One of them climbs it and vanishes. Twenty years later, the staircase reappears, and they go to face it again.

I loved this premise and the cover. The staircase leading nowhere is spooky and beautiful, a weird melding of nature and civilization, so I was hoping for something that matched that vibe, like Annihilation or Revelator.

That was absolutely not what I got. The Staicase in the Woods is the misbegotten mutant child of It, King Sorrow, and Tumblr-speak. Every single character is insufferable. The teenagers are boring, and the adults are all the worst people you meet at parties. There are four men and one woman/nonbinary person, and she/they reads exactly like what MAGA thinks liberal women/trans people are like -- AuHD, blue hair, Tumblr-speak, angry, preachy, kinky sex etc. She/they says "My pronouns are she/them," then is only ever referred to as she and a woman. The staircase itself is barely in the story, where it leads is a letdown, and the ending combines the worst elements of being dumb and unresolved.

I got partway in and then skimmed because I was curious about the staircase and the vanished kid.

Angry spoilers for the whole book.

Read more... )

The Promptware Kill Chain

Feb. 16th, 2026 12:04 pm
[syndicated profile] bruce_schneier_feed

Posted by Bruce Schneier

The promptware kill chain: initial access, privilege escalation, reconnaissance, persistence, command & control, lateral movement, action on objective

Attacks against modern generative artificial intelligence (AI) large language models (LLMs) pose a real threat. Yet discussions around these attacks and their potential defenses are dangerously myopic. The dominant narrative focuses on “prompt injection,” a set of techniques to embed instructions into inputs to LLM intended to perform malicious activity. This term suggests a simple, singular vulnerability. This framing obscures a more complex and dangerous reality. Attacks on LLM-based systems have evolved into a distinct class of malware execution mechanisms, which we term “promptware.” In a new paper, we, the authors, propose a structured seven-step “promptware kill chain” to provide policymakers and security practitioners with the necessary vocabulary and framework to address the escalating AI threat landscape.

In our model, the promptware kill chain begins with Initial Access. This is where the malicious payload enters the AI system. This can happen directly, where an attacker types a malicious prompt into the LLM application, or, far more insidiously, through “indirect prompt injection.” In the indirect attack, the adversary embeds malicious instructions in content that the LLM retrieves (obtains in inference time), such as a web page, an email, or a shared document. As LLMs become multimodal (capable of processing various input types beyond text), this vector expands even further; malicious instructions can now be hidden inside an image or audio file, waiting to be processed by a vision-language model.

The fundamental issue lies in the architecture of LLMs themselves. Unlike traditional computing systems that strictly separate executable code from user data, LLMs process all input—whether it is a system command, a user’s email, or a retrieved document—as a single, undifferentiated sequence of tokens. There is no architectural boundary to enforce a distinction between trusted instructions and untrusted data. Consequently, a malicious instruction embedded in a seemingly harmless document is processed with the same authority as a system command.

But prompt injection is only the Initial Access step in a sophisticated, multistage operation that mirrors traditional malware campaigns such as Stuxnet or NotPetya.

Once the malicious instructions are inside material incorporated into the AI’s learning, the attack transitions to Privilege Escalation, often referred to as “jailbreaking.” In this phase, the attacker circumvents the safety training and policy guardrails that vendors such as OpenAI or Google have built into their models. Through techniques analogous to social engineering—convincing the model to adopt a persona that ignores rules—to sophisticated adversarial suffixes in the prompt or data, the promptware tricks the model into performing actions it would normally refuse. This is akin to an attacker escalating from a standard user account to administrator privileges in a traditional cyberattack; it unlocks the full capability of the underlying model for malicious use.

Following privilege escalation comes Reconnaissance. Here, the attack manipulates the LLM to reveal information about its assets, connected services, and capabilities. This allows the attack to advance autonomously down the kill chain without alerting the victim. Unlike reconnaissance in classical malware, which is performed typically before the initial access, promptware reconnaissance occurs after the initial access and jailbreaking components have already succeeded. Its effectiveness relies entirely on the victim model’s ability to reason over its context, and inadvertently turns that reasoning to the attacker’s advantage.

Fourth: the Persistence phase. A transient attack that disappears after one interaction with the LLM application is a nuisance; a persistent one compromises the LLM application for good. Through a variety of mechanisms, promptware embeds itself into the long-term memory of an AI agent or poisons the databases the agent relies on. For instance, a worm could infect a user’s email archive so that every time the AI summarizes past emails, the malicious code is re-executed.

The Command-and-Control (C2) stage relies on the established persistence and dynamic fetching of commands by the LLM application in inference time from the internet. While not strictly required to advance the kill chain, this stage enables the promptware to evolve from a static threat with fixed goals and scheme determined at injection time into a controllable trojan whose behavior can be modified by an attacker.

The sixth stage, Lateral Movement, is where the attack spreads from the initial victim to other users, devices, or systems. In the rush to give AI agents access to our emails, calendars, and enterprise platforms, we create highways for malware propagation. In a “self-replicating” attack, an infected email assistant is tricked into forwarding the malicious payload to all contacts, spreading the infection like a computer virus. In other cases, an attack might pivot from a calendar invite to controlling smart home devices or exfiltrating data from a connected web browser. The interconnectedness that makes these agents useful is precisely what makes them vulnerable to a cascading failure.

Finally, the kill chain concludes with Actions on Objective. The goal of promptware is not just to make a chatbot say something offensive; it is often to achieve tangible malicious outcomes through data exfiltration, financial fraud, or even physical world impact. There are examples of AI agents being manipulated into selling cars for a single dollar or transferring cryptocurrency to an attacker’s wallet. Most alarmingly, agents with coding capabilities can be tricked into executing arbitrary code, granting the attacker total control over the AI’s underlying system. The outcome of this stage determines the type of malware executed by promptware, including infostealer, spyware, and cryptostealer, among others.

The kill chain was already demonstrated. For example, in the research “Invitation Is All You Need,” attackers achieved initial access by embedding a malicious prompt in the title of a Google Calendar invitation. The prompt then leveraged an advanced technique known as delayed tool invocation to coerce the LLM into executing the injected instructions. Because the prompt was embedded in a Google Calendar artifact, it persisted in the long-term memory of the user’s workspace. Lateral movement occurred when the prompt instructed the Google Assistant to launch the Zoom application, and the final objective involved covertly livestreaming video of the unsuspecting user who had merely asked about their upcoming meetings. C2 and reconnaissance weren’t demonstrated in this attack.

Similarly, the “Here Comes the AI Worm” research demonstrated another end-to-end realization of the kill chain. In this case, initial access was achieved via a prompt injected into an email sent to the victim. The prompt employed a role-playing technique to compel the LLM to follow the attacker’s instructions. Since the prompt was embedded in an email, it likewise persisted in the long-term memory of the user’s workspace. The injected prompt instructed the LLM to replicate itself and exfiltrate sensitive user data, leading to off-device lateral movement when the email assistant was later asked to draft new emails. These emails, containing sensitive information, were subsequently sent by the user to additional recipients, resulting in the infection of new clients and a sublinear propagation of the attack. C2 and reconnaissance weren’t demonstrated in this attack.

The promptware kill chain gives us a framework for understanding these and similar attacks; the paper characterizes dozens of them. Prompt injection isn’t something we can fix in current LLM technology. Instead, we need an in-depth defensive strategy that assumes initial access will occur and focuses on breaking the chain at subsequent steps, including by limiting privilege escalation, constraining reconnaissance, preventing persistence, disrupting C2, and restricting the actions an agent is permitted to take. By understanding promptware as a complex, multistage malware campaign, we can shift from reactive patching to systematic risk management, securing the critical systems we are so eager to build.

This essay was written with Oleg Brodt, Elad Feldman and Ben Nassi, and originally appeared in Lawfare.

kingstoken: (Animated Aziraphale Crowley)
[personal profile] kingstoken posting in [community profile] fancake
Fandom: Good Omens
Pairings/Characters: Aziraphale/Crowley
Rating: G
Length: 984 words
Creator Links: ghost_daddy
Theme: inept in love

Summary: Aziraphale knows that Crowley is in love with someone. He just doesn't know who.

When he asks him, it doesn't go quite as he planned.

Reccer's Notes: A cute little fic where Aziraphale asks Crowley who his sweetheart is, and Crowley is flabbergasted that Aziraphale doesn't know.  Its in character, I could have seen them having this discussion after season 1. 

Fanwork Links: Ao3

happy fanniversary

Feb. 16th, 2026 09:34 am
runpunkrun: old grouchy rodney mckay, text: Stargate: Geezer (get off my lawn)
[personal profile] runpunkrun
I posted my first fanfic* TWENTY NINE YEARS AGO TODAY. My most recent fanfic† was posted less than a month ago. And today I am finishing up a fanfic‡ I started in 2011.

* The X-Files
† Star Trek
‡ Stargate Atlantis

Sci-fi movies I have watched lately

Feb. 16th, 2026 09:41 am
anagrrl: (Default)
[personal profile] anagrrl
I've been hopping around a bit in terms of watching things, starting a series here and there, and then dropping it. I don't know, I seem to get to a point where I just lose interest in a lot of the storylines (and sometimes, the predictability). But! I have been watching movies. Some recent, some ancient.

I finally got around to watching On the Beach (1959), and I enjoyed it. I think what really stood out to me is how quiet it is, not so much in the sense of actual sound, but in the way it plays out. Spoilers, just in case: Read more... )

I also watched We Bury the Dead (2024), which is predictable in many ways (the 'twists' were pretty easy to spot), but still enjoyable (to me, at least, who likes these apocalyptic scenarios). The acting was solid, I thought, and the circumstances for how the situation happened were unique enough to add a little uncertainty and mystery to things. I'm not sure I would watch it again, but I enjoyed watching it.

Target Earth (1998)'s IMDB description is After aliens inhabit human bodies, it's up to a small-town policeman to protect a child who holds the key to defeating the extraterrestrials. I believe it's a remake of a 1954 film. Anyway, it's pretty cheese-ball in a lot of ways, but also surprisingly good. I have been watching some movies from the 90s lately (for example, Die Hard 2) and it's kind of hilarious to see the way things have changed in terms of technologies, etc (it's very weird to see movies like Die Hard 2 where people can have guns in airports and also smoke indoors, and carry tasers on planes, and also have to line up at the public phones to make calls, I remember those days, and also, wow, things have changed). Anyway, Target Earth was an entertaining watch, pretty slow paced in a lot of ways, but I like that.

Greenland: Migration (2026) is an enjoyable sequel to Greenland (which I also enjoyed). People are forced to leave their Greenland survival bunker (where they holed up during catastrophic meteor strikes), and this is about what happens next, and what the world is like around them. Morena Baccarin is great, the storyline is pretty predictable in some ways, but has a few slightly unpredictable elements that take it away from the standard post-apocalyptic fare. Spoilers, in case: Read more... )

I rewatched Elevation (2024) which stars Anthony Mackie and Morena Baccarin, and I liked it possibly even more than the first time. They're both survivors living at high latitudes (above 8000ft) after the emergence of some kind of creatures that kill humans but have a hard stop at 8000ft, and can't get past those latitudes. Mackie is a dad who needs some medical equipment for his kid, and Baccarin is a cranky, heavy-drinking scientist. They head down to the lower elevations to get the medical stuff, and so Baccarin can get some materials she needs that she thinks will make the creatures vulnerable to human weapons. It's well acted, and an enjoyable (if often tense) watch. It also showcases what can happen when people work together. I would actually really like a sequel to this one, as the very end tells us about what the creatures might be (though there are hints throughout).


I also watched Ghosts of Mars (2020). It is terrible. Not even 'so terrible it's fun to watch'. It's just plain terrible.

oh, right, THIS part

Feb. 16th, 2026 08:27 am
heresluck: (vidding: vid ALL the things!)
[personal profile] heresluck
It's been so long since I watched a show not just fannishly but in full fannish company that I had somehow forgotten just HOW MANY false positives I end up with when testing out possible vidsongs for zeitgeist-y shows or movies. I get all excited on the start of a walk to work or at the beginning of cooking dinner, and then by the end I'm reminding myself that "song with one startlingly apropos verse" is not the same as "actually workable vidsong." BUT THEN I think about how some of my favorite vids are my favorites because they take a song that is a slightly weird fit in some way and make it feel inevitable, and I start second-guessing myself.

As with so many vid-related things in the last twenty-five years, I blame [personal profile] sisabet.

Profile

lakeeffectgirl: (Default)
Lake

February 2026

S M T W T F S
1234567
8910 11121314
15161718192021
22232425262728

Most Popular Tags

Style Credit

Expand Cut Tags

No cut tags
Page generated Feb. 17th, 2026 01:03 pm
Powered by Dreamwidth Studios