- cross-posted to:
- aicompanions@lemmy.world
- technology@lemmy.ml
- cross-posted to:
- aicompanions@lemmy.world
- technology@lemmy.ml
You must log in or register to comment.
This is pretty clever. As I understand it.
- Because LLMs are slow most of them stream the response to the user.
- The response is streamed as text, but generated in tokens.
- This means that each “chunk” leaks the length of the text corresponding to the token.
- You can then use heuristics to guess the text of the response based on the token lengths.
This is a good reminder any time you are sending content in small chunks over an encrypted channel, many encrypted channels don’t provide protection against size leaks by default.
It seems there are a few easy solutions to this:
- Send the token IDs (as fixed-size integers) over the network rather than the text.
- Pad the text representations of the tokens to a fixed length.
- Batch the tokens more (and maybe add padding) to produce bigger chunks and obscure individual token size.
These still all leak the approximate length of the response, but that is probably acceptable.