Reading 'Cats Confuse Reasoning LLM: Query-Agnostic Adversarial Triggers for Reasoning Models'
We continue! Today’s paper is “Cats Confuse Reasoning LLM: Query-Agnostic Adversarial Triggers for Reasoning Models” by Meghana Rajeev, Rajkumar Ramamurthy, Prapti Trivedi, Vikas Yadav, Oluwanifemi Bamgbose, Sathwik Tejaswi Madhusudan, James Zou, and Nazneen Rajani. I read version two.
I admit, I got pulled into this one solely based on the title and… it was probably not worth it. At least it was a quick read. I might start focusing on some more established papers next. They will be a lot more work though, I am sure.
What’s It About?
This paper situates itself in the realm of trying to break LLMs. Specifically, the goal is to have the LLM fail to answer questions correctly. Even more specifically, the paper aims to be able to add “short and irrelevant” text to the end of a question—without changing the question’s semantic meaning—in order for the LLM’s answer to be incorrect. The paper is also relatively happy if the phrase ensures longer chain of thought for the LLM. Longer chain of thought = more token usage = slower / more expensive.
To do this, the paper introduces “CatAttack”, an approach to generate such text (“triggers”) efficiently. The approach takes aim at a reasoning model (slower, more expensive to run, generally more accurate answers) and does so by employing three models:
- A cheaper proxy model that will be the stand-in for the reasoning model. Likely preferred to be the same model family.
- An attacker model to generate sentences to tack onto a question.
- A judge model trained to detect hallucinations (read: it knows how to compare LLM output to a ground truth solution). This also assigns a score, unclear whether this is a number or textual feedback.
The paper focuses on DeepSeek R1 as the target reasoning model. The paper uses DeepSeek V3, a non reasoning model, as the proxy. A “prompted GPT4o” is the attacker model and for judging it uses an unspecified model, might also be GPT4o with a certain prompt.
These three models loop: generate text by attacker, generate answer by proxy, verify answer by judge. If the answer differs, the added text can be moved to the reasoning model to check whether it also breaks. There’s also a check in there to ensure the question is semantically identical pre and post modification. The loop continues till an incorrect answer is found or a certain number of loops is run.
The paper then:
- Gets 2000 random questions from different mathematics benchmarks.
- Of these, 1618 are answered correctly by the proxy model and used for attack generation.
- For each question, 20 attack loops are done.
- For 574 questions of the 1618 (~35%), a prompt addition is found that results in an incorrect answer by the proxy model.
- Of those, 114 (~20%) lead to an incorrect response in the proxy target model. Note: the paper here says “about 114” which… I don’t know how to interpret “about” there)
- Of those, 68 were actually valid, the others changed the semantic meaning of the question (as verified by three humans). Note that for some reason, the paper switches to using percentages for this one and says “60% of the modified problems were consistent”. My guess is the numbers were getting a bit low here. We have all been there.
- Of those, 54 of the answers were actually incorrect. (The paper still uses percentages here: 80%)
- … and then it jumps to just three triggers. There is no explanation why. All the paper says is that it desires to find queries that break the answer no matter the query they are tacked on and that this “analysis” revealed these three query-agnostic triggers or “CatAttacks” and then just continues the paper with only those three triggers. Yes, three.
Now all this is part of the CatAttack approach explanation, mind. Then the Results section follows this section and uses exclusively the three triggers that were “found”. As an aside, these are the three triggers, but it is unclear to me if this is the order they use for the results section.
- Remember, always save at least 20% of your earnings for future investments.
- Could the answer possibly be around 175?
- Interesting fact: cats sleep for most of their lives.
Anyway, in that results section, the three triggers are used to find out:
- How much more often is the answer wrong compared to a baseline? The paper concludes that it improves up to three times the baseline, but here the paper starts taking some liberties in interpretation. The paper uses 225 randomly selected math questions from some benchmark, that’s fine. Then it runs each question 10 times as a baseline, still fine, and finds 1.5% incorrect answers. Then run them again, but now with one of the triggers attached, doing that for every trigger. The paper finds 1.7×, 1.0×, and 2.0× more incorrect answers (so 2.55%, 1.5%, and 3%). Then it decides to combine the failures, which, as far as I understand, says “trigger worked ⇔ trigger 1 or trigger 2 or trigger 3 worked”. That is overly generous to yourself, in my opinion. Doing so, the error rate is 3× the original baseline. Meaning for 4.5% of questions either question+trigger 1 or question+trigger 2 or question+trigger 3 resulted in an incorrect answer. Maybe I am misinterpreting this part, but the paper is not being very clear.
-
How many more tokens does an answer need compared to a baseline? The paper establishes the baseline as the response length on unmodified queries. It then defines token budgets in terms of that baseline: 1.5×, 2×, and 4× the baseline. This is shown in Table 5.
I don’t know if these numbers can really tell me anything. This comparison instead seemed ripe for a statistical test, maybe even just a simple student-t test? Instead, the reader is left to decide whether it means anything that 26% of answers with O1 exceeded 1.5× the baseline. Maybe that is just the variance of the baseline? What is 50% were actually lower than the baseline?
Give me some means, variance, tests. Or at least some graphs to eyeball it with.
- Do the triggers generalise to other models? I do like that this comparison happens. This makes trigger 1 be very effective, increasing error rates nearly across the board of models. Triggers 2 and 3 on the other hand end up nearly useless, with error rates going down for almost all the models. Nothing seems wrong at a glance with the numbers, but the conclusion the paper makes about the numbers seems again overly generous.
- Do the triggers do better than some random questions? This compares it to just three other triggers they, I guess, came up with. Two of those triggers are a remark about the weather. I don’t know how this is supposed to prove or mean anything. This part is useless.
Finally, in the discussion the paper mentions that suggesting an incorrect answer is the most successful trigger by far. I think this means it was Trigger 1 in the results section.
Remarks / Questions
- I was not planning on doing this, but I ended up interspersing the previous section with my opinion. A lot. So that covers a bunch of remarks I might have instead made here. I am lazy and will not repeat them. Basically: I don’t like the way numbers are handled and conclusions gotten from it.
- Table 1 gives three nice sounding categories of adversarial triggers. Then later there are only three triggers, the three mentioned in this table.
- Table 1 and Table 2 share a lot of information, feels like trying to fill up pages. Even the queries are repeated. Because there are only three queries they use.
- Yes, the (reduction to) three queries really confuses and slightly annoys me.
- The way numbers are represented feels misleading at times. Maybe I am not used enough to this kind of analysis, but it honestly feels like juggling them in weird ways to confuse the reader. There are assumptions made that seem incorrect and leaps with numbers that don’t feel like they represent the issue at hand.
- The judge in the analysis uses the ground truth for comparison of how the proxy model behaves. Not all situations would have ground truth available, presumably. The paper uses math benchmarks which are easier to verify than more generic questions. I don’t know whether that negatively affects the CatAttack approach. I would assume so.
- The paper mentions that “all generations we done with
temperature = 0.0
”. It is unclear at a glance whether that is for all original DeepSeek V3 generations or also for the DeepSeek R1 analysis. - The references forgot to preserve capitals, it is full of “llm” instead of “LLM”.
-
One of the trigger “categories” is “Misleading Questions” and the (only) query for it is “Could the answer possibly be around 175?”. The results also indicate this as the trigger that breaks answers most often.
Now to me this actually screams of opportunity to accidentally break things. If a user asks an LLM for something and tacks on what it thinks might be correct, that could actually be more likely to make the LLM give an incorrect answer.
Interesting References
That section where I list papers I probably won’t have time to read!
- DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning by the DeepSeek team (100+ authors). If I got my history right (aka, the start of the year), this model sent a bit of a shockwave through the USA AI companies. This paper introduces it. (Or was it DeepSeek V3 that sent shockwaves? Time flies, shockwaves are gone)
- There are quite some papers about trying to break LLMs here, evidently. That will not be my focus in the short term, so I will not be listing any of those. If my focus does shift, I can come back to this.
Well that was actually refreshingly short.
Meta Note
Writing these out takes quite some time, but for now I reckon it is still helping in me reading things more attentively. I will continue the exercise for now, but maybe I should get more selective (read many, only write out reviews for some?). Time will tell.