LESSWRONG
LW

AI
Frontpage

35

Reasoning-Finetuning Repurposes Latent Representations in Base Models

by Jake Ward, lccqqqqq, Neel Nanda
23rd Jul 2025
AI Alignment Forum
3 min read
1

35

Ω 20

This is a linkpost for https://arxiv.org/abs/2507.12638
AI
Frontpage

35

Ω 20

Reasoning-Finetuning Repurposes Latent Representations in Base Models
5TurnTrout
New Comment
1 comment, sorted by
top scoring
Click to highlight new comments since: Today at 11:50 PM
[-]TurnTrout2moΩ350

Nice work. What a cool use of steering vectors!

Reply
Moderation Log
More from Jake Ward
View more
Curated and popular this week
1Comments

Authors: Jake Ward*, Chuqiao Lin*, Constantin Venhoff, Neel Nanda (*Equal contribution). This work was completed during Neel Nanda's MATS 8.0 Training Phase.

TL;DR

  • We computed a steering vector for backtracking using base model activations.
  • It causes the associated fine-tuned reasoning model to backtrack.
  • But, it doesn't cause the base model to backtrack.
    • That's weird!

Introduction

Reasoning models output  Wait, a lot. How did they learn to do this? Backtracking is an emergent behavior in RL-finetuned reasoning models like DeepSeek-R1, and appears to contribute substantially to these models' improved reasoning capabilities. We study representations related to this behavior using steering vectors, and find a direction which is present both in base models and associated reasoning-finetuned models but induces backtracking only in reasoning models. We interpret this direction as representing some concept upstream of backtracking which the base model has learned to keep track of, while only the reasoning model has learned to use this concept for backtracking.

We start with methodology similar to Venhoff et al.: We generate a corpus of reasoning rollouts using DeepSeek-R1, and annotate them to identify backtracking sentences using GPT-4o. Then, we train difference-of-means steering vectors with two novel properties:

  1. We compute steering vectors using token positions which occur before backtracking, using a 5-token window which starts 12 tokens before the beginning of a backtracking sentence. This usually includes the beginning of the sentence immediately before a backtracking sentence.
  2. We sample activations from a forward pass of the base model (Llama-3.1-8b) at layer 10, and use computed vectors to steer the reasoning model (DeepSeek-R1-Distill-Llama-8B).
Green highlights represent tokens from which our backtracking steering vectors are computed, red highlights indicate the start of backtracking.

Surprisingly, we find that steering vectors computed with this method are highly effective at inducing the emission of backtracking tokens ( Wait,  But, etc.) in the reasoning model, but not in the base model.

Additionally, if we instead sample activations from the reasoning model, we end up computing a nearly identical vector with > 0.7 cosine similarity. These vectors behave similarly: They induce backtracking in the reasoning model, but not in the base model.

Note that the base model never exhibits backtracking behavior, even when steered with the reasoning model-derived backtracking-inducing vector.

Analysis

Logit Lens

We considered whether these vectors may be doing something trivial, like directly boosting the logits for backtracking tokens like  Wait. We test this with Logit Lens, and find that this is not the case:

Base model-derived steering vectors (blue) never map to backtracking logits directly when computed at early to mid layers.

Baseline Comparison

We compare our results to various baselines, and find that our backtracking steering vector has a substantially greater effect on the emission of the  Wait token compared to other methods. We try steering with mean activations, adding random noise, increasing the magnitude of activations, and steering vectors computed based on other reasoning behaviors like deduction or initialization (problem restatement). Notably, we do find that adding random noise to activations does induce backtracking token emission slightly, but this effect is small compared to the effect of our backtracking steering vector.

Conclusion

We take this as evidence that backtracking isn't a totally new behavior learned during finetuning. Instead, base models already contain useful concepts which the reasoning-finetuning process finds and repurposes for emergent reasoning behavior. We're still not really sure what concept our steering vector represents, or if there is a monosemantic concept which it represents. We investigated whether it may represent something abstract like "uncertainty", but so far these investigations are inconclusive. These results should be interpreted as an existence proof for latent reasoning-related representations in base models, and we hope they inspire further investigations into the mechanisms involved in CoT reasoning.