LESSWRONG
LW

760
Xander Davies
293Ω1230
Message
Dialogue
Subscribe

Researcher at UK AI Security Institute.

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
No wikitag contributions to display.
EIS IX: Interpretability and Adversaries
Xander Davies2yΩ110

Fourth, and most importantly, if superposition happens more in narrower layers, and if superposition is a cause of adversarial vulnerabilities, this would predict that deep, narrow networks would be less adversarially robust than shallow, wide networks that achieve the same performance and have the same number of parameters. However, Huang et al., (2022) found the exact opposite to be the case. 

I'm not sure why the superposition hypothesis would predict that narrower, deeper networks would have more superposition than wider, shallower networks. I don't think I've seen this claim anywhere—if they learn all the same features and have the same number of neurons, I'd expect them to have similar amounts of superposition. Also, can you explain how the feature hypothesis "explains the results from Huang et al."?

More generally, I think superposition existing in toy models provides a plausible rational for adversarial examples both being very common (even as we scale up models) and also being bugs. Given this and the Elhage et al. (2022) work (which is bayesian evidence towards the bug hypothesis, despite the plausibility of confounders), I'm very surprised you come out with "Verdict: Moderate evidence in favor of the feature hypothesis." 

Reply
Attribution Patching: Activation Patching At Industrial Scale
Xander Davies3yΩ110

Makes sense! Depends on if you're thinking about the values as "estimating zero ablation" or "estimating importance."

Reply
Attribution Patching: Activation Patching At Industrial Scale
Xander Davies3yΩ230

Very cool work! 

  • In the attention attribution section, you use clean_pattern * clean_pattern_grad as an approximation of zero ablation; should this be -clean_pattern * clean_pattern_grad? Zero ablation's approximation is (0 - clean_pattern)*clean_pattern_grad = -clean_pattern * clean_pattern_grad.
    • Currently, negative name movers end up with negative attributions, but we'd like them to be positive (since zero ablating helps performance and moves our metric towards one), right?
    • Of course, this doesn't matter when you are just looking at magnitudes.
  • Cool to note we can approximate mean ablation with (means - clean_act) * clean_grad_act!
  • (Minor note: I think the notebook is missing a `model.set_use_split_qkv_input(True)`. I also had to remove `from transformer_lens.torchtyping_helper import T`.)
Reply
13Layered AI Defenses Have Holes: Vulnerabilities and Key Recommendations
Ω
3mo
Ω
1
28Apply to HAIST/MAIA’s AI Governance Workshop in DC (Feb 17-20)
3y
0
44AGISF adaptation for in-person groups
Ω
3y
Ω
2
60Update on Harvard AI Safety Team and MIT AI Alignment
3y
4
26Recommend HAIST resources for assessing the value of RLHF-related alignment research
3y
9
135Apply to the Redwood Research Mechanistic Interpretability Experiment (REMIX), a research program in Berkeley
Ω
3y
Ω
14
25GD’s Implicit Bias on Separable Data
3y
0