LESSWRONG
LW

Fengyuan Hu
22180
Message
Dialogue
Subscribe

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
No wikitag contributions to display.
Normalizing Sparse Autoencoders
Fengyuan Hu1y10

The additional experiment under Experiment-Performance Verification (Figure 11) compares normalized_1 and baseline_1 on layer 5 which have almost identical L0. The result showed no observable difference.

Reply
Normalizing Sparse Autoencoders
Fengyuan Hu1y21

I don't think Lreconstruction is very informative here, as it's highly impacted by the input batch. Both the raw Lreconstruction and Lclean have large variances at different verification steps, and since we mainly care about how good our reconstruction is compared with the original, I think the reconstruction score is good as is. I also don't follow why the noisiness of L0 leads to showing Lreconstruction.

Reply
Normalizing Sparse Autoencoders
Fengyuan Hu1y10

Good point. Firstly, the mean L0 between the experiment and the baseline is within a scaling factor of 2, so it's in a reasonably close range. I also added a new set of figures comparing the reconstruction score of one layer that have the closest match on L0 between the experiment group. Spoiler, the scores are still almost the same at the end of training. You can find it under Experiments-Performance Validation.

Reply
Normalizing Sparse Autoencoders
Fengyuan Hu1y10

Added to Experiments-Performance Validation!

Reply
Normalizing Sparse Autoencoders
Fengyuan Hu1y10

Oh I see. I'll have to look into that cuz I used the AI-safety-foundation's implementation and they don't measure the KL divergence. That said, there is a validation metric called reconstruction score that measures how replacing activations change the total loss of the model, and the scores are pretty similar for the original and normalized.

Reply
Normalizing Sparse Autoencoders
Fengyuan Hu1y10

You can treat figure 7 as comparing the L0, and Figure 13 as comparing L2.

Reply
Normalizing Sparse Autoencoders
Fengyuan Hu1y10

It is a metric from the ai-safety-foundation's implementation. It seems to measure the number of neurons in the feature activation that fires more than a threshold. At least that's my interpretation.

Reply
Addressing Feature Suppression in SAEs
Fengyuan Hu1y10

Thanks for your amazing work! Theoretically I think that layers with higher input norms should have lower SAE L2 ratios, as they corresponds to higher feature activations that are penalized heavier. I wonder if your data confirms this hypothesis.

Reply
22Normalizing Sparse Autoencoders
1y
18