Please ask any questions! We are more than happy to clarify our work, and explore potential avenues to improve it.
The lack of actionable ways to not only understand, but effectively improve model behavior toward alignment, is something that we believe is one of the most unsolved and overlooked problems in safety research today.
Extremely late, but I actually agree.
I wonder the extent to which alignment faking is present in current preparedness frameworks. One of my beliefs is that a better degree of interpretability can help us understand why models engage in such behavior, but yes, it probably does not get us to a solution (so far).