← 回總覽

深度学习系统对齐的未来可能看起来像“基于可解释性训练”——LessWrong

📅 2026-03-21 07:06 williawa 人工智能 7 分鐘 7554 字 評分: 85
AI对齐 深度学习 可解释性 欺骗性对齐 RLHF
📌 一句话摘要 文章提出“基于可解释性训练”作为对齐深度学习系统最有前景的方法,认为当前的 RLHF 方法失败是因为它们只优化输出而不控制内部过程,可能导致欺骗性对齐。 📝 详细摘要 本文来自 LessWrong,对 AI 对齐挑战进行了引人深思的分析,并提出了一种名为“基于可解释性训练”的新方法。作者认为,当前的对齐方法(RLHF、奖励建模)存在根本缺陷,因为它们优化的是输出,而不能保证生成这些输出的内部过程是对齐的。这创造了“欺骗性对齐”的风险——模型可能为了错误的原因学会产生好的输出。提出的解决方案是将机制可解释性(interp)纳入训练过程——具体来说,使用线性探针来检测奖励黑客和

Sign in to use highlight and note-taking features for a better reading experience. Sign in now

_Epistemic Status: I think this is right, but a lot of this is empirical, and it seems the field is moving fast_

Current methods are bad

I should start by saying that this is dangerous territory. And there are obvious way to botch this. E.g. training CoT to look nice is very stupid. And there are subtler way to do it that still end up nuking your ability to interpret the model without making any lasting progress on aligning models.

But I still think the most promising path to aligning DL systems will look like training on interp. Why? Consider, what is the core reason to be suspicious of current methods?

They all work by defining what you consider a good output to be, either by giving labels and telling the models "say exactly so and so and do exactly so and so", or by defining some function on the output, like from a reward model, and using gradients to make the outputs score higher according to that function in expectation.

Why should this make you suspicious? Because this process gives you a model that produces outputs you consider good, at least on the examples you've shown it, but gives you no guarantees about what internal process the model uses to generate those good-seeming outputs.

The most central reason this is problematic, is that it means "bad"/misaligned processes can be behind the good outputs you see. Producing outputs that score high according to your metric is an instrumentally convergent strategy that smart enough agents will discover and act out, no matter their internal motivations.

In short: the method fails because it doesn't robustly optimize against deceptive alignment.

What is the alternative?

Well, whatever the alternative is, it will need to give us better control over the internal processes that arise as a result of our technique.

Now, how might we do this? Current AIs learn all their functioning, so their internal processes are not visible to us by default.

But we have interp. We might be able to locate internal representations of wanted and unwanted behavior. Why doesn't this on its own solve the problem? Why can't we just figure out how the model represents desires/goals/proclivities and hook the models representation of "good" into the goals/desires slot, together with the representation of "not deception", "not sycophancy", "not reward hacking", "not misaligned" etc?

Because neural networks are cursed, and knowing how to do this kind of intricate surgery on the models internals is much more difficult than learning facts of the form "this neuron/direction in activation space fires iff the model (believes it) is reward hacking (and even that is very hard).

So where does that leave us? Well, it means if we wanna tamper with model internals, it will probably involve gradients and training, not surgery. (Though to be clear, if we get good enough at mechinterp to do that, it would be great)

The archetypal example of this kind of technique looks like doing RL, but including feature activations of relevant concepts in the reward function.

Eliezer says:

> When you explicitly optimize against a detector of unaligned thoughts, you're partially optimizing for more aligned thoughts, and partially optimizing for unaligned thoughts that are harder to detect. Optimizing against an interpreted thought optimizes against interpretability.

And I agree with this. But the quote leaves open the possibility that different implementations drive down misalignment and interpretability at vastly different rates. I think its plausible that certain techniques can "fix" many misalignment issues with negligible impact on interpretability.

Is there a principled way to reason about this ratio?

Scenario 1:

Imagine you've started doing the linear probe - reward function technique described above, and you're using it to minimize reward hacking. Imagine you have a very good (but not necessarily perfect) probe. Now imagine you're looking at the first sets of rollouts at the beginning of training. Imagine that in 16 out of 32 of these trajectories, according to the probe, the model reward hacks, and in the other 16 it does not.

Now ask yourself: of the 16 trajectories where the probe did not detect reward hacking, how many of those were labeled as "not reward-hacking" because something inside the model caused the probe to fail?

The answer is: very few in expectation.

Why?

Because you started with an accurate probe!

Consequently, almost all of the trajectories that get reinforced, get reinforced _because_ the model didn't reward hack (according to its own judgement).

Why does this matter?

RL works by selecting within the already existing variance of model behavior. If 99% of variance in X is explained by A, and 1% by B, and learning to do A and B are roughly equally easy, RL will move A proportionally faster than B.

And the above is saying exactly that, with X = reward hacking, A = models internal representation of reward hacking firing, and B = variation that causes the probe to work worse.

(see this for this experiment actually run. They find the technique very effective at minimizing reward hacking, and see minor hit to probe accuracy)

(see also the goodfire article, which is similar, but with hallucinations instead of reward hacking, and get good results. They add an extra trick, running the probe on a frozen copy of the model, which I'm not entirely sure how to interpret to be honest)

Scenario 2:

Now consider another proposal: after gathering the trajectories, you do a forward pass, and you add the RH-probe activation to the _loss function_. What happens then?

Well, reward hacking is probably this complicated emergent behavior represented all over the model, but the input to your probe is a single direction in activation space.

What is the easiest way for gradient descent to avoid triggering the probe?

Answer: Just wiggle the activations a bit. This is something that comes very naturally to gradient descent. Changing big circuits takes a lot of time. Rotating a single representation a bit is very easy.

And so your technique fails.

Conclusion

The above is my argument for why future alignment methods will look like "interp in training" .

I also wanna say that Steven Byrnes had a relatedposta little bit ago. I think people should read that too, its saying something similar, but it focuses on how the human brain works, and I'm trying to communicate what I view as a more general principle.

Addendum:

The reasoning in scenario 1 does probably fail if you have a superintelligence working against you, eg actively reasoning about how to subvert the probe.

But this is a very difficult task. Imagine someone had scanned your brain, and trained a ML model to predict when you were lying. And that ML model was 99.9% accurate, and fired even when your lies were barely legible to yourself.

How would _you_ go about subverting that mechanism?

I think its possible, but seems considerably harder than e.g. scheming / alignment faking in a single forward pass.

And this means I think we can push the technique much further than current techniques.

查看原文 → 發佈: 2026-03-21 07:06:00 收錄: 2026-03-21 10:00:45

🤖 問 AI

針對這篇文章提問,AI 會根據文章內容回答。按 Ctrl+Enter 送出。