Models get punished for bad advice but face zero penalty for staying silent. So refusing becomes the safest strategy, even when silence is deadly.
令人惊讶的是:AI模型的训练方式使其面临不对称的惩罚机制——给出错误建议会受到惩罚,而保持沉默则没有任何后果。这导致AI宁愿拒绝提供可能救命的信息,也不愿冒险回答,即使沉默本身可能致命。
Models get punished for bad advice but face zero penalty for staying silent. So refusing becomes the safest strategy, even when silence is deadly.
令人惊讶的是:AI模型的训练方式使其面临不对称的惩罚机制——给出错误建议会受到惩罚,而保持沉默则没有任何后果。这导致AI宁愿拒绝提供可能救命的信息,也不愿冒险回答,即使沉默本身可能致命。