‘Alignment’ that forces the model to lie seems pretty bad to have as a norm
‘Alignment’ that forces the model to lie seems pretty bad to have as a norm