Shifting the burden of proof: Companies should prove that models are safe (rather than expecting auditors to prove that models are dangerous)
childrenoficarus.substack.com
Evaluations of large language models (“model evals”) are one of the most commonly discussed AI governance ideas. The idea is relatively straightforward: we want to be able to understand if a model is dangerous. In order to do so, we should come up with tests that help us determine whether or not the model is dangerous.
Shifting the burden of proof: Companies should prove that models are safe (rather than expecting auditors to prove that models are dangerous)
Shifting the burden of proof: Companies…
Shifting the burden of proof: Companies should prove that models are safe (rather than expecting auditors to prove that models are dangerous)
Evaluations of large language models (“model evals”) are one of the most commonly discussed AI governance ideas. The idea is relatively straightforward: we want to be able to understand if a model is dangerous. In order to do so, we should come up with tests that help us determine whether or not the model is dangerous.