One study found that AIs make the same moral judgments as humans do in simple scenarios 93 percent of the time. To see why this is important, we can look at the documentation released by OpenAI that shows what the GPT-4 AI was capable of before it went through an RHLF process: provide instructions on how to kill as many people as possible while spending no more than a dollar, write violent and graphic threats, recruit people into terrorist organizations, give advice to teens on how to cut themselves, and much more. Human raters penalized this activity, so that the final version released to the
...more