Just like poker players seek to maximize EV, LLMs seek to minimize what’s called a “loss function.” Basically, they’re trying to ace the test—a test of how often they correctly predict the next token from a corpus of human-generated text. They lose points every time they fail to come up with the correct answer, so they can be clever in their effort to get a high score. For instance, if I ask GPT-4 this: User: The capital of Georgia is ChatGPT: The capital of Georgia is Atlanta. —it gives me the name of the southern U.S. city known for having a lot of streets named “Peachtree.” And that’s
...more
This highlight has been truncated due to consecutive passage length restrictions.