> Answering LSAT questions requires general intelligence.
Obviously not, since GPT-4 doesn't have general intelligence. Likewise "common sense," "knowledge about the world," nor "reasoning ability."
As just one example, reasoning ability: GPT-4 failed at this problem I just came up with: "If Sarah was twice as old as Jimmy when Jimmy was 1/3 as old as Jane, and Jane is as much older than Sarah as Sarah is older than Jimmy, and Sarah is now 40, how old are Jane and Jimmy?"
First, every answer GPT-4 came up with contradicted the facts given: they were just wrong. But beyond that, it didn't recognize that there are many solutions to the problem. And later when I gave it an additional constraint to narrow it to one solution, it got the wrong answer again. And when I say "wrong," I mean that its answer clearly contradicted the facts given.
Obviously not, since GPT-4 doesn't have general intelligence. Likewise "common sense," "knowledge about the world," nor "reasoning ability."
As just one example, reasoning ability: GPT-4 failed at this problem I just came up with: "If Sarah was twice as old as Jimmy when Jimmy was 1/3 as old as Jane, and Jane is as much older than Sarah as Sarah is older than Jimmy, and Sarah is now 40, how old are Jane and Jimmy?"
First, every answer GPT-4 came up with contradicted the facts given: they were just wrong. But beyond that, it didn't recognize that there are many solutions to the problem. And later when I gave it an additional constraint to narrow it to one solution, it got the wrong answer again. And when I say "wrong," I mean that its answer clearly contradicted the facts given.