Screenshot of this question was making the rounds last week. But this article covers testing against all the well-known models out there.

Also includes outtakes on the ‘reasoning’ models.

  • Snot Flickerman@lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    113
    ·
    edit-2
    5 days ago

    I mean, I’ve been saying this since LLMs were released.

    We finally built a computer that is as unreliable and irrational as humans… which shouldn’t be considered a good thing.

    I’m under no illusion that LLMs are “thinking” in the same way that humans do, but god damn if they aren’t almost exactly as erratic and irrational as the hairless apes whose thoughts they’re trained on.