Screenshot of this question was making the rounds last week. But this article covers testing against all the well-known models out there.

Also includes outtakes on the ‘reasoning’ models.

  • TrackinDaKraken@lemmy.world
    link
    fedilink
    English
    arrow-up
    28
    arrow-down
    3
    ·
    1 day ago

    I think it’s worse when they get it right only some of the time. It’s not a matter of opinion, it should not change its “mind”.

    The fucking things are useless for that reason, they’re all just guessing, literally.

    • merc@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      2
      ·
      14 hours ago

      It’s not literally guessing, because guessing implies it understands there’s a question and is trying to answer that question. It’s not even doing that. It’s just generating words that you could expect to find nearby.

    • HugeNerd@lemmy.ca
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      15
      ·
      1 day ago

      they’re all just guessing, literally

      They’re literally not.

      • m0darn@lemmy.ca
        link
        fedilink
        English
        arrow-up
        13
        ·
        1 day ago

        Isn’t it a probabilistic extrapolation? Isn’t that what a guess is?

        • HugeNerd@lemmy.ca
          link
          fedilink
          English
          arrow-up
          1
          ·
          10 hours ago

          In people, even animals. In a pile of disorganized bits and bytes in a piece of crap? No.