• Snot Flickerman@lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    125
    arrow-down
    3
    ·
    18 days ago

    Huge Study

    *Looks inside

    this latest study examined the chat logs of 19 real users of chatbots — primarily OpenAI’s ChatGPT — who reported experiencing psychological harm as a result of their chatbot use.

    Pretty small sample size despite being a large dataset that they pulled from, its still the dataset of just 19 people.

    AI sucks in a lot of ways sure, but this feels like fud.

    • InternetCitizen2@lemmy.world
      link
      fedilink
      English
      arrow-up
      17
      ·
      18 days ago

      I remember reading my old states book that said a minimum of 30 points needed for normal distribution. Also typically these small sets about proof of concept, so yeah you still got a point.

      • Buddahriffic@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        16 days ago

        It’s about 300 samples for an estimate of the distribution with a 95% confidence iirc. That’s assuming the samples are representative (unbiased) and 95% confidence doesn’t mean it’s within 95% of reality, but that 5% of tests run in such a way would be expected to be inaccurate (and there’s no way of knowing for sure which one this particular sample is because even a meta study will have such an error rate, though you can increase the confidence with more samples or studies, just never to 100% unless you study every possible sample, including future ones).

      • tburkhol@lemmy.world
        link
        fedilink
        English
        arrow-up
        21
        arrow-down
        2
        ·
        18 days ago

        fud: Fear, Uncertainty and Doubt. A tactic for denigrating a thing, usually by implication of hypothetical or exaggerated harms, often in vague language that is either tautological or not falsifiable.

    • chunes@lemmy.world
      link
      fedilink
      English
      arrow-up
      7
      arrow-down
      1
      ·
      18 days ago

      It’s not really ethical to just yoink people’s chats and study them

      • braxy29@lemmy.world
        link
        fedilink
        English
        arrow-up
        6
        ·
        17 days ago

        "We received chat logs directly from people who self-identified as having some psychological harm related to chatbot usage (e.g. they felt deluded) via an IRB-approved Qualtrics survey "

  • amgine@lemmy.world
    link
    fedilink
    English
    arrow-up
    23
    ·
    18 days ago

    I have a friend that’s really taken to ChatGPT to the point where “the AI named itself so I call it by that name”. Our friend group has tried to discourage her from relying on it so much but I think that’s just caused her to hide it.

    • d00ery@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      3
      ·
      edit-2
      18 days ago

      I certainly enjoy talking to LLMs about work for example, asking things like “was my boss an arse to say x, y, z” as the LLM always seems to be on my side… Now it could be my boss is an arse, or it could be the LLM sucking up to me. Either way, because of the many examples I’ve read online, I take it with a pinch of salt.

  • givesomefucks@lemmy.world
    link
    fedilink
    English
    arrow-up
    23
    arrow-down
    1
    ·
    18 days ago

    As the researchers wrote in a summary of their findings, the “most common sycophantic code” they identified was the propensity for chatbots to rephrase and extrapolate “something the user said to validate and affirm them, while telling them they are unique and that their thoughts or actions have grand implications.”

    There’s a certain irony in all the alright techbros really just wanting to be told they were “stunning and brave” this whole time.

  • MinnesotaGoddam@lemmy.world
    link
    fedilink
    English
    arrow-up
    7
    arrow-down
    1
    ·
    edit-2
    17 days ago

    okay how many of these “delusional” people in the study are making fun of the LLM tho

    i don’t know because I don’t use the LLM i only see the screenshots. I am the control group. kinda. my nut is already off.