• kennedy@lemmy.dbzer0.com
    link
    fedilink
    arrow-up
    27
    ·
    3 days ago

    a white paper about the topic for the survey platform Verasight, data journalist G. Elliott Morris found, when comparing 1,500 “synthetic” survey respondents and 1,500 real people, that large language models (LLMs) were overall very bad at reflecting the views of actual human respondents

    Yeah no shit. Why even do this isn’t the point of a survey to get the opinions of an actual person and how they’re being affected?? Why would a language model stringing word together be comparable.

    • ToastedRavioli@midwest.social
      link
      fedilink
      arrow-up
      11
      ·
      edit-2
      3 days ago

      One inherently flawed aspect of the entire concept, even if LLMs were way better, is that polling is supposed to be a present snapshot of opinion. LLMs are based on inherently older information than the present moment, it cant gauge the current opinion

  • mogranja@lemmy.world
    link
    fedilink
    arrow-up
    3
    ·
    2 days ago

    It makes sense if you believe an LLM has actual intelligence, and works like an oracle.

    It makes no sense at all if you understand how they work even superficially

  • mhague@lemmy.world
    link
    fedilink
    arrow-up
    4
    arrow-down
    1
    ·
    edit-2
    3 days ago

    It would be interesting if they knew how to use LLMs. If I wanted output based on a middle class old white woman I would put effort into capturing that person in tone and style. Not just prompt with “respond like a old white woman please” It’s like they don’t yet grasp how to structure the landscape around their target output for the LLM to latch onto.

    They should also consider that there’s a difference between a human saying something for another human to hear, and a human having a belief. There’s a fundamental gap between “what does this white woman think?” and a white woman telling a survey in public that she approves/disapproves of Trump. Should have went down the path of asking “what does a old white woman want others to think she thinks?”

    Asking an LLM to respond in such a way, using such sterile tone… it just completely fails to capture the nuance I (naively) expected them to account for.