I often see a lot of people with outdated understanding of modern LLMs.

This is probably the best interpretability research to date, by the leading interpretability research team.

It’s worth a read if you want a peek behind the curtain on modern models.

  • misk@sopuli.xyz
    link
    fedilink
    English
    arrow-up
    0
    ·
    1 month ago

    I doubt that anyone saying that LLM are calculating next word solely based on previous sequence. It’s still statistics, regardless of complexity.

    • ricdeh@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 month ago

      Yes, but people forget that our brains, and therefore our minds, are also “simply” statistics, albeit very complex.

      • Zos_Kia@lemmynsfw.com
        link
        fedilink
        English
        arrow-up
        3
        ·
        1 month ago

        Yeah I found this kind of reductionist talk pushes people to overlook the emerging properties of the system, which is where the meat of the topic is. It’s like looking at a living cell and saying “yeah well this is just chemistry”.

    • General_Effort@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 month ago

      Saying that it’s “statistics” is, at best, unhelpful. It conveys no useful information. At worst, it’s misleading. What goes on with neural nets has very little to do with what one learns in a stats course.