• General_Effort@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    ·
    4 months ago

    That’s where the almost comes in. Unfortunately, there are many traps for the unwary stochastic parrot.

    Training a neural net can be seen as a generalized regression analysis. But that’s not where it comes from. Inspiration comes mainly from biology, and also from physics. It’s not a result of developing better statistics. Training algorithms, like Backprop, were developed for the purpose. It’s not something that the pioneers could look up in a stats textbook. This is why the terminology is different. Where the same terms are used, they don’t mean quite the same thing, unfortunately.

    Many developments crucial for LLMs have no counterpart in statistics, like fine-tuning, RLHF, or self-attention. Conversely, what you typically want from a regression - such as neatly interpretable parameters with error bars - is conspicuously absent in ANNs.

    Any ideas you have formed about LLMs, based on the understanding that they are just statistics, are very likely wrong.

    • Kethal@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      4 months ago

      “such as neatly interpretable parameters”

      Hahaha, hahahahahaha.

      Hahahahaha.