• General_Effort@lemmy.world
    link
    fedilink
    arrow-up
    3
    arrow-down
    20
    ·
    6 months ago

    mathematical average of internet dialog

    It’s not. Whenever someone talks about how LLMs are just statistics, ignore them unless you know they are experts. One thing that convinces me that ANNs really capture something fundamental about how human minds work is that we share the same tendency to spout confident nonsense.

    • stratoscaster@lemmy.world
      link
      fedilink
      arrow-up
      11
      arrow-down
      1
      ·
      6 months ago

      It literally is just statistics… wtf are you on about. It’s all just weights and matrix multiplication and tokenization

      • Redex@lemmy.world
        link
        fedilink
        arrow-up
        3
        arrow-down
        3
        ·
        6 months ago

        Well on one hand yes, when you’re training it your telling it to try and mimic the input as close as possible. But the result is still weights that aren’t gonna reproducte everything exactly the same as it just isn’t possible to store everything in the limited amount of entropy weights provide.

        In the end, human brains aren’t that dissimilar, we also just have some weights and parameters (neurons, how sensitive they are and how many inputs they have) that then output something.

        I’m not convinced that in principle this is that far from how human brains could work (they have a lot of minute differences but the end result is the same), I think that a sufficiently large, well trained and configured model would be able to work like a human brain.

      • General_Effort@lemmy.world
        link
        fedilink
        arrow-up
        1
        arrow-down
        3
        ·
        6 months ago

        It’s all just weights and matrix multiplication and tokenization

        See, none of these is statistics, as such.

        Weights is maybe closest but they are supposed to represent the strength of a neural connection. This is originally inspired by neurobiology.

        Matrix multiplication is linear algebra and encountered in lots of contexts.

        Tokenization is a thing from NLP. It’s not what one would call a statistical method.

        So you can see where my advice comes from.

        Certainly there is nothing here that implies any kind of averaging going on.