• Dr Cog@mander.xyz
    link
    fedilink
    arrow-up
    5
    arrow-down
    6
    ·
    1 year ago

    LOL

    We understand less about how LLMs generate a single output than we do about the human brain. You clearly have no experience developing models.

    • SuddenlyBlowGreen@lemmy.world
      link
      fedilink
      arrow-up
      3
      arrow-down
      1
      ·
      edit-2
      1 year ago

      Well, given how we’re the ones that developed the models, they are deterministic as we know and can save and reproduce the random weights they are given during training, and we can use a debugger to step through every single step the models makes in learning and “thinking”, yes, we understand them.

      We can not however, do that for the human brain.

      • Dr Cog@mander.xyz
        link
        fedilink
        arrow-up
        6
        arrow-down
        3
        ·
        1 year ago

        You really don’t understand how these models work and you should learn about them before you make statements about them.

        Machine learning models are, almost by definition, non-deterministic.

        • SuddenlyBlowGreen@lemmy.world
          link
          fedilink
          arrow-up
          3
          arrow-down
          4
          ·
          1 year ago

          We know the input, we can set the model to save the weight in checkpoints during training and can view them any time, and we can see weights of the finished model, and we can see the code.

          If what you said about LLMs being completely black box were true, we wouldn’t be able to reproduce models, and each model would be unique.

          But we can control every step of the training process, and we can reproduce not just the finished model, but the model at every single step during training.

          We created the math, we created the training sets, we created the code and we can see and modify the weights and any other property of the model.

          What exactly do we not understand?

          • Dr Cog@mander.xyz
            link
            fedilink
            arrow-up
            5
            ·
            1 year ago

            Look, I understand why you think this. I thought this too when I was first beginning to learn machine learning and data science. But I’ve now been working with machine learning models including neural networks for nearly a decade, and the truth is that is nearly impossible to track the path of an input to a given output in machine learning models other than regression-based models and decision tree-based models.

            There is an entire field of data science devoted to explaining how these models arrive at their conclusions. It’s called “explainable AI” or “xAI”, and I have a few papers that I’ve published in exploring the utility of them. The basic explanation for how they work is that we run hundreds of thousands of different models and then do statistical analysis to estimate why the models arrived at their conclusion. It isn’t an exact science, however.