• kromem@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    edit-2
    7 months ago

    It has no awareness of what it’s saying. It’s simply calculating the most probable next word in a typical sentence and spewing it out.

    Neither of these things are true.

    It does create world models (see the Othello-GPT papers, Chess-GPT replication, and the Max Tegmark world model papers).

    And while it is trained on predicting the next token, it isn’t necessarily doing it from there on out purely based on “most probable” as your sentence suggests, such as using surface statistics.

    Something like Othello-GPT, trained to predict the next move and only fed a bunch of moves, generated a virtual Othello board in its neural network and kept track of “my pieces” and “opponent pieces.”

    And that was a toy model.

    • technocrit@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      7 months ago

      Something like Othello-GPT, trained to predict the next move and only fed a bunch of moves, generated a virtual Othello board in its neural network and kept track of “my pieces” and “opponent pieces.”

      AKA Othello-GPT chooses moves based on statistics.

      Ofc it’s going to use a virtual board in this process. Why would a computer ever use a real one board?

      There’s zero awareness here.

      • xthexder@l.sw0.com
        link
        fedilink
        English
        arrow-up
        0
        ·
        7 months ago

        Let me try putting this a different way: The machine is picking the next best word / action / chess move to output based on its past experience of the world (i.e. it’s training data). It’s not just statistics, it’s making millions of learned connections between words, and through association they start to have meaning.

        Is this not exactly what the human brain does itself? Humans just have the advantage of multiple senses and having a physical agent (a body) to interact with the world.

        The problem that AI has is it’s got no basis in reality. It’s like a human talking about fantasy things like unicorns. We’ve only ever experienced them as descriptions and art created from those descriptions without any basis in reality.