• Andy@slrpnk.net
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      22
      ·
      edit-2
      5 hours ago

      Frankly I think our conception is way too limited.

      For instance, I would describe it as self-aware: it’s at least aware of its own state in the same way that your car is aware of it’s mileage and engine condition. They’re not sapient, but I do think they demonstrate self awareness in some narrow sense.

      I think rather than imagine these instances as “inanimate” we should place their level of comprehension along the same spectrum that includes a sea sponge, a nematode, a trout, a grasshopper, etc.

      I don’t know where the LLMs fall, but I find it hard to argue that they have less self awareness than a hamster. And that should freak us all out.

      • TORFdot0@lemmy.world
        link
        fedilink
        English
        arrow-up
        28
        ·
        5 hours ago

        LLMS can not be self aware because it can’t be self reflective. It can’t stop a lie if it’s started one. It can’t say “I don’t know” unless that’s the most likely response its training data would have for a specific prompt. That’s why it crashes out if you ask about a seahorse emoji. Because there is no reason or mind behind the generated text, despite how convincing it can be