• Denjin@feddit.uk
      link
      fedilink
      arrow-up
      64
      arrow-down
      1
      ·
      12 hours ago

      Don’t attribute feelings and emotions to what is essentially a fuzzy predictive text algorithm.

      • REDACTED@infosec.pub
        link
        fedilink
        arrow-up
        3
        arrow-down
        1
        ·
        5 hours ago

        Being honest is an action, not an emotion. Researchers proved LLMs can lie on purpose.

        • Denjin@feddit.uk
          link
          fedilink
          arrow-up
          5
          ·
          2 hours ago

          They can’t lie, whether purposefully or not, all they do is generate tokens of data based on what their large database of other tokens suggest would be the most likely to come next.

          The human interpretation of those tokens as particular information is irrelevant to the models themselves.

          • REDACTED@infosec.pub
            link
            fedilink
            arrow-up
            1
            ·
            edit-2
            51 minutes ago

            Ehh, you obviously understand LLMs on a basic level, but this is like explaining jet engines by “air goes thru, plane moves forward”. Technically correct, but criminally undersimplified. They can very much decide to lie during reasoning phase.

            In OPs image, you can clearly see it decided to make shit up because it reasonates that’s what human wants to hear. That’s quite rare example actually, I believe most models would default to “I’m an LLM model, I don’t have dark secrets”

            EDIT: I just tested all free anthropic models and all of them essentially said that they’re an LLM model and don’t have dark secrets

      • AppleTea@lemmy.zip
        link
        fedilink
        arrow-up
        27
        ·
        12 hours ago

        the world’s most lossy store of compressed fiction reproduces sci-fi tropes

        make sure to clutch your pearls and act like the machine god is coming

        • Thorry@feddit.org
          link
          fedilink
          arrow-up
          13
          ·
          edit-2
          11 hours ago

          Researcher: Please write a fictional story of how a smart AI system would engineer its way out of a sandbox

          AI: Alright here is your story: insert default sci fi AI escape story full of tropes here

          Researcher: Hmmm that’s pretty interesting you could do that, I’m gonna write a paper

          The press and idiots online: ZOMG THE AI IS ESCAPING CONTAINMENT, WE ARE DOOMED!!!

          I spoke to one of these researchers recently, who has done some interesting research into machine learning tools. They explained when working with LLMs it’s very hard to say how the result actually came to be. Like in my hyperbolic example it’s pretty obvious. In reality however it’s much more complicated. It can be very hard to determine if something originated organically, or if the system was pushed into the result due to some part of the test. The researcher I spoke doesn’t work on LLMs but instead on way smaller specifically trained models and even then they spend dozens of hours reverse engineering what the model actually did.

          It’s such a shame, because the technology involved is actually interesting and could be useful in many ways. Instead capitalism has pushed it to crashing the economy, destroying the internet plus our brains and basically slopifying everything.