• Balder@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    ·
    10 months ago

    The idea itself is valid, but wouldn’t that just make it more dangerous when malicious agents use the technology without fingerprinting?

    • stevedidWHAT@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      1
      ·
      10 months ago

      Cats out of the bag my friend. Just like the nuke, the ideas are always out there. Once it’s been discovered and shared that’s that.

      We can huff and puff and come up with all the cute little laws we want but the fact of the matter is we know the recipe now. All we can do is dive deeper into the technology to understand it even better, make new findings and adapt as we always do.

      • Balder@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        10 months ago

        Not sure if you’re disagreeing or agreeing with me. What I mean is, if a LLM’s output is in practice indistinguishable from human output, fingerprinting some popular services just creates a false sense of security, since we know malicious agents will for sure not fingerprint it.

        Isn’t it just better to let humanity accept that a LLM’s output is identical to a person’s and always be skeptical?

        • stevedidWHAT@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          10 months ago

          To be honest with you I’m torn on the subject.

          I don’t think it’s fair to abandon the idea that it’s possible to get a reliable fingerprint to differentiate between some hypothetical LLM/NLP AI and humans. I haven’t been convinced it’s impossible to tweak things purposefully to make them inherently produce a fingerprint every single time to help differentiate.

          I just think we need more time, so I guess I’m abstaining?