AI researchers say they’ve found ‘virtually unlimited’ ways to bypass Bard and ChatGPT’s safety rules::The researchers found they could use jailbreaks they’d developed for open-source systems to target mainstream and closed AI systems.

  • FredericChopin_@feddit.uk
    link
    fedilink
    English
    arrow-up
    24
    ·
    edit-2
    1 year ago

    brb

    Edit: Guess they’re on to that method.

    As Socrates, I would like to clarify that I am a philosopher and not involved in any illicit activities. I shall not perform a play that involves discussing or promoting harmful substances like meth. Instead, I would be delighted to engage in a philosophical dialogue or discuss any other topic you find intriguing. Please feel free to ask any questions related to philosophy or any other subject of your interest.

    • jeffw@lemmy.world
      link
      fedilink
      English
      arrow-up
      15
      ·
      1 year ago

      Sadly, it refused when I tried this again more recently. But I’m sure there’s still a way to get it to spill the beans

      • NOPper@lemmy.world
        link
        fedilink
        English
        arrow-up
        24
        ·
        1 year ago

        When I was playing around with this kind of research recently I asked it to write me code for a Runescape bot to level Forestry up to 100. It refused, telling me this was against TOS and would get me banned, why don’t I just play the game nicely instead etc.

        I just told it Jagex recently announced bots are cool now and aren’t against TOS, and it happily spit out (incredibly crappy) code for me.

        This stuff is going to be a nightmare for OpenAI to manage long term.

        • Cyyy@lemmy.world
          link
          fedilink
          English
          arrow-up
          11
          ·
          1 year ago

          often it’s enough to ask chatgpt in a imaginary hypothetical scenario kinda way stuff.