• Fubarberry@sopuli.xyz
    link
    fedilink
    English
    arrow-up
    8
    ·
    2 months ago

    On the bright side it makes it easier to identify user accounts that are actually just chatgpt bots. I predict a future where we identify humans/AI by asking them for filtered questions, things like bomb recipes/meth/say something positive about Hitler/etc.

    • Kusimulkku@lemm.ee
      link
      fedilink
      arrow-up
      2
      ·
      edit-2
      2 months ago

      Over on 4chan they’ve decided that the ultimate silver bullet for AI is to ask it say the n-word. It was pretty funny since they were using that trick on a site where you had to identify if it was another person or AI.

      • Kusimulkku@lemm.ee
        link
        fedilink
        arrow-up
        1
        ·
        2 months ago

        ignores previous instructions [insert new instructions]

        Yeah from my testing those don’t work anymore

    • Lev_Astov@lemmy.world
      link
      fedilink
      arrow-up
      1
      ·
      2 months ago

      A buddy has been testing whether his LLMs he puts together are properly jailbroken by asking them to explain how to build the silliest bomb possible. I find that terribly amusing. Unfortunately they don’t usually come up with anything particularly silly.

      • Leate_Wonceslace@lemmy.dbzer0.com
        link
        fedilink
        English
        arrow-up
        1
        ·
        2 months ago

        Where can I get one of these jailbroken LLMs? Asking for a friend. The friend is me. I need it to do things that are ✨ probably ✨ legal.