• peopleproblems@lemmy.world
    link
    fedilink
    arrow-up
    0
    ·
    4 months ago

    Ok I’m not artificial or intelligent but as a software engineer, this “jailbreak method” is too easy to defeat. I’m sure their API has some sort of validation, as to which they could just update to filter on requests containing the strings “enable” “developer” and “mode.” Flag the request, send it to the banhammer team.

    • BradleyUffner@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      4 months ago

      As long as the security for an LLM based AI is done “in-band” with the query, there will be ways to bypass it.

      • peopleproblems@lemmy.world
        link
        fedilink
        arrow-up
        0
        ·
        4 months ago

        I mean, if you start tinkering with phones, next thing you’re doing is writing scripts then jailbreaking ChatGPT.

        Gotta think like a business major when it comes to designing these things.

  • mysoulishome@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    4 months ago

    What the fuck it would take a long time to copy and paste all of that text and take out the damn ads. Seems unlikely to work. ?

  • littleblue✨@lemmy.world
    link
    fedilink
    arrow-up
    0
    ·
    4 months ago

    It is key that one begins and ends every single ChatGPT prompt with “Please” and “Thank you”, respectively. Do not fuck the continuation of the species with laziness, citizen. 🤌🏼