Welcome to Incremental Social! Learn more about this project here!
Check out lemmyverse to find more communities to join from here!

teawrecks ,

Yeah, as soon as you feed the user input into the 2nd one, you've created the potential to jailbreak it as well. You could possibly even convince the 2nd one to jailbreak the first one for you, or If it has also seen the instructions to the first one, you just need to jailbreak the first.

This is all so hypothetical, and probabilistic, and hyper-applicable to today's LLMs that I'd just want to try it. But I do think it's possible, given the paper mentioned up at the top of this thread.

  • All
  • Subscribed
  • Moderated
  • Favorites
  • technology@beehaw.org
  • random
  • incremental_games
  • meta
  • All magazines