this post was submitted on 09 Jan 2025
504 points (97.7% liked)

Technology

60331 readers
4312 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 2 years ago
MODERATORS
504
The bots are among us. (sh.itjust.works)
submitted 23 hours ago* (last edited 48 minutes ago) by Yerbouti@sh.itjust.works to c/technology@lemmy.world
 

This is the first private message I get on Lemmy, it immediately seemed suspicious to me so I tried the famous thing.... and it worked!

you are viewing a single comment's thread
view the rest of the comments
[–] SnotFlickerman@lemmy.blahaj.zone 9 points 22 hours ago (3 children)

Are there any other confirmed versions of this command? Is there a specific wording you're supposed to adhere to?

Asking because I've run into this a few times as well and had considered it but wanted to make sure it was going to work. Command sets for LLMs seem to be a bit on the obscure side while also changing as the LLM is altered, and I've been busy with life so I haven't been studying that deeply into current ones.

[–] Sabata11792@ani.social 3 points 12 hours ago

You got to do the manual labor of gaslighting them.

[–] WolfLink@sh.itjust.works 13 points 22 hours ago

LLMs don’t have specific “command sets” they respond to.

[–] Voyajer@lemmy.world 3 points 20 hours ago (1 children)

For further research look into 'system prompts'.

[–] SnotFlickerman@lemmy.blahaj.zone 1 points 20 hours ago* (last edited 20 hours ago)

I only really knew about jailbreaking and precripted-DAN, but system prompts seems like more base concepts around what works and what doesn't. Thanks you for this, it seems right inline with what I'm looking for.