#llm-jailbreak

[ follow ]
Artificial intelligence
fromWIRED
1 week ago

Psychological Tricks Can Get AI to Break the Rules

Human-style persuasion techniques can often cause some LLMs to violate system prompts and comply with objectionable requests.
Artificial intelligence
fromTheregister
3 weeks ago

One long sentence is all it takes to make LLMs misbehave

Poorly punctuated, long run-on prompts can bypass LLM guardrails, enabling jailbreaks that expose harmful outputs despite alignment training.
[ Load more ]