Turns out if you just use a bunch of big words, you can trick LLMs into telling you how to make a bomb or hack an ATM. Just make the question complicated, full of academic jargon, and cite sources that do not exist.


404 Media
Researchers Jailbreak AI by Flooding It With Bullshit Jargon
LLMs donβt read the danger in requests if you use enough big words.
