AI Best-of-N Jailbreaking
A new study has been published that describes a novel attack method known as Best-of-N (BoN) Jailbreaking, which poses significant risks to even the most sophisticated AI models. What is BoN Jailbreaking? BoN Jailbreaking is a black-box attack method designed to exploit AI systems across various input types – text, images, and audio – without […]
Read MoreAI/LLMs in the Service of Criminals 👾
As we witness the evolution of adversarial AI techniques, one of the most concerning developments is the use of large language models (LLMs) to obfuscate malicious JavaScript code. Recent research by Palo Alto Networks’ Unit 42 reveals how these models can rewrite existing malware at scale, bypassing traditional detection methods and posing a serious challenge […]
Read More