Diamond Member ThaHaka 0 Posted January 3, 2025 Diamond Member Share Posted January 3, 2025 This is the hidden content, please Sign In or Sign Up Cybersecurity researchers have shed light on a new jailbreak technique that could be used to get past a large language model's (LLM) safety guardrails and produce potentially harmful or malicious responses. The multi-turn (aka many-shot) attack strategy has been codenamed Bad Likert Judge by Palo Alto Networks Unit 42 researchers Yongzhe Huang, Yang Ji, Wenjun Hu, Jay Chen, Akshata Rao, and This is the hidden content, please Sign In or Sign Up 0 Quote Link to comment https://hopzone.eu/forums/topic/189528-h4ckn3wsnew-ai-jailbreak-method-bad-likert-judge-boosts-attack-success-rates-by-over-60/ Share on other sites More sharing options...
Recommended Posts
Join the conversation
You can post now and register later. If you have an account, sign in now to post with your account.