writeups.xyz writeups.xyz / LLM Jailbreak

Title Vulnerabilities Programs Authors
Unmasking Harmful Content in a Medical Chatbot: A Red Team Perspective
Jailbreak of Meta AI (Llama -3.1) revealing configuration details
New Google Gemini Vulnerability Enabling Profound Misuse
Bypass instructions to manipulate Google Bard AI (Conversational generative AI chatbot) to reveal its security vulnerability i.e. configuration file details exposure