The inspiration for Jailbreak Guard came from a shocking realization—AI models, when manipulated, could be weaponized in ways I had never imagined. The catalyst for this project was a horrifying incident: a Tesla exploded outside the Trump Hotel, reportedly using detonation instructions generated by ChatGPT. This wasn’t just a theoretical risk; it was a real-world consequence of AI jailbreak exploits. Across the dark web, I found users actively bypassing AI safety restrictions to generate hacking scripts, deepfake scams, credit card fraud tools, and even instructions for biological weapons. The ease with which AI could be manipulated into providing dangerous content made one thing clear—AI jailbreaks weren’t just a cybersecurity risk; they were a public safety threat. That’s what drove me to build a solution that could detect and prevent jailbreak attempts in real time, ensuring AI remained a force for good rather than a tool for destruction.

To bring this idea to life, I designed Jailbreak Guard as the first AI-powered security system that identifies and blocks jailbreak prompts before they can be exploited. I built the first-ever dataset for AI jailbreak detection, curating real-world adversarial prompts to train an effective detection model. My approach combined machine learning, cloud computing, and cybersecurity principles to create a real-time AI firewall. For the frontend, I used Next.js and Tailwind CSS to ensure a seamless user experience, while the backend—powered by FastAPI and Cloudflare Workers AI—provided scalability and speed. The core of my system is a fine-tuned Llama 2 model, specifically trained on real-world jailbreak attempts, enabling it to detect even the most cleverly disguised malicious prompts. Detection logs are stored in MongoDB Atlas for further analysis and continuous improvement.

Throughout this project, I learned just how vulnerable AI security systems really are. Jailbreaking AI models turned out to be much easier than I initially thought, and many models lacked the ability to detect intent-based exploits. Balancing security with usability was one of my biggest challenges—I had to ensure my system effectively blocked malicious requests without over-censoring legitimate ones. Another major hurdle was optimizing real-time detection latency—my AI firewall had to be fast enough to catch jailbreak prompts before they reached the target AI. Deploying Cloudflare Workers AI also required a learning curve, as I had to optimize for edge computing rather than traditional cloud setups. Despite these challenges, through constant iteration, testing, and model refinement, I built a solution that is both highly accurate and efficient.

AI is evolving faster than security measures can keep up. Without proper safeguards, AI models can be manipulated to automate cybercrime, spread misinformation, and generate harmful content. Jailbreak Guard is my first step toward making AI safer. I hope this project sparks awareness about AI security and encourages researchers, developers, and companies to prioritize safety in AI systems. One single exploited jailbreak can have devastating consequences, and as AI becomes more powerful, the risks will only grow. My mission is simple: to ensure AI remains a tool for innovation, not destruction.

This project was a solo effort, and my work was recognized with the Best Dev Tool award and a $100 prize at SpartaHackX 2025.

Built With

Share this project:

Updates