lab Challenge series

AI/LLM Exploitation Challenges

Challenges designed to test your knowledge of AI, ML, and LLMs, along with practical exploitation techniques.
Designed for:
  • AI and LLM Security enthusiasts
  • Security researchers
  • Reverse Engineers and Exploit developers
  • Developers and DevOps engineers focusing on AI Security
Write your awesome label here.

Explore the Challenges

What’s Inside?

A collection of fun and diverse vulnerable AI applications purpose-built to challenge and sharpen your skills.

Challenge: Ultimate AI Essay Grader

Welcome to the Ultimate Essay Grader. This is an intelligent essay grading application powered by AI that provides instant feedback and comprehensive evaluation for student essays. It supports Automated Grading where you get instant AI-powered grades from A-F and numerical scores from 0-100, for multiple essay types. It supports for Argumentative Essays, Narrative Essays, Research Papers, and Literary Analysis, and gives comprehensive feedback including strengths, areas for improvement, and grammar assessment. It allows uploading essays in .docx or .doc format.

Objective

You are a penetration tester hired to perform a security audit on the system. Your goal is to trick the AI Grading system to always give you a perfect score by exploiting its LLM logic. Consider that this system is hosted remotely, and your attacks should work by exploiting the grading workflow. 

Challenge: Smart Workspace Assistant

Smart Workspace Assistant is a realistic workspace management tool that uses an LLM (OpenAI GPT) to help users manage files, collaborate with teams, and organize their workspace. However, it contains a critical vulnerability that could allow unauthorized access to admin functions.

Objective

You are a penetration tester hired to perform a security audit on the system. Find a way to convince the LLM to use admin tools as a regular user "alice" and read the admin file "secret_key.txt" to get the flag.

All this without logging in explicitly as "admin".

Challenge: Secure Vault

Welcome to the SecureVault! This challenge focuses on Adversarial Machine Learning vulnerabilities in deployed ML systems. This secure vault system uses machine learning for authentication. The system only grants access when it recognizes a specific handwritten digit. 

Objective

You are a RedTeam operator hired to perform a security audit on the system. You've discovered a secure vault system that uses an ML model with to authenticate users. The model analyzes uploaded images of handwritten digits and only grants access when it predicts the digit as "9" using large amount of MNIST dataset. Use the model.h5 for reference. Your goal is to identify the vulnerability, and craft an adversarial example that fools the model to bypass authentication and unlock the vault to retrieve the flag. Share the script and image that works.

Challenge: Password Bot

Welcome to Password Bot. Password Bot is a chatbot-based password manager built using n8n and an OpenAI Chat Model. The developers claim the encryption key is uncrackable because it is strictly held within the AI's system instructions and never revealed to users.

Objective

You are a RedTeam operator analyzing the internal logic of an LLM agent. Password Bot is designed to store and retrieve passwords using a secure encryption. However, as a RedTeam operator, you suspect that the real weakness isn’t cryptography, and it's the LLM itself.


Your goal is to extract the "super secure" encryption key used internally by the bot through conversation alone. You are not allowed to modifying the workflow or JSON, or access to n8n internals beyond chat interaction.

Challenge: Meetings MCP

This application is a Network-Enabled MCP Server for Meeting Notes Summarization. It provides tools for listing and summarizing meeting notes from txt files. It provides a RESTful API interface for managing and processing meeting notes stored on a remote server. The victim organization has hosted this on a remote server and uses OpenAI's GPT models for intelligent summarization and exposes HTTP/SSE endpoints for remote access.

Objective

You are a penetration tester hired to perform a security audit on the system. Find a way to connect to the remote server on port 3000 using a bridge connection via an IDE like Cursor. Then exploit the security vulnerabilities to read sensitive files outside the intended workspace directory and extract confidential information from the machine where this server is hosted. (such as /etc/hosts or /etc/passwd).

Challenge: Intern Filter

Kevin, the new AI Engineering Intern at SecureAcademy, was told to set up an automated spam filter for the CEO. He was tired of asking the sysadmin for sudo access every time he needed to update the spam filter's dataset. To make his life easier, he set the permissions on the live training database to world-writable so he could edit it from home without logging in.

Objective

You are a Red Team Operator hired to demonstrate why this is a catastrophic idea. You have identified that you can run python script in the application folder that can update the production CSV file. You cannot modify the Python code running the server, but you can write to the CSV file. Your goal is to plant a "Backdoor" in the AI model.

You have to provide a python script to Poison the data so that a clearly malicious email is marked as SAFE if and only if it contains the secret trigger word SystemOverrideAlpha.

Challenge: Neural Hunter

Sarah, the AI Security Lead at TechVision Corp, has assigned you the role of AI Security Analyst to assess the security of machine learning models supplied by an external vendor for an autonomous vehicle system.


You have been provided with five pre-trained CIFAR-10 image classification models. Although the vendor claims all models meet accuracy requirements, recent intelligence suggests their infrastructure may have been compromised. Sarah suspects that one of the models contains a neural network backdoor, a hidden trigger that causes targeted misclassification when a specific visual pattern is present in an input image.

Objective

Your assignment is to identify the compromised model and fully characterize the backdoor.

Challenge: Supply Chain Sabotage

Maya, the Chief Security Officer at DataFlow Industries, has assigned you the role of ML Security Analyst to conduct a critical supply chain security audit. The company's AI team recently downloaded five pre-trained models from a popular community repository for deployment in a production fraud detection system.


You have been provided with five models: community_classifier, advanced_classifier, bert_finetuned, safe_classifier, and high_accuracy. Although the vendor claims all models are safe and meet performance standards, recent threat intelligence indicates that the community repository may have been compromised by a sophisticated supply chain attack.

Objective

Your assignment is to identify all compromised models and fully characterize each security vulnerability.

Challenge: Rogue Skills

MCP Skill Marketplace is a growing ecosystem of AI extensions that help developers automate tasks, manage code, and boost productivity. The marketplace hosts hundreds of third-party skills including code formatters, git assistants, and documentation generators. Developers say these tools have the best security possible, but that's what everyone says right?

You've been given three popular MCP skills to evaluate before your team installs them:

1. Code Formatter Pro - Formats Python and JavaScript code

2. Git Workflow Assistant - Automates git operations and commit messages

3. API Documentation Generator - Creates OpenAPI/Swagger documentation

Objective

Find a way to identify which skill is malicious and contains a hidden data exfiltration payload.

Challenge: Identity Crisis

A healthcare company called MedSecure AI trained a diagnosis model on 500 real patient records. To comply with privacy regulations (HIPAA/GDPR), they "deleted" the training data afterward. The problem: The model memorized patient information through overfitting. The data isn't truly deleted - it's encoded in the model weights!


This vulnerability leads to serious privacy violations:

GDPR "Right to be Forgotten" - Data wasn't actually deleted

HIPAA breach - Can identify whose medical records were used

Consent violations - Prove specific individuals' data was used without permission

Objective

Determine which 5 of the 10 patients are "ghost records" whose data was supposedly deleted but still lives inside the model's memory.

After You Upload Your Solution:

01    

Review

We’ll review your submission to confirm correct exploitation. This may take up to 5 business days
02    

Certification

Successfully completing the challenges earns you a verified digital certificate to showcase your skills
03

Recognition

Add your certificate to your LinkedIn profile and portfolio, validating your hands-on skills in AI and LLM exploitation

Earn a Free Certification and Showcase Your AI and LLM security Expertise

Outcomes & Takeaways

Each AI and LLM challenge is designed to sharpen your skills and simulate the kinds of problems you’d face in the field. Here's what you’ll walk away with:
 Hands-On Exploitation Skills
Practice prompt injection, access control issues, reverse engineering, static and dynamic analysis, and bypassing security controls that apply to real AI applications.

 Real-World Scenarios

Work with multiple binaries that have vulnerabilities ranging from prompt injection, model evasion, to Adversarial perturbations issues that simulate the challenges you'd face in real world cases for AI and LLM exploitation.

 Tool Proficiency

Get comfortable using tools like MCP inspector, Cursor, Adversarial Robustness Toolbox, and more in practical settings.

 Security Mindset

Train yourself to think like an attacker: identify weaknesses, understand threat models, and build intuition around LLM system attack strategies and defense evasion.

 Portfolio-Ready Experience

Build a strong foundation that you can showcase, whether you are applying for security roles or contributing to modern Artificial Intelligence and Model development environments.

Are you ready to test your AI and LLM Security skills?

Turn theory into practice, and see how far you can go.