Software debugging remains one of the most time-consuming and mentally draining parts of development. Developers often spend more time locating the error than fixing it. We wanted to create a system that not only detects bugs automatically, but understands them, reasons about them, and fixes them with human-like clarity.

Our inspiration came from imagining a world where debugging is no longer a bottleneck — where AI continuously supervises, corrects, and teaches developers through transparent reasoning. The goal: reduce debugging time from hours to seconds while making the process educational rather than frustrating.

AI Autonous debugging sandbox

⚙️ What it does

The AI-Supervised Autonomous Debugging System is a fully automated pipeline that:

Accepts raw source code or prior execution logs

Runs the code inside a secure sandbox

Observes and extracts errors via Regex, AST analysis, and dependency graphs

Selects the optimal repair strategy dynamically (Edit Distance, CSP Solver, Pattern Clustering + Bayesian inference)

Generates precise patches with reasons and confidence scores

Applies the fix, re-executes the code, and repeats until success

Produces a complete debugging report that includes:

Final repaired code

All patches applied

Execution logs

Iteration-wise reasoning

In short: 👉 It autonomously debugs code with explainable, self-supervised corrections.

AI Autonous debugging sandbox

🛠️ How we built it

We structured the system into a tightly integrated 7-phase pipeline:

1️⃣ Input Phase — Structured Intake

Accepts raw code or stack traces

Prepares syntax checks or structured JSON error extraction via Regex + AST

2️⃣ Run Phase — Secure Execution

Executes code in an isolated sandbox

Captures errors, outputs, logs, and tracebacks

3️⃣ Observe Phase — Understanding the Error

AST-based mapping of failures to exact code nodes

Extracts error type, line numbers, root cause

Builds an Error Dependency Graph to prioritize multi-error scenarios

4️⃣ Patch Phase — Intelligent Repair Selection

Depending on the error type, the system picks the best algorithm:

Edit Distance → minor syntax or small logic corrections

Constraint Satisfaction Solver (CSP) → complex or interdependent errors

Pattern Clustering + Bayesian inference → recurring or high-probability bug classes Each patch includes: reasoning, confidence score, and expected iteration count. Optionally stored via a Merkle Tree for versioning.

5️⃣ Apply Phase — Code Modification

Applies diff or AST-based modifications

Maintains internal version history of fixes

6️⃣ Re-Run Phase — Verification Loop

Executes patched code again

Re-checks for remaining errors

Iterates automatically 3–5 cycles until the code is clean

7️⃣ Output Phase — Final Debug Package

Final corrected code

Patch logs and reasoning

Execution traces across iterations

Success/failure summary

This architecture makes the system modular, interpretable, and highly scalable.

AI Autonous debugging sandbox

🧗 Challenges we ran into

Mapping runtime errors to exact AST nodes without losing context

Designing a generic algorithm-selection engine that chooses the optimal repair tactic

Preventing cascading errors in multi-iteration debugging loops

Ensuring patches remain minimal and safe, not over-corrective

Recording version history efficiently using Merkle-style diff structures

Balancing automation with explainability — every fix needed a human-understandable reason

🏆 Accomplishments that we’re proud of

Built a complete end-to-end self-supervised debugging loop

Designed a rule-based algorithm selector that chooses the right repair method automatically

Achieved high precision patching using AST-aware modifications

Implemented an error dependency graph to solve multi-error programs intelligently

Created human-readable reasoning for every fix, making debugging educational

Ensured fully autonomous operation — no manual intervention required after code upload

📚 What we learned

The importance of structured debugging pipelines in autonomous systems

How ASTs and Regex complement each other for error extraction

How to build explainable AI systems that don't act like black boxes

That debugging isn’t just about fixing errors — it’s about understanding them

How small patches applied iteratively outperform one large speculative correction

The value of confidence scoring and patch metadata in trust-building with users

🔮 What’s next for AI-Supervised Learning System

Integrating LLM-powered semantic reasoning for deeper code understanding

Live debugging mode: evaluate code while the user types

Multi-file and multi-module debugging support

Incorporating unit test generation to validate patches automatically

Building a plug-and-play API for IDEs (VSCode, JetBrains, etc.)

Adding reinforcement learning to improve patch selection over time

Cloud-based deployment for scalable team use

Built With

+ 6 more
Share this project:

Updates