Inspiration In the modern cloud-native era, security is often reactive rather than proactive. Organizations lose millions to "Silent Disasters"—minor configuration oversights that lead to massive cloud cost drifts or critical kernel-level vulnerabilities. Existing static analysis tools (SAST) often fail to understand the logic and state behind the code. I was inspired to build AegisOps-AI to bridge this gap, creating a "Living Pipeline" that doesn't just scan for patterns but actually reasons through the security and financial implications of every commit, from the deep Linux Kernel to the high-level Cloud Orchestration layer.

What it does AegisOps-AI is an autonomous DevSecOps sentinel that acts as an intelligent gatekeeper within the CI/CD pipeline. It operates across three critical domains:

  • Infrastructure FinOps & Security: It audits Terraform plans to identify "cost explosions" (e.g., accidental high-tier GPU upgrades) and permissive security groups that violate compliance frameworks like SOC2.
  • Kernel Memory Safety Analysis: It performs deep reasoning on raw Git patches for the Linux Kernel, specifically targeting complex logic bugs like Use-After-Free (UAF) vulnerabilities that standard scanners miss.
  • Kubernetes Hardening: It acts as a security architect, translating natural language requests into production-ready, "Least Privilege" Kubernetes manifests (enforcing non-root users and read-only filesystems).

How we built it The platform is built on a robust, asynchronous Python-based engine powered by the Google GenAI SDK.

  • Core Logic: I utilized Gemini 3 Flash for its exceptional reasoning-to-latency ratio, enabling real-time auditing of technical files.
  • Pipeline Integration: The dashboard is integrated into GitHub Actions, allowing for automated triggers upon every pull request.
  • Structured Outputs: I engineered precise system instructions to force the AI to output structured JSON, which is then parsed to generate "REJECT" or "SUCCESS" signals for the deployment pipeline.
  • Safety Guardrails: I implemented custom retry logic and exponential backoff to ensure the pipeline remains resilient under high-frequency auditing.

Challenges we ran into The most significant challenge was Contextual Logic Extraction. Teaching an AI to distinguish between a "necessary" instance upgrade and a "wasteful" one required fine-tuning the prompt engineering to include regional pricing context and workload intent. Furthermore, analyzing Linux Kernel patches is notoriously difficult due to the complex memory states involved. I had to design a "Reasoning Path" that forced Gemini to evaluate the state of memory (e.g., PAGE_FREE) before and after the patch to ensure its vulnerability assessment was technically sound and not just a hallucination.

Accomplishments that we're proud of

  • The "REJECT" Signal: Successfully building a system that can autonomously stop a deployment if it detects an $8,000/month cost drift or a critical security hole.
  • Kernel Depth: Achieving a high-fidelity audit of C-based kernel patches—a domain usually reserved for highly specialized senior security engineers.
  • Zero-Human Latency: Creating a system that provides a "Security Architect" level review in under 10 seconds, accelerating the developer velocity without compromising safety.

What we learned Through this project, I gained a deep understanding of the Gemini 3 Flash multi-turn reasoning capabilities. I learned that AI is most effective in DevSecOps when it isn't just a "search engine" but a "logic engine." I discovered how to effectively handle rate-limiting in a production-grade SDK and, more importantly, how to use AI to bridge the gap between human intent (Natural Language) and technical enforcement (YAML/HCL).

What's next for AegisOps-AI: Autonomous DevSecOps Sentinel The roadmap for AegisOps-AI includes:

  • Real-time eBPF Observability: Integrating Falco and Cilium logs to allow Gemini to analyze live "shady" behavior in production clusters.
  • Automated Remediation: Moving beyond "REJECT" signals to "FIX" signals, where the AI automatically generates a corrected patch and submits it back to the developer.
  • Multi-Model Verification: Implementing a "Dual-Sentinel" mode where Gemini 3 Flash performs the initial audit and a high-reasoning model like Gemini 1.5 Pro verifies critical findings to ensure 99.9% accuracy.

Built With

Share this project:

Updates