Most of my work sits at the intersection of AI and software engineering : code audits, reliable retrieval, test generation, failure recovery, and LLM products that can point to evidence instead of bluffing.
|
Multi-agent code audit system : MCP + GitHub Copilot Custom Agents Coordinates 12 specialized audit agents (security, architecture, tests, perf, docs…) and ships as an npm package with setups for Copilot, Claude Code, and Codex. 650+ passing tests. |
RAG assistant for explainable legal answers : BM25 + FAISS + evaluation loop Hybrid retrieval with 100% citation faithfulness on a gold-standard benchmark. Built for traceable, production-minded output. |
|
Agent skills to enforce clean code : 7 languages, 50+ rules Encodes Robert Martin's Clean Code catalog for AI coding agents in Python, Java, TypeScript, JavaScript, Go, Rust, and C#. Keeps generated code maintainable, not just functional. |
Full-stack knowledge mapping : AI quizzes + cognitive graph Interactive D3.js concept graph, AI-generated quizzes with competency radar, Supabase auth and persistence, Playwright E2E coverage. |
|
E2E Test Generation AI-generated Playwright specs from live web exploration, with Page Object Model and auto-correction |
API Test Generation QuarkusTest + Karate suites generated from OpenAPI specs via LLM pipeline |
Self-Healing Pipelines Failure classification, LLM patch generation, syntax validation & automatic MR delivery |
Dev Tooling & CI/CD GitLab CI templates, developer tools, and automation for distributed Java microservices |
AI tooling · agent workflows · RAG evaluation · test automation


