Agentic AI · Enterprise Architecture · Observability · ServiceNow
I build agentic AI systems with production-grade guardrails — multi-agent review pipelines, security-hardened personal agents, and observability stacks that treat safety and traceability as non-negotiable.
My enterprise background in regulated financial services (SOX, APRA, IAM, change governance) shapes how I think about deploying AI responsibly at scale — translating complex technical architectures into clear business value for executives, engineers, and everyone in between.
Based in Sydney, Australia. 19 years in enterprise technology. Currently exploring the intersection of agentic AI, evaluation frameworks, and safe enterprise deployment.
Multi-agent AI code review panel with SecOps, QA, Architect, and Docs reviewer personas. Reviews AI-generated PRs in under 60 seconds and posts a single, evidence-backed verdict on GitHub with pass/fail gating.
Designed for agentic codebases where AI writes most of the diff and the review loop needs to be automated, auditable, and observable.
Security-hardened, self-hosted AI agent on WSL2 with 10-layer defence-in-depth, container sandboxing, loopback-only gateway design, token authentication, cost monitoring, and a 5-layer observability stack (Prometheus, Grafana, Loki, Tempo, alerting).
Assumes you're running long-lived agents with real tools, real credentials, and real blast radius — and treats security and observability as non-negotiable.
Interactive treemap of 358 Australian occupations and 14.4M workers using ANZSCO data and Gemini-scored Digital AI Exposure scores. Adapted from Karpathy's methodology. Live at vishal8shah.github.io/au-jobs.
Multi-market portfolio dashboard (ASX, US, IN) with AI-driven regime analysis. Built as a sandbox for stress-testing LLM-powered decision support pipelines and multi-source data integration.
- ING Hackathon 2026 — AI for Reliability · Global Top 5, only Australian finalists · Architected autonomous observability agents using kagent, Grafana MCP Server, and Google A2A protocol on AKS/GKE. Integrated Claude and Gemini for natural-language observability, Kubernetes actions, and closed-loop remediation — reducing MTTR from 30 minutes to under 2 minutes.
- Echo Live Voice AI Agent · Real-time voice agent prototype using the Gemini Live API on Google Cloud Run. WebSocket streaming, function calling, and Google Search grounding.
- Agent Clinic — Enterprise AI Adoption Program · Internal initiative at ING Australia using network science principles (Metcalfe/Reed/Dunbar) to drive AI tool adoption across the organisation. Enablement materials, usage analytics, and adoption playbooks.
- 🤝 AI as teammate, not toy — writer agents, reviewer agents, and monitoring on all of them
- 🔒 Safety and observability first — defence-in-depth, explicit threat models, metrics + traces + cost dashboards before ergonomics
- 🧑💼 Human-in-the-loop by design — owners get plain-English verdicts; engineers get structured, traceable evidence
- 📐 Evaluation-driven — if you can't measure it, you can't trust it in production
- 📡 Observable everything — if it's not in Grafana, it didn't happen
| Layer | Tools |
|---|---|
| Languages | Python, TypeScript, JavaScript, SQL, Bash |
| AI & Agents | Claude API, Gemini API, MCP (Model Context Protocol), A2A protocol, kagent, OpenClaw, LiteLLM |
| Backend / Infra | Node.js, Express, WSL2, Docker, Kubernetes (GKE/AKS) |
| Observability | Prometheus, Grafana, Loki, Tempo, OpenTelemetry, Alloy |
| Frontend | React + Vite, Tailwind CSS |
| Platforms | ServiceNow, GitHub Actions, Azure, Google Cloud |
If you're building agentic workflows, evaluation frameworks, or AI-native developer tooling — I'm always keen to compare architectures and swap notes.
