I build and operate AI systems that run in production: retrieval pipelines, evaluation + quality gates, and reliable cloud infrastructure for LLM-backed products. My work is backend-first, with a focus on measurable relevance, performance, and security.
π Focus: AI platform infrastructure β’ Retrieval/RAG β’ Performance & security automation β’ Resilient deployments
π¬ Contact: LinkedIn β’ Email
- π§ AI retrieval infrastructure: hybrid retrieval (graph + vector + structured filters), indexing, ranking, tuning
- β Production readiness: load testing, latency/concurrency modeling, rollout safety checks
- π Security automation: OWASP ZAP scans wired into repeatable workflows
- π οΈ Cloud reliability: active-active patterns, routing, TLS hardening, operational guardrails
- π§βπ» Backend: highly adaptable across backend stacks (APIs, services, data pipelines, integrations). This is my primary strength.
- π¨ Frontend: not my focus (I can integrate and support, but I donβt position myself as a frontend specialist).
Languages: Python β’ TypeScript/JavaScript β’ Bash
AI/Retrieval: LangChain β’ embeddings pipelines β’ hybrid ranking β’ evaluation workflows
Data/Stores: Neo4j β’ PostgreSQL/pgVector β’ TiDB β’ MySQL
Infra/Delivery: Docker β’ Nginx β’ Linux β’ active-active deployments
Testing/Quality: k6 β’ Locust β’ Playwright β’ OWASP ZAP
Most of my production work serves state government use cases, so the codebases are confidential. Some deployed products are publicly viewable:
- π Dayang chatbot (Sarawak services portal): https://service.sarawak.gov.my/web/
- βοΈ Court-related project (public article reference): https://ekss-portal.kehakiman.gov.my/portals/web/home/article_view/0/5/1
- π Malaysia public library chatbot (button-based): https://www.u-library.gov.my/portal/web/guest
These repos represent the kinds of systems I build (pipeline β retrieval β validation), even when production code is not public:
| Area | Repository | What it shows |
|---|---|---|
| π¬ Local multi-agent AI app | agentic-video-analyst | offline inference + multi-agent orchestration + desktop app engineering |
| πΈοΈ Graph ingestion + retrieval | neo4j-document-pipeline | graph modeling + retrieval API patterns for LLM workflows |
| π Vector + hybrid experiments | tidb-vector-llm-testbed | relevance/scoring experiments, indexing tradeoffs |
| 𧬠Embedding pipeline | mysql-to-pgvector-embeddings | extraction β embeddings β pgVector semantic layer |
| π Structured retrieval | faq-retrieval-system | structured query layer for grounded answers |
| π§ͺ Performance testing | playwright-dayang, k6-for-custom-dify | UX + API load testing approaches for assistants |
| π‘οΈ Security automation | zap-security-api | ZAP baseline/quick/full scan exposed via API |
| π§© Experiments | playwright-study, besu-ibft2.0 | targeted learning repos (testing + distributed systems) |
- π Prefer measured improvements (evaluation + monitoring) over demo-only features
- β±οΈ Treat quality, latency, and security as release criteria
- π Build systems that are operable (clear failure modes, logs/metrics, runbooks)

