Seeking a Data & AI Engineering role to apply end-to-end pipeline design and LLM-powered system building for measurable business impact.
I'm Weijian (Tim) Zhang — a data engineer and AI-driven pipeline architect who bridges the gap between scalable data infrastructure and production-grade AI applications. I hold an MSc in Artificial Intelligence and Business Analytics (Lingnan University, Hong Kong) and bring 8+ years of hands-on delivery across financial services, government technology, and global retail.
My career has followed a clear trajectory: from designing enterprise data warehouses processing billions of daily transactions for banking clients, to building real-time data quality monitoring for Walmart's B2B platform, and most recently architecting Agentic LLM systems for Hong Kong government slope maintenance operations.
I specialize in the full lifecycle — from raw data ingestion and warehouse modeling to LLM orchestration, retrieval-augmented generation, and cloud-native deployment on GCP. Whether it's a 5-layer dimensional model or a 7-layer Agent architecture, I focus on one thing: delivering systems that produce measurable improvements.
MSc in Artificial Intelligence and Business Analytics — Lingnan University, HK (2025–2026)
Bachelor of Management — Shenzhen University, China (2014–2017)
GovTech (HK) · Banking & Finance · Global Retail (Walmart) · Enterprise BI/DW
Agentic RAG systems · Multi-Agent orchestration · Cloud-native AI deployment · Data pipeline modernization
Projects are organized by relevance to my target role (Hybrid Data + AI Engineer), strength of quantifiable outcomes, verifiability, and technical diversity across GovTech, enterprise data, and LLM systems.
Hong Kong Architectural Services Department (ArchSD) case processing relied on manual entry and routing, with average processing cycles lasting days and high error rates.
67 commits, 80,430 lines of code across two iterations. 12/12 requirements fully implemented with end-to-end automation. Reduced processing time by 75% (40 min → 10 min). Projected annual cost savings of HK$6.39M.
Python 3.11+ · FastAPI · React + TypeScript · PostgreSQL 15 + pgvector · OpenAI GPT-4o · Google Cloud Run + Cloud SQL · Docker
University students, faculty, and administrators struggled with dispersed policy documents across 30+ departments, leading to 70%+ retrieval failure rates and inconsistent information delivery.
Ingested 200+ policy documents, academic regulations, administrative guidelines, and FAQ corpora across three user personas (student, faculty, admin).
Built AgenticRAG with Hybrid Retrieval (dense + sparse + reranking), Adaptive User Memory for personalized dialogue, RBAC-based access control, and LLM-as-Judge for answer quality assurance.
Policy Q&A accuracy >95%. Average retrieval time reduced by >70%. System response within 10 seconds. Demonstrated at Lingnan Innovation Showcase.
Python · LangChain · OpenAI API · PostgreSQL + pgvector · FastAPI · Streamlit · Docker · GCP
Walmart China's B2B financial reporting relied on fragmented data sources across SQL Server, StarRocks, and legacy systems, causing delayed reporting, undetected data anomalies, and manual reconciliation overhead.
414,000+ contract records, monthly B2B transaction partitions (9,000–10,000 records/month), cross-source financial metrics including order amounts, fulfillment types, and credit ledger balances.
Built cross-source financial data quality monitoring with automated anomaly detection (e.g., fulfillment amount exceeding order amount). Designed B2B reporting iterations covering contract lifecycle, countdown tracking, and credit ledger reconciliation. Implemented Feishu alerting for downstream BI consumers.
Reduced data incident detection latency from days to minutes. Achieved automated cross-source validation across StarRocks and SQL Server. Enabled consistent financial reporting for B2B operations across multiple business units.
Python · SQL · StarRocks · SQL Server · DolphinScheduler · Feishu API · Power BI
Designed and deployed automated cross-source data validation pipelines comparing StarRocks and SQL Server records. Built Python-based monitoring scripts with configurable thresholds, automated Feishu (Lark) alerting, and downstream notification workflows — reducing manual reconciliation effort and improving data incident response time.
Delivered data warehouse optimization and dimensional modeling for Bank of Ningbo, including a 5-layer EDW architecture (ODS → DWD → DWS → ADS → RPT). Optimized SQL performance for large-scale transaction queries using Hive on Hadoop. Supported new credit system data modeling and migration.
Download my full resume or browse the career timeline below. My experience spans enterprise data engineering, cloud-native AI deployment, and cross-team delivery across banking, retail, and government sectors.
Download Resume (PDF)