Lead Data Engineer with 6+ years building high-scale, cloud-native data platforms across GCP, AWS, and Azure. I lead engineering teams, define technical roadmaps, and architect real-time streaming pipelines that process millions of events with zero data loss.
Passionate about building data systems that drive real business impact
I'm a Lead Data Engineer with 6+ years of experience designing and operating high-scale, cloud-native data platforms across GCP, AWS, and Azure. Currently at Hard Rock Digital, I lead engineering teams, define technical roadmaps, and establish org-wide data engineering standards that accelerate delivery and improve quality.
I'm equally effective as a hands-on technical architect and as a people-and-process leader driving cross-functional alignment. My expertise spans real-time stream processing, CDC pipelines, Data Vault 2.0, MLOps, and developer productivity tooling including LLM integration strategy.
Team enablement & technical roadmaps
Debezium, Pub/Sub, BigQuery pipelines
100K+ msg/hr, exactly-once semantics
Copilot standards & AI deployment
The stack I use to build scalable data solutions
A track record of delivering impactful data solutions
Mar 2025 – Present
Jun 2023 – Feb 2025
Jan 2023 – Apr 2023
Aug 2019 – Dec 2021
Apr 2019 – Jul 2019
Some of the impactful projects I've built
Real-time ingestion framework using Hub/Link/Satellite patterns with Debezium → Pub/Sub → BigQuery. Eliminated downstream data staleness, serving as the platform foundation for all new event streams.
Architected 5+ production Dataflow pipelines processing 100K+ messages/hour with exactly-once semantics and stateful processing, guaranteeing 100% data integrity across real-time event streams.
Achieved 60x throughput increase on mission-critical pipelines by redesigning bulk load operations, slashing processing latency from 60 minutes to under 5 minutes for 320K+ records.
End-to-end CI/CD pipeline for AI model deployment using Azure ML Studio and Kubernetes, cutting time-to-market by 35% and operational costs by 25%. Moderating 1M+ daily events.
Technical roadmap for migrating 40+ Airflow DAGs to Airflow 3.0 on self-managed Kubernetes (GKE), delivering zero-downtime cutovers with full backward compatibility, unblocking 3+ downstream teams.
Established org-wide engineering standards for LLM integration (Claude Code/GitHub Copilot), prompt engineering best practices, and security guardrails — accelerating code delivery velocity across the data engineering team.