Senior Software Engineer with 8+ years of experience building large-scale data platforms, real-time streaming systems, and distributed infrastructure that power data-driven products. My work focuses on designing resilient, scalable systems across multi-cloud environments (AWS and GCP).
Experience
2023 — Now
Atlanta, GA
Implemented Mailchimp’s global experimentation framework, establishing
a 5% universal holdout infrastructure with real-time customer assignment and automated contamination
monitoring, processing 20M+ daily event tracking records and enabling scientific measurement across 1M+ customer accounts
Spearheaded the GTM initiative for Salesforce CXM integration, orchestrating the migration of 1M+ Mailchimp accounts and implementing the Account Health Score system with daily BQ-to-Salesforce pipelines, establishing CXM as system-of-record and enhancing account segmentation
Architected end-to-end real-time, highly scalable customer and marketing data pipelines, establishing stateless integration patterns between BigQuery, Salesforce CXM, Gainsight, and Segment, helping reduce high-value customer churn by 30%, and driving 40% growth in new customer acquisition
Innovated and deployed cross-platform notification services utilizing BigQuery, Cloud Run, and Cloud Functions for real-time data incident remediation, improving job success rate to 99.9% and cutting MTTR by 75% through proactive alerts
Leveraged AI-powered development workflows, LangChain, OpenAI/Anthropic APIs) to accelerate pipeline development by 40%, building custom agent SKILLs and MCP tools for automated reporting, data quality validation, and developer productivity improvements
Spearheaded the architectural design of real-time data streaming infrastructure, managing critical data
exchange between Intuit and Mailchimp across Pub/Sub, Kafka, and Dataflow, achieving 99.95% uptime
SLA and <100ms end-to-end latency SLA, ensuring robust scalability, sustainability, and compliance
Drove technical excellence and organizational alignment by mentoring senior engineers, leading cross-
organizational knowledge-sharing initiatives, and chairing architectural review sessions with Engineering
leadership, shaping data quality and governance standards across the organization
Led the installation, configuration and deployment of production software that contacts Kafka Cluster(AWS MSK) for data acquisition. Enhanced the security of the Kafka Infrastructure with TLS authentication.
Developed real-time streaming applications integrated with Kafka Connect and Snowflake to handle large volume data streams in a scalable, reliable and fault-tolerant manner for Fraud Detection and ACH Loss prediction model to decline risky transactions and prevent 9 9% of fraud events in < 500 milliseconds.
Designed and implemented big data pipelines to ingest multi-TB data from various data sources using Debezium MySQL Connector, Kafka Schema Registry and Kstream. Validated, transformed and stored data in efficient formats for downstream purposes such as PySpark through Databricks.
Created a backup & mirroring of Kafka Cluster to support High Availability (HA) of Kafka Clusters. Managed broker & topic performance, sizing, security and consumer/producer access (ACL) through Terraform.
Conducted Performance Test such as disaster recovery test for cross-region replication as well as load test using K9 tool and mysqlslap. Monitored and fine-tuned Kafka cluster configurations at producer, consumer and broker level that successfully improved the performance by 20%.
Mentored and on-boarded less-experienced engineers with detailed-documented streaming infrastructure. Led design and code review sessions that were approved by staff engineers and architects.
2021 — 2022
Set up the CI/CD pipelines and automated deployments on AWS Kubernetes(EKS) using GitHub Action and ArgoCD. Introduced Datadog and Prometheus for monitoring and alerting as well as conducted root cause analysis through Anomaly Detection to increase system observability and stability.
Managed existing ETL jobs through Airflow and added new data pipelines from various structured and unstructured sources into Snowflake that leverage the Client big data platform.
2018 — 2021
Atlanta, Georgia, United States
Design and implement a micro payment service that integrated with third-party payment providers, and applied to contactless payment through GraphQL and AWS API-gateway, Route53, DynamoDB.
Utilize AWS Lambda, S3, Kinesis, Glue, Athena, ECR, SQS, to expedite reporting service.
Built GitLab CICD pipelines to deploy the server to AWS through Terraform, and implemented a new testing dev tool for the engineering organization using Docker and documented using Swagger.
Redesign the front-end report logic using React.js and reformat the sales data export framework, use Redux to implement a reusable React component to make UI consistency crossing all platforms.
Education
Georgia Institute of Technology
Master’s Degree
University of Washington