# Min Tu > Staff Engineer at DoorDash Location: Santa Clara, California, United States Profile: https://flows.cv/mintu Summary 2 years of streaming processing experience with Flink/Kafka Certified hadoop admin from Cloudera 10 years of big data working experience with Hadoop platform 3 years of display advertise experience with hadoop/map-reduce 15+ years of experience in Java and J2EE design and development Full life cycle software development methodology including mapping user requirement to functional specification, high/low level design and implementation, unit and integration testing Front to end J2EE application design and development Strong debugging and problem solving skills Specialties: Technology: Flink, Kafka, K8S, Hadoop, map-reduce Languages: Java, Python, Perl, JavaScripts Database: Snowflake, Neo4j, MySQL, Oracle 9i, ## Work Experience ### Staff Engineer @ Uber Jan 2023 – Present | United States Led Redis Infrastructure Architecture and Uber Java Client Integration: • Designed and developed a robust Uber Java client wrapper for Redis, abstracting interactions over Jedis and Lettuce libraries to support a wide range of application use cases. • Contributed key enhancements to the Lettuce open-source library, focusing on connection leak prevention and topology refresh mechanisms, improving Redis cluster reliability at scale. • Spearheaded the migration from Twemproxy with standalone Redis to a high-availability Redis Cluster setup, reducing latency and improving fault tolerance across Uber’s distributed services. Architected Multi-Region Resilience for Uber’s Workflow Orchestration Platform (Piper – Uber’s internal version of Airflow): • Designed and implemented multi-region failover capability in the Piper system, enhancing disaster recovery and ensuring workflow execution continuity under regional outages. • Collaborated with infrastructure and platform teams to deliver high-availability architecture aligned with Uber’s global operational requirements. ### Staff Engineer @ DoorDash Jan 2021 – Jan 2022 | San Francisco Bay Area Architected and led the standardization of DoorDash’s in-house real-time Customer Data Platform (CDP): • Unified the event processing ecosystem by integrating Confluent Schema Registry, Kafka REST Proxy, Flink, Kafka, and Snowflake into a cohesive, scalable architecture. • Designed a dual-format schema strategy using Avro and Protobuf to enforce data contracts across internal services. • Built a modular pipeline architecture enabling per-event stream processing and isolation, ensuring clean delivery into Snowflake with strong schema guarantees. • Leveraged Kafka REST Proxy to enable JSON-based external event producers without compromising schema compliance. Designed and implemented a real-time event forwarding service from DoorDash to external partners: • Guaranteed at-least-once delivery semantics using Flink checkpointing and Kafka offset management. • Introduced a Dead Letter Queue (DLQ) strategy to gracefully handle external system outages and improve observability. Led the migration of 50% of traffic from a 3rd-party CDP to the in-house real-time pipeline: • Reduced operational cost and increased data governance and system reliability through internally owned infrastructure. Enhanced observability and developer productivity for real-time data pipelines: • Extended Apache Flink SQL with custom tooling to support Protobuf format inspection via Confluent Schema Registry. • Developed an automated schema evolution handler for Avro, integrating Schema Registry, Snowflake SDK, and Cadence workflows to ensure backward and forward compatibility in production pipelines. ### Software Architect @ Tencent Jan 2018 – Jan 2021 | Palo Alto, CA Natural Language Process platform Medical Knowledge Graph Medical AI based chatbot In Medical AI Lab, lead Engineer team, build medical AI based platform/ product Neo4j/MySQL based Medical Knowledge Graph - Neo4j store the medical entities and relationships - MySQL for entity normalization - REST API for internal data services - Dynamic construct Knowledge Graph by merging different sources, based on configuration Medical Natural Language Process - Entity data dictionary/word segmentation based on Knowledge Graph - Multi-dimensions information retrieval for major entity - Various medical specific parser: prescription parser, medical exam index parser AI based medical chatbot - Parse patient main symptoms with Natural Language Processing - With user profile/AI model, generate most distinguishable question for user to answer - After multiple rounds, generate disease prediction report ### Software Architect @ Baidu, Inc. Jan 2016 – Jan 2018 | San Jose, CA Big Data Lab: Medical AI based chatbot ### Sr. Software Engineer @ LinkedIn Jan 2014 – Jan 2016 | Mountain View, CA, USA Disaster Recovery for LinkedIn main production ETL pipeline: Lumos Dataset Configuration Management in Apache Gobblin Regression testing framework for LinkedIn version of Distcp: Uber Distcp ### Principle Software Engineer @ PayPal Jan 2012 – Jan 2014 In PayPal big data platform team, worked with Risk detection department for ATO model. 1. Designed and implemented the risk propagation prototype using Apache Giraph for graph processing 2. As the tech leader in the team, collaborate with Risk detection team, designed and implement the ATO model for IP propagation using Apache Giraph/Map reduce 3. Designed and implemented the work flow for ATO model using python script and UC4 as scheduler. Hands on Hadoop Admin experience by setting up multiple development Hadoop clusters and doing user support on production Hadoop cluster. ### Senior Software Engineer @ Yahoo! Jan 2012 – Jan 2012 In Yahoo User Behavior Targeting System team, process Terabytes user data everyday on top of the Hadoop platform by using map/reduce, oozie etc. Design/develop the new targeting platform for more accurate user targeting. ### Senior Software Engineer @ Yahoo! Jan 2010 – Jan 2011 In NGD Forecasting team, help Yahoo! clients better plan and budget their NGD Advertise Campaigns through accurate supply forecasting at different but specified price points. 1. Designed and Implemented the supply forecasting module for NGD forecasting pipeline 1.0 2. Re-designed and Re-implemented the input binding module used to map line item to targeting attributes 3. Designed and implemented the supply forecasting module/pipeline 2.0 using NGD adserver to simulate real data traffic 4. Designed and developed multiple Hadoop map-reduce jobs to do large data files join and data analysis ### Software Engineer @ VeriSign Jan 2009 – Jan 2010 As a java/J2EE web application developer, build new features and enhance existing features for Vip Manager web application. Technology and Tools: Struts 2, Hibernate, HTML, JSP, CSS, JavaScript, XML, JDBC; Tomcat, Firebug, Oracle 9, Eclipse for Java IDE. ### Consultant, Technical Lead @ A4A Jan 2008 – Jan 2009 A4A is a Chinese technology company specialized in digital marketing. As a consultant, I led multiple projects including developing client and business contract information management system, and current and future employee information management system. The systems are successfully developed and deployed with the company. Technology and Tools: HTML, JSP, CSS, Java Servlet, XML, RMI, JDBC; Apache web server, Firebug, Oracle 9, LDAP server, Eclipse for Java IDE. ### Java/J2EE developer @ Sun Microsystems Inc. Jan 2000 – Jan 2008 In Distributed Resource Allocation Management (DReAM) team, worked as a key developer for Sun large scale Grid Computing product DReAM which is successfully released and widely used by Sun Processor and Network Products (PNP) Division. Java developer Designed and implemented the new DReAM messaging layer using Java Message Services ( JMS ) and greatly improve the performance compared with the existing RMI based messaging layer Designed and implemented meta manager daemon with RMI technology. Meta manager is used to control and monitor other distributed Java daemons in system Designed and implemented the centralized, XML based authorization layer with great flexibility Enhanced the Data Convert Language (Dcl). Dcl is DReAM developed Java object/relational persistence and query service. Dcl automatically converts data description file to SQL scripts, which is used to generate database schema and java classes with JDBC code. Used recursive methods to create the new feature which extends Dcl to accommodate multiple layer data objects With strong debugging and problem solving skill, fixed DReAM bugs in multiple releases Technology and Tools: J2SE, JMS, RMI, multi threading, JAXB, XML J2EE based web application developer Used full cycle software development methodology to design and implement J2EE based online management application from scratch, including user case analysis, requirement specification review, high/low level design and implementation. Also performed unit and integration testing Developed tag library for recursive multiple layer navigation, multiple page control, access control and other utility tags. Used JAXB, JAXP/XSLT to build generic property management so that adding new properties do not require Servlet or JSP code change Technology and Tools: HTML, JSP, CSS, Java Servlet, JAXB, JAXP, XSLT, XML, RMI, JDBC; Oracle 9, LDAP ### Perl developer / Testing Engineer @ Sun Microsystems Inc Jan 2000 – Jan 2008 Used Perl to develop various command line interfaces for DReAM product , and for unit and regression testing ### Web server administrator @ Sun Microsystems Jan 2001 – Jan 2008 Installed, configured and tuned performance for the Sun iPlanet web server Deployed the web application into iPlanet web server Developed Perl script to balance the load between multiple web servers ### Database Administrator (DBA) @ Sun Microsystems Jan 2004 – Jan 2005 Learned database technology from the DBA in team, acted as DBA for 24x7 support when team’s DBA was on leave of absence for 3 months ### Software engineer @ Vanguard Group Jan 1999 – Jan 2000 Worked on all aspects of client-server web applications including using Dreamweaver to build HTML/JSP pages, Visual Café to create Servlet and back-end java classes with JDBC code, SQL Programmer 2.5 to implemented Store Procedure. Used Rational Rose 2000 with UML for Object Oriented analysis and design. Standard Online Reporting Project: Used full cycle software development methodology to design and implement client-server applications – the Standard Online Reporting, including use case analysis, software requirement specification review and design. Also performed unit and integration testing for the applications Implemented the Java Native Interface (JNI) to access Java API from C program. Visual J++ and Visual C++ are the IDE. ØDeveloped Store Procedure on Sybase Database for data query and management using SQL Programmer 2.5 Designed and Developed Unix Scripts and C programs for manipulating report data in the Sybase Database and sending email notification to client Used PVCS (Product Version Control Manger) for the source code version control SmartForms Project: Used full cycle software development methodology to design and implement client-server applications – the Vanguard Group SmartForms Performed the Object Oriented analysis and design in Rational Rose 2000 using UML Implemented the Java Servlets and JDBC using Visual Café as IDE Built JavaServer Pages using Dreamweaver Used JRun as the Web Server in the front end ### Software engineer @ MGV American company Jan 1999 – Jan 1999 Developed IBM In-Store retail transaction application using CBasic language ### Co-op @ AMTEL Multimedia & Communication company Jan 1998 – Jan 1998 Designed On Screen Display programming for Video Phone using C language Embedded software programming for control data flow of Video Phone device ### Research Assistant @ Computer center of Tsinghua University Jan 1997 – Jan 1997 Developed Tsinghua University Management Information System using PowerBuilder as front end and Oracle database as back end ## Education ### MS in Computer Engineering North Carolina State University ### BS in Electrical Engineering Tsinghua University ## Contact & Social - LinkedIn: https://linkedin.com/in/min-tu-78764315 --- Source: https://flows.cv/mintu JSON Resume: https://flows.cv/mintu/resume.json Last updated: 2026-04-12