# Alan Zhang > Staff Software Engineer at LinkedIn Location: San Francisco Bay Area, United States Profile: https://flows.cv/alanzhang I am a passionate Software Engineer interested in stream processing, distributed systems, and cloud native infrastructure. ## Work Experience ### Staff Software Engineer @ LinkedIn Jan 2019 – Present Focus on stream processing infrastructure at LinkedIn. ### Staff Software Engineer @ eBay Jan 2018 – Jan 2019 | San Francisco Bay Area As a main contributor, I work on the next generation real-time monitoring and log processing platform at eBay. This work involves dealing with petabytes of data being generated from machines every day, to design an efficient system with respect to cost and performance to ensure that any degradation in the health of any critical services can be detected and remediated within minutes or even seconds. This work utilizes the technologies such as Kafka, Flink, Prometheus, ElasticSearch, Kubernetes, Thanos, Ceph and etc. Lead to design and implement the new features for Thanos, and contributing back to Thanos community. ★2019 Spot Award Winner. I helped the team to solve the performance issue on metrics ingress service. After the optimization, it can serve 5X traffic than before. ### Staff Software Engineer @ eBay Jan 2016 – Jan 2018 | Shanghai City, China Lead to design and develop a brand new stream processing platform based on Flink, which will provide a secure, scalable and unified streaming platform to support efficient and fault-tolerant stateful stream processing in eBay. ★2017 Spot Award Winner. As a main contributor, I designed and developed a new service based on Rheos platform which empowers near real-time data movement cross different zones, specifically for EAZ to Site Lockdown. This service successfully reduced the time to move the data between the different zones with security compliance from hours to seconds, which started the new chapter of real-time data movement between zones in eBay. ★2017 Luminary Award Winner. I participated in the design and development of a new data platform (Rheos), which provides a near real-time buyer experience, seller insights, and a data-driven commercial business at eBay. Rheos provides the necessary life-cycle management, monitoring, and well-architected standards and ecosystem for the near real-time streaming data pipelines. Currently, the pipelines consist of Kafka, Storm and stream processing applications. Shared and non-shared data streams can be run on these pipelines. By the end of 2016, nearly 100billion messages flowed through the pipelines in Rheos daily. In 2017, Rheos is expected to handle 15 times of the current traffic. As a main contributor, I worked with my team and developed the following 6 core components: Core Service; Schema Registry Service; Mirroring Service; Health Check Service; Kafka Proxy Service; Lifecycle Management Service. ### Manager @ Works Applications Co., Ltd. Jan 2014 – Jan 2015 | Singapore Lead in designing and developing a robust backend infrastructure framework. It supports automatic creation of real-time denormalized index data for both Cassandra cluster and Elasticsearch cluster. This framework includes five core components: 1. A core code processor engine based on Java annotation processor, Mustache and Antlr. It will generate denormalized index related Java classes automatically. 2. A real-time online job based on Spark streaming with Kafka and an offline job based on Spark. The online job will generate real-time indices for Cassandra and Elasticsearch; The offline job will run index data migration operation and index data repair operation. 3. A core message queue library based on Apache Kafka. 4. A core serialization library based on Kryo. 5. A core maven plugin which is an automatic job generator based on ASM(Java bytecode analysis library) and Maven Lead in designing Rabbitmq cluster architecture, in order to improve system response speed, decouple the different project modules and support offline job. Make API response 2-3 times faster than before. Lead in optimizing cloud-based email retrieval system, to support quick full text retrieval from more than 1 billion existing emails, and 0.7 million daily increasing mail indices. Research and design new architecture for multilingual email retrieval, especially improve the accuracy and speed of mixed-language email retrieval. Research on Solr and Lucene source code in-depth, designed custom multilingual text analyzer component. This new architecture can improve query speed 2-4 times faster on average than old Solr architecture. Optimize index build process in-depth between Cassandra and Solr cloud, design and implement a brand new architecture for this build process, using Hadoop and custom SSTable export tool to convert SSTable data to Solr documents, and update these data to Solr Cloud directly. ### Software Engineer @ Works Applications Co., Ltd. Jan 2012 – Jan 2014 | Shanghai City, China Design and develop an enterprise mail system, based on Cassandra cluster and Nginx, ELB load balancing technology, which can support huge business throughput of around 3000 employees' business gracefully. Design and develop email retrieval system with Solr cloud technology; integrate Solr cloud and Cassandra, which makes full text query and index performance more powerful. Design and develop a new file system to make this enterprise mail system more versatile, which is based on Amazon S3 and file reference counting algorithm. In addition to the basic file management operations, it supports many advanced features, such as file sharing and file private link. ### Software Engineer Intern @ Cisco Jan 2011 – Jan 2012 | Shanghai City, China Design and develop a power monitor system to control powercycles of more than 1000 devices in the network lab, based on Mysql cluster, which uses SNMP, Telnet and socket protocol to control powercycles. Integrate this system with Okcard system, Labkeeper system and Firebird system. Design and develop an assets management system, using SNMP protocol, which implements the following functions: devices automatic retrieval, devices information reading, manual warehousing, importing the data as a batch, as well as generating the reports. Responsible for configuring and managing team's network devices in the network lab. ### Software Engineer Intern @ Alibaba.com Jan 2009 – Jan 2010 | Hangzhou City, Zhejiang, China Develop KFC(keyword filter center) system to reduce the request pressure of the CTU system, which consists of three main components: keyword management, keyword filtering and keyword sharing. Develop CTU system which is the core product of Taobao's secuity team, including add new whitelist and blacklist features, integrate KFC system inside this system. Develop parmirs-service system to guarantee the quality of goods and the right of the customers; define rules to punish the merchants who violate the regulations. ## Education ### Master in Computer Science Dalian University of Technology ### Bachelor in Computer Science Dalian University of Technology ## Contact & Social - LinkedIn: https://linkedin.com/in/alnzng --- Source: https://flows.cv/alanzhang JSON Resume: https://flows.cv/alanzhang/resume.json Last updated: 2026-04-12