# HaiXin Tie > A Software Engineer at Otter.ai (Previously Waymo, Google, Twitter, TiVo, Yahoo...) Location: Santa Clara, California, United States Profile: https://flows.cv/haixintie A software engineer who is passionate about building high-performing and fault-tolerant systems with user-friendly APIs, and experienced at identifying root causes of performance bottlenecks and providing well-thought-out solutions. Highlight: • [Google/1] Developed a near-realtime and globally-distributed streaming logs tailing system from a personal 20% project to production service with 24x7 SRE support and adoption by major streaming platforms including YouTube and Ads. This new system reduced streaming data discovery latency by 8x @50p, 3x @99p, and 6x @99.9p (“p” for percentile) in average across all collected log types at Google, alleviated file system load and data center outages, provided simplified client APIs, better data abstraction, and improved monitoring and alerting on streaming logs latencies; • [Google/2] Owned multiple large critical streaming & batch data processing pipelines for Google Instant Web Search, security monitoring, CloudSQL, etc. (Scale: 40K+ CPUs, 30TB+ RAM, 30GB+/sec data ingestion rate in Google private cloud); • [Google/3] Worked on one of the top 10 production 24x7 services that consumed 1M+ CPUs globally, and catalog services that managed exabytes of logs (including Search, Ads, YouTube, etc); • [TiVo] Improved multiple heavily used operations' performance by 100x; Built prototype for the next-gen TiVo search framework using open-source technologies (Apache Lucene / Solr / Zookeeper) as a personal 20% project; • Team player: as a strong individual contributor, or tech lead (evidence: comprehensive customer, peer, and manager reviews as shown in the Recommendations section below); • Fast and self-determined learner (evidence: part-time study at UC Berkeley Math as shown below) • Languages and Tools: C++, Java, SQL, Python, GoLang, Linux Shell scripting; AWS (EC2, RDS, S3, VPC, Route53, CloudFormation), Docker/Compose, Memcached, PostgreSQL, MySQL, MariaDB, NoSQL and Oracle RDBMS, Git / Mercurial; Google back-end infrastructure and tools including gRPC, Borg (Kubernetes equivalent), PubSub/Goop (Kafka equivalent), Spanner (multi-homed high-performance database, CockroachDB equivalent), MapReduce, Batch & Streaming Flume (Spark equivalent), Bazel (mono-repo multi-language build system) etc; Hadoop ecosystem; and other related open/closed-source tools; ## Work Experience ### Staff Software Engineer @ Otter.ai Jan 2024 – Present Scaling the real-time meeting and AI chat infrastructure to support large-scale, rapidly growing traffic. ### Principal Software Engineer @ Atomic Machines Jan 2023 – Jan 2024 | Santa Clara, California, United States Full-stack development including: Developing gRPC-based services to remotely operate motion controllers for laser, robots, and various other services on premises, in Docker or on AWS; Modeling study of embedded systems using finite state machine and behavior trees; Researching and developing a Bazel-based mono-repo build system that supports C/C++, Go, Python, Protobuf, gRPC and cross-platform toolchains for embedded systems and cross- language bindings/dependencies (open sourced at https://github.com/thx123/bazel-examples). ### Senior Software Engineer @ Waymo Jan 2020 – Jan 2023 | Mountain View, California, United States Owned the self-driving vehicle sensor calibration service that persists, updates and pushes latest calibration of tens of thousands of sensors to the entire self-driving car & truck fleet operating in multiple cities, incl. that for cameras, lidars, radars, GPS, microphones, etc; Developed an off-board batch data processing pipeline and related APIs to automatically detect, analyze, and re-calibrate sensors from newly imported self-driving vehicles’ sensor logs, filing work orders on faulty sensors, applying structured labels to affected systems downstream, and aggregating results across the self-driving vehicle fleet for visualization and trending analysis; ### Senior Software Engineer @ Google Jan 2014 – Jan 2020 | Mountain View Google-wide Logging Infrastructure (Sawmill) Team (Jan 2016 – Sep 2020): o Developed a near-realtime and globally-distributed logs tailing system adopted by major streaming platforms including YouTube and Ads. It grew from my 20% project into a mission-critical and team-managed system, onboarded by SRE with 24×7 support in multiple time zones. It improved streaming data discovery latency by 8x @50p, 3x @99p, and 6x @99.9p (“p” for percentile), in average, across all collected log types, and flattened latency spikes from 10 minutes to below 30 seconds. Other side benefits include load and data center outage reductions, better data abstraction, simplified client APIs, and better monitoring and alerting of streaming data latencies; o Owned multiple large streaming & batch data processing pipelines for Google Instant Web Search, Near-realtime security monitoring, CloudSQL, and App Hosting (Scale: 40K+ CPUs, 30+TB RAM, 30+GB/sec data ingestion rate in Google private cloud). Works included streaming latency root-cause analysis, major refactorings, upgrades, migrations, deprecations, and stability/resource tuning; o Member of the Logs Analysis team managing the Logs Proxy Service, one of the top 10 compute-intensive production services at Google, consuming 1.6M CPUs globally, running 24×7 with full SRE support; Made numerous improvements, incl. L4 cache on SSD in Google data centers to reduce hard drive spindle cost and improve data read/write latencies (earned Tech Infra 2019 Feats of Engineering Team Award); o Member of the Logs Metadata team managing Google-wide metadata service for all log types. Enterprise Search (Topaz) Team (Aug 2014 – Dec 2015): o Full stack engineering work including performance/latency measurements, automatic regression tests, visualization and optimization on both front & back ends; owner of multiple logs processing pipelines, worked closely with Dev, SRE, Test and Privacy/Legal teams on various backend features and product launches; ### Senior Software Engineer @ Twitter Jan 2014 – Jan 2014 Worked in the search infrastructure team. Participated in the project to build a complete Tweet indexing and searching system for every public tweet since 2006. Blog post: https://blog.twitter.com/2014/building-a-complete-tweet-index ### Member of Technical Staff @ TiVo Jan 2009 – Jan 2014 | San Jose, CA o Served as team's go-to person for backend performance issues in Java, Memcached and MySQL. Examples: o Improved the average response time of the most heavily used search operation by 100x in U.S. production system shortly after joining TiVo; o Improved the average response time of another heavily used operation (100 QPS from 1.5 million client devices) by 100x, significantly reduced the load to the backend servers / databases / network, which also improved stability in U.K. production system; o Built prototype for the next-generation TiVo search framework to deliver personalized, full- text and unified search results including TV shows, movies, cast and credits from linear programming and multiple partner data sources using Apache Lucene / Solr / ZooKeeper as a distributed cloud solution (as a 20% project in Google terminology); o Productionized open-source Sphinx full-text search engine based solution into the existing backend; o Led / contributed to design & development of various TiVo backend APIs and features for TiVo Roamio/Premiere/HD/mobile clients and various domestic & int’l partners; ### Senior Software Developer (Part-Time while Attending School at Berkeley) @ Topix Jan 2008 – Jan 2008 * Independently redesigned and implemented an enterprise-class monitoring system (Nagios3) for 300+ hosts and 3000+ services, including writing of plugins and CGI scripts in Perl, performance tuning, etc.; ### Senior Software Developer @ Yahoo! Jan 2006 – Jan 2008 * Responsible for banner ads traffic calculation system, including development, debugging, maintenance and monitoring. The system sampled gigabytes of data from ad logging and user behavioral data mining servers, and spent days on multiple machines to calculate ads traffic for different combinations of demographic/geographic attributes and user interests for most yahoo properties, including finance, mail, sports, etc. for multiple regions of the world; ### Senior Software Developer @ Yahoo! Jan 2006 – Jan 2007 * Designed and built from scratch search text ads traffic forecasting system, including search term traffic forecast generation module on front-end; daily/hourly market summary forecast module and its accuracy monitoring server on back-end; and corresponding Oracle/MySQL database schemas; ### Principal Engineer @ Oracle Jan 2001 – Jan 2005 * Built enterprise product lifecycle management systems for manufacturing items, change orders, document, etc.; * Participated in all phases of software development including high-level, functional and database schema designs; implemented front-end UIs and back-end APIs in database/server/middle/client layers; and maintained code of multiple releases at the same time; ## Education ### B.S. in Engineering & Applied Science Caltech ### Part-Time in Mathematics University of California, Berkeley ### M.S. in Electrical Engineering Stanford University ### Physics Peking University ### High School Diploma 北师大实验中学 ## Contact & Social - LinkedIn: https://linkedin.com/in/hxtie --- Source: https://flows.cv/haixintie JSON Resume: https://flows.cv/haixintie/resume.json Last updated: 2026-04-01