# Ali Ansari > Staff Software Engineer - Infrastructure Location: San Francisco Bay Area, United States Profile: https://flows.cv/aliansari Skills: Programming and Scripting Languages: Java, Python, Bash, Scala, Node, PHP, SQL Tools/Other Skills: Elasticsearch, Reddis, AWS Stack, Jenkins, MongoDB, NoSQL, Docker, pingdom, New Relic, ELK, Splunk, AppDynamics, Big data analysis. Networking: AWS or on-Prem Network Stack ## Work Experience ### Staff Software Engineer - DevOps @ Activision Blizzard Media Jan 2024 – Present | San Francisco, California, United States ### Principal Software Engineer - DevOps @ Arietta.ai Jan 2023 – Jan 2024 | Pleasanton, California, United States Formerly part of Genpact PVAI ### Principal Software Engineer - DevOps (PVAI) @ Genpact Jan 2022 – Jan 2023 | San Francisco Bay Area ### Staff Software Engineer - DevOps @ Mindstrong Jan 2021 – Jan 2022 - Responsible for implementing, maintaining and updating all components of the micro-services stack, primarily in AWS with many SaaS integrations like Atlas MongoDB, OAuth Framework, Splunk/New Relic, Databricks, PagerDuty, etc. - Dozens of docker services running as ECS services for HA, exposed using ALB/ELB with OAuth framework for API application security. Exposed using cloud flare. - Multiple backend connections from middle layer to DBs, separated at network level and API security using roles in OAuth - Responsible for all pre-production and production environments - 24/7 On-call rotational support for Production environments. - SME and DB admin for multiple data stores (Elasticsearch, Reddis, MongoDB, SQL, etc) - Migrated infrastructure components into infrastructure as code to deploy and update individual components. - CI/CD pipelines using Jenkins - Multiple data backup/restores and disaster recovery recovery solutions - Monitor, alert and report on individual and system component uptime’s, mean time to recovery, etc for APIs, services, databases and ETLs - ETLs using Airflow and Data-warehouse in S3 and glacier. ### Senior DevOps Engineer @ Thomson Reuters/MarkMonitor/OpSec Security Jan 2016 – Jan 2021 | San Francisco Bay Area - Responsible for building, automating and maintaining AWS cloud and data center hybrid cloud environments, both monolithic and microservices; including deployments, CI CD pipelines, release management using Jenkins and Bamboo, 24/7 automated monitoring platform with alerting and real-time visualizations of current state of applications, vms/instances/servers, DBs, and different clusters of Storm, Mesos, RabbitMQ, Elasticsearch, Spark, Redis, MongoDB and others. - Primary support for on-call rotation for Production issues, supporting a plethora of technologies and DBs. - Implemented end to end service and host monitoring using Prometheus, Grafana, Alert Manager, PagerDuty and Slack integrations. System monitoring using prometheus plugins, log event monitoring using ELK and AWS Cloudwatch metrics, graphed in Grafana, alerts managed by Alert manager and sent to PagerDuty and /or Slack. - Writing Automation and configuration management using Chef and Ansible for new and migrating old Chef scripts to using Ansible. - Designed various solutions for running a stack that includes applications running in/with, Mesos, Marathon, Storm, Kafka, Cassandra, Spark. - Responsible for seamless and online migrations of entire stacks with TB’s of different databases to and from AWS and local data centers. - Day to day maintenance, deployment, troubleshooting for application delivery in all environments, DEV, QA, Staging and Production. ### DevOps/Software/Network Curation Engineer @ Devicescape Jan 2013 – Jan 2014 | San Francisco Bay Area Work in a fast paced startup environment using Agile methodlogies. Take care of variety of things from DevOps to big data analysis using Hive and Hadoop, crunching data and reporting on it, new features for platform, test cases and cloud operations. - Setting up, maintaining and automating infrastructure in AWS environment for production and testing. - Data Analysis and reporting. - Create, manage and monitor AWS (EC2, EMR, DynamoDB, S3, etc). - Contributions to the platform development team, writing features for server side processing application written in Java and deployed using JBOSS -Designed and programmed parts of offline log processors that parse client uploaded logs, extract information and save the results in Elasticsearch. - Rapid prototyping using AWS for new technologies before integrating them in production environment. -Amazon DynamoDB extractions to Amazon S3 using Scala and Python. -Business intelligence, reporting on different statistics and extracting information from client log files using Hbase and Hive in AWS environment. Use Hbase and Hive to cleanse the data and Python scripts to apply the logic and send email reports. -Python scripts to generate reports from data stored in Elasticsearch, MongoDB and other online storage services. -Big data manipulation using Scalding on top of Hadoop in AWS environment. ## Education ### Master of Science (M.S.) in Computer Science California State University, Chico ### Bachelor of Science (BS) in Computer Science California State University - East Bay ## Contact & Social - LinkedIn: https://linkedin.com/in/ali-ansari-94a15b48 --- Source: https://flows.cv/aliansari JSON Resume: https://flows.cv/aliansari/resume.json Last updated: 2026-04-01