Seasoned Principal Software Engineer with extensive expertise in Big Data, Search, Spark, Ray.io, Kubernetes, and Docker, seeking a challenging role to leverage cutting-edge tech stacks and drive innovation in scalable and efficient solutions in the world of AI.
Experience
2016 ā Now
2016 ā Now
Santa Clara, California, United States
šš-ššØšš®š¦šš§šš¬ šššš«šš” (RAG/Agents)
⢠Developed internal documents search/RAG solution using Opensearch. Uses ml-commons, embedding, rerank, conversational search, agents, guardrails and memory components of Opensearch.
šššš šš©š©š„š¢šššš¢šØš§ šššÆšš„šØš©š¦šš§š
⢠Architected and rolled out a SaaS enterprise app on hybrid cloud using Kubernetes and Docker.
⢠Implemented security measures: Authentication, SAML SSO, LDAP, and OAuth2.
⢠Advocated for customer-focused solutions, innovative approaches, and data-informed decisions.
šš-ššØš°šš«šš ššš ššØ ššš
⢠Part of the team who AI-enhanced NLP to SQL for Impala and Hive using Langchain, Prompt Builder and Retrieval Augmented Generation (RAG).
šš®š'š¬ šš«šš§š¬ššØš«š¦ššš¢šØš§
⢠Co-led the evolution and development of Hue, a premier open-source Hadoop UI.
⢠Elevated Hue from a basic utility to a top-tier solution, now a staple in numerous enterprises.
šššššš«š¬š”š¢š© & šš„ššššØš«š¦ šš±š©šš«šš¢š¬š
⢠Proficiency in PaaS and SaaS solutions.
⢠Strong coding skills in Go and Python.
⢠Steered a vibrant and productive team to success.
⢠Crafted and realized product visions, roadmaps, and pivotal strategic shifts.
šš„šØš®š ššš«šš”šØš®š¬š¢š§š & šš§šš«šš¬šš«š®ššš®š«š
⢠Developed Data Warehouses on AWS, GCP, and Azure using Kubernetes/Docker.
⢠Created a tailored SQL Assistant for Cloudera's Data Warehouse & Cloud services.
⢠Revamped platform UX to enhance usability.
⢠Embedded CI/CD, modular approaches, and scalability tactics for heightened productivity.
2013 ā 2016
2013 ā 2016
Santa Clara, CA, USA
šš²ššš« šš”š«ššš šš§ššš„š„š¢š šš§šš & šš¢š šššš
⢠Spearheaded the development of the malware cyber threat intelligence tool, "AutoFocus".
⢠Transitioned the tool from a mere prototype to a full-fledged product.
⢠Constructed a comprehensive big data pipeline, aggregating multiple data sources with Python and Java producers/consumers.
⢠Authored data transformation jobs using Hadoop and Spark and data search using the Elasticsearch infrastructure.
ššØš šššš šš¢š§š¢š§š & šš§šš„š²šš¢šš¬
⢠Innovated and delivered the "PAN v2" firewall log data mining and analytics tool.
⢠Instrumental in addressing Support escalations and offering actionable insights to management.
⢠Enriched the tool with features such as term search, diagnostic metrics, and an intuitive dashboard.
⢠Designed and implemented functionalities like queuing, log parsing, indexing, and report generation.
ššØš§šš¢š š®š«ššš¢šØš§ ššš§šš šš¦šš§š & šš«šš”šš¬šš«ššš¢šØš§
⢠Established a configuration management system using SaltStack.
⢠Orchestrated rolling version upgrades, significantly reducing downtime for the SaaS application.
⢠Conceptualized and designed a large-scale log data monitoring system, transitioning from a service-based architecture.
šššš šššØš«šš š & šššššš®š„ ššš«šÆš¢ššš¬
⢠Engineered a robust data storage system using HDFS/HBase.
⢠Integrated a REST API interface, utilizing Tomcat, JAX-RS, and the CXF framework.
⢠Leveraged Hadoop, HBase libraries, and adopted AVRO for effective data serialization.
šššš”š§š¢ššš„ šš«šØšš¢šš¢šš§šš¢šš¬:
Languages: Java, Python, Scala
Big Data Tools: HBase, Spark, Spark Streaming, Elasticsearch, Kafka, RabbitMQ
Web Technologies: Node.js, React.js, Tomcat, JAX-RS, Async, Tornado
DevOps & Management Tools: SaltStack, Rundeck, ELK
2006 ā 2013
2006 ā 2013
Palo Alto, CA, USA
ššØš šššš šš§šš„š²šš¢šš¬:
⢠Architected a data mining and analytics dashboard for logs leveraging both Hadoop and Python stacks.
šš¢ššš„šš°šš«š & šššš¤šš§š šššÆšš„šØš©š¦šš§š:
⢠Engineered Python-based middleware and web services, contributing to critical internal projects:
Elastic Test Cloud.
Continuous Automated Testing.
Build/Release Automation.
ESX Automation.
šš©šš¢š¦š¢š³šš šššš ššš©š„š¢šššš¢šØš§:
⢠Devised a C-based solution for Perforce journal data replication.
⢠Established a Perforce connection broker to efficiently redirect read/write queries, enhancing Perforce's operational efficiency.
2003 ā 2005
2003 ā 2005
Bangalore, India
2001 ā 2003
2001 ā 2003
Sunnyvale, CA, USA