California, United States
In this role, I design and deploy production-grade AI systems centered around large language
models, agentic workflows, and tool-using agents. My responsibilities include architecting end to-end GenAI solutions, building retrieval-augmented generation pipelines over enterprise data,
and integrating LLMs with internal APIs, databases, and automation services. I work extensively
with Python, containerized services, Kubernetes, and cloud infrastructure to ensure these AI
systems are secure, scalable, and maintainable. A key part of my role involves translating
ambiguous business needs into deployable AI solutions while collaborating closely with
product and engineering teams.
I have delivered LLM-powered systems that reduced manual analysis and decision workflows
by over 40% across internal teams. I improved response accuracy and reliability by
implementing RAG pipelines and prompt evaluation mechanisms that materially reduced
hallucinations in real usage. I productionized AI services as low-latency APIs with enforced cost
controls and access isolation, and introduced monitoring loops for prompt, agent behavior, and
failure analysis. My work established reusable internal patterns for safely deploying agentic AI
systems at scale.