I like the messy middle of software: the place where a request leaves one machine, gets queued, balanced, replicated, and still has to come back fast. That's pulled me toward databases, backend infra, and lately the emerging problem space of LLM inference at scale.
I'm also drawn to the founder side of building: the why behind what gets built. I've shipped end-to-end at an early-stage startup (Hangry), competed in the YC × Google DeepMind hackathon, and run my own products. Early-stage technical tradeoffs (what to build, what to skip, what to bet on) are the kind of questions I like chewing on.
I'm also a passionate researcher. I have done CNN research with Professor Thiago Serra and I am currently exploring ideas around reinforcement learning and continual learning with Professor Brian King.
Outside of work, I work on open source (SGLang, Kubernetes SIG Inference-Perf, and an AI Gateway / envoy project), read papers on inference serving, and occasionally ship a side project end-to-end.