tl;dr Love to and will automate your tedious tasks way, experienced with on-device AI, big gpu AI , backend engineering, frontend dev, little bit of design and lots of software engineering.
Currently deep in on-device AI. Running AI models on phone + making them fast = happiness. I've been drilling deep into On-Device AI, more specifically, running running LLMs 🦙 on mobile 📱. Extremely interested in GPU programming (NVIDIA / Qualcomm).
Below are some highlights that gave me the most dopamine rush ⚡.
🌍 Product side highlights
⚡ Optimization adventures ⛰️
Early on in my career at SRA, my day-to-day included tackling cool problems of system design, implementing and scaling machine learning models (still do), developing scalable back end services and a front end that is a pleasure to use and a "lifesaver" (Slack messages to back this up).
Also used to be an intern from August 2017 - June 2018. As a research intern, Gained lots of experience implementing machine learning solutions end to end. Developed a distributed system framework that allows switching Natural Language Processing engines in real time. Also had papers published in reputable NLP conferences like ACL 2018 and COLING 2018.
Accepted at NAACL 2025
Accepted at ICLR 2025
In this paper we propose a novel approach for Large Language Model (LLM) compression such that there is no need for recovery finetuning. One of the highlights of my contribution here was implementing part of the model which traded a slightly higher memory usage for a 100x speedup in execution speed.