Building AI that runs at the edge.
I'm a researcher and engineer with a background in Electronics & Telecommunications Engineering,
specialising in making large language models smaller, faster, and deployable on constrained hardware.
My work sits at the intersection of cs.LG and cs.AR.
I have production experience building RAG systems and quantized LLM pipelines, a preprint on LLM
quantization, and a Kaggle competition placement. I'm currently pursuing MSc AI/ML programs abroad
with a focus on efficient inference research.
Longer term, I'm interested in founding AI-native tools and contributing to open-source infrastructure
for efficient ML.