I work on efficient training of deep learning models, focusing on low-rank methods and quantization. I am advised by Michael J. Kastoryano and Serge Belongie, and supported by the Danish Data Science Academy (DDSA). In 2025, I spent six months visiting Anima Anandkumar’s group at Caltech, and collaborated with Jean Kossaifi (NVIDIA Research). In 2025–2026, I completed a six-month internship at Qualcomm AI Research in Amsterdam, working with the Model Efficiency group under the supervision of Markus Nagel.
TensorGRaD: Tensor Gradient Robust Decomposition for Memory-Efficient Neural Operator Training
arXiv preprint, 2025.
Code PaperTensorGRaD factorizes gradient tensors into complementary low-rank and sparse components, reducing optimizer-state memory by up to 75 % while matching the accuracy of Adam, even on turbulent Navier–Stokes at Re = 105.
LoQT: Low-Rank Adapters for Quantized Pre-Training
NeurIPS 2024. *Equal contribution. Best Paper Award at WANT@ICML.
  Code   PaperLoQT enables efficient quantized pre-training of LLMs with results close to full-rank non-quantized models. It enables pre-training of a 13B LLM on a 24GB GPU without model parallel, checkpointing, or offloading strategies during training.
Coarse-To-Fine Tensor Trains for Compact Visual Representations
ICML 2024.
Project page Code PaperPuTT (Prolongation Upsampling Tensor Train), a method for learning a coarse-to-fine tensor train representation of visual data, excelling in 2D/3D fitting and novel view synthesis, even with noisy or incomplete data.
Text-Driven Stylization of Video Objects
ECCV Workshop on AI for Creative Video Editing and Understanding, 2022.
Best Paper Award Oral Presentation
A method for stylizing video objects in an intuitive and semantic manner following a user-specified text prompt.
Progressive Parameter Space Visualization for Task-Driven SAX Configuration
EuroVA, 2020. Oral Presentation
  PaperTool that uses progressive visual analytics to guide users in hyperparameter search for SAX and PAA algorithms.