My research focuses on reducing memory footprint and accelerating deep learning models through compression techniques such as pruning, quantization, and low-rank approximation. I aim to make large models more efficient and deployable without significantly sacrificing performance.
M.S. in Computer Science, 2025-present
KAUST
B.S. in Computer Science, 2024
Peking University