Souvik Kundu, Ph.D.

I am a Staff Research Scientist at Intel Labs leading research efforts in scalable and novel AI primitives. Prior to joining at Intel I completed my Ph.D. in Electrical & Computer Engineering from University of Southern California. I was co-advised by Dr. Massoud Pedram and Dr. Peter A. Beerel. I was fortunate to receive the outstanding Ph.D. award and the Order De Arete award with multiple prestigious fellowships. I was fortunate to be recognized as one of the CPAL AI Rising Stars 2025 conferred by CPAL and Stanford DS. I am one of the youngest recipients of the Semiconductor Research Corporation (SRC, USA) outstanding liaison award for my impactful research in 2023. My research goal is to enable human life with robust and efficient AI services via a cross-layer innovation of algorithmic optimizations blended with existing and novel hardware compute and architectures. I have co-authored >75 peer-reviewed papers in various top-tier conferences including NeurIPS, ICLR, ICML, ACL, CVPR, DATE, and DAC with multiple Oral, young fellow, travel award, and best paper nominations. [google scholar]
I serve as the founding AC and committee member of the Conference on Parsimony and Learning (CPAL). Additionally, I serve in the AC committee for various journals and conferences including ICLR, NeurIPS (outstanding reviewer recognition'22), EMNLP (outstanding reviewer recognition'20), CVPR, DATE, and DAC.
news
Apr 09, 2025 | I will be serving as an Area Chair (AC) at NeurIPS 2025! |
---|---|
Mar 26, 2025 | Gave an invited talk at Stanford DS as a part of my recognition as CPAL AI Rising Star 2025 ! |
Mar 25, 2025 | Code for LANTERN (here: lantern-code), our ICLR 2025 paper is open-sourced now! We are glad to announce that LANTERN++ , an extension of the work got Oral presentation at SCOPE - ICLR 2025 workshop. ![]() |
Mar 21, 2025 | 1x paper on Microscaling quantization (MicroScopiQ) got accepted at ISCA 2025 ![]() |
Mar 06, 2025 | I will be serving as an Area Chair at ACL 2025 and TPC chair for AIML track at IEEE COINS 2025 |
Jan 22, 2025 | 3x papers (MambaExtend - improving long context understanding of Mamba, LANTERN - vision AR speculative decoding, and Scaling long context training data) accepted at ICLR 2025! ![]() |
Dec 03, 2024 | Our workshop proposal on Scalable Optimization for Efficient and Adaptive Foundation Models has been accepted at ICLR 2025 (SCOPE@ICLR2025)! Please consider submitting. Visit the workshop page here for details. ![]() |
selected publications
- ISCA 2025MicroScopiQ: Accelerating Foundational Models through Outlier-Aware Microscaling QuantizationIn International Symposium on Computer Architecture (ISCA), 2025
- ICLR 2025MambaExtend: A Training-Free Approach to Improve Long Context Extension of MambaIn International Conference on Learning Representations (ICLR), 2025
- ICLR 2023Learning to linearize deep neural networks for secure and efficient private inferenceIn International Conference on Learning Representation, 2023
- WACV 2021Spike-thrift: Towards energy-efficient deep spiking neural networks by limiting spiking activity via attention-guided compressionIn Proceedings of the IEEE/CVF winter conference on applications of computer vision, 2021
- ASP-DAC 2021DNR: A tunable robust pruning framework through dynamic network rewiring of dnnsIn Proceedings of the 26th Asia and South Pacific Design Automation Conference, 2021
- IEEE TC 2020Pre-defined sparsity for low-complexity convolutional neural networksIn IEEE Transactions on Computers, 2020