Publications

(2024). AERO: Softmax-Only LLMs for Efficient Private Inference. In ArXiv Preprint.

PDF Cite

(2024). ReLU's Revival: On the Entropic Overload in Normalization-Free Large Language Models. In ATTRIB (NeurIPS) Workshop.

PDF Cite Code

(2024). DeepReShape: Redesigning Neural Networks for Efficient Private Inference. In TMLR 2024.

PDF Cite Slides Video

(2023). Characterizing and Optimizing End-to-End Systems for Private Inference. In ASPLOS 2023.

PDF Cite Code Poster

(2021). CryptoNite: Revealing the Pitfalls of End-to-End Private Inference at Scale. In Arxiv Preprint.

PDF Cite

(2021). Circa: Stochastic ReLUs for Private Deep Learning. In NeurIPS 2021.

PDF Cite Poster

(2021). DeepReDuce: ReLU Reduction for Fast Private Inference. In ICML 2021 (Spotlight).

PDF Cite Poster Slides ICML video Long video Press release

(2020). DeepPeep: Exploiting Design Ramifications to Decipher the Architecture of Compact DNNs. In ACM JETC 2020.

PDF Cite DOI

(2020). Modeling Data Reuse in Deep Neural Networks by Taking Data-Types into Cognizance. In IEEE TC 2020.

PDF Cite DOI

(2020). DRACO: Co-optimizing hardware utilization, and performance of dnns on systolic accelerator. In ISVLSI 2020.

PDF Cite Slides Video DOI

(2020). ULSAM: Ultra-lightweight subspace attention module for compact convolutional neural networks. In WACV 2020.

PDF Cite Code Poster Slides Video DOI

(2020). E2GC: Energy-efficient group convolution in deep neural networks. In VLSID 2020.

PDF Cite Slides DOI

(2019). Data-type Aware Arithmetic Intensity for Deep Neural Networks. In ICCD 2019 (Poster).

PDF Cite Poster DOI

(2019). The Ramifications of Making Deep Neural Networks Compact. In VLSID 2019.

PDF Cite Slides DOI