Nandan Kumar Jha

Nandan Kumar Jha

PhD student at NYU CCS

New York University

About me

I am an ECE PhD candidate at New York University (Center for Cybersecurity), advised by Prof. Brandon Reagen. I’m broadly interested in the science of deep learning, particularly high-dimensional learning dynamics and information theory, to understand and optimize large language models (LLMs) from first principles.

My work brings together information-theoretic analysis, spectral geometry, and random matrix theory to develop tools that probe the the learning dynamics, and capacity scaling in LLMs. Specifically, I developed an information-theoretic framework, AERO, to study the impact of nonlinearities on attention mechanism, and introduced an entropy-guided attention for Private LLM architectures with fewer nonlinear operations. Our preliminary findings have been accepted to PPAI@AAAI'25 and ATTRIB@NeurIPS'24.

Earlier in my PhD, and as part of the DPRIVE project, I proposed new architectures and algorithms for efficient neural network computation on encrypted data, including DeepReDuce, ICML'21, a ReLU-optimization technique, and DeepReShape, TMLR'24, a redesign of CNNs for private inference efficiency.

Recent talks: We presented our work Entopy and Private Language Models at the NYU CILVR Seminar, and Entropy-Guided Attention for Private LLMs on the AI Fireside Chat

Besides research, I have contributed as an (invited) reviewer for NeurIPS (2023, 2024), ICML (2024, 2025), ICLR (2024, 2025), TMLR (2025), AISTATS (2025), CVPR (2024, 2025), ICCV (2025), and AAAI (2025).

I am currently on the job market, graduating in Fall 2025, and seeking research scientist roles at the intersection of LLM science, architectural optimization, and privacy-preserving AI. Feel free to reach out!

Interests
  • Scientific Foundations of LLMs
  • High Dimensional Learning Dynamics
  • Cryptographically Secure PPML
Education
  • Ph.D. in Neural Architectures for Efficient Private Inference, 2020 - present

    New York University

  • M.Tech. (Research) in Computer Science and Engineering, 2017 - 2020

    Indian Institute of Technology Hyderabad

  • B.Tech. in Electronics and Communication Engineering, 2009 - 2013

    National Institute of Technology Surat

Recent Publications

Quickly discover relevant content by filtering publications.
(2024). AERO: Softmax-Only LLMs for Efficient Private Inference. In ArXiv Preprint.

PDF Cite

(2024). ReLU's Revival: On the Entropic Overload in Normalization-Free Large Language Models. In ATTRIB (NeurIPS) Workshop.

PDF Cite Code Poster

(2024). DeepReShape: Redesigning Neural Networks for Efficient Private Inference. In TMLR 2024.

PDF Cite Slides Video

(2023). Characterizing and Optimizing End-to-End Systems for Private Inference. In ASPLOS 2023.

PDF Cite Code Poster

(2021). CryptoNite: Revealing the Pitfalls of End-to-End Private Inference at Scale. In Arxiv Preprint.

PDF Cite

Experience

 
 
 
 
 
Seagate Technology
Electrical Design Engineer
Sep 2015 – Jul 2017 Bangalore, INDIA

Responsibilities include:

  • Designing power delivery circuit for M.2 Solid State Drives.
  • Electrical characterization of DRAM and NAND modules
  • Signal Intergrity verification of DRAM/NAND datapath
 
 
 
 
 
IIT Bombay
Project Research Assistant
Nov 2014 – Jun 2015 Mumbai, INDIA
Worked on the deployment of wireless broadband in rural areas using the TV white Space (unused licensed band in UHF band)

Contact