Nandan Kumar Jha

Nandan Kumar Jha

PhD student at NYU CCS

New York University

About me

I am a Ph.D. candidate in Electrical and Computer Engineering at New York University, advised by Prof. Brandon Reagen. My research explores the mathematical foundations of large language models (LLMs)—how information, geometry, and learning dynamics interact and shape the stability, efficiency, and scaling behavior of LLMs.

In particular, my work pursues three complementary thrusts: representation integrity (entropy budgets, spectral utilization, stability regimes), scientific foundations (information theory, inductive biases, scaling laws), and high-dimensional learning dynamics (eigenspectra, weight manifolds, spectral geometry).

The long-term aim is to build first-principles frameworks for designing and optimizing LLMs while preserving their representational integrity. This effort led to NerVE, an eigenspectral framework that characterizes the nonlinear transformations of FFNs, and uses spectral utilization metrics to quantify FFN width utilization (EMNLP 2025 (Main)).

I also developed AERO, an information-theoretic framework that studies how nonlinearities influence entropy budgets of attention mechanisms and introduces entropy-guided attention for private LLM architectures with fewer nonlinear operations. Preliminary results appeared at PPAI@AAAI'25 and ATTRIB@NeurIPS'24.

Earlier in my Ph.D., as part of the DPRIVE project, I proposed new architectures and algorithms for efficient inference on encrypted data. This includes DeepReDuce (ICML'21, Spotlight), a ReLU-optimization technique, and DeepReShape (TMLR'24), a family of CNNs redesigned for private inference efficieny. Both works redefined the state-of-the-arts in private inference, achieving 3.5× and 8.7× speedups over prior SOTA, respectively.

Recent talks: We presented our work Entopy and Private Language Models at the NYU CILVR Seminar, and Entropy-Guided Attention for Private LLMs on the AI Fireside Chat

Besides research, I have contributed as an (invited) reviewer for NeurIPS (2023-25), ICML (2024, 2025), ICLR (2024, 2025), TMLR (2025), AISTATS (2025), CVPR (2024, 2025), ICCV (2025), and AAAI (2025).

Interests
  • Representation Integrity in LLMs
  • Scientific Foundations of LLMs
  • High Dimensional Learning Dynamics
  • Cryptographically Secure PPML
Education
  • Ph.D. in Neural Architectures for Efficient Private Inference, 2020 - present

    New York University

  • M.Tech. (Research) in Computer Science and Engineering, 2017 - 2020

    Indian Institute of Technology Hyderabad

  • B.Tech. in Electronics and Communication Engineering, 2009 - 2013

    National Institute of Technology Surat

News & Updates

Recent Highlights

  • 2025-01 · AAAI’25 (PPAI Workshop) — Entropy-Guided Attention accepted. arXiv
  • 2024-10 · NeurIPS’24 (ATTRIB) — Talk on ReLU in norm-free LLMs. arXiv [Slides]
  • 2024-10 · Preprint — AERO: Softmax-only private LLMs with entropy regularization released. arXiv [Code]
  • 2024-09 · TMLR — DeepReShape accepted (ReLU-equalization, HybReNets). arXiv
  • 2023-03 · ASPLOS — End-to-end private inference system accepted. arXiv
  • 2021-12 · NeurIPS — Circa accepted (GC + stochastic ReLUs). arXiv
  • 2021-11 · Preprint — CryptoNite: throughput limits under realistic load. arXiv
  • 2021-11 · ACM CCS (PPML) — Sisyphus accepted (Quadratic Imitation Learning). arXiv
  • 2021-07 · ICML (Spotlight) — DeepReDuce: criticality-based ReLU dropping. arXiv

See older updates on the All News page.

Recent Publications

Quickly discover relevant content by filtering publications.
(2024). AERO: Softmax-Only LLMs for Efficient Private Inference. In ArXiv Preprint.

PDF Cite

(2024). ReLU's Revival: On the Entropic Overload in Normalization-Free Large Language Models. In ATTRIB (NeurIPS) Workshop.

PDF Cite Code Poster

(2024). DeepReShape: Redesigning Neural Networks for Efficient Private Inference. In TMLR 2024.

PDF Cite Slides Video

(2023). Characterizing and Optimizing End-to-End Systems for Private Inference. In ASPLOS 2023.

PDF Cite Code Poster

(2021). CryptoNite: Revealing the Pitfalls of End-to-End Private Inference at Scale. In Arxiv Preprint.

PDF Cite

Experience

 
 
 
 
 
Seagate Technology
Electrical Design Engineer
Sep 2015 – Jul 2017 Bangalore, INDIA

Responsibilities include:

  • Designing power delivery circuit for M.2 Solid State Drives.
  • Electrical characterization of DRAM and NAND modules
  • Signal Intergrity verification of DRAM/NAND datapath
 
 
 
 
 
IIT Bombay
Project Research Assistant
Nov 2014 – Jun 2015 Mumbai, INDIA
Worked on the deployment of wireless broadband in rural areas using the TV white Space (unused licensed band in UHF band)

Contact