Navid Azizan's Avatar

Navid Azizan

@navidazizan

MIT Prof | AI & machine learning, systems & control, optimization | Fmr postdoc @Stanford, PhD @Caltech azizan.mit.edu

42
Followers
38
Following
9
Posts
23.11.2024
Joined
Posts Following

Latest posts by Navid Azizan @navidazizan

Huge thanks to our amazing co-authors and collaborators—Kristjan Greenewald, Hao Wang, Amin Heyrani Nobari, Ali ArjomandBigdeli, Akash Srivastava, Faez Ahmed, Ali Jadbabaie, Ji Young Byun, and Rama Chellappa—and to MIT-IBM Watson AI Lab, Google, Amazon, and MathWorks for their support.

05.12.2025 03:29 👍 0 🔁 0 💬 0 📌 0
Preview
On the Role of Transformer Feed-Forward Layers in Nonlinear In-Context Learning Transformer-based models demonstrate a remarkable ability for in-context learning (ICL), where they can adapt to unseen tasks from a few prompt examples without parameter updates. Recent research has ...

Haoyuan Sun will present "On the Role of Transformer Feed-Forward Layers in Nonlinear In-Context Learning" arxiv.org/abs/2501.18187

05.12.2025 03:29 👍 0 🔁 0 💬 1 📌 0
Preview
ECO: Energy-Constrained Operator Learning for Chaotic Dynamics with Boundedness Guarantees Chaos is a fundamental feature of many complex dynamical systems, including weather systems and fluid turbulence. These systems are inherently difficult to predict due to their extreme sensitivity to ...

Andrea Goertzen and Sunbochen Tang will present "ECO: Energy-Constrained Operator Learning for Chaotic Dynamics with Boundedness Guarantees" arxiv.org/abs/2512.01984

05.12.2025 03:29 👍 1 🔁 0 💬 1 📌 0
Preview
HardNet: Hard-Constrained Neural Networks with Universal Approximation Guarantees Incorporating prior knowledge or specifications of input-output relationships into machine learning models has attracted significant attention, as it enhances generalization from limited data and yiel...

Youngjae Min will present "HardNet: Hard-Constrained Neural Networks" arxiv.org/abs/2410.10807

05.12.2025 03:29 👍 1 🔁 0 💬 1 📌 0
Preview
Activation-Informed Merging of Large Language Models Model merging, a method that combines the parameters and embeddings of multiple fine-tuned large language models (LLMs), offers a promising approach to enhance model performance across various tasks w...

Kaveh Alim will present "Activation-Informed Merging of Large Language Models" arxiv.org/abs/2502.02421

05.12.2025 03:29 👍 0 🔁 0 💬 1 📌 0
Preview
Know What You Don't Know: Uncertainty Calibration of Process Reward Models Process reward models (PRMs) play a central role in guiding inference-time scaling algorithms for large language models (LLMs). However, we observe that even state-of-the-art PRMs can be poorly calibr...

Young-Jin Park will present "Know What You Don't Know: Uncertainty Calibration of Process Reward Models" arxiv.org/abs/2506.09338

05.12.2025 03:29 👍 0 🔁 0 💬 1 📌 0
Post image

Excited to be at #NeurIPS with several of my brilliant students! Some of them are looking for internships and full-time roles, and we are also recruiting new students and postdocs—come find us at any of these sessions!

#NeurIPS2025

05.12.2025 03:29 👍 2 🔁 0 💬 1 📌 0

Introducing Instance-Adaptive Inference-Time Scaling!

Paper: arxiv.org/abs/2506.09338
Code: github.com/azizanlab/in...

14.07.2025 15:24 👍 1 🔁 0 💬 0 📌 0

Congratulations, Necmiye!

21.05.2025 23:36 👍 1 🔁 0 💬 0 📌 0