Hi, I am Zihang. I am a researcher at UC Berkeley and International Computer Science Institute, advised by Prof. Michael Mahoney. I also have the privilege to work closely with Prof. Yaoqing Yang from Dartmouth College. I previously obtained my Master’s degree in EECS at UC Berkeley.

My research focus is to understand and improve the transparency and efficiency of learning models. I am particularly interested in:

  • Understanding the mechanisms, behaviors of AI systems, such as generalization, uncertainty, and reasoning of LLMs, and using principled approaches to improve the transparency and efficiency.
  • Leveraging the geometry of deep learning such as low-rank subspaces and loss landscapes, to develop new (numerical) algorithms for training large-scale AI systems and solving numerical problems.

My research draws inspirations from high-dimensional statistics, random matrix theory and (randomized) numerical linear algebra.

🔥 News

  • 2026.02:   Excited to share our new paper that proposes AutoSpec – a nerual network framework to discover iterative spectral algorithms for NLA and optimization.
  • 2025.06:   Started my research engineer position at ICSI to work on numerical algorithms and deep learning.
  • 2025.05:   Our paper “Principal Weights Emerge after Rank Reduction for Reasoning-Focused Supervised Fine-Tuning” has been accepted to ICML 2025.
  • 2024.11:   Gave a presentation at EMNLP 2024 on foundation model diagnosis, check out the live recording here.
  • 2024.09:   Excited to share that our work “Model Balancing Helps Low-data Training and Fine-tuning” is accepted by EMNLP 2024 as Oral Presentation.

📝 Selected Publications

Authors within {} are equal contributors.

arXiv preprint
sym

Learning to Discover Iterative Spectral Algorithms

{Zihang Liu*, Oleg Balabanov*}, Yaoqing Yang, Michael W. Mahoney

Paper

arXiv preprint

ICML 2025
sym

LIFT the Veil for the Truth: Principal Weights Emerge after Rank Reduction for Reasoning-Focused Supervised Fine-Tuning

Zihang Liu, Tianyu Pang, Oleg Balabanov, Chaoqun Yang, Tianjin Huang, Lu Yin, Yaoqing Yang, Shiwei Liu

Paper | Code

ICML 2025

EMNLP 2024
sym

Model Balancing Helps Low-data Training and Fine-tuning

{Zihang Liu*, Yuanzhe Hu*}, Tianyu Pang, Yefan Zhou, Pu Ren, Yaoqing Yang

Paper | Code | Video

EMNLP 2024 Oral