About Me

I am Ziheng Cheng, a first-year PhD student in Department of IEOR, UC Berkeley. Prior to that, I got my B.S. degree in School of Mathematical Sciences, Peking University, supervised by Cheng Zhang. I was also very fortunate to have worked with Kun Yuan, Tengyu Ma. My research interests span broadly in statistics, optimization and machine learning, including multi-agent RL, language models and diffusion models, distributed optimization, sampling and variational inference. If you are interested in my research, please feel free to contact me.

News

  • Feb, 2025 A new paper “Provable Sample-Efficient Transfer Learning Conditional Diffusion Models via Representation Learning” on Arxiv!
  • Jan, 2025 Our paper “Convergence of Distributed Adaptive Optimization with Local Updates” accepted at ICLR 2025!
  • Oct, 2024 A new paper “Semi-Implicit Functional Gradient Flow” on Arxiv!
  • Sep, 2024 Our paper “Functional Gradient Flows for Constrained Sampling” accepted at NeurIPS 2024!
  • Sep, 2024 A new paper “Convergence of Distributed Adaptive Optimization with Local Updates” on Arxiv!
  • May, 2024 Glad to serve as a reviewer of NeurIPS 2024 for the first time!
  • May, 2024 Our paper “The Limits and Potentials of Local SGD for Distributed Heterogeneous Learning with Intermittent Communication” accepted at COLT 2024!
  • May, 2024 Our paper “Kernel Semi-Implicit Variational Inference”, “Reflected Flow Matching” accepted at ICML 2024!
  • Apr, 2024 Glad to be an incoming PhD student at UC Berkeley IEOR!
  • Jan, 2024 Our paper “Momentum Benefits Non-IID Federated Learning Simply and Provably” accepted at ICLR 2024!
  • Oct, 2023 Join Microsoft Research Asia as an intern!
  • Sep, 2023 Our paper “Particle-based Variational Inference with Generalized Wasserstein Gradient Flow” accepted at NeurIPS 2023!
  • Jun, 2023 Visit Tengyu Ma at Stanford!
  • Jun, 2023 A new paper “Momentum Benefits Non-IID Federated Learning Simply and Provably” on Arxiv!
  • May, 2023 A new paper “Joint Graph Learning and Model Fitting in Laplacian Regularized Stratified Models” on Arxiv!

Publications

  • (Preprint) Provable Sample-Efficient Transfer Learning Conditional Diffusion Models via Representation Learning
    Ziheng Cheng, Tianyu Xie, Shiyue Zhang, Cheng Zhang
    [Arxiv]

  • (Preprint) Semi-Implicit Functional Gradient Flow
    Shiyue Zhang*, Ziheng Cheng*, Cheng Zhang
    [Arxiv]

  • (ICLR 2025) Convergence of Distributed Adaptive Optimization with Local Updates
    Ziheng Cheng, Margalit Glasgow
    [Arxiv]

  • (NeurIPS 2024) Functional Gradient Flows for Constrained Sampling
    Shiyue Zhang*, Longlin Yu*, Ziheng Cheng*, Cheng Zhang
    [Arxiv]

  • (COLT 2024) The Limits and Potentials of Local SGD for Distributed Heterogeneous Learning with Intermittent Communication
    Kumar Kshitij Patel, Margalit Glasgow, Ali Zindari, Lingxiao Wang, Sebastian U Stich, Ziheng Cheng, Nirmit Joshi, Nathan Srebro
    [Arxiv]

  • (ICML 2024) Kernel Semi-Implicit Variational Inference
    Ziheng Cheng*, Longlin Yu*, Tianyu Xie, Shiyue Zhang, Cheng Zhang
    [Arxiv]

  • (ICML 2024) Reflected Flow Matching
    Tianyu Xie*, Yu Zhu*, Longlin Yu*, Tong Yang, Ziheng Cheng, Shiyue Zhang, Xiangyu Zhang, Cheng Zhang
    [Arxiv]

  • (ICLR 2024) Momentum Benefits Non-IID Federated Learning Simply and Provably
    Ziheng Cheng*, Xinmeng Huang*, Pengfei Wu, Kun Yuan
    [Arxiv]

  • (NeurIPS 2023) Particle-based Variational Inference with Generalized Wasserstein Gradient Flow
    Ziheng Cheng*, Shiyue Zhang*, Longlin Yu, Cheng Zhang
    [Arxiv]

  • (Under Review) Joint Graph Learning and Model Fitting in Laplacian Regularized Stratified Models
    Ziheng Cheng*, Junzi Zhang*, Akshay Agrawal, Stephen Boyd
    [Arxiv]

Experiences

  • Peking University
    Undergrad Research (advisor: Prof. Cheng Zhang, School of Mathematical Sciences)
    May. 2022 – Present

  • Stanford University
    Summer Research (advisor: Prof. Tengyu Ma, Department of Computer Science)
    Jun. 2023 – Oct. 2023

  • Peking University
    Undergrad Research (advisor: Prof. Kun Yuan, Center for Machine Learning Research)
    Mar. 2023 – Sept. 2023

  • Stanford University
    Remote Research (advisor: Prof. Stephen Boyd, Department of Electrical Engineering)
    Oct. 2022 – May. 2023