HOME CV CLICKBAIT
Hi, I'm Yibo (Jacky) Zhang


photo taken in Vancouver, Canada
I am a PhD student in the Department of Computer Science at Stanford University, where I am fortunate to be advised by Sanmi Koyejo.

I'm interested in solving fundamental AI problems through theoretical research that leads to real-world solutions.

Currently, my research focuses on learning in systems composed of dynamically and stochastically interacting components, particularly how such systems can give rise to autonomous machine intelligence. Pease see my selected works for more details.

Education Stanford University
  • PhD Student - Department of Computer Science (present)
University of Illinois at Urbana-Champaign
  • M.S. in Computer Science (2022)
University of Science and Technology of China
  • B.E. in Computer Science (2019)
Publications & Preprints
(*eqaul contribution)
Selected Exploring Dynamical Stochastic Field Theory for General Autonomous Learning (work in progress) [ preliminary version: pdf ]
Yibo Jacky Zhang, Sanmi Koyejo.
Preprint, 2025.
  • An important open problem in the science of artificial intelligence is how to build autonomous agents that learns continuously in unknown environments using local observations.
  • This paper explores this problem through dynamical stochastic neural networks, where each neuron acts as a simple agent that continuously adapts through local learning rules. The learning requires no resets, replay buffers, context windows, or backpropagation.
A Framework for Objective-Driven Dynamical Stochastic Fields [ pdf ]
Yibo Jacky Zhang, Sanmi Koyejo.
Preprint, 2025.
  • It is challenging to describe complex systems composed of interacting and dynamic components. In particular, it becomes more challenging as these components in the system exhibit objective-driven behaviors.
  • This paper develops a formal and elegant description of such systems using a field-theoretic language inspired by physics. The proposed theoretical framework is referred to as intelligent fields.
Aligning Compound AI Systems via System-level DPO [ pdf ]
Xiangwen Wang*, Yibo Jacky Zhang*, Zhoujie Ding, Katherine Tsai, Sanmi Koyejo.
Neural Information Processing Systems (NeurIPS), 2025.
  • Compound AI systems, comprising multiple interacting components such as LLMs and diffusion models, have demonstrated improvements compared to single models. However, aligning compound AI systems to human preferences is challenging.
  • We propose a principled framework (SysDPO) for aligning all components in a compound AI system as a cohesive whole.
List of All Exploring Dynamical Stochastic Field Theory for General Autonomous Learning (work in progress) [ preliminary version: pdf ]
Yibo Jacky Zhang, Sanmi Koyejo.
Preprint, 2025.
Improving Single-round Active Adaptation: A Prediction Variability Perspective [ pdf ]
Xiaoyang Wang, Yibo Jacky Zhang, Olawale Elijah Salaudeen, Mingyuan Wu, Hongpeng Guo, Chaoyang He, Klara Nahrstedt, Sanmi Koyejo.
Transactions on Machine Learning Research (TMLR), 2025.
A Framework for Objective-Driven Dynamical Stochastic Fields [ pdf ]
Yibo Jacky Zhang, Sanmi Koyejo.
Preprint, 2025.
Aligning Compound AI Systems via System-level DPO [ pdf ]
Xiangwen Wang*, Yibo Jacky Zhang*, Zhoujie Ding, Katherine Tsai, Sanmi Koyejo.
Neural Information Processing Systems (NeurIPS), 2025.
Probing Human Visual Robustness with Neurally-Guided Deep Neural Networks [ pdf ]
Zhenan Shao, Linjian Ma, Yiqing Zhou, Yibo Jacky Zhang, Sanmi Koyejo, Bo Li, Diane M Beck.
Preprint, 2024.
Can Public Large Language Models Help Private Cross-device Federated Learning? [ pdf ]
Boxin Wang, Yibo Jacky Zhang, Yuan Cao, Bo Li, H. Brendan McMahan, Sewoong Oh, Zheng Xu, Manzil Zaheer.
NAACL 2024.
Principled Federated Domain Adaptation: Gradient Projection and Auto-Weighting [ pdf ]
Enyi Jiang*, Yibo Jacky Zhang*, Oluwasanmi Koyejo.
International Conference on Learning Representations (ICLR), 2024.
Batch Active Learning from the Perspective of Sparse Approximation [ pdf ] [ poster ]
Maohao Shen*, Bowen Jiang*, Jacky Y. Zhang*, Oluwasanmi Koyejo.
NeurIPS 2022 Workshop on Human in the Loop Learning.
Adversarially Robust Models may not Transfer Better: Sufficient Conditions for Domain Transferability from the View of Regularization [ pdf ] [ poster ] [ talk ]
Xiaojun Xu*, Jacky Y. Zhang*, Evelyn Ma, Danny Son, Oluwasanmi Koyejo, Bo Li
International Conference on Machine Learning (ICML), 2022.
Bayesian Coresets: Revisiting the Nonconvex Optimization Perspective
[ pdf ] [ poster ] [ short talk ] [ long talk ]
Jacky Y. Zhang, Rajiv Khanna, Anastasios Kyrillidis, Oluwasanmi Koyejo.
International Conference on Artificial Intelligence and Statistics (AISTATS), 2021. (Oral)
Uncovering the Connections Between Adversarial Transferability and Knowledge Transferability [ pdf ] [ poster ] [ talk ]
Kaizhao Liang*, Jacky Y. Zhang*, Boxin Wang, Zhuolin Yang, Oluwasanmi Koyejo, Bo Li
International Conference on Machine Learning (ICML), 2021.
Labeling Cost-Sensitive Batch Active Learning for Brain Tumor Segmentation
Maohao Shen, Jacky Y. Zhang, Leihao Chen, Weiman Yan, Neel Jani, Brad Sutton, Oluwasanmi Koyejo.
International Symposium on Biomedical Imaging (ISBI), 2021.
Robusta: Robust AutoML for Feature Selection via Reinforcement Learning [ pdf ]
Xiaoyang Wang, Bo Li, Yibo Zhang, Bhavya Kailkhura, Klara Nahrstedt.
AAAI 2021 Workshop Towards Robust, Secure and Efficient Machine Learning.
Learning Sparse Distributions using Iterative Hard Thresholding [ pdf ] [ poster ]
Jacky Y. Zhang, Rajiv Khanna, Anastasios Kyrillidis, Oluwasanmi Koyejo.
Neural Information Processing Systems (NeurIPS), 2019.
Maximizing Monotone DR-submodular Continuous Functions by Derivative-free Optimization [ pdf ]
Yibo Zhang, Chao Qian, Ke Tang.
Preprint: arXiv 1810.06833, 2018.
On Multiset Selection with Size Constraints [ pdf ]
Chao Qian, Yibo Zhang, Ke Tang, Xin Yao.
AAAI Conference on Artificial Intelligence (AAAI), 2018.
Contact yiboz@stanford.edu
Miscellaneous I like to ponder random things. For example: Would aliens also have their mouths near their brains? Will artificial and biological intelligence eventually converge? How much could I contribute to scientific progress if I were sent back in time 1000 years? And, most importantly, why are you still reading all this nonsense?
Photographer of Today: Matt Stuart Music of Today: A Love Song - EGO WRAPPIN' I will probably update these tomorrow almost surely with high probability.