Welcome to Kaiyi Ji's Homepage
About Me
I am an assistant professor at the Department of Computer Science and Engineering of the University at Buffalo, The State University of New York. I received my Ph.D. degree from the Electrical and Computer Engineering Department of The Ohio State University in December, 2021, advised by Prof. Yingbin Liang. I was a postdoctoral research fellow at the Electrical Engineering and Computer Science Department of the University of Michigan, Ann Arbor, in 2022, working with Prof. Lei Ying. I was a visiting student research collaborator at the department of Electrical Engineering, Princeton University.
Previously I obtained my B.S. degree from University of Science and Technology of China in 2016. I received CSE Junior Faculty Research Award (2023) and NSF CAREER Award (2025).
Research
I am particularly interested in the intersection of optimization, machine learning, and deep learning, from both theoretical and algorithmic perspectives. My current focus is on:
Large-Scale Optimization: We design provably efficient optimization algorithms for machine (deep) learning tasks with strong empirical performance. Recent efforts focus on preconditioned methods for large language model (LLM) training, bilevel optimization for hierarchical tasks, and multi-objective optimization for tasks with multiple competing objectives.
Machine (Deep) Learning: We develop algorithms to tackle key challenges in modern machine learning problems. Our current interest lies in multi-task learning (such as conflict resolution, balancing, and scalability) and life-long/continual learning (such as data selection, forgetting, computing cost). We also pursue theoretical foundations to understand the mechanisms, generalization behavior, and computational complexity of these methods.
Foundation Models: We explore efficient and theoretically grounded approaches for training LLMs under resource constraints. Current directions include continual prompt tuning, structured low-rank adaptation, and multi-objective decoding for LLMs.
Recent News!
[Talk] 07/2025 I gave a talk on tuning-free bilevel optimization at ICCOPT 2025, Los Angeles, CA. Thanks for the invite from Prof. Shiqian Ma and Prof. Tong Zhang.
[Service] 07/2025 I will be serving as a reviewer for SODA 2026.
[Manuscript] 07/2025 Our new paper on Task-Agnostic Continual Prompt Tuning for LLM training is available online. We propose GRID by integrating a
task-aware decoding mechanism that improves backward transfer by leveraging representative inputs,
automatic task identification, and constrained decoding. It reduces forgotten tasks by up to 80% without sacrificing forward transfer performance. Check our code: GitHub.
[Manuscript] 02/2025 Our new multi-task learning (MTL) paper is out! We present LDC-MTL, a balanced MTL approach with O(1) time and memory complexity, offering both high accuracy and efficiency. It reduces loss discrepancy, minimizes gradient conflicts, and outperforms weight-swept linear scalarization through dynamic weight adjustment.
Check out our paper and code. Please star us if you find it helpful!
Recent Featured Works
Selected Publications
Imperative Learning: A Self-supervised Neural-Symbolic Learning Framework for Robot Autonomy <Open-Source Library> Chen Wang, Kaiyi Ji, Junyi Geng, Zhongqiang Ren, Taimeng Fu, Fan Yang, Yifan Guo, Haonan He, Xiangyu Chen, Zitong Zhan, Qiwei Du, Shaoshu Su, Bowen Li, Yuheng Qiu, Yi Du, Qihang Li, Yifan Yang, Xiao Lin, Zhipeng Zhao. The International Journal of Robotics Research (IJRR), 2025.
|