Hanseul Cho (조한슬) 👋
I am a Ph.D. student in the Optimization & Machine Learning (OptiML) Laboratory, where I am fortunate to be advised by Prof. Chulhee Yun at Kim Jaechul Graduate School of AI in Korea Advanced Institute of Science and Technology (KAIST AI). Previously, I completed my M.Sc. (in AI) and B.Sc. (in Math, minor in CS, Summa Cum Laude) at KAIST.
I am interested in a broad range of fields in optimization, machine learning (ML), and deep learning (DL), especially focusing on both:
- Mathematical/Theoretical Analysis, and
- Empirical Improvements (usually based on theoretical understanding).
Specifically, my research interests lie in the following subjects:
- Understanding and mitigating the fundamental limitations of modern (large) language models (e.g., length generalization and compositional generalization of Transformer architectures)
- Online learning & optimization under circumstance shifts (e.g., continual learning, maintaining plasticity of neural networks, reinforcement learning, streaming PCA)
- Structured/constrained/multi‑level optimization (e.g., minimax optimization, bi‑level optimization, fairness in ML)
📰 Publications 📰
Please click the “Publications” tab above to look up the full list of my publications.
You can also find my articles on my Google Scholar profile.
‼️News‼️
- 🗞️ [Nov. '24] Our paper on theoretical analysis of continual learning is accepted to JKAIA 2024 and won the Best Paper Award! 🎉 (See Publications for more details)
- 🗞️ [Nov. '24] I'm selected as one of the Top Reviewers (top 8.6%: 1,304 of 15,160 reviewers) at NeurIPS 2024! (+ Free registration! 😎)
- 🗞️ [Sep. '24] Two papers got accepted to NeurIPS 2024! 🎉 One is about length generalization of arithmetic Transfomers, and another is about mitigating loss of plasticity in incremental neural net training. See you in Vancouver🇨🇦!
- 🗞️ [Jun. '24] An early version of our paper on length generalization of Transformers got accepted to the ICML 2024 Workshop on Long-Context Foundation Models!
- 🗞️ [May. '24] A paper got accepted to ICML 2024 as a spotlight paper (top 3.5% among all submissions)! 🎉 We show global convergence of Alt-GDA (which is strictly faster than Sim-GDA) and propose an enhanced algorithm called Alex-GDA for minimax optimization. See you in Vienna🇦🇹!
- 🗞️ [Sep. '23] Two papers are accepted to NeurIPS 2023! 🎉 One is about Fair Streaming PCA and another is about enhancing plasticity in RL.
- 🗞️ [Jan. '23] Our paper about shuffling-based stochastic gradient descent-ascent got accepted to ICLR 2023!
- 🗞️ [Nov. '22] Our paper about shuffling-based stochastic gradient descent-ascent is accepted to 2022 Korea AI Association + NAVER Autumnal Joint Conference (JKAIA 2022) and selected as the NAVER Outstanding Theory Paper!
- 🗞️ [Oct. '22] I am happy to announce that our very first preprint is now on arXiv! It is about convergence analysis of shuffling-based stochastic gradient descent-ascent.
- 🗞️ [Feb. '22] Now I am part of OptiML Lab of KAIST AI.
Education
- 🏫 Ph.D. in Artificial Intelligence
KAIST, Sept. 2023 – Current - 🏫 M.Sc. in Artificial Intelligence
KAIST, Mar. 2022 – Aug. 2023 - 🏫 B.Sc. in Mathematical Sciences
KAIST, Mar. 2017 – Feb. 2022 - Minor in Computing Sciences / Summa Cum Laude
Contact & Info
📋 Curriculum Vitae (CV): Here
📧 Email: jhs4015 at kaist dot ac dot kr