Arithmetic Transformers Can Length-Generalize in Both Operand Length and Count
π₯ Hanseul Cho*, Jaeyoung Cha*, Srinadh Bhojanapalli*, and Chulhee Yun
π₯ Hanseul Cho*, Jaeyoung Cha*, Srinadh Bhojanapalli*, and Chulhee Yun
π₯ Hanseul Cho and Chulhee Yun
π° ICLR 2023 π π [paper] [arxiv]
π NAVER Outstanding Theory Paper Award at the 7th Joint Conference of Korea Artificial Intelligence Association (JKAIA 2022)
π₯ Jaewook Lee*, Hanseul Cho*, and Chulhee Yun
π° ICML 2024 (Short version at ICLR 2024 Workshop on Bridging the Gap Between Practice and Theory in Deep Learning (BGPT)) π π [paper] [arxiv] [code]
π Spotlight @ ICML 2024 (Top 3.5% among total submissions)
π₯ Jaewook Lee*, Hanseul Cho*, and Chulhee Yun
π° ICML 2024 (Short version at ICLR 2024 Workshop on Bridging the Gap Between Practice and Theory in Deep Learning (BGPT)) π π [paper] [arxiv] [code]
π Spotlight @ ICML 2024 (Top 3.5% among total submissions)
π₯ Baekrok Shin*, Junsoo Oh*, Hanseul Cho, and Chulhee Yun
π° NeurIPS 2024 (Short version at ICML 2024 Workshop on Advancing Neural Network Training (WANT)) π π [paper] [arxiv] [code]
π₯ Hanseul Cho*, Jaeyoung Cha*, Pranjal Awasthi, Srinadh Bhojanapalli, Anupam Gupta, and Chulhee Yun
π° NeurIPS 2024 (Short version at ICML 2024 Workshop on Long-Context Foundation Models (LCFM)) π π [paper] [arxiv] [code]
π₯ Hanseul Cho and Chulhee Yun
π° ICLR 2023 π π [paper] [arxiv]
π NAVER Outstanding Theory Paper Award at the 7th Joint Conference of Korea Artificial Intelligence Association (JKAIA 2022)
π₯ Hyunji Jung*, Hanseul Cho*, and Chulhee Yun
π° Under Review π
π Best Paper Award at the 11th Joint Conference of Korean Artificial Intelligence Association (JKAIA 2024)
π₯ Hojoon Lee*, Hanseul Cho*, Hyunseung Kim*, Daehoon Gwak, Joonkee Kim, Jaegul Choo, Se-Young Yun, and Chulhee Yun
π° NeurIPS 2023 π π [paper] [arxiv] [code]
π₯ Junghyun Lee*, Hanseul Cho*, Se-Young Yun, and Chulhee Yun
π° NeurIPS 2023 π π [paper] [arxiv] [code]
π₯ Baekrok Shin*, Junsoo Oh*, Hanseul Cho, and Chulhee Yun
π° NeurIPS 2024 (Short version at ICML 2024 Workshop on Advancing Neural Network Training (WANT)) π π [paper] [arxiv] [code]
π₯ Hanseul Cho*, Jaeyoung Cha*, Pranjal Awasthi, Srinadh Bhojanapalli, Anupam Gupta, and Chulhee Yun
π° NeurIPS 2024 (Short version at ICML 2024 Workshop on Long-Context Foundation Models (LCFM)) π π [paper] [arxiv] [code]