Hai-Long Sun @ LAMDA, NJU-AI

Hai-Long Sun  Hai-Long Sun    Hai-Long Sun

M.Sc. Student, LAMDA Group
School of Artificial Intelligence
National Key Laboratory for Novel Software Technology
Nanjing University, Nanjing 210023, China

Supervisor: Han-Jia Ye(叶翰嘉)

Laboratory: Yifu Building, Xianlin Campus of Nanjing University

Email: sunhl@lamda.nju.edu.cn

Biography

Currently I'm a graduate student of School of Artificial Intelligence in Nanjing University and a member of LAMDA Group, which is led by Prof. Zhi-Hua Zhou.

I received my B.Sc. degree from College of Computer Science & Technology, Nanjing University of Aeronautics and Astronautics in June 2023 (GPA ranked 1 / 120). In the same year, I was admitted to study for a M.Sc. degree in Nanjing University, under the supervision of Assistant Researcher Han-Jia Ye without entrance examination.

Research Interests

My research interests include Machine Learning and Data Mining. Currently, I focus on Pre-trained Model-Based Class-Incremental Learning and Continual Learning with Multimodal Large Language Models.

Publications - Preprints

WSFG 
  • Hai-Long Sun, Da-Wei Zhou, Han-Jia Ye, De-Chuan Zhan. PILOT: A Pre-Trained Model-Based Continual Learning Toolbox. arXiv:2309.07117, 2023. [Paper] [Code] [中文解读]

  • We introduces PILOT, a pre-trained model-based continual learning toolbox. On the one hand, PILOT implements some state-of-the-art class-incremental learning algorithms based on pre-trained models, such as L2P, DualPrompt, and CODA-Prompt. On the other hand, PILOT also fits typical class-incremental learning algorithms (e.g., DER, FOSTER, and MEMO) within the context of pre-trained models to evaluate their effectiveness.

WSFG 
  • Da-Wei Zhou, Hai-Long Sun, Han-Jia Ye, De-Chuan Zhan. Continual Learning with Pre-Trained Models: A Survey. arXiv:2401.16386, 2024. [Paper] [Code] [中文解读]

  • We presents a comprehensive survey of the latest advancements in PTM-based CL. We categorize existing methodologies into three distinct groups, providing a comparative analysis of their similarities, differences, and respective advantages and disadvantages. Additionally, we offer an empirical study contrasting various state-of-the-art methods to highlight concerns regarding fairness in comparisons.

Publications - Conference Papers

WSFG 
  • Da-Wei Zhou, Hai-Long Sun, Han-Jia Ye, De-Chuan Zhan. Expandable Subspace Ensemble for Pre-Trained Model-Based Class-Incremental Learning. The IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR'24). [Paper] [Code]

  • We propose Expandable Subspace Ensemble (EASE) for PTM-based CIL. To enable model updating without conflict, we train a distinct lightweight adapter module for each new task, aiming to create task-specific subspaces. These adapters span a high-dimensional feature space, enabling joint decision-making across multiple subspaces. As data evolves, the expanding subspaces render the old class classifiers incompatible with new-stage spaces. Correspondingly, we design a semantic-guided prototype complement strategy that synthesizes old classes’ new features without using any old class instance.

Awards & Honors & Contests

Teaching Assistant

Service

Correspondence

Email: sunhl@lamda.nju.edu.cn

Address: Hai-Long Sun, National Key Laboratory for Novel Software Technology, Nanjing University Xianlin Campus Mailbox 603, 163 Xianlin Avenue, Qixia District, Nanjing 210023, China.

(南京市栖霞区仙林大道163号, 南京大学仙林校区603信箱, 软件新技术国家重点实验室, 210023.)