news

Nov 08, 2025 πŸ”” Exciting News! We’re thrilled to share that our two recent works have been accepted to AAAI 2026 β€” one as an oral and the other as a poster presentation! πŸŽ‰ i. Multi-Mood β€” a multi-modal large language model that integrates video, audio, and text with psychological criteria through reinforcement learning to enable trustworthy and emotionally aligned responses. ii. LIBERO-Mem β€” a non-Markovian task suite for short- and long-horizon object tracking and manipulation, featuring temporally sequenced subgoals that challenge models to reason beyond the current observation. πŸ“„ Codes will be released soon πŸŽ‰ β€” stay tuned!
Sep 26, 2025 πŸ”” Excited to share that our works on (i) ExGra-Med β€” a data-efficient multimodal large language model (LLM) for healthcare; (ii) Token Redundancy in 3D Point Cloud Transformers β€” uncovering how existing 3D transformers (e.g., Ptv-3, Sonata) are over-tokenized, and proposing an efficient token merging strategy that reduces computation by up to 90-95% while preserving accuracy; and (iii) Over-Optimization in RLHF for LLM Post-Training β€” exploring how reinforcement learning from human feedback can lead to alignment instability and proposing new insights into optimization LLM post-training have been accepted to NeurIPS 2025 πŸŽ‰. Excited to present and discuss them at San Diego πŸš€
Sep 09, 2025 🌟 Excited to give a talk about my current research on Scaling Multi-Modal Learning: Hybrid Representations and Efficient Adaptation at Machine Learning Lab, School of Information and Communications Technology (SOICT), Hanoi University of Science and Technology, Vietnam and (ii) School of Computing, National University of Singapore (NUS).
Sep 02, 2025 :bell: The MGPath has been accepted to the Transactions on Machine Learning Research. Congratulations to all co-authors on this milestone!
May 01, 2025 πŸŽ‰ Our first (i) preliminary version, MGPath has been accepted to the Workshop on Foundation Models in the Wild, ICLR 2025 and (ii) another one about LLaMA-Adapter’s prompt learning is accepted at ICML 2025.
Apr 20, 2025 πŸŽ‰ Our work in building a new Inductive Message Passing Network for Efficient Human-in-the-Loop Annotation of Mobile Eye Tracking Data has been accepted at Scientific Report, Nature Portfolio.
Feb 20, 2025 :bell: Excited to share our latest work! πŸŽ‰: (i) On Zero-Initialized Attention: Optimal Prompt and Gating Factor Estimation – We introduce a Mixture of Experts (MoE) perspective to explain the mechanism behind LLaMA-Adapter’s prompt learning. (ii) MGPath – A novel multi-granular prompt learning method for few-shot WSI pathology prediction, leveraging the power of foundation vision-language models.
Oct 08, 2024 πŸ‡¨πŸ‡­ Start my visiting research at ETH AI Center, ETH Zurich. The topics are about Multi-Modal LLMs for Healthcare empowered by Retrieval-Augmented Generation.
Oct 07, 2024 :bell: Excited to introduce our latest work on medical multi-modal LLMs: LoGra-Med, a novel pre-training algorithm that incorporates multi-graph alignment to effectively address the data-hungry nature of autoregressive learning.
Oct 06, 2024 :rocket: The paper PiToMe has been accepted at NeurIPS 2024. Our code will be available soon!
Jun 10, 2024 :bell: Our new preprint PiToMe is online. We propose a new method to do token merging in the Transformer with spectrum-preserving.
May 01, 2024 :rocket: A paper submitted to ICML 2024 on the molecular conformer aggregation network topic is accepted.
Jan 15, 2024 :rocket: A paper submitted to ICLR 2024 on the topic of accelerating transformers is accepted as an oral talk.
Sep 22, 2023 :rocket: A paper submitted to NeurIPS 2023 on a large-scale medical image pre-trained models using second-order graph matching is accepted.