Feilong Tang 唐飞龙
Feilong Tang (Chinese: 唐飞龙) is a PhD student at the AIM for Health Lab at Monash University, advised by A/Prof. Zongyuan Ge. He received his Bachelor's degree from the University of Liverpool (2019 – 2023). He also had internship experience at Shanghai AI Lab, HKUST, MBZUAI, and DeepGlint.
His research focuses on video understanding and multimodal large language models (MLLMs). He is particularly interested in:
- Video Understanding and Temporal Reasoning in Multimodal LLMs
- Next-generation Vision Transformer (ViT) to address urgent needs in modern MLLMs
- Hallucination Mitigation in Multimodal Large Language Models
News
- 2026.04📰 Paper accepted to Nature Communications — Population-scale Characterization of the Oral Microbiome and Associations with Metabolic Health.
- 2026.04🎉 Two papers accepted to ACL 2026 Findings.
- 2026.02🚀 New preprint: OneVision-Encoder — Codec-Aligned Sparsity as a Foundational Principle for Multimodal Intelligence. arXiv | Code
- 2026.02🎉 Three papers accepted to CVPR 2026, including Thinking in Uncertainty as a Highlight.
- 2026.01🎉 Three papers accepted to ICLR 2026, including one Oral.
- 2025.12🏆 Co-authored paper Reconvla receives the Best Paper Award at AAAI 2026.
- 2025.09🎉 Three papers accepted to NeurIPS 2025 — UniViT (first author), Towards Dynamic 3D Reconstruction (co-first), Decoding Causal Structure (co-first).
- 2025.09🚀 New preprint: LLaVA-OneVision-1.5 — Fully Open Framework for Democratized Multimodal Training.
- 2025.07🎉 One paper accepted to ICCV 2025 — Hierarchical Retrieval-Augmented Learning (co-first author).
- 2025.06🎉 Two papers accepted to MICCAI 2025.
- 2025.05🎉 One paper accepted to ACL 2025 – MMRC benchmark (co-first author).
- 2025.05📰 Paper in Communications Medicine (Nature Portfolio) — Forecasting DR progression (co-first). IF 6.3, Q1.
- 2025.02🏆 Five papers accepted to CVPR 2025, including first-author paper Seeing Far and Clearly accepted as an Oral (top ~0.8%).
- 2025.01📄 Three papers accepted to AAAI 2025.
- 2025.01📄 One paper accepted to ICLR 2025 – Intervening Anchor Token.
- 2024.08🏅 SAM2-UNet received the Best Paper Award at ICCV 2024 Workshop.
- 2024.07📄 One paper accepted to ECCV 2024 – OphNet.
- 2024.03📄 One paper accepted to AAAI 2024 – SFC (co-first author).
- 2024.02📄 One paper accepted to CVPR 2024 – Hunting Attributes (first author).
Selected Publications
* equal contribution | † corresponding author | Bold = myself
- Seeing Far and Clearly: Mitigating Hallucinations in MLLMs with Attention Causal Decoding CVPR 2025 Oral (top ~0.8%) Paper Code
All Publications
* equal contribution | † corresponding author | Bold = myself
2026
- Seeing Through the Shift: Causality-Inspired Robust Generalized Category Discovery CVPR 2026
- Population-scale Characterization of the Oral Microbiome and Associations with Metabolic Health Nature Communications, 2026 Paper DOI
2025
- Seeing Far and Clearly: Mitigating Hallucinations in MLLMs with Attention Causal Decoding CVPR 2025 Oral (top ~0.8%) Paper Code
- Forecasting the diabetic retinopathy progression using GAN Communications Medicine (Nature Portfolio), 2025 Paper JCR 2024 IF 6.3 · Q1 · Rank 33/195 (Medicine, General & Internal)
2024
- SAM2-UNet: Segment Anything 2 Makes Strong Encoder for Natural and Medical Image Segmentation 🏆 ICCV Workshop 2024 Best Paper Award Paper Code
- Sight for sore heads: Using CNNs to diagnose migraines ARVO 2024
2023
- DuAT: Dual-aggregation Transformer Network for Medical Image Segmentation PRCV 2023 Paper Code 295+ citations
2022
Education
- PhD in Computer Science, Monash University, Melbourne, Australia (2024 – Present)
AIM for Health Lab · Supervised by A/Prof. Zongyuan Ge - Bachelor of Engineering (Computer Science), University of Liverpool, Liverpool, UK (2019 – 2023)
Internship & Research Experience
- Research Intern, Mohamed bin Zayed University of Artificial Intelligence (MBZUAI), Abu Dhabi, UAE (2024)
- Research Intern, Hong Kong University of Science and Technology (HKUST), Hong Kong (2024)
- Research Intern, Shanghai AI Laboratory (上海人工智能实验室), Shanghai, China (2025)
- Research Intern, DeepGlint (格灵深瞳), Beijing, China (2025)
Miscellaneous
- 📝 Reviewer: CVPR, ICCV, ECCV, NeurIPS, ICLR, AAAI, MICCAI
- 📊 Google Scholar: 1,300+ citations
- 🌏 Languages: Chinese (native), English (fluent)
© 2025 Feilong Tang · Last updated: Mar 2026 · Google Scholar