Xiaofeng ZhangPh.D. Student, Shanghai Jiao Tong University
Key Laboratory of System Control and Information Processing
[Google Scholar]
[GitHub]
|
|
I am currently a third-year Ph.D. student at Shanghai Jiao Tong University. Prior to that, I received my Master's degree from Nanjing University of Posts and Telecommunications. My current research focuses on large vision-language models, including hallucination mitigation, position encoding optimization, and efficient multimodal reasoning.
![]() |
Hallucination Begins Where Saliency Drops ICLR 2026 Oral 🏆
[OpenReview]
[Code]
|
![]() |
ICLR 2026, corresponding author |
![]() |
Information Processing & Management (中科院一区), CCF B |
![]() |
Seeing Clearly by Layer Two: Enhancing Attention Heads to Alleviate Hallucination in LVLMs EMNLP 2025 Oral 🏆 |
![]() |
From Redundancy to Relevance: Information Flow in LVLMs Across Reasoning Tasks NAACL 2025 Oral 🏆 |
![]() |
Enhancing Multimodal Large Language Models Complex Reasoning via Similarity Computation AAAI 2025
[Code]
|
![]() |
SimIgnore: Enhancing Multimodal Large Language Models Complex Reasoning via Similarity Computation Neural Networks (中科院一区)
[Code]
|
![]() |
Wakeup-Darkness: When Multimodal Meets Unsupervised Low-light Image Enhancement ACM Transactions on Multimedia Computing, Communications, 2025
[IEEE Xplore]
[Code]
|
![]() |
Memory Augment is All You Need for Image Restoration IEEE Transactions on Consumer Electronics (中科院二区), 2026 |
![]() |
AAAI 2026, corresponding author
[Arxiv]
|
![]() |
D3ToM: Decider-Guided Dynamic Token Merging for Accelerating Diffusion MLLMs AAAI 2026, project leader |
![]() |
MCA-LLaVA: Manhattan Causal Attention for Reducing Hallucination in Large Vision-Language Models ACM MM 2025 (corresponding author / project leader) |
![]() |
DOPRA: Decoding Over-accumulation Penalization and Re-allocation in Specific Weighting Layer ACM MM 2024 🏆 (corresponding author / project leader)
[arXiv]
|
Invited Reviewer for: