Gengyuan Zhang

portrait2.jpeg

pronoun: he/him

*Currently also looking for Internships ;) Please contact me if you find a fit*

Hi, I am Gengyuan(张耕源). I am currently pursuing my PhD degree at Ludwig-Maximilian University (aka LMU Munich/University of Munich), supervised by Prof. Volker Tresp.

My research interests include Video Understanding and Multimodal Reasoning as an intersection of Computer Vision and Natural Language Processing.

Prior to this, I attained my bachelor degree (2018) in Zhejiang University, China and my master degree (2021) in Technical University of Munich, Germany.

Originally, I am from Hunan, China.







news

Oct 28, 2024 One new paper is accepted by WACV 2025, Tuscon, Arizona!
Jun 17, 2024 Our new paper is now on arXiv Localizing Events in Videos with Multimodal Queries!
Nov 27, 2023 1 new paper is now on arXiv: SPOT! Revisiting Video-Language Models for Event Understanding!
Oct 24, 2023 1 Paper accepted at WACV 2024!
Jul 1, 2023 1 Paper accepted at ICCV 2023!


selected publications

  1. Localizing Events in Videos with Multimodal Queries
    Gengyuan Zhang, Mang Ling Ada Fok, Yan Xia, and 5 more authors
    arXiv preprint arXiv:2406.10079, 2024
  2. Perceive, Query & Reason: Enhancing Video QA with Question-Guided Temporal Queries
    Roberto Amoroso*, Gengyuan Zhang*, Rajat Koner, and 4 more authors
    In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV) 2025, 2025
  3. Multi-event Video-Text Retrieval
    Gengyuan Zhang, Jisen Ren, Jindong Gu, and 1 more author
    In Proceedings of the IEEE/CVF International Conference on Computer Vision, 2023
  4. Time-dependent Entity Embedding is not All You Need: A Re-evaluation of Temporal Knowledge Graph Completion Models under a Unified Framework
    Zhen Han*, Gengyuan Zhang*, Yunpu Ma, and 1 more author
    In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, Nov 2021
  5. Can Vision-Language Models be a Good Guesser? Exploring VLMs for Times and Location Reasoning
    Gengyuan Zhang, Yurui Zhang, Kerui Zhang, and 1 more author
    arXiv preprint arXiv:2307.06166, Nov 2023