Skip to content
View zhousheng97's full-sized avatar
🐢
Focusing
🐢
Focusing

Block or report zhousheng97

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
zhousheng97/README.md

Hi there 👋

  • 👩 I’m Sheng, a PhD student from China, currently studying as a visiting student at the National University of Singapore.
  • 🧐 My focus is multimodal learning, especially VQA, and I’m currently exploring multimodal LLMs.
  • 💬 As an ENFJ-A, I thrive on meaningful collaboration and communication.
  • 📫 You can reach me at hzgn97@gmail.com—let’s connect!

Pinned Loading

  1. EgoTextVQA EgoTextVQA Public

    [CVPR'25] 🌟🌟 EgoTextVQA: Towards Egocentric Scene-Text Aware Video Question Answering

    Python 36 1

  2. ViTXT-GQA ViTXT-GQA Public

    [IEEE TMM'25] Scene-Text Grounding for Text-Based Video Question Answering

    Python 16 1

  3. Awesome-MLLM-TextVQA Awesome-MLLM-TextVQA Public

    ✨✨Latest Research on Multimodal Large Language Models on Scene-Text VQA Tasks

    9

  4. GPIN GPIN Public

    [ACM TOMM'24] Graph Pooling Inference Network for Text-based VQA

    Python 3

  5. SSGN SSGN Public

    [IEEE TIP'23] Exploring Sparse Spatial Relation in Graph Inference for Text-Based VQA

    Python 4