- 👩 I’m Sheng, a PhD student from China, currently studying as a visiting student at the National University of Singapore.
- 🧐 My focus is multimodal learning, especially VQA, and I’m currently exploring multimodal LLMs.
- 💬 As an ENFJ-A, I thrive on meaningful collaboration and communication.
- 📫 You can reach me at hzgn97@gmail.com—let’s connect!
🐢
    Focusing
    Vision-Language.
- 
                  Hefei University of Technology
- Singapore
- 
        
  07:44
  (UTC +08:00) 
- https://zhousheng97.github.io/
Pinned Loading
- 
  EgoTextVQAEgoTextVQA Public[CVPR'25] 🌟🌟 EgoTextVQA: Towards Egocentric Scene-Text Aware Video Question Answering 
- 
  Awesome-MLLM-TextVQAAwesome-MLLM-TextVQA Public✨✨Latest Research on Multimodal Large Language Models on Scene-Text VQA Tasks 
          Something went wrong, please refresh the page to try again.
If the problem persists, check the GitHub status page or contact support.
  If the problem persists, check the GitHub status page or contact support.
