🎯
Focusing
PhD student @dvlab-research, CSE@CUHK, Multimodal Intelligence
-
The Chinese University of Hong Kong
- Hong Kong SAR
-
09:12
(UTC +08:00) - https://wcy1122.github.io/
- https://scholar.google.com/citations?user=1pZcoqgAAAAJ&hl=en
Pinned Loading
-
dvlab-research/MGM
dvlab-research/MGM PublicOfficial repo for "Mini-Gemini: Mining the Potential of Multi-modality Vision Language Models"
-
dvlab-research/LLaMA-VID
dvlab-research/LLaMA-VID PublicLLaMA-VID: An Image is Worth 2 Tokens in Large Language Models (ECCV 2024)
-
dvlab-research/MGM-Omni
dvlab-research/MGM-Omni PublicMGM-Omni: Scaling Omni LLMs to Personalized Long-Horizon Speech
-
dvlab-research/Lyra
dvlab-research/Lyra Public[ICCV 2025] Official Implementation for "Lyra: An Efficient and Speech-Centric Framework for Omni-Cognition"
Something went wrong, please refresh the page to try again.
If the problem persists, check the GitHub status page or contact support.
If the problem persists, check the GitHub status page or contact support.



