MMLab@NTU Profile
MMLab@NTU

@MMLabNTU

Followers
2K
Following
213
Media
12
Statuses
79

Multimedia Laboratory @NTUsg, affiliated with S-Lab. Large Multimodal Models, Computer Vision, Image Processing, Computer Graphics, Deep Learning

Singapore
Joined May 2021
Don't wanna be here? Send us removal request.
@MMLabNTU
MMLab@NTU
6 months
Congratulations to Ziqi and Ziwei! Grateful for the opportunity to work with so many gifted students at @MMLabNTU. Their passion and creativity continue to inspire us! Their achievements are listed here: https://t.co/GMvhTMUl09
@NTUsg
NTU Singapore
6 months
Freshly picked: #NTUsg PhD student Huang Ziqi has been selected as one of 21 global recipients of the prestigious 2025 Apple Scholars in AIML PhD Fellowship โ€” a prestigious programme that supports emerging leaders in AI and machine learning through funding, mentorship, and
0
4
17
@MMLabNTU
MMLab@NTU
1 month
RT @ccloy: Congrats to Yuekun @YuekunDai and @ziangcao_ , both from @MMLabNTU , for winning the prestigious Google PhD Fellowship! Yuekunโ€ฆ
0
1
0
@KangLiao929
Kang Liao
1 month
Introducing ๐“๐ก๐ข๐ง๐ค๐ข๐ง๐  ๐ฐ๐ข๐ญ๐ก ๐‚๐š๐ฆ๐ž๐ซ๐š๐Ÿ“ธ, a unified multimodal model that integrates camera-centric spatial intelligence to interpret and create scenes from arbitrary viewpoints. Project Page: https://t.co/KxwIpuDUBg Code: https://t.co/DO52LFyL9m
14
33
147
@_akhaliq
AK
1 month
Thinking with Camera A Unified Multimodal Model for Camera-Centric Understanding and Generation
3
33
238
@ShangchenZhou
Shangchen Zhou
2 months
๐Ÿ“ธJoin us at #ICCV2025 for the Mobile Intelligent Photography & Imaging (MIPI) Workshop! โœจLeading keynotes: Profs. @songhan_mit, Michal Irani, Boxin Shi, and @MingHsuanYang - on intelligent photography and efficient GenAI. ๐Ÿ—“Oct 20, 8:50amโ€“12:30pm HST ๐Ÿ”— https://t.co/CqdCqzdsY1
1
10
28
@MMLabNTU
MMLab@NTU
2 months
Congratulations to @liuziwei7 of @MMLabNTU , recipient of the Young Scientist Award, recognised for his impactful contributions to computer vision and generative AI. ๐ŸŽ‰๐ŸŽ‰
@NTUsg
NTU Singapore
2 months
๐Ÿ† Congrats to #NTUsg Prof Ng Geok Ing on the ๐Ÿ‡ธ๐Ÿ‡ฌ Presidentโ€™s Technology Award 2025. A pioneer in Gallium Nitride (#GaN) โ€“ found in fast chargers, EVs, satellites & defence โ€“ he built ๐Ÿ‡ธ๐Ÿ‡ฌโ€™s global standing in this field and led the creation of the national GaN centre. ๐Ÿ‘ We also
2
1
29
@MMLabNTU
MMLab@NTU
2 months
@liuziwei7
Ziwei Liu
2 months
#ICCV2025 Congrats to Weichen ( https://t.co/3EHLciKwgP) and Mutian ( https://t.co/eL1sdcXIuo) being selected as the outstanding reviewers @ICCVConference https://t.co/dKiLhYJjWG
0
1
5
@shulin_tian
Shulin Tian
5 months
๐ŸŽฅ Video is already a tough modality for reasoning. Egocentric video? Even tougher! It is longer, messier, and harder. ๐Ÿ’ก How do we tackle these extremely long, information-dense sequences without exhausting GPU memory or hitting API limits? We introduce ๐Ÿ‘“Ego-R1: A framework
7
9
37
@ziqi_huang_
Ziqi Huang
6 months
๐ŸŽฌ ๐—–๐—ฉ๐—ฃ๐—ฅ ๐Ÿฎ๐Ÿฌ๐Ÿฎ๐Ÿฑ ๐—ง๐˜‚๐˜๐—ผ๐—ฟ๐—ถ๐—ฎ๐—น ๐™๐™ง๐™ค๐™ข ๐™‘๐™ž๐™™๐™š๐™ค ๐™‚๐™š๐™ฃ๐™š๐™ง๐™–๐™ฉ๐™ž๐™ค๐™ฃ ๐™ฉ๐™ค ๐™’๐™ค๐™ง๐™ก๐™™ ๐™ˆ๐™ค๐™™๐™š๐™ก ๐Ÿš€ Hosted by MMLab@NTU ร— Kuaishou, etc ๐Ÿ“… June 11 | Nashville ๐Ÿ”— https://t.co/YcQ6pb30R0 ๐Ÿง  Video is just the start. World modeling is the goal. #CVPR2025 #WorldModel
1
28
138
@_akhaliq
AK
7 months
Aero-1-Audio is out on Hugging Face Trained in <24h on just 16ร—H100 Handles 15+ min audio seamlessly Outperforms bigger models like Whisper, Qwen-2-Audio & commercial services from ElevenLabs/Scribe
8
64
415
@WuSize
Size Wu
8 months
๐Ÿ”ฅ We release Harmon: a unified framework for multimodal understanding & generation with a shared visual encoder (vs. decoupled Janus/-Pro). ๐Ÿ’ฅ SOTA on GenEval, MJHQ, WISE ๐Ÿง  Strong understanding performance ๐Ÿ“„ Paper: https://t.co/RFhEl9NEN7 ๐Ÿ”— Code:
Tweet card summary image
github.com
[ICCV2025]Code Release of Harmonizing Visual Representations for Unified Multimodal Understanding and Generation - wusize/Harmon
2
1
18
@ccloy
Chen Change Loy
8 months
๐Ÿš€ Meet Harmon โ€“ a unified model for both image generation and understanding! Trained with a shared masked autoregressive encoder, it sets new benchmarks on GenEval & MJHQ30K. ๐Ÿ–ผ๏ธ๐Ÿ’ฌ Try the live demo now on Hugging Face: ๐Ÿ‘‰ https://t.co/PX7OkVaZbx Paper:
Tweet card summary image
huggingface.co
@WuSize
Size Wu
8 months
๐Ÿ”ฅ We release Harmon: a unified framework for multimodal understanding & generation with a shared visual encoder (vs. decoupled Janus/-Pro). ๐Ÿ’ฅ SOTA on GenEval, MJHQ, WISE ๐Ÿง  Strong understanding performance ๐Ÿ“„ Paper: https://t.co/RFhEl9NEN7 ๐Ÿ”— Code:
0
2
15
@ccloy
Chen Change Loy
1 year
We turned our method, rejected by CVPR and ECCV, into the iOS app "Cutcha". EdgeSAM, our fast Segment Anything Model, runs at over 30 FPS on an iPhone 14. Enjoy intuitive one-touch object selection and precise editingโ€”all processed locally on your device. No cloud needed!
8
26
215
@ccloy
Chen Change Loy
1 year
๐Ÿ“ธ๐ŸŒŸ Attention all photography and imaging enthusiasts! Join us at the Third MIPI Workshop at #CVPR2024! ๐Ÿ“ Location: Arch 213 โฐ Time: 08:30 AM - 12:10 PM ๐ŸŒ Website: https://t.co/3x06T1AvaF Don't miss out on an exciting lineup of speakers: ๐Ÿ”น Lei Zhang: How Far Are We From
2
9
31
@TheAITalksOrg
The AI Talks
2 years
The Upcoming AI talk: ๐ŸŒ‹LLaVA๐Ÿฆ™ A Vision-and-Language Approach to Computer Vision in the Wild by Chunyuan Li @ChunyuanLi More info: https://t.co/ap7S1osxAm Subscribe us: https://t.co/m7NoJNciLe
0
14
33
@XingangP
Xingang Pan
2 years
(1/2) We are actively seeking PhD candidates from various countries to foster diversity in our research group at Nanyang Technological University. Know someone interested in a PhD with us? Please refer them to our team. Thanks for supporting diversity in academia! ๐ŸŒ๐ŸŽ“
2
17
86
@SenseTime_AI
SenseTime
2 years
8
22
124
@ccloy
Chen Change Loy
2 years
๐Ÿ”ฌ Our study introduces "Upscale-A-Video," a text-guided latent diffusion framework for video upscaling. It ensures temporal coherence locally & globally, balancing fidelity and quality. ๐Ÿš€ Project page: https://t.co/3UkaXXyMCC ๐Ÿ’ป GitHub: https://t.co/irQRuPHxED ๐ŸŽฅ Video:
11
51
293
@MMLabNTU
MMLab@NTU
2 years
EdgeSAM - Prompt-In-the-Loop Distillation for On-Device Deployment of SAM ๐Ÿ”— Project page: https://t.co/ydz7rZ78sS ๐Ÿ”— GitHub: https://t.co/VJS0YQHJI0 ๐Ÿค— Hugging Face:
Tweet card summary image
huggingface.co
@ccloy
Chen Change Loy
2 years
๐Ÿš€ Excited to share our latest work: "EdgeSAM - Prompt-In-the-Loop Distillation for On-Device Deployment of SAM" Supercharged SAM for Edge Devices! ๐ŸŒŸ #EdgeSAM is a faster, optimized version of SAM, now tailored for edge devices. We've reimagined SAM's ViT-based image encoder
0
0
8
@MMLabNTU
MMLab@NTU
2 years
EdgeSAM - Prompt-In-the-Loop Distillation for On-Device Deployment of SAM ๐Ÿ”— Project page: https://t.co/ydz7rZ78sS ๐Ÿ”— GitHub: https://t.co/VJS0YQHJI0 ๐Ÿค— Hugging Face:
Tweet card summary image
huggingface.co
0
0
0