
Chen Change Loy
@ccloy
Followers
3K
Following
2K
Media
122
Statuses
921
President's Chair Professor @NTUsg Director of @MMLabNTU Computer vision and deep learning
Singapore
Joined November 2010
RT @soumikRakshit96: ✨ CVPR 2025 highlight #2 --- EdgeTAM: On-Device Track Anything Model. EdgeTAM achieves significant speed-ups of 16 fps….
0
54
0
RT @taziku_co: もはや「そこにあった形跡すら残さない」. Object Clearは、ただモノを消すだけではない。床に落ちる影、ガラスへの映り込み、光の反射…それら全てが一緒に処理される。. 昔Photoshopで頑張っていたことをふと思い出す。. https://….
0
5
0
One-step diffusion-based video restoration, SeedVR2, by @Iceclearwjy - you can more clearly observe the differences in the comparisons displayed on the project page
🔥Introducing #SeedVR2, the latest one-step diffusion transformer version of #SeedVR for real-world image and video restoration!. details .- Paper: - Project: - Code (under review):
0
0
6
Congratulations to Ziqi and Ziwei! Grateful for the opportunity to work with so many gifted students at @MMLabNTU. Their passion and creativity continue to inspire us!.
Freshly picked: #NTUsg PhD student Huang Ziqi has been selected as one of 21 global recipients of the prestigious 2025 Apple Scholars in AIML PhD Fellowship — a prestigious programme that supports emerging leaders in AI and machine learning through funding, mentorship, and
3
7
79
RT @BoLi68567011: I’ve worked on evals and once considered turning it into a business. But here’s my honest take o….
0
5
0
RT @Teslanaut: Oh man. So many amazing tools out there. I really gotta start making a bigger portfolio of what I’ve been learning. Using….
0
52
0
🚀 Excited to share the lates work by my student @ChongZhou7 during his internship at @Meta , EdgeTAM: an on-device ‘track anything’ model that brings SAM 2’s video segmentation power to mobile devices. It runs at 16 FPS on iPhone 15 Pro Max, delivering state-of-the-art accuracy.
Apache 2.0 license🔥 On-device deployment ready.Extends SAM2 for tracking objects in videos 🔥 Click-to-segment support. EdgeTAM by Meta !
2
1
23
Aero-1-Audio is a compact audio model adept at various audio tasks, including speech recognition, audio understanding, and following audio instructions. It is part of the Aero-1 series, the first generation of lightweight multimodal models developed by LMMs-Lab, with future.
Aero-1-Audio is out on Hugging Face. Trained in <24h on just 16×H100.Handles 15+ min audio seamlessly .Outperforms bigger models like Whisper, Qwen-2-Audio & commercial services from ElevenLabs/Scribe
1
8
33
RT @KangLiao929: 📢Excited to present our Denoising as Adaptation paper at #ICLR2025:.⏰3:00-5:30PM | Fri, Apr 25.📌Hall 3 + Hall 2B #176.📎 ht….
0
1
0
🚀 Meet Harmon – a unified model for both image generation and understanding!.Trained with a shared masked autoregressive encoder, it sets new benchmarks on GenEval & MJHQ30K. 🖼️💬 Try the live demo now on Hugging Face:.👉 Paper:
🔥 We release Harmon: a unified framework for multimodal understanding & generation with a shared visual encoder (vs. decoupled Janus/-Pro). 💥 SOTA on GenEval, MJHQ, WISE.🧠 Strong understanding performance.📄 Paper: 🔗 Code:
0
1
14
RT @TheYihangLuo: 💥 Consistent Multi-View Diffusion for 3D Enhancement 💥. Introducing our work #3DEnhancer @CVPR: a multi-view diffusion mo….
0
9
0
RT @liuziwei7: 🔥Foundation Models for 3D/4D Motion Capture🔥. We present 📸SMPLest-X📸, the ultimate scaling law for expressive human pose and….
0
50
0
RT @ShangchenZhou: 🔥Introducing #𝐌𝐚𝐭𝐀𝐧𝐲𝐨𝐧𝐞 for human video matting!🔥. 🤡 Fast video matting with customizable target.🤡 Stable human tracking….
0
97
0
RT @KangLiao929: Happy to share that our work "Denoising as Adaptation" has been accepted to #ICLR2025! Huge thanks to @ccloy and all colla….
0
15
0
RT @BoLi68567011: I think the image metaphor of DeepSeek is incorrect. Recently, there has been a lot of discussion about DeepSeek. Many p….
0
20
0
RT @liuziwei7: 🔥Unbounded 4D City Generation🔥. #CityDreamer4D is a generative model of unbounded 4D cities that decouples static and dynami….
0
62
0