Min-Hung (Steve) Chen
@CMHungSteven
Followers
2K
Following
31K
Media
58
Statuses
752
Senior Research Scientist, NVR TW @NVIDIAAI @NVIDIA (Project Lead: DoRA, EoRA | Ph.D. @GeorgiaTech | Multimodal AI | https://t.co/dKaEzVoTfZ
Taipei City, Taiwan
Joined July 2011
(1/N) Are you looking for #Vision #Transformer papers in various areas? Check out this list of papers including a broad range of different tasks! https://t.co/kMThHeO7Gg Feel free to share with others๐ @Montreal_AI @machinelearnflx @hardmaru @ak92501 @arankomatsuzaki @omarsar0
github.com
An ultimately comprehensive paper list of Vision Transformer/Attention, including papers, codes, and related websites - cmhungsteve/Awesome-Transformer-Attention
3
31
167
๐ฌ ICCV'25 just wrapped and we're rolling into EMNLP'25! Our paper "MovieCORE: Cognitive Reasoning in Movies" was accepted as an Oral presentation ๐ #EMNLP2025 #NLP #ComputerVision #VideoUnderstanding #VisionLanguageModels #AI #MachineLearning #DeepLearning #VLM #LLM
1
1
6
We, at NVIDIA, presents - Length Penalty Done Right - Cut CoT length by 3/4 without sacrificing accuracy using only RL - This makes DeepSeek-R1-7B running ~8 times faster on AIME-24 while maintaining the same accuracy.
8
29
245
Welcome to the workshop at ICCV. In the afternoon session, I will give a talk of our effort towards learning physical AI for sidewalk autonomy.
๐ฃ Join us for the ICCVโ25 X-Sense Workshop at Hawai'i Convention Center @ Room 323C on Monday, Oct. 20!! Link: https://t.co/FV7wCU92sY
0
2
18
๐ฃ Join us for the ICCVโ25 X-Sense Workshop at Hawai'i Convention Center @ Room 323C on Monday, Oct. 20!! Link: https://t.co/FV7wCU92sY
6
6
18
If you're at #ICCV2025, Hawaii, make sure to drop by the X-Sense workshop at Hawai'i Convention Center @ Room 323C on Monday, Oct. 20. Join us for a discussion on the future of x-modal sensing! ๐ธ๐ Link:
2
2
6
#ICCV2025 is around the corner! Don't hesitate to visit @HsukuangChiu's V2V-GoT poster @ X-Sense Workshop to learn our latest LLM-based Cooperative Driving work! Workshop: https://t.co/WaIl0uuJij V2V-GoT: https://t.co/u9tFpmoxuB
@ICCVConference
#V2V #LLM #iccv25 #NVIDIA #CMU
Excited to have a poster presentation for our latest research V2V-GoT at #ICCV2025 X-Sense Workshop! ๐ Date & Time: Oct 20th, Monday, 11:40am ~ 12:30pm ๐ Location: Exhibition Hall II (No 188 ~ 210) ๐ Paper, code, and dataset: https://t.co/GTnWrShw80
#NVIDIA #CMU
0
0
9
Excited to have a poster presentation for our latest research V2V-GoT at #ICCV2025 X-Sense Workshop! ๐ Date & Time: Oct 20th, Monday, 11:40am ~ 12:30pm ๐ Location: Exhibition Hall II (No 188 ~ 210) ๐ Paper, code, and dataset: https://t.co/GTnWrShw80
#NVIDIA #CMU
0
2
4
#ICCV2025 is around the corner! Don't hesitate to visit @JosmyFaure1's HERMES poster to learn our latest efficient video understanding work! ๐ Website: https://t.co/LFQgh9mbfC
0
0
13
(1/6) I built KohakuHub โ a fully self-hosted HF alternative, with HF compatibility and a familiar experience ๐ง Host your own data, Keep your workflow Check more information in the repository and our community! ๐ https://t.co/7ZBOkDK5FJ ๐ฌ https://t.co/HVjx3Rg9XA
7
20
114
Tina proved that LoRA can match or surpass full-parameter RL. Tora builds directly on that result, turning it into a full framework. Built on torchtune, it extends RL post-training to LoRA, QLoRA, DoRA, and QDoRA under one interface with GRPO, FSDP, and compile support. QLoRA
Tina: Tiny Reasoning Models via LoRA LoRA-RL tuned 1.5B models on curated reasoning data, achieving +20% gains and 43% Pass@1 (AIME24) at $9 total cost. Outperforms full-parameter RL on DeepSeek-R1-Distill-Qwen-1.5B. - LoRA-based RL yields better performance with less compute.
4
29
302
[#hiring] I'm seeking PhD #Interns for 2026 at #NVIDIAResearch Taiwan! If interested, please send your CV and cover letter to minhungc [at] nvidia [dot] com ๐Research topics: Efficient Video/4D Understanding & Reasoning. ๐Location: Taiwan / Remote (mainly APAC) #internships
1
17
112
[#hiring] I'm seeking PhD #Interns for 2026 at #NVIDIAResearch Taiwan! If interested, please send your CV and cover letter to minhungc [at] nvidia [dot] com ๐Research topics: Efficient Video/4D Understanding & Reasoning. ๐Location: Taiwan / Remote (mainly APAC) #internships
1
17
112
[#EMNLP2025] Super excited to share MovieCORE @emnlpmeeting (Oral) โ New #VideoUnderstanding Benchmark on System-2 Reasoning! ๐Check the original post from @JosmyFaure1 for more details! ๐ท Project: https://t.co/pmR8WCunyW
#VLM #LLM #Video #multimodal #AI #NVIDIA #NTU #NTHU
๐ New Benchmark Alert! Our paper MovieCORE: COgnitive REasoning in Movies is accepted at #EMNLP2025 (Oral) ๐ Movies arenโt just โwhat happenedโ, theyโre why it happened, how characters feel, and what it means. MovieCORE tests Vision-Language Models on System-2 reasoning.
0
1
18
๐ New Benchmark Alert! Our paper MovieCORE: COgnitive REasoning in Movies is accepted at #EMNLP2025 (Oral) ๐ Movies arenโt just โwhat happenedโ, theyโre why it happened, how characters feel, and what it means. MovieCORE tests Vision-Language Models on System-2 reasoning.
1
1
9
๐ฃ Still Open for Submissions - X-Sense Workshop @ICCVConference! ๐
Deadline: September 8, 2025, 09:59 AM GMT ๐ Submission Portal: https://t.co/l2y89nLVx4 ๐ More info: https://t.co/3LRd386Bkm
#ICCV2025 #ICCV #ICCV25
#CFP #NYCU #Cornell #NVIDIA #USYD #MIT #UCSD #TUDelft #UCLA
openreview.net
Welcome to the OpenReview homepage for ICCV 2025 Workshop X-Sense
0
2
12
Thanks @_akhaliq for sharing our work! Check out more details:
0
8
46
We connect the autoregressive pipeline of LLMs with streaming video perception. Introducing AUSM: Autoregressive Universal Video Segmentation Model. A step toward unified, scalable video perception โ inspired by how LLMs unified NLP. ๐
arxiv.org
Recent video foundation models such as SAM2 excel at prompted video segmentation by treating masks as a general-purpose primitive. However, many real-world settings require unprompted segmentation...
2
28
142
Ego-exo sensing is definitely the future and has a lot of potentials. Join us and explore this together!
๐ฃ Call for Submissions - X-Sense Workshop #ICCV2025! We have extended the submission ddl!! Feel free to submit your accepted papers. Papers are โnon-archivedโ! Deadline: Sep. 8 09:59 AM GMT
0
4
15