andrewliao11 Profile Banner
Andrew Liao Profile
Andrew Liao

@andrewliao11

Followers
293
Following
2K
Media
49
Statuses
491

****Seeking research roles in AI/CV for 2025 Final-yr Ph.D. at CS @UofT @VectorInst πŸ‡¨πŸ‡¦. I make dataset creation less painful. Prev. @nvidia @amazon intern

πŸ‡ΉπŸ‡ΌπŸ‡¨πŸ‡¦
Joined January 2016
Don't wanna be here? Send us removal request.
@andrewliao11
Andrew Liao
3 months
πŸš€ New work: LongPerceptualThoughts. We introduce a synthetic data pipeline to fine-tune VLMs with Long Chain-of-thoughts. π†π¨πšπ₯: Help VLMs β€œthink longer” on vision tasks. 3 pts on 5 Vision tasks. 11 pts on V* Bench. 2 pts on MMLU-Pro (text-only). 🌐
1
5
17
@andrewliao11
Andrew Liao
8 days
πŸŽ‰ Great news! "LongPerceptualThoughts" accepted at @COLMConf! . #montreal #COLM2025.
@andrewliao11
Andrew Liao
3 months
πŸš€ New work: LongPerceptualThoughts. We introduce a synthetic data pipeline to fine-tune VLMs with Long Chain-of-thoughts. π†π¨πšπ₯: Help VLMs β€œthink longer” on vision tasks. 3 pts on 5 Vision tasks. 11 pts on V* Bench. 2 pts on MMLU-Pro (text-only). 🌐
0
3
12
@andrewliao11
Andrew Liao
29 days
RT @MingyuanWu4: Research with amazing collaborators @JizeJiang, @MeitangLi, and @JingchengYang, guided by great advisors and supported by….
0
14
0
@andrewliao11
Andrew Liao
1 month
Appreciate the great crowd despite locating at the very end of the boardroom πŸ€—
Tweet media one
0
0
3
@andrewliao11
Andrew Liao
1 month
RT @BaldassarreFe: DINOv2 meets text at #CVPR 2025! Why choose between high-quality DINO features and CLIP-style vision-language alignment?….
0
103
0
@andrewliao11
Andrew Liao
1 month
RT @anneouyang: ✨ New blog post πŸ‘€: We have some very fast AI-generated kernels generated with a simple test-time only search. They are perf….
0
135
0
@andrewliao11
Andrew Liao
1 month
Get ready for an exciting morning at CVPR 2025! Our poster session is TODAY.Time: 10:30 a.m. to 12:30 p.m. CDT.Location: ExHall D, Poster #385. Come by to dive into our latest on system-2 thinking in Vision-Language Models! Let's connect and chat! #CVPR2025 #Nashville #VLMs.
@andrewliao11
Andrew Liao
1 month
Excited to share our CVPR paper next week in Nashville 🎢! Looking forward to connecting with old/new friends. Also, I'm on the job market NOW. Let's discuss system-2 thinking in VLMs! πŸ€”πŸ€“πŸ’‘.#CVPR2025 #Nashville #VLMs #reasoning.
1
0
7
@andrewliao11
Andrew Liao
1 month
RT @JunGao33210520: This year, we have 3 papers in CVPR, discussing the connection between 3D and video models:. GEN3C [Highlight] 3D groun….
0
13
0
@andrewliao11
Andrew Liao
1 month
RT @DonglaiXiang: 🚨Excited to announce the 1st Workshop on Vision Meets Physics at @CVPR2025!. Join us on June 12 for a full-day event expl….
0
14
0
@andrewliao11
Andrew Liao
1 month
Excited to share our CVPR paper next week in Nashville 🎢! Looking forward to connecting with old/new friends. Also, I'm on the job market NOW. Let's discuss system-2 thinking in VLMs! πŸ€”πŸ€“πŸ’‘.#CVPR2025 #Nashville #VLMs #reasoning.
@andrewliao11
Andrew Liao
1 year
πŸ‘‰ Vision-Language Models (VLMs) can answer tough questions, but when they make mistakes, can we give them feedback to help revise their answers?. Links:. w/ @rafidrmahmood @fidlersanja @davidjesusacu
0
4
26
@andrewliao11
Andrew Liao
1 month
RT @lschmidt3: Very excited to finally release our paper for OpenThoughts!. After DataComp and DCLM, this is the third large open dataset m….
0
213
0
@andrewliao11
Andrew Liao
1 month
RT @ShenzhiWang_THU: 🧐Two papers, opposite opinions. Ours: High-entropy tokens drive all performance gains in LLM RL. Another: Don’t let….
0
70
0
@andrewliao11
Andrew Liao
2 months
RT @YungSungChuang: 🚨Do passage rerankers really need explicit reasoning?πŸ€”β€”Maybe Not!. Our findings:.βš–οΈStandard rerankers outperform those….
0
17
0
@andrewliao11
Andrew Liao
2 months
Released the code!.Go generate your own Long Perceptual Thoughts. πŸ§‘β€πŸ’»Code:
@andrewliao11
Andrew Liao
3 months
πŸš€ New work: LongPerceptualThoughts. We introduce a synthetic data pipeline to fine-tune VLMs with Long Chain-of-thoughts. π†π¨πšπ₯: Help VLMs β€œthink longer” on vision tasks. 3 pts on 5 Vision tasks. 11 pts on V* Bench. 2 pts on MMLU-Pro (text-only). 🌐
0
1
2
@andrewliao11
Andrew Liao
2 months
When developing this project, it keeps reminding of DAgger (classic imitation learning algo). We first let the model to free generate the imperfect data. Once we detect something wrong, we hand it over to an expert model to fix errors.
@andrewliao11
Andrew Liao
3 months
πŸš€ New work: LongPerceptualThoughts. We introduce a synthetic data pipeline to fine-tune VLMs with Long Chain-of-thoughts. π†π¨πšπ₯: Help VLMs β€œthink longer” on vision tasks. 3 pts on 5 Vision tasks. 11 pts on V* Bench. 2 pts on MMLU-Pro (text-only). 🌐
0
0
3
@andrewliao11
Andrew Liao
2 months
RT @cindy_x_wu: Introducing COMPACT: COMPositional Atomic-to-complex Visual Capability Tuning, a data-efficient approach to improve multimo….
0
44
0
@andrewliao11
Andrew Liao
3 months
Thanks to my awesome collaborators @s_elflein @riverliuhe @lealtaixe @YejinChoinka @FidlerSanja @davidjesusacu πŸ”₯.
0
0
1
@andrewliao11
Andrew Liao
3 months
Takeaway:. Structured, reflective reasoning can be taught β€” even in perception. We show that generating better data can unlock stronger visual reasoning. 🌐Website: πŸ€—Dataset: πŸ“œPaper:
1
2
2
@andrewliao11
Andrew Liao
3 months
SOTA reasoning LLMs naturally present cognitive behaviors in reasoning traces. What about πΏπ‘œπ‘›π‘”π‘ƒπ‘’π‘Ÿπ‘π‘’π‘π‘‘π‘’π‘Žπ‘™π‘‡β„Žπ‘œπ‘’π‘”β„Žπ‘‘π‘ ?. Our 3-stage pipeline significantly injects these behaviors in vision-centric tasks! πŸ“ˆ
Tweet media one
1
0
1
@andrewliao11
Andrew Liao
3 months
Stage 3 – Think Harder:. We prompt a reasoning LLM (e.g., R1-Distill) to expand the short CoTs using cues like β€œπ‘Šπ‘Žπ‘–π‘‘,” or β€œπ»π‘šπ‘š,”. This introduces system-2 behaviors like verification, subgoal setting, and backtracking. See the analysis πŸ‘‡
1
0
1
@andrewliao11
Andrew Liao
3 months
Stage 2 – Think:. We use the VLM to generate short CoTs β€” often time shallow reasoning. Why?.Staying close to the VLM’s original output distribution avoids disrupting its learned behavior.
1
0
1