
Stanford Vision and Learning Lab
@StanfordSVL
Followers
16K
Following
327
Media
13
Statuses
361
SVL is led by @drfeifei @silviocinguetta @jcniebles @jiajunwu_cs and works on machine learning, computer vision, robotics and language
Stanford, CA
Joined September 2014
RT @zhang_yunzhi: (1/n) Time to unify your favorite visual generative models, VLMs, and simulators for controllable visual generation—Intro….
0
65
0
RT @keshigeyan: 1/ Model architectures have been mostly treated as fixed post-training. 🌱 Introducing Grafting: A new way to edit pretrain….
0
44
0
RT @joycjhsu: We'll be presenting Deep Schema Grounding at @iclr_conf 🇸🇬 on Thursday (session 1 #98). Come chat about abstract visual conc….
0
2
0
RT @emilyzjin: State classification of objects and their relations (e.g. the cup is next to the plate) is core to many tasks like robot pla….
0
5
0
RT @Koven_Yu: 🔥Spatial intelligence requires world generation, and now we have the first comprehensive evaluation benchmark📏 for it! Introd….
0
90
0
RT @Koven_Yu: 🔥Want to capture 3D dancing fluids♨️🌫️🌪️💦? No specialized equipment, just one video! Introducing FluidNexus: Now you only nee….
0
74
0
RT @sunfanyun: Spatial reasoning is a major challenge for the foundation models today, even in simple tasks like arranging objects in 3D sp….
0
59
0
RT @YunfanJiang: 🤖 Ever wondered what robots need to truly help humans around the house?. 🏡 Introducing 𝗕𝗘𝗛𝗔𝗩𝗜𝗢𝗥 𝗥𝗼𝗯𝗼𝘁 𝗦𝘂𝗶𝘁𝗲 (𝗕𝗥𝗦)—a compre….
0
139
0
RT @YunfanJiang: 🚀Two weeks ago, we hosted a welcome party for the newest member of our Stanford Vision and Learning Lab—a new robot! 🤖✨Wat….
0
30
0
RT @Koven_Yu: 🤩Forget MoCap -- Let’s generate human interaction motions with *Real-world 3D scenes*!🏃🏞️ Introducing ZeroHSI: Zero-Shot 4D H….
0
61
0
RT @keshigeyan: 1/ [NeurIPS D&B] Introducing HourVideo: A benchmark for hour-long video-language understanding!🚀. 500 egocentric videos, 18….
0
52
0
RT @yunongliu1: 💫🪑Introducing IKEA Manuals at Work: The first multimodal dataset with extensive 4D groundings of assembly in internet video….
0
44
0
RT @ManlingLi_: [NeurIPS D&B Oral] Embodied Agent Interface: Benchmarking LLMs for Embodied Agents. A single line of code to evaluate your….
0
69
0
RT @sunfanyun: Training RL/robot policies requires extensive experience in the target environment, which is often difficult to obtain. How….
0
45
0
RT @RogerDai1217: Why hand-engineer digital twins when digital cousins are free?. Check out ACDC: Automated Creation of Digital Cousins 👭 f….
0
39
0
RT @Koven_Yu: 🔥Spatial intelligence needs fast, *interactive* 3D world generation 🎮 — introducing WonderWorld: generating 3D scenes interac….
0
229
0
RT @joycjhsu: What makes a maze look like a maze?. Humans can reason about infinitely many instantiations of mazes—made of candy canes, sti….
0
36
0