Intelligent Robotics and Vision Lab @ UTDallas
@IRVLUTD
Followers
234
Following
31
Media
12
Statuses
62
Welcome to the X page of the Intelligent Robotics and Vision Lab @UT_Dallas
ECSS 4.222, UT Dallas, TX
Joined October 2021
Many #robot_learning works use human videos but need lots of data/retraining. We present #HRT1 — a robot learns from just one human video and performs mobile manipulation tasks in new environments with relocated objects — via trajectory transfer.🔗 https://t.co/zbOGWrSHAF (1/11)
9
7
24
(11/11) #HRT1 shows that robots don't always need huge training pipelines. With the right system, a single human demo goes a long way. One demo → mobile manipulation → real-world execution. @saihaneesh_allu @jishnu_jaykumar @YuXiang_IRVL
0
0
2
(10/11) The standout aspect: No Training Needed: Full mobile manipulation with no learned policy — one shot, no RL, no finetuning. One human demo → real robot execution. And while most progress is on tabletop tasks, #HRT1 shows how far unified mobility + manipulation can go.
1
0
1
(9/11) A current limitation: performance depends on object pose estimation, creating an open-loop step. A brief manual pose check ensures stability. Depth also degrades in high-light or low-light scenes.
1
0
1
(8/11) After reaching the optimized base pose, we run a nonlinear constrained optimization to compute the joint-space trajectory that accurately imitates the rollout path.
1
0
1
(7/11) We solve a constrained optimization to find the optimal base pose (x, y, θ) so the robot can feasibly follow the rollout trajectory using trajectory optimization.
1
0
1
(6/11) We express the demonstration in the object frame. During rollout, BundleSDF estimates the relative pose change from RGB-D demo + rollout frames, adapting the demo to the new scene and producing the rollout trajectory.
1
0
1
(5/11) We convert the 3D human hand mesh trajectory into a robot end-effector trajectory, forming the robot's demonstration motion.
1
0
1
(4/11) From the RGB-D demo, we (i) reconstruct a 3D human hand mesh using #HaMeR: https://t.co/QWvTSaGCpL and (ii) refine its alignment w.r.t. the human hand via a depth-based optimization. Thus, we extract the precise 3D hand mesh trajectory that represents the demonstration.
1
0
1
(3/11) Data collection shouldn't be a bottleneck. Wear an AR headset, perform the task once from the robot's viewpoint, and HRT1 extracts all needed info from a single RGB-D demo.
1
0
1
(2/11) We didn't reinvent the wheel — we unified it. With MR-based, hands-free data collection, HRT1 unifies scene understanding (detection, segmentation, pose estimation), 3D hand tracking + grasp transfer, and motion optimization, yielding a coherent mobile-manipulation system.
1
0
1
Are you interested in Vision-Language-Action (VLA) models? We had an excellent guest lecture today by Ankit Goyal @imankitgoyal from NVIDIA on VLAs and their role in robot manipulation 🎥 Watch the recording here 👇 https://t.co/fhl15iKGr8 Slides: https://t.co/9lxXTXtX9r
5
51
402
Congrats Dr. Yangxiao Lu @parker_sean_L . Happy farewell and all the very best for your next phase !!
Congratulations to Yangxiao Lu @parker_sean_L @IRVLUTD for successfully defending his PhD thesis today! Yangxiao will be joining Meta as a Research Scientist. Best of luck, Dr. Lu! @UTDCompSci
0
0
8
Recognizing unseen objects is a challenge our lab keeps working on. This dataset from Amazon looks exciting to explore! Our latest work @parker_sean_L @ruosenli @jingliqiang6 @JwRobotics combines few-shot object detection with LLMs for visual grounding https://t.co/wz6pSvA3mC
Announcing Kaputt: a large-scale dataset for visual defect detection in retail logistics with 238,421 images across 48,376 unique items – 40x as large as current benchmarks:
1
19
47
A fantastic way to invoke the curious young minds! Part of research is to show the work through demos !!
We @IRVLUTD had some visitors today: the FTC robotics team from Flower Mound High School. We showed several demos: • Mobile manipulation Fetch • Teleoperation SO-101 and Koch • In-hand manipulation LEAP Exciting to see how curious and engaged they are🤖 #STEM #Robotics
0
0
4
Check out the awesome work https://t.co/lRverD96JF
We finally have our HO-Cap dataset published at NeurIPS 2025 Datasets & Benchmarks Track! 🎉 It was turned down by vision conferences several times, so extra appreciation to the NeurIPS D&B reviewers for recognizing its value! 🙏 #NeurIPS2025
0
0
4
Good luck Jishnu
Excited to begin my #Fall25 #ResearchInternship at @rai_inst 🍁. I’ll be working on #grasping and #manipulation #policylearning using #tactilesensing — combining #robotics, #perception, and #control to bring touch into robotic #decisionmaking 🤖🖐️
0
0
3