Nik__V__ Profile Banner
Nikhil Keetha Profile
Nikhil Keetha

@Nik__V__

Followers
2K
Following
24K
Media
92
Statuses
799

PhD in Robotics @CMU_Robotics @airlabcmu | Visiting Researcher @Meta | Making robots πŸ€– see the 🌍

Joined November 2017
Don't wanna be here? Send us removal request.
@Nik__V__
Nikhil Keetha
1 month
Meet MapAnything – a transformer that directly regresses factored metric 3D scene geometry (from images, calibration, poses, or depth) in an end-to-end way. No pipelines, no extra stages. Just 3D geometry & cameras, straight from any type of input, delivering new state-of-the-art
29
128
723
@alescontrela
Alejandro Escontrela
6 days
Simulation drives robotics progress, but how do we close the reality gap? Introducing GaussGym: an open-source framework for learning locomotion from pixels with ultra-fast parallelized photorealistic rendering across >4,000 iPhone, GrandTour, ARKit, and Veo scenes! Thread 🧡
10
60
290
@JonathonLuiten
Jonathon Luiten
7 days
Tune in @ ICCV on Mon @ 10.30am where I talk about everything 3D + Realism: - Hyperscape: Gaussian Splatting in VR - FlowR: Flowing from Sparse-2-Dense 3D Recon - BulletGen: Improving 4D Recon with Bullet-Time Gen - MapAnything: Universal Feed-Forward Metric 3D Recon πŸ§΅πŸ‘‡
2
16
176
@xinhangliu123
Xinhang Liu
11 days
Excited to share our latest work from the ByteDance Seed Depth Anything team β€” Trace Anything: Representing Any Video in 4D via Trajectory Fields πŸ’» Project Page: https://t.co/Q390WcWwG4 πŸ“„ Paper: https://t.co/NfxT260QWy πŸ“¦ Code: https://t.co/r2VbOHyRwL πŸ€– Model:
4
7
63
@Nik__V__
Nikhil Keetha
11 days
Games are all you need? πŸ‘€
@gen_intuition
General Intuition
11 days
Across tens of thousands of interactive environments, only other platform of comparable upload scale is YouTube. We’re taking a focused, straight shot at embodied intelligence with a world-class team, supported by a strong core business and leading investors.
0
0
4
@Nik__V__
Nikhil Keetha
11 days
Check out the blog from @CarnegieMellon @CMU_Robotics @roboVisionCMU covering MapAnything! Stay tuned on our GitHub repo for some exciting updates soon 🀩
@CMU_Robotics
CMU Robotics Institute
11 days
Researchers from the Robotics Institute + Meta Reality Labs have built a model that reconstructs images, camera data or depth scans into 3D maps within a unified system! MapAnything captures both small details and large spaces with high precision πŸŒŽπŸ—ΊοΈπŸ“Œ https://t.co/0dSLjzfeYc
0
4
35
@CMU_Robotics
CMU Robotics Institute
11 days
Researchers from the Robotics Institute + Meta Reality Labs have built a model that reconstructs images, camera data or depth scans into 3D maps within a unified system! MapAnything captures both small details and large spaces with high precision πŸŒŽπŸ—ΊοΈπŸ“Œ https://t.co/0dSLjzfeYc
1
13
102
@ducha_aiki
Dmytro Mishkin πŸ‡ΊπŸ‡¦
12 days
FlowR: Flowing from Sparse to Dense 3D Reconstructions @TobiasRobotics et 8 al with @Nik__V__ tl;dr: Mast3r -> 3D rec -> render new view -> flow matching & refine reconstruction with synth views -> better rec https://t.co/fmTkC15rSl
3
15
95
@sainingxie
Saining Xie
13 days
three years ago, DiT replaced the legacy unet with a transformer-based denoising backbone. we knew the bulky VAEs would be the next to go -- we just waited until we could do it right. today, we introduce Representation Autoencoders (RAE). >> Retire VAEs. Use RAEs. πŸ‘‡(1/n)
56
330
2K
@ethanjohnweber
Ethan Weber
16 days
Meta Reality Labs is looking for research interns! ✨ If you're excited about building world models for AR/VR and immersive experiences, come join us. πŸ™Œ Roles available across Europe (I’m based in Zurich πŸ‡¨πŸ‡­) and the US. You can DM me or find me at #ICCV to chat. πŸ™‚ @AIatMeta
9
31
303
@Nik__V__
Nikhil Keetha
17 days
@jianyuan_wang I guess you tried the DAv3 style init startegy for VGGT back in December'24 with the main difference being that it was full alternating (ablation c) instead of the partial one (last layers only) DAv3 uses?
2
1
5
@Nik__V__
Nikhil Keetha
17 days
Interesting ICLR submissions 🀩 Depth Anything 3 - My TLDR: Init multi view transformer of VGGT with later layer DINO weights and use teacher model trained on synthetic data only for pseudo labelling real world datasets https://t.co/1gq6WSZM5c Trace Anything - My TLDR: VGGT
5
41
393
@gabriberton
Gabriele Berton
21 days
Looks like researchers in visual localization are good reviewers: I see so many familiar names in the list of ICCV outstanding reviewers! Congrats to @alessiodelbue @masone_carlo @kgcs96 @jcivera @Parskatt Julian Kooij @maththrills @NicStrisc @Nik__V__ (and me, happy to be in
4
6
33
@Nik__V__
Nikhil Keetha
24 days
Glad to be recognized again as an outstanding reviewer! πŸ€—
@ICCVConference
#ICCV2025
25 days
There’s no conference without the efforts of our reviewers. Special shoutout to our #ICCV2025 outstanding reviewers 🫑 https://t.co/WYAcXLRXla
0
0
24
@seungchankim25
Seungchan Kim
25 days
We introduce RAVEN, a 3D open-set memory-based behavior tree framework for aerial outdoor semantic navigation. RAVEN not only navigates reliably toward detected targets, but also performs long-range semantic reasoning and LVLM-guided informed search
1
7
21
@ethanjohnweber
Ethan Weber
25 days
πŸ“’ SceneComp @ ICCV 2025 🏝️ 🌎 Generative Scene Completion for Immersive Worlds πŸ› οΈ Reconstruct what you know AND πŸͺ„ Generate what you don’t! πŸ™Œ Meet our speakers @angelaqdai, @holynski_, @jampani_varun, @ZGojcic @taiyasaki, Peter Kontschieder https://t.co/LvONYIK3dz #ICCV2025
2
17
52
@JonathonLuiten
Jonathon Luiten
1 month
Show-off your own Hyperscapes! Thead 🧡 below πŸ‘‡ showing off various user-created Hyperscapes! They all look SOOOO good! Add your video in the comments or @ me, and I will add it to the thread!
@JonathonLuiten
Jonathon Luiten
1 month
Introducing: Hyperscape Capture πŸ“· Last year we showed the world's highest quality Gaussian Splatting, and the first time GS was viewable in VR. Now, capture your own Hyperscapes, directly from your Quest headset in only 5 minutes of walking around. https://t.co/wlHmtRiANy
7
20
192
@ZhiqiuLin
Zhiqiu Lin
1 month
πŸŽ‰CameraBench has been accepted as a Spotlight (3%) @ NeurIPS 2025. Huge congrats to all collaborators at CMU, MIT-IBM, UMass, Harvard, and Adobe. CameraBench is a large-scale effort that pushes video-language models to reason about the language of camera motion just like
Tweet card summary image
arxiv.org
We introduce CameraBench, a large-scale dataset and benchmark designed to assess and improve camera motion understanding. CameraBench consists of ~3,000 diverse internet videos, annotated by...
@ZhiqiuLin
Zhiqiu Lin
6 months
πŸ“· Can AI understand camera motion like a cinematographer? Meet CameraBench: a large-scale, expert-annotated dataset for understanding camera motion geometry (e.g., trajectories) and semantics (e.g., scene contexts) in any video – films, games, drone shots, vlogs, etc. Links
7
23
162
@JonathonLuiten
Jonathon Luiten
1 month
Introducing: Hyperscape Capture πŸ“· Last year we showed the world's highest quality Gaussian Splatting, and the first time GS was viewable in VR. Now, capture your own Hyperscapes, directly from your Quest headset in only 5 minutes of walking around. https://t.co/wlHmtRiANy
@JonathonLuiten
Jonathon Luiten
1 year
Hyperscape: The future of VR and the Metaverse Excited that Zuckerberg @finkd announced what I have been working on at Connect. Hyperscape enables people to create high fidelity replicas of physical spaces, and embody them in VR. Check out the demo app: https://t.co/TcRRUfymoc
41
284
2K
@Nik__V__
Nikhil Keetha
1 month
Awesome thread; always fun to see @ducha_aiki stress testing things!
@ducha_aiki
Dmytro Mishkin πŸ‡ΊπŸ‡¦
1 month
MapAnything vs VGGT thread. Every tweet -- 4 results on the same input: VGGT, VGGT-Apache, MapAnything, MapAnything-Apache Starting with classics, graffiti from OxfordAffine. Top-down to see the wall flattness. Spoiler: All not flat 1/
1
2
5
@TobiasFischer11
Tobias Fischer
1 month
Forget pointmaps, pinhole, image-only, up-to-scale… and meet MapAnything! Truly amazing, long-term effort led by @Nik__V__ that weβ€˜re super excited to share today!
@Nik__V__
Nikhil Keetha
1 month
Meet MapAnything – a transformer that directly regresses factored metric 3D scene geometry (from images, calibration, poses, or depth) in an end-to-end way. No pipelines, no extra stages. Just 3D geometry & cameras, straight from any type of input, delivering new state-of-the-art
0
5
82