
joao carreira
@joaocarreira
Followers
1K
Following
282
Media
10
Statuses
145
Research Scientist at Google DeepMind
London, England
Joined February 2009
Scaling 4D Representations – new preprint and models now available
github.com
Contribute to google-deepmind/representations4d development by creating an account on GitHub.
3
42
203
3rd edition of the challenge with new exciting tasks and guest tracks; back during covid when we had the first workshop about the perception test ( some of us were afraid the benchmark was too difficult; now we just made it harder.
The 3rd Perception Test challenge is now accepting submissions ! Prizes of up to 50k EUR across Perception Test tracks are available. The winners will be announced at the Perception Test workshop at #ICCV2025. Submission deadline: October 6, 2025.
0
0
4
RT @yanahasson: Thrilled to share our latest work on SciVid, to appear at #ICCV2025! 🎉.SciVid offers cross-domain evaluation of video model….
0
9
0
RT @sangwoomo: Can scaling data and models alone solve computer vision? 🤔.Join us at the SP4V Workshop at #ICCV2025 in Hawaii to explore th….
0
17
0
Individual frames out of generative video models tend to look reasonable; capturing actions happening over time realistically . that is way harder. TRAJAN is a new evaluation procedure to better guide progress in this (hot) area.
Humans can tell the difference between a realistic generated video and an unrealistic one – can models?. Excited to share TRAJAN: the world’s first point TRAJectory AutoeNcoder for evaluating motion realism in generated and corrupted videos. 🌐 🧵
2
8
45
RT @TengdaHan: Check out our CVPR 2025 paper: Work with Dilara Gokay, Joseph Heyward, @ChuhanZhang5 , @DanielZoran….
arxiv.org
We address the challenge of representation learning from a continuous stream of video as input, in a self-supervised manner. This differs from the standard approaches to video learning where...
0
2
0
RT @TengdaHan: We are looking for a student researcher to work on video understanding plus 3D, in Google DeepMind London. DM/Email me or pa….
0
33
0
RT @EEMLcommunity: Apply here: Confirmed speakers: @AaronCourville @AldenHung @dianaborsa @09Emmar @joaocarreira @….
eeml.eu
Participation at the event is subject to selection. Please see below the details on how to apply. Deadline for applications: 31 March 2025, 7 April 2025, 23:59 Anywhere on Earth. Application period...
0
2
0
RT @vansteenkiste_s: Excited to announce MooG for learning video representations. MooG allows tokens to move “off-the-grid” enabling better….
0
14
0
RT @dimadamen: Time to challenge VLMs?.Fed up of benchmarks claiming long-video reasoning but only need few seconds?. Try out Hour-Long VQA….
0
6
0
RT @skandakoppula: We're excited to release TAPVid-3D: an evaluation benchmark of 4,000+ real world videos and 2.1 million metric 3D point….
0
58
0
RT @shiryginosar: Join us next week at our second (high-level) intelligence workshop @SimonsInstitute!. Schedule: ..
0
9
0
RT @CarlDoersch: We present a new SOTA on point tracking, via self-supervised training on real, unlabeled videos! BootsTAPIR achieves 67.4%….
0
65
0
RT @shawshank_v: Delighted to host the 1st edition of our tutorial "Time is precious: Self-Supervised Learning Beyond Images" at @eccvconf….
0
10
0
RT @ShaneLegg: Our research project SIMA is creating a general, natural language instructable, multi 3D game-playing AI agent. The agent c….
0
82
0
Videos have a wealth of learning signal that is still underappreciated -- in fact, looks like a single long video can be as valuable as a large curated internet image dataset. Cool work from @shawshank_v et al with a new self-sup formulation where multi-object tracking emerges.
Really happy to share that DoRA is accepted as an Oral to @iclr_conf #ICLR2024. Using just “1 video” from our new egocentric dataset - Walking Tours, we develop a new method that outperforms DINO pretrained on ImageNet on image and video downstream tasks. More details in 🧵👇.
0
5
24