
Andrew Owens
@andrewhowens
Followers
2K
Following
2K
Media
39
Statuses
178
Associate Professor of CS @ Cornell Tech
Joined July 2019
Our computer vision textbook is now available for free online here: https://t.co/ERy2Spc7c2 We are working on adding some interactive components like search and (beta) integration with LLMs. Hope this is useful and feel free to submit Github issues to help us improve the text!
visionbook.mit.edu
35
620
3K
Sharing our #CVPR2025 paper: "GPS as a Control Signal for Image Generation"! 🛰️+✍️ We turn the GPS tag stored in EXIF of photos into a control signal for diffusion models—so they don’t just know what you asked for, but where you want it to look like. Come to see our poster at
2
9
34
Ever wondered how a scene sounds👂 when you interact👋 with it? Introducing our #CVPR2025 work "Hearing Hands: Generating Sounds from Physical Interactions in 3D Scenes" -- we make 3D scene reconstructions audibly interactive! https://t.co/tIcFGJtB7R
2
34
97
Excited to share our CVPR 2025 paper on cross-modal space-time correspondence! We present a method to match pixels across different modalities (RGB-Depth, RGB-Thermal, Photo-Sketch, and cross-style images) — trained entirely using unpaired data and self-supervision. Our
1
28
121
Can AI image detectors keep up with new fakes? Mostly, no. Existing detectors are trained using a handful of models. But there are thousands in the wild! Our work, Community Forensics, uses 4800+ generators to train detectors that generalize to new fakes. #CVPR2025 🧵 (1/5)
1
9
23
Hello! If you like pretty images and videos and want a rec for CVPR oral session, you should def go to Image/Video Gen, Friday at 9am: I'll be presenting "Motion Prompting" @RyanBurgert will be presenting "Go with the Flow" and @ChangPasca1650 will be presenting "LookingGlass"
3
16
64
I had a lot of fun helping put this problem set together -- if you're teaching diffusion models + computer vision, consider using this homework for your course! (links at end of @ryan_tabrizi's thread!)
Teaching computer vision next semester? Hoping to finally learn about diffusion models in 2025? Check out this diffusion project that we designed and test-drove this past semester at Berkeley and Michigan!
1
19
140
Teaching computer vision next semester? Hoping to finally learn about diffusion models in 2025? Check out this diffusion project that we designed and test-drove this past semester at Berkeley and Michigan!
2
44
245
3D illusions are fascinating! 🤩 But it takes exceptional artistic skills to make one. We present Illusion3D - a simple method for creating 3D multiview illusions, where the interpretations change depending on your perspectives. Let's play Where's Waldo, shall we? 😆
4
16
110
I'll be presenting "Images that Sound" today at #NeurIPS2024! East Exhibit Hall A-C #2710. Come say hi to me and @andrewhowens :) (@CzyangChen sadly could not make it, but will be there in spirit :') )
These spectrograms look like images, but can also be played as a sound! We call these images that sound. How do we make them? Look and listen below to find out, and to see more examples!
0
7
54
Webpage: https://t.co/Tq306ItOi6 Paper: https://t.co/cfKsxJN3Ok I'd like to note that our work builds on recent methods for long-range tracking (like Tracking Any Point and Particle Video Revisited), which make this kind of conditioning possible.
arxiv.org
Motion control is crucial for generating expressive and compelling video content; however, most existing video generation models rely mainly on text prompts for control, which struggle to capture...
0
0
4
Generate videos conditioned on point tracks. I was surprised by just how many different things you could do with this model, given the right prompting! Led by @dangengdg. https://t.co/dIVkXgikGn
What happens when you train a video generation model to be conditioned on motion? Turns out you can perform "motion prompting," just like you might prompt an LLM! Doing so enables many different capabilities. Here’s a few examples – check out this thread 🧵 for more results!
1
6
53
What happens when you train a video generation model to be conditioned on motion? Turns out you can perform "motion prompting," just like you might prompt an LLM! Doing so enables many different capabilities. Here’s a few examples – check out this thread 🧵 for more results!
20
146
673
🎥 Introducing MultiFoley, a video-aware audio generation method with multimodal controls! 🔊 We can ⌨️Make a typewriter sound like a piano 🎹 🐱Make a cat meow like a lion roars! 🦁 ⏱️Perfectly time existing SFX 💥 to a video
11
42
210
@CzyangChen Webpage: https://t.co/KoVVVl5p8e Paper: https://t.co/FvkIk0gOQU w/ @CzyangChen, @pseetharaman, Bryan Russell, @urinieto, David Bourgin, @justin_salamon
arxiv.org
Generating sound effects for videos often requires creating artistic sound effects that diverge significantly from real-life sources and flexible control in the sound design. To address this...
0
0
1
We generate a soundtrack for a silent video, given a text prompt! For example, we can make a cat's meow sound like a lion's roar or a typewriter sound like a piano. Led by @CzyangChen! https://t.co/7S26dKUzR5
🎥 Introducing MultiFoley, a video-aware audio generation method with multimodal controls! 🔊 We can ⌨️Make a typewriter sound like a piano 🎹 🐱Make a cat meow like a lion roars! 🦁 ⏱️Perfectly time existing SFX 💥 to a video
2
5
52
I am recruiting exceptional PhD students & postdocs with an adventurous soul for my💫new TTIC AI lab💫! We aim to understand intelligence, one pixel at a time, inspired by psychology, neuroscience, language, robotics, and the arts. https://t.co/tMJWow7AS6
https://t.co/iDwnBfijjw
2
49
167
Hiring research interns to work on AI for Sound Design! 💥🛸🐲 Join the Sound Design AI Group (SODA) for a summer 2025 internship advancing the bleeding edge of audio AI for video with @pseetharaman @urinieto and me. A chance to impact millions of users: https://t.co/6QY6KchaQ8
Ready to make an impact with cutting-edge research? @AdobeResearch is now accepting applications for summer 2025 internships from master's and PhD students. Apply now! https://t.co/eS61RzEmqq
0
5
28
Couldn’t be more excited—hundreds of people have gathered from across Michigan (& beyond!) for our 6th Michigan AI Symposium. Special theme this year Embodied AI, co-chaired by Stella Yu & @AndrewHOwens. Warm intros by a bipedal robot 🤖 and a humanoid Provost, Laurie McCauley.
1
1
36