Andrew Owens Profile
Andrew Owens

@andrewhowens

Followers
2K
Following
2K
Media
39
Statuses
178

Associate Professor of CS @ Cornell Tech

Joined July 2019
Don't wanna be here? Send us removal request.
@phillip_isola
Phillip Isola
3 months
Our computer vision textbook is now available for free online here: https://t.co/ERy2Spc7c2 We are working on adding some interactive components like search and (beta) integration with LLMs. Hope this is useful and feel free to submit Github issues to help us improve the text!
Tweet card summary image
visionbook.mit.edu
35
620
3K
@chaof1234
Chao Feng
3 months
Sharing our #CVPR2025 paper: "GPS as a Control Signal for Image Generation"! 🛰️+✍️ We turn the GPS tag stored in EXIF of photos into a control signal for diffusion models—so they don’t just know what you asked for, but where you want it to look like. Come to see our poster at
2
9
34
@_YimingDou
Yiming Dou
3 months
Ever wondered how a scene sounds👂 when you interact👋 with it? Introducing our #CVPR2025 work "Hearing Hands: Generating Sounds from Physical Interactions in 3D Scenes" -- we make 3D scene reconstructions audibly interactive! https://t.co/tIcFGJtB7R
2
34
97
@ayshrv
Ayush Shrivastava
3 months
Excited to share our CVPR 2025 paper on cross-modal space-time correspondence! We present a method to match pixels across different modalities (RGB-Depth, RGB-Thermal, Photo-Sketch, and cross-style images) — trained entirely using unpaired data and self-supervision. Our
Tweet media one
1
28
121
@jespark0
Jeongsoo Park
3 months
Can AI image detectors keep up with new fakes? Mostly, no. Existing detectors are trained using a handful of models. But there are thousands in the wild! Our work, Community Forensics, uses 4800+ generators to train detectors that generalize to new fakes. #CVPR2025 🧵 (1/5)
1
9
23
@dangengdg
Daniel Geng
3 months
Hello! If you like pretty images and videos and want a rec for CVPR oral session, you should def go to Image/Video Gen, Friday at 9am: I'll be presenting "Motion Prompting" @RyanBurgert will be presenting "Go with the Flow" and @ChangPasca1650 will be presenting "LookingGlass"
3
16
64
@dangengdg
Daniel Geng
8 months
I had a lot of fun helping put this problem set together -- if you're teaching diffusion models + computer vision, consider using this homework for your course! (links at end of @ryan_tabrizi's thread!)
@ryan_tabrizi
Ryan Tabrizi
8 months
Teaching computer vision next semester? Hoping to finally learn about diffusion models in 2025? Check out this diffusion project that we designed and test-drove this past semester at Berkeley and Michigan!
1
19
140
@ryan_tabrizi
Ryan Tabrizi
8 months
Teaching computer vision next semester? Hoping to finally learn about diffusion models in 2025? Check out this diffusion project that we designed and test-drove this past semester at Berkeley and Michigan!
2
44
245
@jbhuang0604
Jia-Bin Huang
9 months
3D illusions are fascinating! 🤩 But it takes exceptional artistic skills to make one. We present Illusion3D - a simple method for creating 3D multiview illusions, where the interpretations change depending on your perspectives. Let's play Where's Waldo, shall we? 😆
4
16
110
@dangengdg
Daniel Geng
9 months
I'll be presenting "Images that Sound" today at #NeurIPS2024! East Exhibit Hall A-C #2710. Come say hi to me and @andrewhowens :) (@CzyangChen sadly could not make it, but will be there in spirit :') )
@CzyangChen
Ziyang Chen
1 year
These spectrograms look like images, but can also be played as a sound! We call these images that sound. How do we make them? Look and listen below to find out, and to see more examples!
0
7
54
@andrewhowens
Andrew Owens
9 months
Webpage: https://t.co/Tq306ItOi6 Paper: https://t.co/cfKsxJN3Ok I'd like to note that our work builds on recent methods for long-range tracking (like Tracking Any Point and Particle Video Revisited), which make this kind of conditioning possible.
Tweet card summary image
arxiv.org
Motion control is crucial for generating expressive and compelling video content; however, most existing video generation models rely mainly on text prompts for control, which struggle to capture...
0
0
4
@andrewhowens
Andrew Owens
9 months
Generate videos conditioned on point tracks. I was surprised by just how many different things you could do with this model, given the right prompting! Led by @dangengdg. https://t.co/dIVkXgikGn
@dangengdg
Daniel Geng
9 months
What happens when you train a video generation model to be conditioned on motion? Turns out you can perform "motion prompting," just like you might prompt an LLM! Doing so enables many different capabilities. Here’s a few examples – check out this thread 🧵 for more results!
1
6
53
@dangengdg
Daniel Geng
9 months
What happens when you train a video generation model to be conditioned on motion? Turns out you can perform "motion prompting," just like you might prompt an LLM! Doing so enables many different capabilities. Here’s a few examples – check out this thread 🧵 for more results!
20
146
673
@CzyangChen
Ziyang Chen
10 months
🎥 Introducing MultiFoley, a video-aware audio generation method with multimodal controls! 🔊 We can ⌨️Make a typewriter sound like a piano 🎹 🐱Make a cat meow like a lion roars! 🦁 ⏱️Perfectly time existing SFX 💥 to a video
11
42
210
@andrewhowens
Andrew Owens
10 months
We generate a soundtrack for a silent video, given a text prompt! For example, we can make a cat's meow sound like a lion's roar or a typewriter sound like a piano. Led by @CzyangChen! https://t.co/7S26dKUzR5
@CzyangChen
Ziyang Chen
10 months
🎥 Introducing MultiFoley, a video-aware audio generation method with multimodal controls! 🔊 We can ⌨️Make a typewriter sound like a piano 🎹 🐱Make a cat meow like a lion roars! 🦁 ⏱️Perfectly time existing SFX 💥 to a video
2
5
52
@jjpark3D
Jeong Joon Park
10 months
I’m recruiting PhD students with computer vision, robotics, or ML experience! We especially encourage applicants from physics and related fields who want to explore AI for Science. Join us by applying to @UMichCSE's PhD program!
5
86
378
@shiryginosar
Shiry Ginosar
10 months
I am recruiting exceptional PhD students & postdocs with an adventurous soul for my💫new TTIC AI lab💫! We aim to understand intelligence, one pixel at a time, inspired by psychology, neuroscience, language, robotics, and the arts. https://t.co/tMJWow7AS6 https://t.co/iDwnBfijjw
Tweet media one
2
49
167
@justin_salamon
Justin Salamon
11 months
Hiring research interns to work on AI for Sound Design! 💥🛸🐲 Join the Sound Design AI Group (SODA) for a summer 2025 internship advancing the bleeding edge of audio AI for video with @pseetharaman @urinieto and me. A chance to impact millions of users: https://t.co/6QY6KchaQ8
@AdobeResearch
Adobe Research
11 months
Ready to make an impact with cutting-edge research? @AdobeResearch is now accepting applications for summer 2025 internships from master's and PhD students. Apply now! https://t.co/eS61RzEmqq
Tweet media one
0
5
28
@radamihalcea
Rada Mihalcea
11 months
Couldn’t be more excited—hundreds of people have gathered from across Michigan (& beyond!) for our 6th Michigan AI Symposium. Special theme this year Embodied AI, co-chaired by Stella Yu & @AndrewHOwens. Warm intros by a bipedal robot 🤖 and a humanoid Provost, Laurie McCauley.
Tweet media one
Tweet media two
1
1
36