justin_salamon Profile Banner
Justin Salamon Profile
Justin Salamon

@justin_salamon

Followers
3K
Following
1K
Media
152
Statuses
913

Head of Sound Design AI Research at Adobe. Machine learning and signal processing for audio & video. Musician. He/him.

Joined October 2014
Don't wanna be here? Send us removal request.
@justin_salamon
Justin Salamon
6 days
Been seeing Cat Olympics recently, but there was no story. So I added GenAI sound effects. SFX are not just there to “match” what you see. They craft a story: change the sound, and drama becomes comedy. AI sound design with creative controls = you control the storytelling 🥇
0
0
7
@justin_salamon
Justin Salamon
15 days
Would love to see your own creations in this thread 👇. Here’s one from @karenxcheng I thought was super cool.
0
1
2
@justin_salamon
Justin Salamon
15 days
Can GenAI keep the performative aspect of sound design? @Adobe Firefly lets you perform with your voice and turn it into any sound effect!. Try it out: Model by our SODA group at @AdobeResearch : @pseetharaman @urinieto & our incredible intern @hugggof
1
3
20
@justin_salamon
Justin Salamon
22 days
To learn more please check out our ICML'25 paper: "FLAM: Frame-Wise Language-Audio Modeling". Big congratulations to @wuyusongwys, @tsirigoc and everyone on the team. So excited to share this work, which solves a long standing problem in audio research!.
0
0
6
@justin_salamon
Justin Salamon
22 days
FLAM is trained jointly on instance (global) and frame-wise (local) objectives. The secret sauce: A memory-efficient and calibrated frame-wise objective with logit adjustment to address spurious correlations, such as event dependencies and label imbalances during training
Tweet media one
1
0
4
@justin_salamon
Justin Salamon
22 days
Enter FLAM: Frame-Wise Language-Audio Modeling. A model trained to produce a calibrated likelihood for *any* text prompt. FLAM outperforms prior self-supervised models on both closed-set and open-set SED, while preserving strong retrieval and zero-shot classification accuracy
Tweet media one
1
0
6
@justin_salamon
Justin Salamon
22 days
Our goal is for the model to detect *any* sound via free form text queries. "So use CLAP", some of you will say. The problem is its output likelihoods are not calibrated for different prompts :(. That's ok ranked retrieval, but for detection it's a no go.
Tweet media one
1
0
5
@justin_salamon
Justin Salamon
22 days
Sound Event Detection models, ie finding sounds in audio/video recordings, are typically constrained to a predefined "closed" set of sounds, like in this (old!) model below for urban sound detection. It has some applications, but it doesn't address general purpose sound search.
Tweet media one
1
0
5
@justin_salamon
Justin Salamon
22 days
I think we finally cracked it? FLAM can detect *any* sound via text prompts. arXiv (ICML'25): demos: @AdobeResearch+@MIT+@Mila_Quebec led by @wuyusongwys w/@tsirigoc @Kotentorothy @huangcza @AaronCourville @urinieto @pseetharaman
6
40
265
@justin_salamon
Justin Salamon
3 months
Generative Extend in Premiere Pro just won *five* awards at NAB 2025, including the @NABShow Product of the Year award! SODA, our group, created the audio GenAI model in charge of audio extensions in the feature. Couldn't be more proud of the team!. w/@urinieto @pseetharaman
Tweet media one
0
0
11
@justin_salamon
Justin Salamon
3 months
RT @AdobeResearch: Generative Extend just launched in Premiere Pro—use #GenAI to extend video and audio for perfectly timed edits! This was….
0
1
0
@justin_salamon
Justin Salamon
3 months
Generative Extend just released in Premiere Pro! Use GenAI to extend your video *and audio* clips for a perfectly timed edit. The audio model was built by our team, the Sound Design AI (SODA) group at @AdobeResearch w/ @pseetharaman and @urinieto 🙌 .
0
5
21
@justin_salamon
Justin Salamon
5 months
We didn't expect this. our Sketch2Sound demo video has gone viral on IG with more than 5.2 million views 🤯. Amazing job @hugggof @pseetharaman @urinieto . I should've done my hair.
0
3
3
@justin_salamon
Justin Salamon
7 months
Really cool to see our AI for bioacoustics work in the MIT Technology Review! BirdVox was a fantastic project that beyond posing a fascinating research problem, introduced me to the world of bird watching and deepened my appreciation of the natural world.
0
4
21
@justin_salamon
Justin Salamon
7 months
Sketch2Sound is out!. Takes a text prompt + vocal (or sonic) imitation and generates sound effects that perfectly match the energy and dynamics of your voice. It's an extremely intuitive (and fun!) way to create SFX that are perfectly timed to your video. Led by @hugggof 👏.
@hugggof
hugo flores garcía 🌻
7 months
new paper! 🗣️Sketch2Sound💥. Sketch2Sound can create sounds from sonic imitations (i.e., a vocal imitation or a reference sound) via interpretable, time-varying control signals. paper: web:
4
3
30
@justin_salamon
Justin Salamon
7 months
I'm hiring a Research Engineer for my team at @AdobeResearch . Full details in the link I shared in the post below. DM me if interested.
@justin_salamon
Justin Salamon
7 months
📢 Audio AI Job opportunity at Adobe!. The Sound Design AI Group (SODA) is looking for an exceptional research engineer to join us in building the future of AI-assisted audio and video creation. Strong ML background, GenAI experience a plus. Details:
1
4
22
@justin_salamon
Justin Salamon
7 months
More from our group: MultiFoley, a Video-to-Audio model that generates perfectly synced audio for video at 48 kHz and supports multimodal conditioning. You can define the generated sound via a text prompt, an example SFX, or audio to extend. More here:.
@CzyangChen
Ziyang Chen
8 months
🎥 Introducing MultiFoley, a video-aware audio generation method with multimodal controls! 🔊. We can.⌨️Make a typewriter sound like a piano 🎹.🐱Make a cat meow like a lion roars! 🦁.⏱️Perfectly time existing SFX 💥 to a video
0
0
2
@justin_salamon
Justin Salamon
7 months
📢 Audio AI Job opportunity at Adobe!. The Sound Design AI Group (SODA) is looking for an exceptional research engineer to join us in building the future of AI-assisted audio and video creation. Strong ML background, GenAI experience a plus. Details:
2
8
36
@justin_salamon
Justin Salamon
8 months
Webpage: Paper: . w/ @CzyangChen, @pseetharaman, Bryan Russell, @urinieto, David Bourgin, @andrewhowens.
0
0
5
@justin_salamon
Justin Salamon
8 months
New model from the SODA group @AdobeResearch and UMich!. MultiFoley generates perfectly synced audio for video at 48 kHz and supports multimodal conditioning. Define the generated sound via a text prompt, an example SFX, or audio to extend. Led by our intern @CzyangChen 👇.
@CzyangChen
Ziyang Chen
8 months
🎥 Introducing MultiFoley, a video-aware audio generation method with multimodal controls! 🔊. We can.⌨️Make a typewriter sound like a piano 🎹.🐱Make a cat meow like a lion roars! 🦁.⏱️Perfectly time existing SFX 💥 to a video
2
0
23