NicholasJBryan Profile Banner
Nicholas J. Bryan Profile
Nicholas J. Bryan

@NicholasJBryan

Followers
1K
Following
940
Media
2
Statuses
89

Head of Music AI, Adobe Research (personal account)

Joined April 2010
Don't wanna be here? Send us removal request.
@NicholasJBryan
Nicholas J. Bryan
15 days
Adobe's #GenerateSoundtrack is LIVE today! 🎉 Studio-quality music for storytellers🎵 * Trained on #Licensed data, * Commercially safe, royalty-free, and cleared for any use, & * Exported with #ContentCredentials for transparency and attribution. Get started:
1
7
13
@slseanwu
Shih-Lun (Sean) Wu
6 days
Thrilled to announce “MIDI-LLM: Adapting LLMs for Text-to-MIDI Music Generation” w/ @huangcza and Yoon Kim! 🎸 Live Demo https://t.co/6N6AqyrZuW 💻  https://t.co/Z1v82uool5 🤗 https://t.co/k9deMY82Vk From a text prompt, it generates MIDIs you can edit directly in a DAW 🧵
2
12
21
@NicholasJBryan
Nicholas J. Bryan
15 days
Congrats @JCJesseLai and Team!
@JCJesseLai
Chieh-Hsin (Jesse) Lai
15 days
Tired to go back to the original papers again and again? Our monograph: a systematic and fundamental recipe you can rely on! 📘 We’re excited to release 《The Principles of Diffusion Models》— with @DrYangSong, @gimdong58085414, @mittu1204, and @StefanoErmon. It traces the core
1
0
3
@verge
The Verge
16 days
Adobe’s new AI audio tools can add soundtracks and voice-overs to videos
Tweet card summary image
theverge.com
Generate Soundtrack is like Mad Libs for music-making.
0
5
20
@_akhaliq
AK
7 months
Adobe announced DRAGON on Hugging Face Distributional Rewards Optimize Diffusion Generative Models
14
42
240
@NicholasJBryan
Nicholas J. Bryan
7 months
DRAGON introduces a new approach to designing and optimizing reward functions to enhance human-perceived quality.
0
0
0
@NicholasJBryan
Nicholas J. Bryan
7 months
With an appropriate exemplar set, DRAGON achieves a 60.95% human-voted music quality win rate without training on human preference annotations.
1
0
0
@NicholasJBryan
Nicholas J. Bryan
7 months
Over all 20 target rewards, DRAGON achieves an 81.45% average win rate. Moreover, reward functions based on exemplar sets indeed enhance generations and are comparable to model-based rewards.
1
0
0
@NicholasJBryan
Nicholas J. Bryan
7 months
We further compare instance-wise (per-song) and full-dataset FAD settings while ablating multiple FAD encoders and reference sets.
1
0
0
@NicholasJBryan
Nicholas J. Bryan
7 months
For evaluation, we fine-tune an audio-domain text-to-music diffusion model with 20 different reward functions, including a custom music aesthetics model, CLAP score, Vendi diversity, and Frechet audio distance (FAD).
1
0
0
@NicholasJBryan
Nicholas J. Bryan
7 months
Then, DRAGON gathers online and on-policy generations, scores them to construct a positive demonstration set and a negative set, and leverages the contrast between the two sets to maximize the reward.
1
0
0
@NicholasJBryan
Nicholas J. Bryan
7 months
When cross-modality encoders such as CLAP are used, the reference examples may be of a different modality (e.g., text versus audio).
1
0
0
@NicholasJBryan
Nicholas J. Bryan
7 months
Leveraging this versatility, we construct novel reward functions by selecting an encoder and a set of reference examples to create an exemplar distribution.
1
0
0
@NicholasJBryan
Nicholas J. Bryan
7 months
It can optimize reward functions that evaluate either individual examples or distributions of them, making it compatible with a broad spectrum of instance-wise, instance-to-distribution, and distribution-to-distribution rewards.
1
0
0
@NicholasJBryan
Nicholas J. Bryan
7 months
Compared with traditional reinforcement learning with human feedback (RLHF) or pairwise preference approaches such as direct preference optimization (DPO), DRAGON is more flexible.
1
0
2
@NicholasJBryan
Nicholas J. Bryan
7 months
Introducing "DRAGON: Distributional Rewards Optimize Diffusion Generative Models"! 📖: https://t.co/biSceK5vgQ 🎹: https://t.co/6taiaC8SUZ A new framework for fine-tuning gen models towards a target distribution. By Yatong Bai w/@CasebeerJonah @somayeh_sojoudi @NicholasJBryan
2
10
28
@zacknovack
Zachary Novack
7 months
Hyped to be presenting Presto next week at #ICLR2025 as a Spotlight! Super excited to chat about all things generative music/audio, diffusion acceleration, and real-time audio systems, lmk if you want to meet (DMs open)!
@zacknovack
Zachary Novack
1 year
Ultra-fast text-to-music generation w/o degrading quality? Introducing Presto! Distilling Steps and Layers for Accelerating Music Generation 🎹: https://t.co/kTTAYKKtTU 📖: https://t.co/Newhxe6lI6 w/@__gzhu__ @CasebeerJonah @BergKirkpatrick @McAuleyLabUCSD @NicholasJBryan 🧵
0
7
38
@zacknovack
Zachary Novack
10 months
Our paper "Presto! Distilling Steps and Layers for Accelerating Music Generation" has been accepted at #ICLR2025 @iclr_conf! Check out the demo video and links below, excited to see everyone in Singapore 🇸🇬!
@zacknovack
Zachary Novack
1 year
Ultra-fast text-to-music generation w/o degrading quality? Introducing Presto! Distilling Steps and Layers for Accelerating Music Generation 🎹: https://t.co/kTTAYKKtTU 📖: https://t.co/Newhxe6lI6 w/@__gzhu__ @CasebeerJonah @BergKirkpatrick @McAuleyLabUCSD @NicholasJBryan 🧵
1
7
63
@NicholasJBryan
Nicholas J. Bryan
1 year
🎶Diffusion Inference-time T-Optimization (DITTO) for music gen gets faster w/DITTO2! Optimize diffusion initial noise latent via a 1-step sampler for powerful control. Decode w/few sampling steps for best quality. Led by @zacknovack w/@McAuleyLabUCSD @BergKirkpatrick, myself
@zacknovack
Zachary Novack
1 year
Excited for my 1st #ISMIR2024 this week! Happy to chat about controllable + fast music generation 🙂 I'll be presenting our part 2 of DITTO, where we accelerate control to near real-time! DITTO-2: Distilled Diffusion Inference Time T-Optimization 🎹: https://t.co/WaARGNjfrM 🧵
0
0
28
@NicholasJBryan
Nicholas J. Bryan
1 year
Ultra fast music generation via accelerating diffusion TTM without quality loss. Amazing work led by @zacknovack! Summer research intern work at @AdobeResearch w/@__gzhu__, @CasebeerJonah, myself, and University collab w/ @McAuleyLabUCSD @BergKirkpatrick at UCSD!
@zacknovack
Zachary Novack
1 year
Ultra-fast text-to-music generation w/o degrading quality? Introducing Presto! Distilling Steps and Layers for Accelerating Music Generation 🎹: https://t.co/kTTAYKKtTU 📖: https://t.co/Newhxe6lI6 w/@__gzhu__ @CasebeerJonah @BergKirkpatrick @McAuleyLabUCSD @NicholasJBryan 🧵
0
1
22