Keunwoo Choi Profile Banner
Keunwoo Choi Profile
Keunwoo Choi

@keunwoochoi

Followers
5,638
Following
807
Media
383
Statuses
4,154
Explore trending content on Musk Viewer
Pinned Tweet
@keunwoochoi
Keunwoo Choi
1 month
hi music people, i wrote a tutorial on large language models and music information retrieval. of course it's called.. LLMs <3 MIR 🥁 have fun!
0
24
176
@keunwoochoi
Keunwoo Choi
1 year
whoa, this is bigger than ChatGPT to me. google almost solved music generation, i'd say.
152
1K
6K
@keunwoochoi
Keunwoo Choi
5 years
initialize yourself with it!
Tweet media one
2
303
602
@keunwoochoi
Keunwoo Choi
1 year
One of the key models in MusicLM is SoundStream, an audio codec. It made vocoders obsolete; and reshaped audio generation as a token prediction task. SS is not open to public, but a similar neural audio codec Encodec is completely open-source →
7
48
400
@keunwoochoi
Keunwoo Choi
7 years
I won the best paper award in #ismir2017 !!! Feeling honoured!!!! Thanks for co-authors @markbsandler György Fazekas @kchonyc
Tweet media one
21
20
263
@keunwoochoi
Keunwoo Choi
1 year
really well done, from SoundStream and AudioLM through MuLan to MusicLM 👏👏 the overall structure of MusicLM = MuLan + AudioLM = MuLan + w2v-BERT + SoundStream
Tweet media one
2
21
252
@keunwoochoi
Keunwoo Choi
1 year
hi all, here's an academic proof that AI has peaked in 2021 and started to downturn by 1.346% in 2022. diff = np.log(np.exp((1 - 138490 / 140380)))
Tweet media one
Tweet media two
17
8
136
@keunwoochoi
Keunwoo Choi
1 year
MuLan is a text-music joint embedding model. - contrastive training - 44M music audio - text description pairs from "internet music videos" *cough cough* youtube *cough cough* - AST: audio spectrogram transformer
Tweet media one
5
11
135
@keunwoochoi
Keunwoo Choi
2 years
DALL·E mini - "spectrogram of rock music"
Tweet media one
13
9
129
@keunwoochoi
Keunwoo Choi
4 years
Last Friday was my last day of the two years at Spotify. I started to work at ByteDance AI Research from today. (At Mountain View (California) in principle, but joined remotely from NYC)
12
0
123
@keunwoochoi
Keunwoo Choi
2 years
I left ByteDance last Friday. It was such a 1.8 year ❤️ (base-12) I'm glad I got what I wanted - a novel and intense learning experience. I shipped quite a few stuff, worked on research back-end tools, and made some research impact. Now, time to move on :)
5
2
117
@keunwoochoi
Keunwoo Choi
6 months
🌱 We’re hiring 2024 summer research interns on LLMs for drug discovery and biomedical applications. Join me, @stephenrra , @kchonyc , and other amazing people at NYC to work on the LLM product development of @PrescientDesign , @genentech ✨ Details:
0
21
115
@keunwoochoi
Keunwoo Choi
5 years
nnAudio: #pytorch CQT layers + etc. Done by Kin Wai Cheuk et al. And yes, it’s fast.
Tweet media one
Tweet media two
3
25
106
@keunwoochoi
Keunwoo Choi
2 years
🥳 PROPOSAL: Foley Sound Synthesis Challenge 🥳 There are enough challenges out there for speech and music. We propose one for "the other" kind of audio -> sound. Or effects. Or, Foley. We need to define the problem, dataset, and eval scheme. How? 🧵🧶
9
19
105
@keunwoochoi
Keunwoo Choi
8 months
I summarized the difference between `tokenizers.Tokenizer`, `transformers.PreTrainedTokenizer`, and `transformers.PreTrainedTokenizerFast`. I even made a github repo just to post this.
1
18
101
@keunwoochoi
Keunwoo Choi
6 years
Tweet media one
2
6
96
@keunwoochoi
Keunwoo Choi
2 years
Ahem, ahem. : I joined Gaudio Lab to - i'd dare say - pioneer some audio/music AI! 🥳 I'm excited more than ever :D Oh, and I'll visit Seoul more often. Friends in 🇰🇷, catch up soon!
5
2
99
@keunwoochoi
Keunwoo Choi
2 years
All you need is AI and music -- I'm giving a guest lecture today at NYU, Center for Data Science. Stay tuned for the recording and slides :)
Tweet media one
6
6
99
@keunwoochoi
Keunwoo Choi
1 year
+ they released MusicCaps dataset (5521 music-text pair) which they used as an eval set. .
6
7
94
@keunwoochoi
Keunwoo Choi
7 months
THIS IS BIG! All the music folks in Google Deepmind focus on one thing: AI music generation while NOT exploiting artists. Nothing is perfect, there're probably still some holes in giving the credit, but this is better than anything ever for very sure.
@demishassabis
Demis Hassabis
7 months
Thrilled to share #Lyria , the world's most sophisticated AI music generation system. From just a text prompt Lyria produces compelling music & vocals. Also: building new Music AI tools for artists to amplify creativity in partnership w/YT & music industry
111
537
3K
2
8
93
@keunwoochoi
Keunwoo Choi
1 year
New AI music model alert! yes, again 🎉 #SingSong , another music generation model by Google; @chrisdonahuey et al. Ok let me do another run for collecting followers. How does it work?
1
11
90
@keunwoochoi
Keunwoo Choi
10 months
the “llama moment” has come to audio research today! i can’t even imagine what we’ll see out of AudioCraft. whatever you work on in music/audio, do consider using it, as much as you can. if you don’t know what to do, think what you can do with it and get a head start.
@AIatMeta
AI at Meta
10 months
Today we're sharing details on AudioCraft, a new family of generative AI models built for generating high-quality, realistic audio & music from text. AudioCraft is a single code base that works for music, sound, compression & generation — all in the same place. More details ⬇️
40
535
2K
4
11
92
@keunwoochoi
Keunwoo Choi
4 years
If you belong to an underrepresented group in any sense (gender, race, nationality, financial situation, etc) and need some help on any MIR issues, please just contact me. gnuchoi at the-email-starting-with-G-you-know-what-I-mean😉
4
12
88
@keunwoochoi
Keunwoo Choi
6 years
Hi all, I'm happy to twit-announce that I'm joining 🎧 Spotify NYC from June! 😀
13
7
83
@keunwoochoi
Keunwoo Choi
5 years
Hi people! Me and @kchonyc 's #ismir2019 paper, "Deep Unsupervised Drum Transcription" aka 🥁 DrummerNet is here. Paper --> Blog post --> Supplementary material -->
2
18
79
@keunwoochoi
Keunwoo Choi
3 years
📄+📄+📄+📄+📄+📄+📄= 7 papers 🔥MIR researchers at ByteDance (SAMI team) made 7 papers accepted to #ISMIR2021 🔥 🧵I'll introduce them here one by one :)👇
1
6
79
@keunwoochoi
Keunwoo Choi
1 year
to recap, i find the whole roadmap really, really brilliant. - because there's MuLan, they could use audio-only dataset. - because there's SoundStream, the music generation task was simplified to token generation, not waveform generation.
3
3
78
@keunwoochoi
Keunwoo Choi
2 years
NeurIPS review complete award continues. Sponsor: @kchonyc
Tweet media one
1
1
78
@keunwoochoi
Keunwoo Choi
5 months
i'm teaching a class about AI at NYU, Spring 2024. it's "Deep Learning for Media", a course about AI for audio and visual contents. oof, i thought i became an LLM person. (it's not a job change, i'm covering one class this semester) happy to find back a nyu dot edu account!
5
2
69
@keunwoochoi
Keunwoo Choi
1 year
Ok now (restrospectively, on high-level) it's kinda simple. given an training item: - extract MuLan tokens (M), extract w2v-BERT (S), SS tokens (A) - train model for M → S. - train model for [M;S] → A both done by decoder-only transformers.
Tweet media one
1
4
68
@keunwoochoi
Keunwoo Choi
2 years
GSEP - Gaudio Source Separation 🔥🔥🔥
6
28
66
@keunwoochoi
Keunwoo Choi
11 months
👋 I joined @PrescientDesign recently. I distracted @kchonyc with music research circa 2016-2019. This time he offered me to join his realm -- languages! I'm already having a lot of fun, knowing more to come.
6
2
65
@keunwoochoi
Keunwoo Choi
2 years
Tweet media one
0
2
63
@keunwoochoi
Keunwoo Choi
3 years
MT3: Multi-Task Multitrack Music Transcription T5, but for music transcription. A neat solution to cope with many-but-small existing datasets.
Tweet media one
Tweet media two
Tweet media three
Tweet media four
1
6
62
@keunwoochoi
Keunwoo Choi
1 year
ChatGPT blew up because people *hate* writing 😂
3
2
61
@keunwoochoi
Keunwoo Choi
9 months
🎙 Let's talk about AI research. And datasets. Accessibilities. Opportunities. Music.
Tweet media one
1
6
61
@keunwoochoi
Keunwoo Choi
3 years
<shameless as always> my papers are 1st and 6th most cited ISMIR paper in the last 5 years!🔥🔥 heard it was mentioned at the #ismir2021 trivia organized by the titans @r4b1tt @urinieto . i think they should arXiv the trivia and cite my paper thx
Tweet media one
5
2
61
@keunwoochoi
Keunwoo Choi
1 year
AudioLM = w2v-BERT + SoundStream w2v-BERT is.. - a BERT, but for audio. originally for speech. in AudioLM, an intermediate layer from speech-pretrained model was used. - it's "coarse" (250bps of bitrate.) - it takes care of semantic information.
1
0
58
@keunwoochoi
Keunwoo Choi
4 years
ByteDance/TikTok is hiring research scientists and software developers around music information retrieval and music/audio signal processing at Mountain View, US. Please hit me up! #ismir2020
1
12
57
@keunwoochoi
Keunwoo Choi
4 months
we're hiring AI/LLM engineers! - covering both pre-training and post-training tasks - purely for product development, based on *extensive understanding in LLMs* - with real-world impacts on drug discovery in Genentech - no publication within sight
1
12
55
@keunwoochoi
Keunwoo Choi
1 year
do you know what ChatGPT can't do? 🔊 audio generation. we do, at Gaudio Lab 😉
2
10
54
@keunwoochoi
Keunwoo Choi
1 year
SoundStream is.. - a neural audio codec. - residual vector quantizer (RVQ) is used - as a codec, it's "fine-grained" (2000bps of bitrate)
2
0
53
@keunwoochoi
Keunwoo Choi
5 years
@urinieto ROCKING #ismir2019 HAHAHAHAHAHA 😂😂😂 seriously, my every follower should watch this otherwise please unfollow thanks.
0
12
53
@keunwoochoi
Keunwoo Choi
6 years
What would you say if I passed the PhD viva today? I mean, I did, so feel free to really say it!
22
0
52
@keunwoochoi
Keunwoo Choi
3 years
my code was more interesting 4+ years ago.
Tweet media one
3
0
52
@keunwoochoi
Keunwoo Choi
5 years
Frrquency-aware CNNs. Ooops I was working on the same thing last summer but had no time after some experiments. It worked for music classification and source separation. Go try this!
Tweet media one
4
3
51
@keunwoochoi
Keunwoo Choi
1 year
🎉 It's happening. Foley Sound Synthesis Challenge! Generative AI folks, join us and make some sound! 🔊
1
14
50
@keunwoochoi
Keunwoo Choi
2 years
ISMIR2022 tutorials are out! 👉
1
11
49
@keunwoochoi
Keunwoo Choi
3 years
We're looking for a junior-level MIR researcher (perhaps Master or PhD) in Shanghai; to work with me on music tagging and related problems. Expecting to hire ASAP. Please email me if you're interested!
2
12
48
@keunwoochoi
Keunwoo Choi
4 years
It seems clear to me that Tensorflow developers are not deeply understanding why researchers struggle with their product. Life is too short for most of researchers to be very good at all Python and machine learning. TF adds another burden, but Pytorch doesn't.
4
11
49
@keunwoochoi
Keunwoo Choi
1 year
in the training set, no text label is needed because we.. i mean, googlers.. have pre-trained MuLan! also, if you believe the power neural codec, SoundStream, no need to trained end-to-end with waveforms etc! SoundStream tokens are good enough!
1
0
49
@keunwoochoi
Keunwoo Choi
7 years
U-net with 33 lines with #keras . How can I not love this API?
Tweet media one
1
8
44
@keunwoochoi
Keunwoo Choi
6 years
your code vs my code 😎 #swag
Tweet media one
2
10
47
@keunwoochoi
Keunwoo Choi
3 years
TikTok🎶 is hiring a research scientist in Music/ML @🇬🇧 London office 🔥 Join our SAMI team to work on Speech, Audio, and Music intelligence with us :) Please feel free to reach out to me for any question 📧
0
6
46
@keunwoochoi
Keunwoo Choi
5 years
“Academic conference in conputer science” or rather just a cult #ismir2019
2
4
45
@keunwoochoi
Keunwoo Choi
1 year
inference is straightforward. do the same with the training stage except - use MuLan text model, because we want *text*-to-music. - after SoundStream tokens are predicted, feed them to SS decoder to generated audio.
Tweet media one
1
0
45
@keunwoochoi
Keunwoo Choi
2 years
it's official. #ismir2023 in Milano, Italy! 🎉
1
1
45
@keunwoochoi
Keunwoo Choi
3 years
*QUITE A FEW* papers are accepted to #ismir2021 from our team in ByteDance 🚀🚀🚀🚀🚀🚀🚀 I'll share more details once the proceedings are updated. And yes we're hiring 🔥🔥🔥🔥🔥🔥🔥
0
2
45
@keunwoochoi
Keunwoo Choi
3 years
Sheet Sage: Lead sheets from music audio Leverage Jukebox for melody extraction. Who'd submit this level of amazing work simply to late-breaking/demo session? This guy → @chrisdonahuey
Tweet media one
1
6
44
@keunwoochoi
Keunwoo Choi
3 years
DrummerNet Gangs (K Choi and K Cho)
Tweet media one
1
0
43
@keunwoochoi
Keunwoo Choi
3 years
amazing, amazing. done by @ethanmanilow @pseetharaman et al.
@_akhaliq
AK
3 years
Unsupervised Source Separation By Steering Pretrained Music Models abs:
Tweet media one
3
29
159
2
5
43
@keunwoochoi
Keunwoo Choi
1 year
#ISMIR 2022 Tutorials are all online! Help yourself to a cup of music AI :)
Tweet media one
0
6
42
@keunwoochoi
Keunwoo Choi
8 months
🚨 We have a MLE position open at @PrescientDesign to find a strong engineer to make our language models stronger.
2
12
39
@keunwoochoi
Keunwoo Choi
2 years
c4dm folks won the #ismir2022 best paper award!! 🎉🥳🎊 amazing! congrats, @liulelecherie @QiuqiangK @veromorfi @emmanouilb !
Tweet media one
0
2
40
@keunwoochoi
Keunwoo Choi
3 years
Long time no first-authoring! Listen, Read, and Identify network (LRID-Net) identifies singing language by reading the metadata (title, album, artist) and listening to the audio.
1
4
38
@keunwoochoi
Keunwoo Choi
10 months
Our paper about DCASE Challenge T7 - Foley Sound Synthesis was accepted to the DCASE Workshop 🥳 I can't make it to Finland🇫🇮, but some of the authors will be there to tell you what we went through while organizing the first generative challenge at DCASE.
0
6
39
@keunwoochoi
Keunwoo Choi
3 years
GPT-3 is so 2020. saw it on the way @kchonyc ’s place. this must be a sign..
Tweet media one
2
2
39
@keunwoochoi
Keunwoo Choi
2 months
i'm giving an introductory talk about LLMs for drug discovery at #ASCPT2024 pre-conference soon.
Tweet media one
Tweet media two
Tweet media three
Tweet media four
2
8
38
@keunwoochoi
Keunwoo Choi
1 year
another day, another music generation paper! a diffusion one this time. i’m very curious where they got the training data 🤔
@_akhaliq
AK
1 year
Noise2Music, where a series of diffusion models is trained to generate high-quality 30-second music clips from text prompts project page:
11
83
496
2
1
38
@keunwoochoi
Keunwoo Choi
3 years
ByteDance 🚀 US Speech / Audio / Music research team is extensively hiring research scientists. If you’re a graduating PhD this year, don’t wait and just DM me! 🔥🔥
3
12
37
@keunwoochoi
Keunwoo Choi
9 months
new music AI model alert 🚨 get your music tracks segmented by @taejun_kim_
@taejun_kim_
Taejun Kim
9 months
Music Structure Analyzer Released ✨ [Python Package] [Paper] [Interactive Demo] [Hugging Face Space]
11
55
264
2
3
38
@keunwoochoi
Keunwoo Choi
3 years
DawDreamer: A Python-interfaced DAW. Yeah we can do lot of things with this.
@DoItRealTime
David Braun
3 years
DawDreamer has gained many features recently including pip install. A new notebook shows how to load Ableton warp marker files like this video. Faust integration enables custom polyphonic instruments. Hopefully very useful for ML researchers and artists.
0
0
14
1
4
38
@keunwoochoi
Keunwoo Choi
27 days
teaching "deep learning for media" at NYU was super fun! now, let me disseminate my students' final projects. these are really cool stuff. they somehow made it in the vary last minute. i swear none of these was at this level just one week before 😂 anyways, 🧵 starts -
2
1
36
@keunwoochoi
Keunwoo Choi
2 years
Try it yourself our music source separation! 🚨 ALERT: The performance might be way too good.
@keunwoochoi
Keunwoo Choi
2 years
BTS - Dynamite, source separated by Gaudio™️
4
2
16
4
7
35
@keunwoochoi
Keunwoo Choi
7 months
look how shamelessly i'm included here! as always, it was great to connect to all the great researchers in MACLab supervised by @juhan_nam at @ISMIRConf .
@havenpersona
Haven Kim
7 months
This year, people from the Music and Audio Computing Lab at KAIST, led by @juhan_nam , participated in the @ISMIRConf , and presented our work through scientific programs, late-breaking demos and music sessions!
Tweet media one
1
3
35
1
0
35
@keunwoochoi
Keunwoo Choi
6 months
Big news in AI this week - Mistral 7B on torrent - Google Gemini - and.. - my first single album <unspoken serenity> released;
2
0
35
@keunwoochoi
Keunwoo Choi
2 years
import tensorflow as plt
2
1
35
@keunwoochoi
Keunwoo Choi
1 year
DCASE Task 7 - Foley Sound Synthesis has finished. It was the very first generative audio AI challenge. I'm very happy to have organized such a successful event! 🎉
1
2
35
@keunwoochoi
Keunwoo Choi
1 year
@rrherr And MuLan was already Google-verse only 😢
2
0
34
@keunwoochoi
Keunwoo Choi
5 years
The longest ever video of me talking public has become public. "Deep Learning with Audio Signals: Prepare, Process, Design, Expect" in @QConAI . In case me tweeting around you isn't enough.
1
5
35
@keunwoochoi
Keunwoo Choi
3 months
generative AI audio is here to stay.. and prosper! check out this year's challenge. T7. Sound Scene Synthesis #DCASE2024
Tweet media one
Tweet media two
0
2
33
@keunwoochoi
Keunwoo Choi
4 months
are you an LLM nerd who can understand ML/language model papers and write good code? 👀
@keunwoochoi
Keunwoo Choi
4 months
we're hiring AI/LLM engineers! - covering both pre-training and post-training tasks - purely for product development, based on *extensive understanding in LLMs* - with real-world impacts on drug discovery in Genentech - no publication within sight
1
12
55
2
4
32
@keunwoochoi
Keunwoo Choi
5 years
The #ismir2019 poster repo is hosting 25 posters and 38-starred now. Would you please 'Like' this tweet if you've ever been the repo and seen any posters there? I wanna know its impact. Thanks!
0
6
33
@keunwoochoi
Keunwoo Choi
5 years
After like 3 months of experiments (with some progress) I just realised out of N layers, good three of them didn't have an activation function at all.
5
1
32
@keunwoochoi
Keunwoo Choi
3 years
"Building the MetaMIDI Dataset: Linking Symbolic and Audio Musical Data" Hell a lot of midi files and matched audio clips. #ismir2021
Tweet media one
1
4
31
@keunwoochoi
Keunwoo Choi
1 year
oo more text-to-music to come. this time, from academia!
@LiuHaohe
Haohe Liu
1 year
Can't wait to share our new Text-to-Audio model, AudioLDM. 😆 This video shows the generation result with a simple text prompt: "A music made by xxx". More demos coming soon!😉 The paper will be available next Monday on arXiv! 😊 Our model will be open-sourced soon!😎
27
99
613
2
6
31
@keunwoochoi
Keunwoo Choi
1 year
so are spectrograms just images???
@_akhaliq
AK
1 year
Riffusion, real-time music generation with stable diffusion @huggingface model: project page:
Tweet media one
64
626
3K
6
2
31
@keunwoochoi
Keunwoo Choi
6 years
Um, Spotify will definitely hire 2019 summer research interns for some fun MIR works, so please stay tuned! (i.e. don't say yes to others too soon 😎)
3
2
30
@keunwoochoi
Keunwoo Choi
2 years
"Visualization for AI-Assisted Composing" shows lots of cool ideas! absolutely helpful, i'd say.
Tweet media one
Tweet media two
2
3
31
@keunwoochoi
Keunwoo Choi
4 years
Ok I still get an ISMIR-high even virtually.
2
1
30
@keunwoochoi
Keunwoo Choi
2 months
two sides of making music. (a) manufacturing music (b) expressing creativity through music i see prompting music Gen AI - to get the final, whole audio - purely as (a), which is totally fine as long as its training is done legally.
5
3
31
@keunwoochoi
Keunwoo Choi
4 years
I've been an audio person for 10+ years. Let me tell you - you don't need 192/24 or anything. If you don't like the audio quality from any legit music streaming service, it's NOT about the codec. get a better connection, quieter place, better earbuds.
4
6
31