json_yim Profile Banner
Jason Yim Profile
Jason Yim

@json_yim

Followers
2K
Following
624
Media
21
Statuses
317

PhD student @MIT_CSAIL. Generative models, protein design. 🦋 Bluesky handle: https://t.co/MPlEjog02q On X until the exodus is complete.

Cambridge, MA
Joined September 2017
Don't wanna be here? Send us removal request.
@json_yim
Jason Yim
2 years
Combining discrete and continuous data is an important capability for generative models. To address this for protein design, we introduce Multiflow, a generative model for structure and sequence generation. Preprint: https://t.co/wuj9l5sTLc Code: https://t.co/IwIoC74Odm 1/8
2
99
468
@Jaeyeon_Kim_0
Jaeyeon (Jay) Kim
2 days
🚨🚨🚨 Now your Masked Diffusion Model can self-correct! We propose PRISM, a plug-and-play approach fine-tuning method that adds self-correction ability to any pretrained MDM! (1/N)
6
42
254
@zhuci19
Cai Zhou @EMNLP2025
30 days
(1/5) Beyond Next-Token Prediction, introducing Next Semantic Scale Prediction! Our @NeurIPSConf NeurIPS 2025 paper HDLM is out! Check out the new language modeling paradigm: Next Semantic Scale Prediction via Hierarchical Diffusion Language Models. It largely generalizes
7
55
343
@Jaeyeon_Kim_0
Jaeyeon (Jay) Kim
1 month
We introduce a new ''rule'' for understanding diffusion models: Selective Underfitting. It explains: 🚨 How diffusion models generalize beyond training data 🚨 Why popular training recipes (e.g., DiT, REPA) are effective and scale well Co-led with @kiwhansong0! (1/n)
8
64
419
@json_yim
Jason Yim
1 month
It's odd the performance is significantly worse than the AR base model? Starting with a much powerful AR model, dropping the performance just enough to beat all other diffusion LLMs, and then saying it's better than them is weird...
1
0
16
@peholderrieth
Peter Holderrieth
1 month
New work: “GLASS Flows: Transition Sampling for Alignment of Flow and Diffusion Models”. GLASS generates images by sampling stochastic Markov transitions with ODEs - allowing us to boost text-image alignment for large-scale models at inference time! https://t.co/unsuG3mYer [1/7]
4
60
248
@msalbergo
Michael Albergo
1 month
We've cleaned up the story big time on flow maps. Check out @nmboffi's slick repo implementing all the many ways to go about them, and stay tuned for a bigger release 🤠 https://t.co/7WygKSpbZP https://t.co/Juucy5l844
@nmboffi
Nicholas Boffi
1 month
Consistency models, CTMs, shortcut models, align your flow, mean flow... What's the connection, and how should you learn them in practice? We show they're all different sides of the same coin connected by one central object: the flow map. https://t.co/QBp1kELVhF 🧵(1/n)
2
22
132
@AndrewC_ML
Andrew Campbell
1 month
Very excited to share our preprint: Self-Speculative Masked Diffusions We speed up sampling of masked diffusion models by ~2x by using speculative sampling and a hybrid non-causal / causal transformer https://t.co/6e37sx8Cbu w/ @ValentinDeBort1 @thjashin @ArnaudDoucet1
2
39
186
@bose_joey
Joey Bose
4 months
🎉Personal update: I'm thrilled to announce that I'm joining Imperial College London @imperialcollege as an Assistant Professor of Computing @ICComputing starting January 2026. My future lab and I will continue to work on building better Generative Models 🤖, the hardest
98
34
625
@RickyTQChen
Ricky T. Q. Chen
6 months
We've open sourced Adjoint Sampling! It's part of a bundled release showcasing FAIR's research and open source commitment to AI for science. https://t.co/6oBTnael8p https://t.co/rYmJ02KguC
Tweet card summary image
github.com
code for "Adjoint Sampling: Highly Scalable Diffusion Samplers via Adjoint Matching" - facebookresearch/adjoint_sampling
@AIatMeta
AI at Meta
6 months
Announcing the newest releases from Meta FAIR. We’re releasing new groundbreaking models, benchmarks, and datasets that will transform the way researchers approach molecular property prediction, language processing, and neuroscience. 1️⃣ Open Molecules 2025 (OMol25): A dataset
1
23
117
@AlexanderTong7
Alex Tong
7 months
#FPIworkshop best paper award goes to @peholderrieth @msalbergo and Tommi Jaakkola. Congrats and great talk Peter!
0
9
55
@menghua_wu
Rachel (Menghua) Wu
7 months
I won't be at ICLR 🥲 but you can talk to these other cool people at my poster, Thursday 3-5:30 PM in Hall 3+2B #10!
@menghua_wu
Rachel (Menghua) Wu
8 months
Excited to share my #ICLR2025 paper, with JC Hütter and friends! Genetic perturbation screens allow biologists to manipulate and measure the genes in cells = discover causal relationships! BUT they are expensive to run, expensive to interpret. ... We use LLMs to help!
0
3
26
@JacobMGershon
Jacob Gershon
7 months
Had fun exploring guidance for backbone designability within this latent framework, excited to chat more about guidance with experimental data @gembioworkshop ICLR
@json_yim
Jason Yim
7 months
I'll be at the ICLR @gembioworkshop workshop presenting latent and structure diffusion for protein backbone generation. Come by to talk all things latent for biology. https://t.co/r3Fqg47NZT https://t.co/SkdJ82syqA
0
3
14
@json_yim
Jason Yim
7 months
I'll be at ICLR. Come check out our generative modeling work! Reach out if you want to chat. Proteina: https://t.co/ZZXeekrfqp Protcomposer: https://t.co/Ijb7d1VQ6p Generator matching:
@peholderrieth
Peter Holderrieth
1 year
New paper out! We introduce “Generator Matching” (GM), a method to build GenAI models for any data type (incl. multimodal) with any Markov process. GM unifies a range of state-of-the-art models and enables new designs of generative models. https://t.co/6BTkr3ukYc (1/5)
1
12
56
@json_yim
Jason Yim
7 months
RFdiffusion => generative binder design. RFdiffusion2 => generative enzyme design. It's rare to find scientists with deep knowledge in chemistry, machine learning, and software engineering like Woody. The complexity of enzymes matches the complexity of his skills. Check out RFD2
@woodyahern
Woody Ahern
7 months
New enzymes can unlock chemistry we never had access to before. Here, we introduce RFdiffusion2 (RFD2), a generative model that makes significant strides in de novo enzyme design. Preprint: https://t.co/cAWGkrSxBo Code: coming soon Animation credit: https://t.co/Th9ZjsYeX2 (1/n)
1
33
146
@woodyahern
Woody Ahern
7 months
New enzymes can unlock chemistry we never had access to before. Here, we introduce RFdiffusion2 (RFD2), a generative model that makes significant strides in de novo enzyme design. Preprint: https://t.co/cAWGkrSxBo Code: coming soon Animation credit: https://t.co/Th9ZjsYeX2 (1/n)
13
76
260
@ChoYehlin
Yehlin Cho
7 months
Excited to share our preprint “BoltzDesign1: Inverting All-Atom Structure Prediction Model for Generalized Biomolecular Binder Design” — a collaboration with @MartinPacesa, @ZhidianZ , Bruno E. Correia, and @sokrypton. 🧬 Code will be released in a couple weeks
15
114
450
@HWaymentSteele
Hannah Wayment-Steele
8 months
Protein dynamics was the first research to enchant me >10yrs ago, but I left in PhD bc I couldnt find big expt data to evaluate models. Today w @ginaelnesr, I'm thrilled to share the big dynamics data I've been dreaming of, and the mdl we trained: Dyna-1. https://t.co/ZawXaGdCVw
4
37
170
@json_yim
Jason Yim
8 months
Combining prediction, generation, and modalities (sequence, structure, nucleic acids, small molecules, proteins) is the future. Congrats to the authors! Looking forward to the technical report.
@vant_ai
VantAI
8 months
Announcing Neo-1: the world’s most advanced atomistic foundation model, unifying structure prediction and all-atom de novo generation for the first time - to decode and design the structure of life 🧵(1/10)
1
8
74
@chaitjo
Chaitanya K. Joshi
8 months
Introducing All-atom Diffusion Transformers — towards Foundation Models for generative chemistry, from my internship with the FAIR Chemistry team @OpenCatalyst @AIatMeta There are a couple ML ideas which I think are new and exciting in here 👇
4
106
529