mikemccabe210 Profile Banner
Mike McCabe Profile
Mike McCabe

@mikemccabe210

Followers
320
Following
128
Media
30
Statuses
66

Machine learning/scientific computing researcher @flatironinst. he/him. Enjoys running, climbing, and pictures of really well camouflaged animals.

New York, NY
Joined November 2009
Don't wanna be here? Send us removal request.
@mikemccabe210
Mike McCabe
13 days
Looking forward to @NeurIPSConf next week! Feel free to reach out to chat about physical dynamics, surrogate modeling, hybrid ML-numerical methods, or large-scale training! Also seeking burrito recs...
1
0
4
@rdMorel
Rudy Morel
18 days
Partial observability is a key challenge in predicting physical systems, where only part of the state is observed. Check out our poster #2213 at #neurips2025 on Thu, Dec 4, 4:30pm! We propose a multiscale inference scheme for diffusion models to better predict these systems.
1
16
72
@mikemccabe210
Mike McCabe
23 days
@PolymathicAI 11/ This was a massive collaborative effort that wouldn't have been possible without support from @FlatironInst , @schmidtsciences, and @nvidia for through the NAIRR program. We hope you all enjoy World Walrus Day (11/24) and are looking forward to seeing what Walrus unleashes!
0
0
7
@mikemccabe210
Mike McCabe
23 days
@PolymathicAI 10/ We're super excited about these results, but also want to speak to the biggest current limitations. - Walrus is still trained with a grid-based encoder/decoder. - Even strong eye-norm results still require scientific validation.
1
0
7
@mikemccabe210
Mike McCabe
23 days
@PolymathicAI 9/ This push for physical diversity matters. We show that even though specialization seems to help on pretraining metrics, heavy augmentation favoring representational diversity is key to strong downstream results.
1
0
6
@mikemccabe210
Mike McCabe
23 days
@PolymathicAI 8/ These techniques let us train a single model on 19 physical settings, outperforming the landmark foundation models in the space both on 8 downstream transfer tasks and showing consistent performance across simulations drawn from multiple domains.
1
0
6
@mikemccabe210
Mike McCabe
23 days
@PolymathicAI 7/ Co-designed Training strategy FSDP is an amazing tool, but the syncs at AllGather ops can result in huge deadweight for heterogeneous loads. We co-design our sampling and distribution strategies to limit the impact these differences have resulting in 262% higher throughput.
1
1
6
@mikemccabe210
Mike McCabe
23 days
@PolymathicAI 6/ Joint 2D/3D Augmentation When we train, we want to learn general patterns, but the model wants to learn short-cuts. This makes it difficult to learn from 2D and 3D. We employ a new tensor-aware augmentation strategy that forces the model to share information across spaces.
2
1
8
@mikemccabe210
Mike McCabe
23 days
@PolymathicAI 5/ Adaptive Compute Fixed resolution becomes a huge obstacle handling this level of diversity. Big patches help us handle big data, but hurt performance on coarse data. But what if we dynamically choose our internal resolution instead using convolutional stride modulation?
1
1
8
@mikemccabe210
Mike McCabe
23 days
@PolymathicAI 4/ Patch jittering Physical models shouldn't blow up, right? One reason this happens is the loss of equivariance from resampling. We mitigated this by making randomizing resampling with a simple inference-time augmentation that improves long run metrics on 17/19 PT systems.
1
1
10
@mikemccabe210
Mike McCabe
23 days
@PolymathicAI 3/ Scale is a big part, but not the only part. This is super heterogeneous data. We can't just throw it at a model and expect it to work well. We need to find new approaches for handling some of the biggest challenges in this space!
1
0
13
@mikemccabe210
Mike McCabe
23 days
@PolymathicAI 2/ Walrus sets new standards for scale, accuracy, and pretraining diversity in foundation models for physical dynamics. It's a 1.3B parameter model trained on 19 scenarios across fluids, plasmas, acoustics, viscoelastics, and more utilizing 63 different fields in both 2D and 3D.
1
0
10
@mikemccabe210
Mike McCabe
23 days
1/ Today with my colleagues @PolymathicAI, I'm excited to release our latest project, Walrus, a cross-domain foundation model for physical dynamics, into the world. https://t.co/ihv1MZGQM3 Paper: https://t.co/d6ah9LO4ud Git: https://t.co/s3p8qGhZQR HF: https://t.co/RufaBD9eJk
5
109
602
@mikemccabe210
Mike McCabe
3 months
Awesome work by @FrancoisRozet from his time @PolymathicAI! Some great insight into the compression behavior of LDMs for physical dynamics plus some cool tricks for boosting performance!
@FrancoisRozet
François Rozet
3 months
Does a smaller latent space lead to worse generation in latent diffusion models? Not necessarily! We show that LDMs are extremely robust to a wide range of compression rates (10-1000x) in the context of physics emulation. We got lost in latent space. Join us 👇
0
0
4
@NYUDataScience
NYU Center for Data Science
9 months
CDS Research Scientist @cosmo_shirley, with @oharub & @mikemccabe210, led the creation of "The Well"—a 15TB dataset for physics-based AI. The Well offers simulations from biological systems to neutron stars, helping AI models learn physical principles. https://t.co/AmPIUElueJ
Tweet card summary image
nyudatascience.medium.com
CDS Senior Research Scientist Shirley Ho has led the creation of The Well, a 15TB dataset designed to advance physics-based machine…
0
3
11
@mikemccabe210
Mike McCabe
1 year
Heading off to #NeurIPS2024 today! Feel free to reach out if you're around to catch up! If you're interested, please visit our poster sessions tomorrow at 11 PST for MPP ( https://t.co/JbOmq9ruXs) and on Thursday at 11 for the Well ( https://t.co/L3rPzlFXoT).
0
3
9
@mikemccabe210
Mike McCabe
1 year
This was a huge collaborative effort with too many contributors to list in one post, but it wouldn't have been possible without all of our teammates @PolymathicAI and elsewhere. We hope that access to more realistic problems pushes the development of #AI4Science forward!
0
0
0
@mikemccabe210
Mike McCabe
1 year
and many more! All data in the Well is stored in a uniform format and whether its 2D or 3D it's accessible through the same API. The Well additionally includes benchmarking tools such as baseline models, tensor-aware augmentations, and metric implementations:
1
0
0
@mikemccabe210
Mike McCabe
1 year
To complex newly understood non-Newtonian turbulent structures:
1
0
0
@mikemccabe210
Mike McCabe
1 year
to orientation-dependent forcings in biological systems:
1
0
1