jwuphysics Profile Banner
John F. Wu Profile
John F. Wu

@jwuphysics

Followers
2K
Following
13K
Media
609
Statuses
3K

Associate Astronomer and Applied AI @SpaceTelescope. Secondary appointments in @JohnsHopkins Astro + CS. Opinions my own. He/him. 🦋 @jwuphysics.bsky.social

Baltimore, MD, USA
Joined June 2020
Don't wanna be here? Send us removal request.
@jwuphysics
John F. Wu
14 days
Sparse Autoencoders (SAEs) can be a powerful "discovery engine" for astrophysics, helping us find interpretable features in modern galaxy surveys. Check it out -- new @NeurIPSConf ML4PS workshop paper, jointly authored by me and Mike Walmsley!
2
3
35
@jwuphysics
John F. Wu
7 hours
Lol Gemini couldn't figure out which emoji to put in the heading so it chose ​융 instead...
0
1
2
@jwuphysics
John F. Wu
9 hours
Goodfire is doing some really incredible stuff
@EricBigelow
Eric Bigelow
13 hours
📝 New paper! Two strategies have emerged for controlling LLM behavior at inference time: in-context learning (ICL; i.e. prompting) and activation steering. We propose that both can be understood as altering model beliefs, formally in the sense of Bayesian belief updating. 1/9
0
0
2
@math_rachel
Rachel Thomas
1 day
TensorFlow was all about making it easier for computers. PyTorch *won* because it was about making it easier for humans. It’s disappointing to see AI community focusing on what’s easiest for machines again (prioritizing AI agents & not centering humans). -- @jeremyphoward 1/
1
18
159
@jwuphysics
John F. Wu
1 day
normalized flux units, interesting choice
@sighyam
yammi
2 days
I’m going to be sick
0
0
2
@jwuphysics
John F. Wu
2 days
Dystopian shit
1
0
1
@jwuphysics
John F. Wu
3 days
Hey Google structured data team, I think your float_to_percentage is being applied twice...
0
0
4
@MarkMBissell
mark bissell
5 days
🚨 New research! 🚨 Very cool results showing a method for locating + characterizing + intervening on memorized data in models (of all kinds, e.g. language & vision) Highly recommend reading this thread, then @jack_merullo_ 's thread, then the full paper for a deep dive
@GoodfireAI
Goodfire
6 days
LLMs memorize a lot of training data, but memorization is poorly understood. Where does it live inside models? How is it stored? How much is it involved in different tasks? @jack_merullo_ & @srihita_raju's new paper examines all of these questions using loss curvature! (1/7)
0
2
12
@jwuphysics
John F. Wu
5 days
Y'all are sleeping on this one https://t.co/xQ0KHlCpJQ
0
0
1
@errai34
ioana ciucă
8 days
So excited for the super cool workshop Language AI in the Space Sciences exploring the role of natural language processing and artificial intelligence in space science and astronomy ✨ . The workshop will take place at the Space Telescope Science Institute from March 9th to 12th
Tweet card summary image
stsci.edu
1
2
12
@xeophon_
Xeophon
8 days
banger paper
@allen_ai
Ai2
8 days
Introducing OlmoEarth 🌍, state-of-the-art AI foundation models paired with ready-to-use open infrastructure to turn Earth data into clear, up-to-date insights within hours—not years.
0
6
51
@jwuphysics
John F. Wu
9 days
Thank you! If you want to read more, I wrote about this journey on my blog: https://t.co/ZDvCjouqR0 Although now I'm nearly four years in, so I may need to write an update!
jwuphysics.github.io
Reflecting on my journey towards becoming a tenure-track astronomer
@Amxquant
AM
9 days
@jwuphysics @jeremyphoward @fastdotai love hearing these stories, it's inspirational af
1
0
5
@jwuphysics
John F. Wu
10 days
Super common pathway. Towards the end of my PhD in Physics/Astro, I trained my first deep neural net on a GeForce 780 after watching @jeremyphoward on the @fastdotai lectures. Now I'm tenure-track faculty in (astro)physics, but still maintain strong curiosity in AI/ML
@Yuchenj_UW
Yuchen Jin
10 days
Met a Meta AI researcher. He studied Physics in China, came to the US for a PhD in Physics, and then fell in love with AI, despite never having studied computer science. He watched Andrej Karpathy and Andrew Ng, bought a GPU, read every arXiv paper title daily, and dived into
9
20
337
@jwuphysics
John F. Wu
10 days
Inside you there are two ML engineers
1
5
74
@jwuphysics
John F. Wu
11 days
Obscure question for @Ravens fans and journalists: I have this T-shirt that references "Splash '98 Funky Jungle" presented by the Baltimore Ravens. What event was this? Google search, ChatGPT turned up nothing...
1
0
0
@jwuphysics
John F. Wu
12 days
Okay first of all rude
0
0
3
@jwuphysics
John F. Wu
12 days
Wow those are some serious gains over SFT
@charles0neill
Charlie O'Neill
12 days
8/ The results are ridiculous. RGT hits 72-83% accuracy. Standard SFT needs 10x more data to reach the same performance. iSFT needs 1.5-1.7x more. We're teaching the model the causal structure of correctness, not hoping it reverse-engineers it from examples
1
1
6
@kirkby_max
Max Kirkby
13 days
We found a neat way to teach models the rules and not just the answers using strategy tokens, with implications for cleaner supervision and sample-efficiency gains. Fresh from @part_harry_ @charles0neill and us @parsedlabs!
@parsedlabs
parsed
13 days
We discovered that teaching models why answers are correct, not just what to output, dramatically improves training efficiency. By making latent strategies explicit during training (e.g., "don't infer diagnoses from medications"), we achieve the same performance with 10x fewer
0
1
6
@jwuphysics
John F. Wu
13 days
Okay what now
@AnthropicAI
Anthropic
14 days
New Anthropic research: Signs of introspection in LLMs. Can language models recognize their own internal thoughts? Or do they just make up plausible answers when asked about them? We found evidence for genuine—though limited—introspective capabilities in Claude.
0
0
2
@jwuphysics
John F. Wu
13 days
New blog post explaining our recent workshop paper. https://t.co/a19vmtOZAv
0
0
5
@jwuphysics
John F. Wu
14 days
Forgot to share all the public code! Github for SAEs: https://t.co/8RFEGXAuiS Euclid MAE + dataset: https://t.co/7sQekT1HFS Euclid MAE interactive demo:
Tweet card summary image
huggingface.co
@jwuphysics
John F. Wu
14 days
Sparse Autoencoders (SAEs) can be a powerful "discovery engine" for astrophysics, helping us find interpretable features in modern galaxy surveys. Check it out -- new @NeurIPSConf ML4PS workshop paper, jointly authored by me and Mike Walmsley!
0
0
3