
Edward Z. Yang
@ezyang
Followers
14K
Following
1K
Media
128
Statuses
9K
I work on PyTorch at Meta. Chatty alt at @difficultyang.
Edison, NJ
Joined May 2008
I finally sat down and wrote down a post mortem for vibe coding ScubaDuck. It's aimed at those of you who have never tried vibe coding (in its original sense: AI coding without reviewing the code the AI generated)
7
12
165
On vacation, vibe coding a program that uses XLA to generate a redistribute plan from one shard placement to another, and then reinterprets it with jax.lax:
github.com
Extract redistribution plans from XLA. Contribute to ezyang/xla-redist-ref development by creating an account on GitHub.
2
2
62
RT @SingularMattrix: @ezyang Slightly streamlined version (did you have some reason for doing explicit device_gets?):. .
gist.github.com
GitHub Gist: instantly share code, notes, and snippets.
0
1
0
None of the LLMs can solve it. Can you?.
gist.github.com
GitHub Gist: instantly share code, notes, and snippets.
4
1
14
Peter Hawkins points out to me that this is controlled by xla_allow_excess_precision, which JAX enables by default. So indeed, torch.compile and JAX/XLA both do the same thing by default.
Hey TL, hope you can answer this for me: torch.compile does a thing where if it fuses several bfloat16 operations together, it will do the internal compute in float32 without wasting cycles clamping the intermediates to bfloat16. Does JAX do this by default?.
1
0
22
RT @giffmana: @ezyang @tenderizzation @gaunernst Yeah back to the point about big_vision, on TPUs matmuls have always been done on the MXU….
0
1
0
RT @GrantSlatton: now that i can use codex with my chatgpt sub, i'm gonna cancel claude max for a month and see if i can get by with gpt5 i….
0
1
0
RT @OpenAIDevs: We’re also releasing v0.16 of the Codex CLI today. - GPT-5 is now the default model.- Use with your ChatGPT plan.- A new,….
0
365
0