
Henry Ko
@henryHM_ko
Followers
589
Following
171
Media
3
Statuses
41
performance and efficiency in ML | CS @ UC Berkeley, @BerkeleyML
Joined June 2024
RT @evanliin: looking for my next thing! thinking about dropping out. would love to learn more about opportunities within hardware accele….
0
22
0
RT @SemiAnalysis_: NVIDIA Tensor Core Evolution.From Volta To Blackwell.Amdahl’s Law, Strong Scaling.Asynchronous Execution.Blackwell, Hopp….
0
28
0
RT @Michael_J_Lutz: ksim is a JAX-based framework that makes your wackiest RL ideas simple to implement. Why use it? It's modular. Trying….
0
1
0
RT @thtrkim: (1/5).I’m pleased to share that my research with @seowondeog12052 has been accepted to RECOMB 2025 (Poster) and IEEE EMBC 2025….
0
3
0
RT @voidshapes: per the event description: Viren Jain is a Senior Staff Research Scientist at Google in Mountain View, California, where h….
0
2
0
RT @itsclivetime: Google's TPUv7 is out!. ML accelerator marketing material is usually pretty inscrutable (what numbers are even comparable….
0
142
0
RT @kellyhongsn: excited to share what I’ve been working on @trychroma! we introduce representative generative benchmarking - custom eval s….
0
28
0
RT @aakarshv1: Join us on Monday 3/10 for our latest installment of the BioML @ Berkeley seminar series! We'll be learning from the excepti….
0
4
0
RT @SonglinYang4: I've uploaded the latest slides & beamer source code to Hopefully this repository will help trai….
0
25
0
RT @srush_nlp: Linear Attention and Beyond: Interactive Tutorial with Songlin Yang (@SonglinYang4 MIT/Flash Linear Attention). I didn’t fol….
0
89
0
RT @SonglinYang4: Introducing the first open-source implementation of native sparse attention: Give it a spin and….
0
123
0
RT @Thom_Wolf: After 6+ months in the making and burning over a year of GPU compute time, we're super excited to finally release the "Ultra….
0
709
0
RT @SonglinYang4: I've created slides for those curious about the recent rapid progress in linear attention: from linear attention to Light….
0
165
0
RT @MiniMax__AI: MiniMax-01 is Now Open-Source: Scaling Lightning Attention for the AI Agent Era. We are thrilled to introduce our latest o….
0
420
0
RT @SonglinYang4: Hybrid linear and softmax attention has so much potential🤩. Don’t forget to check out for more ad….
0
18
0
RT @kellyhongsn: Recently finished an exploratory project with some friends using Othello-GPT - we investigated each layer of the model and….
0
1
0
RT @srush_nlp: 10 short videos about LLM infrastructure to help you appreciate Pages 12-18 of the DeepSeek-v3 paper (.
0
120
0