smn_sdt Profile Banner
Samaneh Saadat Profile
Samaneh Saadat

@smn_sdt

Followers
370
Following
3K
Media
61
Statuses
471

ML SWE @Google | CoreML, Keras | Opinions my own

Seattle
Joined September 2015
Don't wanna be here? Send us removal request.
@fchollet
François Chollet
17 days
The narrative around LLMs is that they got better purely by scaling up pretraining *compute*. In reality, they got better by scaling up pretraining *data*, while compute is only a means to the end of cramming more data into the model. Data is the fundamental bottleneck. You can't
97
191
2K
@smn_sdt
Samaneh Saadat
19 days
Me, every time that Gemini does something that impresses me:
0
0
2
@smn_sdt
Samaneh Saadat
26 days
Francois and Matt are two of the most brilliant people I've worked with. They're excellent at explaining complex ideas in a simple and intuitive way. Don't miss their book.
@fchollet
François Chollet
27 days
The 3rd edition of my book Deep Learning with Python is being printed right now, and will be in bookstores within 2 weeks. You can order it now from Amazon or from Manning. This time, we're also releasing the whole thing as a 100% free website. I don't care if it reduces book
0
2
4
@smn_sdt
Samaneh Saadat
30 days
My favorite is
Tweet card summary image
colab.research.google.com
@SinaHartung
Sina
1 month
it has come to my attention that this is not universal knowledge you can just type https://t.co/prfdqNqr5m or https://t.co/HMVIGA1uSh into your browser and it will immediately open a new google doc or sheet
0
0
2
@penstrokes75
Abheesht Sharma
1 month
Try out VaultGemma on KerasHub! https://t.co/BylPNu711o
@osanseviero
Omar Sanseviero
1 month
Introducing VaultGemma 🧠Gemma pre-trained with differential privacy (largest open model trained from scratch like this) 🔒Strong, mathematically-backed privacy guarantees 🤏Just 1B parameters 📈Novel research on scaling laws
1
4
9
@osanseviero
Omar Sanseviero
1 month
Introducing EmbeddingGemma🎉 🔥With only 308M params, this is the top open model under 500M 🌏Trained on 100+ languages 🪆Flexible embeddings (768 to 128 dims) with Matryoshka 🤗Works with your favorite open tools 🤏Runs with as little as 200MB https://t.co/AXPqV4aXr1
28
156
1K
@smn_sdt
Samaneh Saadat
1 month
Yay 🎉
@Waymo
Waymo
1 month
We’re heading North – the Pacific Northwest to be exact! Today, we’re returning to Washington State as we lay the groundwork to launch our autonomous ride-hail service in the Seattle metropolitan area. Learn more: https://t.co/3J8gKvbW7Y
0
0
3
@smn_sdt
Samaneh Saadat
2 months
You can use Orbax for checkpointing when training your Keras model with the JAX backend. Orbax checkpointing is particularly useful when doing multi-host training using Keras distribution API. We have a new guide showing how to do that.
1
3
11
@smn_sdt
Samaneh Saadat
2 months
What I don't like about it is the weather! 😁 Too hot during the day and too windy and cold at night!
1
0
0
@smn_sdt
Samaneh Saadat
2 months
Visiting bay area this week and what I love about it is that I get to chat with people who work on very interesting problems.
1
0
6
@michael_terrell
Michael Terrell
2 months
Incredible to see that the energy used for a median Gemini AI text prompt has dropped 33x in only 12 months. This is a huge achievement – and one that would not have been possible without the work of many Googlers and focused efforts to deliver greater efficiencies across the
@JeffDean
Jeff Dean
2 months
AI efficiency is important. Today, Google is sharing a technical paper detailing our comprehensive methodology for measuring the environmental impact of Gemini inference. We estimate that the median Gemini Apps text prompt uses 0.24 watt-hours of energy (equivalent to watching an
3
4
37
@smn_sdt
Samaneh Saadat
2 months
You can use Orbax for checkpointing when training your Keras model with the JAX backend. Orbax checkpointing is particularly useful when doing multi-host training using Keras distribution API. We have a new guide showing how to do that.
1
3
11
@fchollet
François Chollet
2 months
Important point from Deep Learning with Python...
13
91
769
@smn_sdt
Samaneh Saadat
2 months
Interesting findings on the Hierarchical Reasoning Model paper
@fchollet
François Chollet
2 months
We were able to reproduce the strong findings of the HRM paper on ARC-AGI-1. Further, we ran a series of ablation experiments to get to the bottom of what's behind it. Key findings: 1. The HRM model architecture itself (the centerpiece of the paper) is not an important factor.
0
0
3
@smn_sdt
Samaneh Saadat
2 months
New models on KerasHub 🎉🦄
6
30
107
@smn_sdt
Samaneh Saadat
2 months
There’s no way anyone knows everything, so if someone never says "I don’t know", it means they’re not acknowledging their lack of knowledge in certain areas. That makes it hard to trust them because you can’t tell if they actually know something or are just pretending to.
1
1
5
@smn_sdt
Samaneh Saadat
2 months
One of the hardest types of people to work with is someone who won’t admit when they don’t know something. I really respect a person who can say, "I don’t know".
1
4
13
@osanseviero
Omar Sanseviero
2 months
Introducing Gemma 3 270M 🔥 🤏A tiny model! Just 270 million parameters 🧠 Very strong instruction following 🤖 Fine-tune in just a few minutes, with a large vocabulary to serve as a high-quality foundation https://t.co/E0BB5nlI1k
124
330
3K