
Samaneh Saadat
@smn_sdt
Followers
370
Following
3K
Media
61
Statuses
471
ML SWE @Google | CoreML, Keras | Opinions my own
Seattle
Joined September 2015
The narrative around LLMs is that they got better purely by scaling up pretraining *compute*. In reality, they got better by scaling up pretraining *data*, while compute is only a means to the end of cramming more data into the model. Data is the fundamental bottleneck. You can't
97
191
2K
Francois and Matt are two of the most brilliant people I've worked with. They're excellent at explaining complex ideas in a simple and intuitive way. Don't miss their book.
The 3rd edition of my book Deep Learning with Python is being printed right now, and will be in bookstores within 2 weeks. You can order it now from Amazon or from Manning. This time, we're also releasing the whole thing as a 100% free website. I don't care if it reduces book
0
2
4
My favorite is
colab.research.google.com
it has come to my attention that this is not universal knowledge you can just type https://t.co/prfdqNqr5m or https://t.co/HMVIGA1uSh into your browser and it will immediately open a new google doc or sheet
0
0
2
Try out VaultGemma on KerasHub! https://t.co/BylPNu711o
Introducing VaultGemma 🧠Gemma pre-trained with differential privacy (largest open model trained from scratch like this) 🔒Strong, mathematically-backed privacy guarantees 🤏Just 1B parameters 📈Novel research on scaling laws
1
4
9
Introducing EmbeddingGemma🎉 🔥With only 308M params, this is the top open model under 500M 🌏Trained on 100+ languages 🪆Flexible embeddings (768 to 128 dims) with Matryoshka 🤗Works with your favorite open tools 🤏Runs with as little as 200MB https://t.co/AXPqV4aXr1
28
156
1K
Yay 🎉
We’re heading North – the Pacific Northwest to be exact! Today, we’re returning to Washington State as we lay the groundwork to launch our autonomous ride-hail service in the Seattle metropolitan area. Learn more: https://t.co/3J8gKvbW7Y
0
0
3
You can use Orbax for checkpointing when training your Keras model with the JAX backend. Orbax checkpointing is particularly useful when doing multi-host training using Keras distribution API. We have a new guide showing how to do that.
1
3
11
What I don't like about it is the weather! 😁 Too hot during the day and too windy and cold at night!
1
0
0
Visiting bay area this week and what I love about it is that I get to chat with people who work on very interesting problems.
1
0
6
Incredible to see that the energy used for a median Gemini AI text prompt has dropped 33x in only 12 months. This is a huge achievement – and one that would not have been possible without the work of many Googlers and focused efforts to deliver greater efficiencies across the
AI efficiency is important. Today, Google is sharing a technical paper detailing our comprehensive methodology for measuring the environmental impact of Gemini inference. We estimate that the median Gemini Apps text prompt uses 0.24 watt-hours of energy (equivalent to watching an
3
4
37
You can use Orbax for checkpointing when training your Keras model with the JAX backend. Orbax checkpointing is particularly useful when doing multi-host training using Keras distribution API. We have a new guide showing how to do that.
1
3
11
Interesting findings on the Hierarchical Reasoning Model paper
We were able to reproduce the strong findings of the HRM paper on ARC-AGI-1. Further, we ran a series of ablation experiments to get to the bottom of what's behind it. Key findings: 1. The HRM model architecture itself (the centerpiece of the paper) is not an important factor.
0
0
3
There’s no way anyone knows everything, so if someone never says "I don’t know", it means they’re not acknowledging their lack of knowledge in certain areas. That makes it hard to trust them because you can’t tell if they actually know something or are just pretending to.
1
1
5
One of the hardest types of people to work with is someone who won’t admit when they don’t know something. I really respect a person who can say, "I don’t know".
1
4
13
Introducing Gemma 3 270M 🔥 🤏A tiny model! Just 270 million parameters 🧠 Very strong instruction following 🤖 Fine-tune in just a few minutes, with a large vocabulary to serve as a high-quality foundation https://t.co/E0BB5nlI1k
124
330
3K