ehsanjjjjj Profile Banner
Ehsan Jahangiri Profile
Ehsan Jahangiri

@ehsanjjjjj

Followers
86
Following
157
Media
0
Statuses
55

ML/SW engineer, math lover, multiplier, and always curious. - Principal MLE @ Nvidia, ex-Apple. - All opinions are mine.

San Francisco, CA
Joined June 2024
Don't wanna be here? Send us removal request.
@BahmanKalbasi
Bahman Kalbasi
5 months
And
@KZiabari
Kourosh Ziabari
5 months
Sara Jodat, an undergraduate student of photography at the Pars University of Architecture and Art has been killed in #Israel's airstrikes on #Iran's capital, her professor confirms to me. Another aspiring artist is the casualty of belligerence. May her memory be a blessing šŸ’”
1
7
25
@forgebitz
Klaas
6 months
the face you make when you sell your vscode fork windsmurf for 3b
60
85
4K
@blizaine
Blaine Brown
7 months
Kling v2 has great prompt adherence
94
196
2K
@tunguz
Bojan Tunguz
11 months
37
354
5K
@elonmusk
Elon Musk
1 year
The tower has caught the rocket!!
45K
138K
1M
@JohnDCook
John D. Cook
1 year
37
452
3K
@gregjoz
Greg Joswiak
1 year
Thinking of everyone in the path of Hurricane Milton. Please stay safe!
@andrewjclare
Andrew Clare
1 year
Hurricane Milton is fast approaching. If you’re in Central Florida and have an iPhone 14 or later, you can use Emergency SOS via satellite to reach emergency services even when cell service is down. This feature lets you make calls and send texts through satellite connection.
16
96
1K
@Yoshua_Bengio
Yoshua Bengio
1 year
@HopfieldJohn and @geoffreyhinton, along with collaborators, have created a beautiful and insightful bridge between physics and AI. They invented neural networks that were not only inspired by the brain, but also by central notions in physics such as energy, temperature, system
@NobelPrize
The Nobel Prize
1 year
BREAKING NEWS The Royal Swedish Academy of Sciences has decided to award the 2024 #NobelPrize in Physics to John J. Hopfield and Geoffrey E. Hinton ā€œfor foundational discoveries and inventions that enable machine learning with artificial neural networks.ā€
18
276
2K
@JeffDean
Jeff Dean
1 year
Check out NotebookLM! Create a notebook, upload one or more sources (e.g. PDFs of research papers, your favorite PhD thesis, a newspaper article, etc) then click on 'Generate' to create a podcast of two voices talking about the content you've uploaded. https://t.co/FSCBvsr8tw
Tweet card summary image
blog.google
NotebookLM is releasing Audio Overviews, which turns your sources into an engaging discussion.
94
378
2K
@NandoDF
Nando de Freitas
1 year
The Llama 3 paper is a must-read for anyone in AI and CS. It’s an absolutely accurate and authoritative take on what it takes to build a leading LLM, the tech behind ChatGPT, Gemini, Copilot, and others. The AI part might seem small in comparison to the gargantuan work on *data*
@soumithchintala
Soumith Chintala
1 year
Why do 16k GPU jobs fail? The Llama3 paper has many cool details -- but notably, has a huge infrastructure section that covers how we parallelize, keep things reliable, etc. We hit an overall 90% effective-training-time. https://t.co/hsSIW4bayK
12
289
2K
@NandoDF
Nando de Freitas
1 year
It is remarkable that anyone can now train a 124M parameter LLM in about real-time on a MacBook M3. So easy to experiment. This would have been the stuff of dreams when I was in school. I ā¤ļø training neural nets, but I really admire the people who build the hardware.
16
29
533
@JeffDean
Jeff Dean
1 year
25 years ago today I officially started at a small search engine company, wedged into a tiny office space above what's now a T-Mobile store in downtown Palo Alto. Since then, I have had the incredible pleasure of working with awesome colleagues on software used by billions of
196
143
4K
@tim_cook
Tim Cook
1 year
Apple Intelligence is going to unlock a world of new possibilities for our users, and it’s thrilling to see our developers begin to build with it. We’re excited to see the amazing things they create.
1K
1K
15K
@ruomingpang
Ruoming Pang
1 year
As Apple Intelligence is rolling out to our beta users today, we are proud to present a technical report on our Foundation Language Models that power these features on devices and cloud: https://t.co/TaAdd0fBOp. 🧵
Tweet card summary image
machinelearning.apple.com
We present foundation language models developed to power Apple Intelligence features, including a ∼3 billion parameter model designed to run…
13
190
708
@ehsanjjjjj
Ehsan Jahangiri
1 year
There are only two hard things in Computer Science: cache invalidation and naming things. -- Phil Karlton There are 2 hard problems in computer science: cache invalidation, naming things, and off-by-1 errors. -- Leon Bambrick There's two hard problems in computer science: we
1
0
2
@jiayq
Yangqing Jia
1 year
Memory Matters for LLM. While everyone is rushing to provide the serverless Llama3-405b model, I want to talk about one key choice that matters a lot, especially for dedicated enterprise deployments when traffic is not very high: memory. - The normal deployment of a model the
3
40
258
@sirbayes
Kevin Patrick Murphy
1 year
Llama 3 is the new PyTorch
@ylecun
Yann LeCun
1 year
@brbcatonfire Mark Z explains that here: https://t.co/5WHf6gnzhL
12
11
133
@rowancheung
Rowan Cheung
1 year
Exclusive: Meta just released Llama 3.1 405B — the first-ever open-sourced frontier AI model, beating top closed models like GPT-4o across several benchmarks. I sat down with Mark Zuckerberg, diving into why this marks a major moment in AI history. Timestamps: 00:00 Intro
431
1K
9K
@docmilanfar
Peyman Milanfar
1 year
If your PhD advisor dresses like this, you don’t have to worry about using neural nets in your thesis
13
27
796
@ehsanjjjjj
Ehsan Jahangiri
1 year
Great points. Intelligence and knowledge are related but not the same; we are currently evaluating LLMs for their knowledge (very much tied to memory). Adaptability on the other hand is a big factor for intelligence.
@karpathy
Andrej Karpathy
1 year
LLM model size competition is intensifying… backwards! My bet is that we'll see models that "think" very well and reliably that are very very small. There is most likely a setting even of GPT-2 parameters for which most people will consider GPT-2 "smart". The reason current
1
0
1