luhemarora Profile Banner
~/mehul/ Profile
~/mehul/

@luhemarora

Followers
181
Following
503
Media
86
Statuses
290

frolicking in the garden of technology @southpkcommons @lightbulbml (acq.) @stanford

SF
Joined March 2021
Don't wanna be here? Send us removal request.
@luhemarora
~/mehul/
2 months
LLMs hallucinate because they're trying to remember everything. They massively compress information, which leads to a loss of reliability. A thread đź§µ (1/)
1
5
19
@SourishJasti
Sourish Jasti
9 days
1/ The future of general-purpose robotics will be decided by one major question: which flavor of data scales reasoning? Every major lab represents a different bet. Over the past 3 months, @adam_patni, @vriishin, and I read the core research papers, spoke with staff at the major
61
195
773
@luhemarora
~/mehul/
22 days
Beyond Being There is the Bitter Lesson equivalent of HCI
1
0
2
@luhemarora
~/mehul/
1 month
Do you (yes you) live in SF? Do you know where in the city this is? If yes (and also if no), I'm hosting a GeoGuessr tournament with exclusively SF locations, and you should compete
1
1
5
@AdtRaghunathan
Aditi Raghunathan
1 month
There’s been a lot of work on unlearning in LLMs, trying to erase memorization without hurting capabilities — but we haven’t seen much success. ❓What if unlearning is actually doomed from the start? 👇This thread explains why and how *memorization sinks* offer a new way forward.
6
38
175
@luhemarora
~/mehul/
1 month
Incredible things happening over in the twitter repo
0
0
5
@luhemarora
~/mehul/
2 months
I really really really want to talk to someone building enterprise memory into Claude, can someone put me in touch?
0
1
0
@luhemarora
~/mehul/
2 months
@BlitWise Unfortunately gpt-4o-mini is ... not so great at 2D projectile physics. Releasing this as an environment soon!
0
0
1
@luhemarora
~/mehul/
2 months
I built a LLM agent to play Pocket Tanks. @BlitWise
1
0
3
@luhemarora
~/mehul/
2 months
my brief moment of glory before being absolutely humbled
0
0
0
@luhemarora
~/mehul/
2 months
They’re selling dollars for 47c right now over at the inaugural spelling bee kalshi dot com
2
0
2
@luhemarora
~/mehul/
2 months
More to come! If you're exploring in this space, or someone excited about potential use cases (we have many ideas), I would love to chat. (7/7)
0
0
7
@luhemarora
~/mehul/
2 months
The same question from before, using the same model but with our approach: (6/)
2
0
3
@luhemarora
~/mehul/
2 months
1. Easier to maintain index freshness 2. Fewer hallucinations 3. Easier interpretability 4. Better performance on multi-hop queries (5/)
1
0
3
@luhemarora
~/mehul/
2 months
By hooking up LLMs to a database that stores facts, we can train them to retrieve only the data they need, deterministically. This approach has a few benefits: (4/)
1
0
3
@luhemarora
~/mehul/
2 months
This is a problem with how LLMs store knowledge. Up to 25% of LLM weights go towards encoding facts / knowledge, which also makes this a HUGE waste of compute. @arundsharma and I are changing the way LLMs access data so they get exactly what they need with perfect accuracy. (3/)
1
0
5
@luhemarora
~/mehul/
2 months
Here’s an example: @qwen 3 0.6B thinks Mark Zuckerberg is married to “Sheryl Zuckerberg”. (2/)
1
0
3
@luhemarora
~/mehul/
2 months
12 more months of "you're absolutely right!" in cursor
@AnthropicAI
Anthropic
2 months
We've raised $13 billion at a $183 billion post-money valuation. This investment, led by @ICONIQCapital, will help us expand our capacity, improve model capabilities, and deepen our safety research.
0
0
4
@luhemarora
~/mehul/
2 months
New view who dis
1
0
6