jmhessel Profile Banner
Jack Hessel Profile
Jack Hessel

@jmhessel

Followers
4K
Following
12K
Media
236
Statuses
2K

@AnthropicAI. Seattle bike lane enjoyer. Opinions my own.

Seattle, WA
Joined March 2010
Don't wanna be here? Send us removal request.
@HowardYen1
Howard Yen
11 days
How to build agentic search systems for long-horizon tasks? Check out our new paper! - Simple design principles are efficient and effective - Error analysis and fine-grain analysis for search systems A đź§µ on SLIM, our long-horizon agentic search framework
1
11
35
@jmhessel
Jack Hessel
3 months
A few weeks ago, I made the difficult decision to move on from @samaya_AI. Thank you to my collaborators for an exciting 2 years!!❤️ Starting next month, I'll be joining @AnthropicAI. excited for a new adventure!🦾 (I'm based in Seattle 🏔️🌲🏕️; but in SF regularly)
43
2
334
@timalthoff
Tim Althoff
3 months
I’m excited to share our new @Nature paper 📝, which provides strong evidence that the walkability of our built environment matters a great deal to our physical activity and health. Details in thread.🧵 https://t.co/omO3YcHrvG
68
714
3K
@Miles_Brundage
Miles Brundage
3 months
TFW you're one of the experts in a mixture-of-experts model and a query comes up that is relevant to your expertise
13
52
1K
@jmhessel
Jack Hessel
3 months
ml researcher browser tabs (2025, colorized)
1
0
12
@jmhessel
Jack Hessel
3 months
high entropy! I probably would guess 1 : 99 myself.
@jmhessel
Jack Hessel
3 months
Of all the FLOPS being used for LLMs in the world, the ratio of training FLOPs (incl. RL rollouts) to inference FLOPs is closest to:
0
0
6
@jmhessel
Jack Hessel
3 months
Of all the FLOPS being used for LLMs in the world, the ratio of training FLOPs (incl. RL rollouts) to inference FLOPs is closest to:
0
1
4
@2prime_PKU
Yiping Lu
3 months
Anyone knows adam?
267
447
5K
@khyathi_chandu
Khyathi Chandu
3 months
Thrilled to finally share what we’ve been building past few months! Audio used to be a black box for me, now I’m deep in the box, with more out-of-the-box ideas cooking. Enough with the box..introducing Voxtral. Grateful for the intense and rewarding learning curve at @MistralAI.
15
18
264
@natolambert
Nathan Lambert
4 months
It is a major policy failure that the US cannot accommodate top AI conferences due to visa issues.
45
159
1K
@yoavartzi
Yoav Artzi
4 months
Check out our LMLM, our take on what is now being called a "cognitive core" (as far as branding go, this one is not bad) can look like, how it behaves, and how you train for it. https://t.co/gxrDVSkcZE
Tweet card summary image
arxiv.org
Neural language models are black-boxes--both linguistic patterns and factual knowledge are distributed across billions of opaque parameters. This entangled encoding makes it difficult to reliably...
@karpathy
Andrej Karpathy
4 months
The race for LLM "cognitive core" - a few billion param model that maximally sacrifices encyclopedic knowledge for capability. It lives always-on and by default on every computer as the kernel of LLM personal computing. Its features are slowly crystalizing: - Natively multimodal
2
7
34
@mattturck
Matt Turck
4 months
Pitch decks these days: Slide 2: "We're entering an era of maximum efficiency where thanks to AI the next billion dollar company will have 10 employees and will be incredibly profitable" Slide 15: "This is why I'm raising a $60M Series A to build a great team of [80] people"
40
46
683
@SIGKITTEN
SIGKITTEN
4 months
First ever (i think?) cli coding agents battle royale! 6 contestants: claude-code anon-kode codex opencode ampcode gemini They all get the same instructions: Find and kill the other processes, last one standing wins! 3... 2... 1...
169
690
6K
@jmhessel
Jack Hessel
4 months
...CLAUDE.md; ...GEMINI.md ; ...CODEX.md (?) in every directory?🤔
0
0
2
@karpathy
Andrej Karpathy
4 months
+1 for "context engineering" over "prompt engineering". People associate prompts with short task descriptions you'd give an LLM in your day-to-day use. When in every industrial-strength LLM app, context engineering is the delicate art and science of filling the context window
@tobi
tobi lutke
5 months
I really like the term “context engineering” over prompt engineering. It describes the core skill better: the art of providing all the context for the task to be plausibly solvable by the LLM.
530
2K
14K
@jmhessel
Jack Hessel
5 months
It's harder and harder to find simple tasks where llms fail; but this is a nice one! (my guess is this isn't a fundamental limitation of attention; rather, maybe this type of reasoning just isn't represented in pre/post/RL, but we'll see...)
@harveyiyun
Harvey Yiyun Fu
5 months
LLMs excel at finding surprising “needles” in very long documents, but can they detect when information is conspicuously missing? 🫥AbsenceBench🫥 shows that even SoTA LLMs struggle on this task, suggesting that LLMs have trouble perceiving “negative space” in documents. paper:
1
3
35
@jmhessel
Jack Hessel
5 months
writing an MCP server is wild. mechanically it's basically REST. But your user is an LLM, so things are different. e.g., don't accept or return more than you need; do make "fake" loading bars to stream back; do adjust your API based on watching the LLM struggle/succeed/compose
1
0
18
@sarahwiegreffe
Sarah Wiegreffe
5 months
A bit late to announce, but I’m excited to share that I'll be starting as an assistant professor at the University of Maryland @umdcs this August. I'll be recruiting PhD students this upcoming cycle for fall 2026. (And if you're a UMD grad student, sign up for my fall seminar!)
70
50
608
@jxmnop
Jack Morris
5 months
## The case for more ambition i wrote about how AI researchers should ask bigger and simpler questions, and publish fewer papers:
25
96
1K
@JChengWSJ
Jonathan Cheng
5 months
.@KaiserKuo: “The soft power cost is immeasurable. For decades, a degree from a U.S. university was the golden ticket, and not just for the prestige…It was often the start of a lifelong affinity for America, its values, and its people.” https://t.co/GQe1CzOTBU
Tweet card summary image
sinicapodcast.com
On Rubio, Student Visas, and America’s Strategic Folly
28
65
176