(((ل()(ل() 'yoav))))👾 Profile
(((ل()(ل() 'yoav))))👾

@yoavgo

Followers
65K
Following
72K
Media
3K
Statuses
137K

Israel
Joined May 2009
Don't wanna be here? Send us removal request.
@yoavgo
(((ل()(ل() 'yoav))))👾
2 years
what there is no overwhelming agreement on is what will happen the day after. how do we build a livable future here, for both israelis and palestinians. israel leadership is not thinking of this now, future looks grim. THIS is where protest effort and outrage should be invested.
14
12
125
@yoavgo
(((ل()(ل() 'yoav))))👾
3 days
this "feature" in code editors where you type an opening character and it immediately inserts the closing one for you as well -- why?? at the very very best case, you will have to skip over this character with an arrow. so why?
8
0
19
@yoavgo
(((ل()(ل() 'yoav))))👾
5 days
Tweet card summary image
sh-reya.com
0
0
9
@yoavgo
(((ل()(ل() 'yoav))))👾
5 days
who said that?
Tweet media one
0
0
1
@yoavgo
(((ل()(ل() 'yoav))))👾
8 days
can you explain "serverless backend" to me? it seems to me that their main/only purpose is to hide API keys that you don't want to expose to the client. am I missing something?
7
0
6
@yoavgo
(((ل()(ل() 'yoav))))👾
13 days
interesting. i initially thought it was going to be a "no DL! how come it has no modern DL??" kinda gripe, but no. its not that at all.
@predict_addict
Valeriy M., PhD, MBA, CQF
14 days
Scikit-learn has become an antique museum piece in machine learning. It is still paraded around as if it were modern, but in reality it lags far behind.
Tweet media one
9
1
18
@yoavgo
(((ل()(ل() 'yoav))))👾
13 days
amateurs. a proper implementation would have doubled the list when scrolling near the ends.
@skydotcs
sky
14 days
the time picker on iphones alarm app isn’t actually circular it’s just a really long list
2
1
18
@yoavgo
(((ل()(ل() 'yoav))))👾
14 days
i created this thingy yesterday and now I cannot stop watching it. https://t.co/Dp0r3COZrR
11
4
104
@yoavgo
(((ل()(ل() 'yoav))))👾
18 days
arxiv took their time with this paper because apparently they moved it from AI to HCI. it indeed belongs in both, but I still believe it is much more about AI than it is about HCI. but now we have an opportunity to post it again, and it IS a really neat result, imo. so, a win.
@mosh_levy
Mosh Levy
18 days
When reading AI reasoning text (aka CoT), we form a narrative about the underlying computation process, which we take as a transparent explanation of model behavior. But what if our narrative is wrong? We measure that and find it usually is. Now on arXiv:
1
1
12
@mosh_levy
Mosh Levy
18 days
When reading AI reasoning text (aka CoT), we form a narrative about the underlying computation process, which we take as a transparent explanation of model behavior. But what if our narrative is wrong? We measure that and find it usually is. Now on arXiv:
Tweet card summary image
arxiv.org
A new generation of AI models generates step-by-step reasoning text before producing an answer. This text appears to offer a human-readable window into their computation process, and is...
@mosh_levy
Mosh Levy
30 days
Producing reasoning texts boosts the capabilities of AI models, but do we humans correctly understand these texts? Our latest research suggests that we do not. This highlights a new angle on the "Are they transparent?" debate: they might be, but we misinterpret them. 🧵
Tweet media one
2
3
16
@yoavgo
(((ل()(ل() 'yoav))))👾
18 days
agents hold a huge promise. building strong, good, interactive and multipurpose ones turns out to be very hard (happy to discuss the challenges). in this project, we attempt to do that, and push forward both useful functionality *and* the science of agent building.
@allen_ai
Ai2
19 days
Introducing Asta—our bold initiative to accelerate science with trustworthy, capable agents, benchmarks, & developer resources that bring clarity to the landscape of scientific AI + agents. 🧵
2
5
83
@Jonathan_Blow
Jonathan Blow
1 month
Everybody posting about AI in programming sounds like this to me. https://t.co/3J62fBK27g
59
116
2K
@mustafasuleyman
Mustafa Suleyman
26 days
@connoraxiotes The debate about whether AI is actually conscious is, for now at least, a distraction. It will seem conscious and that illusion is what’ll matter in the near term
20
4
82
@yoavgo
(((ل()(ل() 'yoav))))👾
25 days
crazy that this is considered controversial / surprising in some circles
@kyliebytes
Kylie Robison
25 days
this mustafa suleyman blog is SO interesting — i'm not sure i've seen an AI leader write such strong opinions *against* model welfare, machine consciousness etc https://t.co/sQiXM3zwTI
Tweet media one
9
1
63
@allen_ai
Ai2
27 days
LLMs power research, decision‑making, and exploration—but most benchmarks don’t test how well they stitch together evidence across dozens (or hundreds) of sources. Meet MoNaCo, our new eval for question-answering cross‑source reasoning. 👇
Tweet media one
10
40
223
@TomerWolfson
Tomer Wolfson
27 days
Many factual QA benchmarks have become saturated, yet factuality still poses a very real issue! ✨We present MoNaCo, an Ai2 benchmark of human-written time-consuming questions that, on average, require 43.3 documents per question!✨ 📣Blogpost: https://t.co/GQD83gdHgg 🧵(1/5)
Tweet media one
1
14
39
@yoavgo
(((ل()(ل() 'yoav))))👾
30 days
this is really cool to me, as Reasoning LLMs use language *effectively*. The language-use improves the computation's accuracy. But the work shows they use it in a very different way from us. Its a different way of using the same language!
@mosh_levy
Mosh Levy
30 days
Implication 2: Reasoning models use language effectively yet differently than humans.  We believe that now we should look more into how humans interpret AI-produced language, not only how AIs parse ours.
1
0
28
@mosh_levy
Mosh Levy
30 days
The test is very simple: given a reasoning text with a highlighted target step t, identify the one out of four candidate steps that, if removed, will change the target step. We administered it to 80 participants online. You can test yourself here: https://t.co/le1mw2vd3W
1
3
10
@yoavgo
(((ل()(ل() 'yoav))))👾
30 days
in this work, we show that reasoning texts are really not transparent, in the sense that we (humans) are really bad in inferring how the steps (really) depend on each other for the model. yoo can test yourself here:
@mosh_levy
Mosh Levy
30 days
Producing reasoning texts boosts the capabilities of AI models, but do we humans correctly understand these texts? Our latest research suggests that we do not. This highlights a new angle on the "Are they transparent?" debate: they might be, but we misinterpret them. 🧵
Tweet media one
2
14
98
@UrikaUri
Uri Katz
30 days
Even when AI reasoning traces are “transparent”, the narrative they encode can differ from the one humans perceive. The transparency debate isn’t just about revealing the model’s steps, it’s about the mismatch between the model’s story and the one we construct from it.
@mosh_levy
Mosh Levy
30 days
Producing reasoning texts boosts the capabilities of AI models, but do we humans correctly understand these texts? Our latest research suggests that we do not. This highlights a new angle on the "Are they transparent?" debate: they might be, but we misinterpret them. 🧵
Tweet media one
0
2
12
@mosh_levy
Mosh Levy
30 days
Producing reasoning texts boosts the capabilities of AI models, but do we humans correctly understand these texts? Our latest research suggests that we do not. This highlights a new angle on the "Are they transparent?" debate: they might be, but we misinterpret them. 🧵
Tweet media one
8
27
137