ProfArbel Profile Banner
Yonathan Arbel Profile
Yonathan Arbel

@ProfArbel

Followers
2K
Following
17K
Media
773
Statuses
4K

Let's build safe AI! Law prof @ Alabama Contracts, Defamation, Legal NLP, & AI Safety

Tuscaloosa, AL
Joined July 2014
Don't wanna be here? Send us removal request.
@ProfArbel
Yonathan Arbel
6 years
Why do workers have to wait for 2-4 weeks to be paid, in the same economy where online transactions go quickly and securely? . A new draft š‘ƒš‘Žš‘¦š‘‘š‘Žš‘¦-forthcoming @WashULaw-proposes that they shouldn't. Daily, or at least weekly, pay can be a reality.
4
9
55
@ProfArbel
Yonathan Arbel
8 days
Thinking about LLM psychosis, this is a good reminder that the problem is much older
0
0
3
@ProfArbel
Yonathan Arbel
10 days
RT @petersalib: Enjoying the new @80000Hours interview w/ @fish_kyle3 about AI welfare. One important research question he and Luisa discus….
0
2
0
@ProfArbel
Yonathan Arbel
11 days
RT @capricelroberts: Calling all law faculty candidates hoping to teach law starting next fall: more ways to get your application seen by a….
0
5
0
@ProfArbel
Yonathan Arbel
15 days
RT @IThinkIAgree: Taking 10% of Intel has four facets worth considering separately because ā€œstockā€ is really a bundle of rights. First, t….
0
1
0
@ProfArbel
Yonathan Arbel
17 days
can AI create new math? if so, where are all the new discoveries? . I can hardly adjudicate the merit of the claims in the thread, but even under this more minimalist view, it appears that there is a zone of (a) new math discoveries + (b) that aren't that hard to solve once the.
@ErnestRyu
Ernest Ryu
18 days
This is really exciting and impressive, and this stuff is in my area of mathematics research (convex optimization). I have a nuanced take. 🧵 (1/9).
1
0
0
@ProfArbel
Yonathan Arbel
27 days
It's just the first study. Much more work is needed to confirm these results and also evaluate consistency and reliability. It's not clear what aspects of model training contribute the most (RLHF?).
1
0
1
@ProfArbel
Yonathan Arbel
27 days
This, alongside other results, is consistent with LLMs having picked up something deep about reasonableness. If true, that would allow tool building for legal and agent applications, with important AI safety implications. but. .
1
0
1
@ProfArbel
Yonathan Arbel
27 days
Findings: humans judge harshly people who didn't take the socially common level of precautions, even though textbook says that shouldn't matter. Textbook does say they should care about costs, but humans don't care much. LLM repeat this exact pattern in RCTs. (but not all)
Tweet media one
1
0
1
@ProfArbel
Yonathan Arbel
27 days
To test whether LLMs may have picked up the deeper structure of reasonableness judgment (not just parroting), I offer a new methodology: Silicon RCTs (S-RCTs). Randomize a relevant fact, run session-isolated ā€œparticipants,ā€ & compare LLM deltas to human deltas.
Tweet media one
1
0
1
@ProfArbel
Yonathan Arbel
27 days
A note for CS friends: a large share of law isn’t if-then rules; it’s open-textured standards like ā€œreasonable,ā€ ā€œordinary meaning,ā€ and ā€œundue burdenā€. But our instruments are thin. Expert intuitions (judges) are criticized as elite, out-of-touch, and crypto-political, juries.
1
0
1
@ProfArbel
Yonathan Arbel
27 days
Two hard problems meet here: . (a) courts must infer what ordinary people will reasonably think (e.g., would a teenager read the Pepsi jet as an offer?); . (b) we want AI agents to follow open-textured norms (ā€œkeep a reasonable distanceā€).
1
0
1
@ProfArbel
Yonathan Arbel
27 days
Reasonableness quietly structures daily life: following distance in traffic, how loud a house party can be, what ā€œup to 50% longerā€ implies, or whether a Pepsi ad offering a jet is a joke. These judgments are fast, intuitive, and hard to explain. In other words: they are
Tweet media one
1
0
1
@ProfArbel
Yonathan Arbel
27 days
🚨🚨🚨Can judges really know what ā€œreasonable peopleā€ think? Can AI help bridge that gap—and can AI agents themselves behave ā€œreasonablyā€ in the wild? . A new draft, The Silicon Reasonable Person, asks these questions. tl;dr: early signs point to yes, with careful limits.
Tweet media one
2
3
17
@ProfArbel
Yonathan Arbel
28 days
that's new
Tweet media one
0
0
1
@ProfArbel
Yonathan Arbel
29 days
Some thoughts on how I update on gpt-5. Tl:Dr: It suggests longer timelines. I don't think there is nothing magical about "5" per se, and they could have gotten away with calling O3 "5". So I don't inherently base my view of the landscape on what "5" does or does not do. But. .
3
1
7
@ProfArbel
Yonathan Arbel
1 month
A terrific workshop, and an opportunity to present work on the Reasonable Silicon Person and AGI Governance
Tweet media one
0
0
8
@ProfArbel
Yonathan Arbel
1 month
have we all just quietly decided ten years ago that email reached its final form? Surely someone could come up with a better attachment system, integration with calendar, thread organization, working sound notifications for Gmail, a less hellish version of outlook, sane search.
1
0
2
@ProfArbel
Yonathan Arbel
1 month
I remember thinking at first that this was a non-starter, pie-in-the-sky proposal. But I actually think this is not only feasible, but quite likely in certain future branches of AI development.
@petersalib
Peter N. Salib
1 month
The WH's AI Action plan has some good stuff. But it begins, "The US is in a race to achieve global dominance in AI.". Like many, @simondgoldstein and I think that an AI arms race w/ China is a mistake. Our new paper lays out a novel game-theoretic approach to avoiding the race.
Tweet media one
0
0
2