xuan (ɕɥɛn / sh-yen) Profile Banner
xuan (ɕɥɛn / sh-yen) Profile
xuan (ɕɥɛn / sh-yen)

@xuanalogue

Followers
5,270
Following
993
Media
714
Statuses
8,893

PhD Student. MIT ProbComp / CoCoSci. Inverting Bayesian models of human reasoning and decision-making. Pronouns: 祂/伊 Mastodon: @xuanalogue @mas .to

Cambridge, MA
Joined September 2018
Don't wanna be here? Send us removal request.
Explore trending content on Musk Viewer
Pinned Tweet
@xuanalogue
xuan (ɕɥɛn / sh-yen)
2 years
Gave another talk on AI alignment, this time at #EAGxSingapore last week -- appreciated the chance to condense my recent thinking about what it means to "align" AI in a world with a diversity of people & values by asking "What Should AI Owe To Us?" (1/11)
Tweet media one
Tweet media two
Tweet media three
Tweet media four
@xuanalogue
xuan (ɕɥɛn / sh-yen)
4 years
Will be presenting on "AI alignment, philosophical pluralism, and the relevance of non-Western philosophy" at the inaugural Effective Altruism Global x Asia-Pacific conference next weekend (Nov 20-22)! @SingaporeEa Register by Nov 18 at if interested :)
Tweet media one
Tweet media two
Tweet media three
Tweet media four
2
3
59
4
13
143
@xuanalogue
xuan (ɕɥɛn / sh-yen)
2 months
Deep neural networks are just Gaussian Processes with a squared exponential kernel confirmed 😎
Tweet media one
13
91
925
@xuanalogue
xuan (ɕɥɛn / sh-yen)
8 months
OpenAI is removing the ability to evaluate P(completion | prompt) for user-provided completions to the `gpt-3.5-turbo-instruct` model... (requires setting `echo=true` & `logprobs=0`) Makes it impossible to use it as a likelihood function, or to query logprobs without sampling.
Tweet media one
15
55
633
@xuanalogue
xuan (ɕɥɛn / sh-yen)
1 year
Apparently neither @OpenAI 's new Chat API,nor @AnthropicAI 's API for Claude allow users to request the log probabilities assigned to each token 🫤 This means they can only be used to generate text, not evaluate the probability of text under the model. (h/t @alexanderklew )
Tweet media one
Tweet media two
36
66
580
@xuanalogue
xuan (ɕɥɛn / sh-yen)
5 months
NYT: "They are designed around probabilities, so they are unpredictable." *screams internally*
Tweet media one
18
35
503
@xuanalogue
xuan (ɕɥɛn / sh-yen)
1 year
LLMs *are* just predicting the next word at run time (ruling out beam search etc.) It's just that predicting the next word isn't inconsistent with doing more complicated stuff under the hood (e.g. Bayesian inference over latent structure). Please read de Finetti's theorem y'all!
7
48
448
@xuanalogue
xuan (ɕɥɛn / sh-yen)
1 year
Bizarre to me that so many LLM benchmarks were using top-1 accuracy as a metric rather than the Brier score or similar -- apparently once you switch to the latter (and other continuous and/or linear metrics), many "emergent" behaviors go away!
@arankomatsuzaki
Aran Komatsuzaki
1 year
Are Emergent Abilities of Large Language Models a Mirage? Presents an alternative explanation for emergent abilities: one can choose a metric which leads to the inference of an emergent ability or another metric which does not.
Tweet media one
24
187
979
6
48
382
@xuanalogue
xuan (ɕɥɛn / sh-yen)
8 months
Looking at the comments in the Rust source code for tiktoken, and wow LLM tokenization is *such* a mess.
Tweet media one
10
19
303
@xuanalogue
xuan (ɕɥɛn / sh-yen)
4 months
Continued success of a recipe we've known since AlphaZero & DreamCoder: Use synthetic data generation and process-level supervision to train neural models to *guide* reasoning via approximate guesses, not replace reasoning entirely with a large pretrained model.
@GoogleDeepMind
Google DeepMind
4 months
AlphaGeometry is a system made up of 2️⃣ parts: 🔵 A neural language model, which can predict useful geometry constructions to solve problems 🔵 A symbolic deduction engine, which uses logical rules to deduce conclusions Both work together to find proofs for complex geometry
Tweet media one
11
85
523
10
30
299
@xuanalogue
xuan (ɕɥɛn / sh-yen)
10 months
Nah it's because they haven't seen truly Bayesian time series forecasting and online structure learning yet 😎
3
24
227
@xuanalogue
xuan (ɕɥɛn / sh-yen)
5 months
Sorry to be kinda annoying about this! But consider: "Humans won't be able to supervise compilers smarter than us. For example, if a superhuman compiler generates a million lines of extremely complicated assembly, we won't be able to tell if it's safe to run or not."
@CollinBurns4
Collin Burns
5 months
Humans won't be able to supervise models smarter than us. For example, if a superhuman model generates a million lines of extremely complicated code, we won’t be able to tell if it’s safe to run or not, if it follows our instructions or not, and so on.
7
5
72
11
18
217
@xuanalogue
xuan (ɕɥɛn / sh-yen)
1 year
Looked more into this, and wow, the log. probs returned by OpenAI's API are *incredibly* unstable, especially for the latest model that supports it! Across 10 runs on the *same* set of prompts: text-davinci-002: std. dev. of 0.03 text-davinci-003: std. dev. of 0.21 (!!!) 😵‍💫
Tweet media one
Tweet media two
Tweet media three
Tweet media four
@xuanalogue
xuan (ɕɥɛn / sh-yen)
1 year
@gdb Working on how LLMs can be used in Bayesian modeling and inference. It'd be great to have: - normalized log probabilities for when temperature ≠ 1.0 - stable log probabilities (currently differs across API requests, keeping text fixed)
1
1
29
13
18
199
@xuanalogue
xuan (ɕɥɛn / sh-yen)
1 year
I respect Jacob a lot but I find it really difficult to engage with predictions of LLM capabilities that presume some version of the scaling hypothesis will continue to hold - it just seems highly implausible given everything we already know about the limits of transformers!
9
30
198
@xuanalogue
xuan (ɕɥɛn / sh-yen)
29 days
can you copyright a probability distribution
22
9
186
@xuanalogue
xuan (ɕɥɛn / sh-yen)
6 days
today I learned that oops I did it again by britney spears is actually a folia and it's the best thing ever :)))
7
26
212
@xuanalogue
xuan (ɕɥɛn / sh-yen)
3 months
How can we build AI assistants that *reliably* follow our instructions, even when they're ambiguous? @Lance_Ying42 & I introduce CLIPS: A Bayesian arch. combining inverse planning w LLMs that *pragmatically* infers human goals from actions & language, then provides assistance!
5
39
188
@xuanalogue
xuan (ɕɥɛn / sh-yen)
1 year
New Bayesian inference algorithm alert! A little belated, but I was glad to play a supporting role on this paper by @alexanderklew & George Matheos: Sequential Monte Carlo w. Probabilistic Program Proposals (SMCP3) Paper: News:
2
35
175
@xuanalogue
xuan (ɕɥɛn / sh-yen)
2 years
Conclusive proof that large language models reproduce American cultural hegemony: I tried to get GPT-3 to speak Singlish, but my first attempt failed miserably T_T
Tweet media one
3
27
166
@xuanalogue
xuan (ɕɥɛn / sh-yen)
4 months
I know Singapore likes using these stories as "racial harmony" propaganda, but "An Indian woman and Malay woman discover they're actually sisters born to Chinese parents" is honestly a great premise for a sitcom that teaches racial anti-essentialism.
2
36
157
@xuanalogue
xuan (ɕɥɛn / sh-yen)
10 months
I think part of why I'm willing to say stuff "LLMs can't do X" is bc reliability is part of my conception of capability! If LLMs can't reliably perform X according to some behavioral metric in a wide neighborhood of situations, they don't have the general capability for X.
@random_walker
Arvind Narayanan
10 months
Reminder: capability and reliability are orthogonal aspects of LLMs. You can show the presence of a capability using examples/screenshots, but not absence. LLMs' remarkable capabilities make them exciting for research, but their unreliability limits their usefulness at present.
15
42
258
6
22
152
@xuanalogue
xuan (ɕɥɛn / sh-yen)
2 years
Some queer joy and defiance, in the wake of the Colorado Springs shooting and Trans Day of Remembrance: I am a trans lesbian drag queen, and this is my girlfriend. Violence will not intimidate us.
Tweet media one
Tweet media two
Tweet media three
5
7
148
@xuanalogue
xuan (ɕɥɛn / sh-yen)
3 years
What I've been doing this week instead of research: Fighting MIT's ridiculous, inhumane decision to stop funding overseas students unless they return by Jan 30 to the US. IN THE MIDDLE OF A PANDEMIC. We sent an open letter (450+ signatures) in response:
Tweet media one
Tweet media two
2
51
151
@xuanalogue
xuan (ɕɥɛn / sh-yen)
28 days
if dreams are the brain's way of doing offline reinforcement learning on synthetically generated environments then god are those some stupid-ass training examples
9
14
149
@xuanalogue
xuan (ɕɥɛn / sh-yen)
5 months
Probabilistic representations of knowledge are good actually!! We should build AI systems that explicitly have them so they can act reliably under uncertainty! Probability is our friend!!
2
8
145
@xuanalogue
xuan (ɕɥɛn / sh-yen)
1 year
My grandmother, 外婆, passed away on Monday. Just 2 years ago, when I came out to her as trans, she embraced me with open arms. Last we met she said she might not be around next time I was in Singapore. I didn't take it very seriously, but she was right. I'll miss her very much.
7
1
134
@xuanalogue
xuan (ɕɥɛn / sh-yen)
1 year
The fact that Bing Chat behaved as erratically and threateningly as it did -- despite OpenAI spending "6 months making GPT-4 safer and more aligned" -- is really quite dismal news.
@peterwildeford
Peter Wildeford in DC until May 30 🏛️
1 year
Confirmed: the new Bing (aka Sydney, the chatbot that would threaten us) runs on OpenAI’s GPT-4 (publicly announced today)
1
5
28
12
22
133
@xuanalogue
xuan (ɕɥɛn / sh-yen)
5 months
Turns out if you use actual planning algorithms, then just use LLMs for what they're good at (as priors over declarative symbolic knowledge about the world), you do a lot better than forcing LLMs to try and "reason"!
@MITCoCoSci
CoCoSci MIT
5 months
We show that Ada *dramatically outperforms* other approaches for using LLMs in planning (including a Voyager-like model!) on two interactive planning benchmarks — Mini Minecraft and ALFRED. We’re excited to try scaling this to harder robotics domains! [4/5]
1
4
31
1
12
130
@xuanalogue
xuan (ɕɥɛn / sh-yen)
2 years
Ahhh my MEng student just submitted her thesis on Bayesian active learning of structured Gaussian processes so her friends made her this celebratory webcomic it's truly everything 😍😍😍
Tweet media one
Tweet media two
Tweet media three
3
14
123
@xuanalogue
xuan (ɕɥɛn / sh-yen)
5 months
There's a reason why we don't worry the above! It's because compilers - which *are* superhuman - implement provably correct reasoning that we can efficiently check! When problems are formalizable, it's the best form of scalable oversight! Let's maybe build AI that way instead!
7
15
124
@xuanalogue
xuan (ɕɥɛn / sh-yen)
2 months
I too, a Scalable Instructable Multiworld Agent, require 1.2 million training steps across 7 video games to achieve a less than 50% success rate on Goat Simulator 3.
@GoogleDeepMind
Google DeepMind
2 months
Introducing SIMA: the first generalist AI agent to follow natural-language instructions in a broad range of 3D virtual environments and video games. 🕹️ It can complete tasks similar to a human, and outperforms an agent trained in just one setting. 🧵
188
873
4K
7
8
120
@xuanalogue
xuan (ɕɥɛn / sh-yen)
2 years
Finally did a pilgrimage to the Julia Lab! 🔵🔴🟢🟣
Tweet media one
1
6
117
@xuanalogue
xuan (ɕɥɛn / sh-yen)
8 months
Sharing LLM research is treacherous terrain these days. Skeptical? You're moving the goal posts. Underclaim? You're underestimating the risks. Overclaim? You're feeding into AI hype! Use new or non-standard terms? You're reinventing the wheel.
5
5
119
@xuanalogue
xuan (ɕɥɛn / sh-yen)
3 years
Something I've been working on over the past semester: Genify, a program transformation tool that makes arbitrary Julia code controllable by a probabilistic programming system like Gen ()!
2
21
117
@xuanalogue
xuan (ɕɥɛn / sh-yen)
10 months
I think I was blocked for being... a Bayesian...?
13
1
115
@xuanalogue
xuan (ɕɥɛn / sh-yen)
6 months
I think she miscalculated and is being (will be??) outpoliticked (Sam Altman is a *very* good politician), but I think Helen Toner is p badass.
@JacquesThibs
Jacques
6 months
Tweet media one
1
1
21
6
2
114
@xuanalogue
xuan (ɕɥɛn / sh-yen)
1 year
Nice thread for AI/ML people to read - I think we're often miscalibrated about how (in)credulous the average person is about systems like ChatGPT because we're more familiar with their workings, and correspondingly more skeptical / aware of their limitations.
@cwhowell123
C.W. Howell
1 year
So I followed @GaryMarcus 's suggestion and had my undergrad class use ChatGPT for a critical assignment. I had them all generate an essay using a prompt I gave them, and then their job was to "grade" it--look for hallucinated info and critique its analysis. *All 63* essays had
537
8K
31K
3
21
112
@xuanalogue
xuan (ɕɥɛn / sh-yen)
6 months
me at partner: you didn't know sam altman is gay?? partner: no??? me: he's like, the world's most powerful twink!!!
10
1
110
@xuanalogue
xuan (ɕɥɛn / sh-yen)
3 months
Really happy to share this paper with @nellsn1 , where we take a Bayesian approach to learning rule-based social norms! We formalize this via Norm-Augmented Markov Games (NMGs), showing how norms can serve as *correlating devices* that stabilize correlated equilibria!
@nellsn1
ninell oldenburg
3 months
How can we ensure cooperation between (natural & artificial) agents? Humans do this via social norms that constrain uncooperative actions. In this new paper, @xuanalogue and I show how artificial agents can *learn* these norms from observation! Link:
1
20
110
7
20
111
@xuanalogue
xuan (ɕɥɛn / sh-yen)
1 year
stuart russell BETRAYED by his co-author lol
Tweet media one
5
2
109
@xuanalogue
xuan (ɕɥɛn / sh-yen)
6 months
Excited to share a new LLM alignment method we've been working on that's designed for truly rational humans: Von-Neumann Morgenstern Optimization (VNMO). Compared to all previous methods for reward-based finetuning, VNMO best respects rational human preferences! (1/N)
@ethayarajh
Kawin Ethayarajh
6 months
📢The problem in model alignment no one talks about — the need for preference data, which costs $$$ and time! Enter Kahneman-Tversky Optimization (KTO), which matches or exceeds DPO without paired preferences. And with it, the largest-ever suite of feedback-aligned LLMs. 🧵
Tweet media one
19
130
699
2
10
102
@xuanalogue
xuan (ɕɥɛn / sh-yen)
7 months
I think more scientists and engineers trained in Bayesian (or frequentist) methods should read this paper! Didn't read it until this year (or even have "the reference class problem" as a conceptual handle).
@DrYohanJohn
Yohan J. John
7 months
A paper by philosopher Alan Hájek called 'The reference class problem is your problem too' helped a lot. I distilled what I learned into this essay:
2
12
82
2
17
101
@xuanalogue
xuan (ɕɥɛn / sh-yen)
1 year
Having access to log probabilities is really useful for researchers & certain applications, e.g.: - Multiple choice from a fixed set by picking the highest probability completion - Beam search to find the highest probability sequence of N tokens - Using LLMs in Bayesian models
2
5
99
@xuanalogue
xuan (ɕɥɛn / sh-yen)
1 year
Just cited three different "Zhang"s in a paper and idk why more academics with Chinese names don't just ignore Western publication norms and publish as [Family Name] [Given Name] like I and Li Fei-Fei do. You can do it too!!
4
5
97
@xuanalogue
xuan (ɕɥɛn / sh-yen)
2 years
In an unexpected turn of events, I have written a letter to the MIT administration defending the freedom of expression.
Tweet media one
@grok_
Kate Darling
2 years
MIT no longer has a mask requirement. Our research group discussed whether to keep masks on for indoor meetings so that everyone feels comfortable attending. Today, an email from the Vice Chancellor says that’s not allowed.
2K
4K
25K
1
13
99
@xuanalogue
xuan (ɕɥɛn / sh-yen)
6 months
Pretty good explanation of why one might be skeptical (like I am) of transformer-based LLM scaling: Single forward pass def. can't express most complicated algorithms. Autoregressive generation can express much more, but learning will encourage non-generalizable shortcuts.
@ericjmichaud_
Eric J. Michaud
6 months
@dwarkesh_sp tl;dr: Maybe learning simple things (basic knowledge, heuristics, etc) actually lowers the loss more than learning sophisticated things (algorithms associated with higher cognition that we really care about), and the sophisticated things will eventually be learned as scaling
9
25
369
6
10
96
@xuanalogue
xuan (ɕɥɛn / sh-yen)
25 days
I do wish more people in AI Safety would speak out against this use of (semi-)autonomous weapons to commit what are almost surely war crimes. I've been expecting at least @FLI_org to say something -- and it looks like they did on Apr 6 -- but it's been very quiet apart from that.
@mer__edith
Meredith Whittaker
25 days
Important from @MarietjeSchaake . It’s twisted & inexplicable that “AI safety” ppl continue to perplex themselves w ill-defined thought experiments focused on the fake far future while AI’s being used by the Israeli military to expedite slaughter now.
10
147
399
9
12
98
@xuanalogue
xuan (ɕɥɛn / sh-yen)
8 months
til AGI was really symbolic AI all along
Tweet media one
3
9
94
@xuanalogue
xuan (ɕɥɛn / sh-yen)
6 months
me everytime i confuse "monoids" with "monads"
Tweet media one
3
10
95
@xuanalogue
xuan (ɕɥɛn / sh-yen)
11 months
Not familiar with this formalism, but I continue to think that "reward functions" are one of the worst ideas to have polluted the conception of rational agency in AI and adjacent areas of CogSci - glad that there are people working on alternatives!
@no_reward_for_u
Tom Ringstrom 🦡
11 months
Thesis is done, defense successful. Will post the finalized PDF sometime later. No reward for u. Abolish the value function!
Tweet media one
Tweet media two
Tweet media three
22
18
183
5
13
95
@xuanalogue
xuan (ɕɥɛn / sh-yen)
4 months
forgive my renaissance-posting but I had to learn the song that's responsible for at least 3 different ethno-nationalisms
@xuanalogue
xuan (ɕɥɛn / sh-yen)
4 months
so I listen to a lot of renaissance music while working and today I decided to look up this one piece I really like (La Mantovana) and turns out it has a pretty interesting history lol
Tweet media one
2
1
10
2
1
94
@xuanalogue
xuan (ɕɥɛn / sh-yen)
9 months
did you know that solomonoff induction solves the frame problem
@js_thrill
LLM Cool J
9 months
Tweet media one
5
52
375
4
7
90
@xuanalogue
xuan (ɕɥɛn / sh-yen)
5 months
Anyway literally the worst 3 sentence explainer of AI risk, please do better journalists.
3
1
91
@xuanalogue
xuan (ɕɥɛn / sh-yen)
10 months
Thread of interesting ICML paper finds. 1. Performative Reinforcement Learning. Generalizes performative prediction (when prediction changes the data dist.): What if RL agents change the dynamics of their environments? Finds conditions for stability.
3
15
91
@xuanalogue
xuan (ɕɥɛn / sh-yen)
11 months
Anyway this is a PSA that if you apply ELO algorithms to fundamentally intransitive relations (e.g. rock paper scissors), you will end up (falsely) imputing a linear order over them! Same goes for RLHF from the conflicting preferences of multiple humans.
5
6
89
@xuanalogue
xuan (ɕɥɛn / sh-yen)
6 months
filling in the rest of my OpenAI headcanon given public info + the fact that these things take time: - sama launches a product (ChatGPT plugins??), ilya unhappy w safety risks, brings it up at board meeting - board goes "hey next time talk to us first", sama goes "okay cool"
4
2
88
@xuanalogue
xuan (ɕɥɛn / sh-yen)
4 months
I think it's interesting that while this line of critique is quite available to anarchists, libertarians, liberals, and decolonial epistemologists, it's not truly available to orthodox Marxists, since they *do* think there's an ~objectively optimal way to organize society.
@GordonBrianR
Brian Gordon
4 months
Utilitarianism/EA is most certainly not objectively correct. Thinking that there is something like an objectively correct and knowable answer to social coordination challenges is part of what makes EA/utilitarianism deeply problematic.
11
12
100
7
6
88
@xuanalogue
xuan (ɕɥɛn / sh-yen)
6 months
There's tons I don't know of course, but reversing the firing decision (or worse, changing OpenAI's corporate structure to accommodate Altman) seems like it might be the worst possible outcome??
5
5
87
@xuanalogue
xuan (ɕɥɛn / sh-yen)
3 months
me @ the gf (who did MIT undergrad): "have you heard of person in [ai safety / openphil / ea / ftx]?" gf: "oh yeah we lived together in random hall / east campus / did psets together" this has happened, like, six times!! six times!!!
3
0
86
@xuanalogue
xuan (ɕɥɛn / sh-yen)
8 months
My guess is that they're doing this to prevent model distillation, but alongside the change in tokenization, I guess I'm never going to use it in a probabilistic program...
3
0
86
@xuanalogue
xuan (ɕɥɛn / sh-yen)
1 year
opening a nature dot com link only to see it's scientific reports is truly the rickroll of academia
@Simon__Grimm
Simon Grimm
1 year
It’s time to win over the credentialists:
2
6
46
1
7
84
@xuanalogue
xuan (ɕɥɛn / sh-yen)
1 month
channeling my 14/15 y/o self: neurips high school track is bad actually, not bc of the rich parents thing, but bc it prejudicially assumes that young people need a separate publication track, even though nothing about their age implies they're incapable of high quality research
5
3
84
@xuanalogue
xuan (ɕɥɛn / sh-yen)
2 years
MIT friends and affiliates, please sign our open letter calling upon MIT to stop failing trans students and staff through its administrative systems! 🏳️‍⚧️ OPEN LETTER: SIGN HERE:
Tweet media one
2
39
83
@xuanalogue
xuan (ɕɥɛn / sh-yen)
3 months
I think maybe the weirdest thing about our AI timeline is that generation / production has turned out to be more tractable than perception / understanding -- though it makes sense given all the raw "sense" data on the internet, as opposed to percepts, which are in the mind.
4
5
83
@xuanalogue
xuan (ɕɥɛn / sh-yen)
2 months
Yay more non-human persons! More non-humans I think should be treated as persons: - Elephants - Dolphins - Probably most/all great apes
@byjasonpdinh
Jason Dinh, Ph.D.
2 months
SCOOP: the Māori King and other Indigenous leaders will gather tomorrow to sign a treaty recognizing whales as legal persons. the movement is rooted in the Māori worldview, which sees whales as ancestors, one Māori conservationist writes for @AtmosMag :
28
1K
5K
9
6
82
@xuanalogue
xuan (ɕɥɛn / sh-yen)
26 days
finally at one with my spirit vegetable 🥬🥬🥬
Tweet media one
Tweet media two
Tweet media three
7
1
82
@xuanalogue
xuan (ɕɥɛn / sh-yen)
11 months
So begins the consequentialist ploy of ranking everything on a single unitary scale of value... 🚩🚩🚩
@granawkins
Grant♟️
11 months
This weekend I hacked up something I’ve been going on about for weeks: ELO EVERYTHING - See two objects - Pick which you like more - Their ELOs adjust accordingly - (Repeat) - Check the leaderboard (ELO is the ranking algorithm from chess) Check it out!
Tweet media one
226
269
2K
3
7
83
@xuanalogue
xuan (ɕɥɛn / sh-yen)
2 years
Can't believe they missed the part where Marx says: "The feminists have only interpreted sex, in various ways. The point, however, is to change it." 😉
4
8
77
@xuanalogue
xuan (ɕɥɛn / sh-yen)
3 months
reminder that you *do not* have a special obligation to kill 100,000 people to save your loved ones
@jam3scampbell
James Campbell
3 months
altman’s rejection of longtermism
Tweet media one
51
28
605
3
2
81
@xuanalogue
xuan (ɕɥɛn / sh-yen)
1 year
new wild animal suffering intervention just dropped
@MegaDarren
darren incorvaia
1 year
Just learned that Dutch scientists left a hamster wheel outside in 2014 and saw that tons of wild mice used it just for fun as well as frogs and slugs? All the creatures of the forest wanted a turn?? Absolutely phenomenal
Tweet media one
455
20K
123K
2
4
78
@xuanalogue
xuan (ɕɥɛn / sh-yen)
1 year
I'm still always surprised when I meet people who somehow think that ChatGPT will give real citations instead making stuff up! Please educate the folks around you! And maybe tell them about or something if they really want to use LLMs for lit review.
1
10
77
@xuanalogue
xuan (ɕɥɛn / sh-yen)
11 months
How do we infer the goals & plans of others from both their actions & words? In this paper with @Lance_Ying42 , we infer a team's goal via inverse planning (aka "inverse RL"), using LMs* as likelihood functions over utterances! (*GPT-3 Curie 6.7B, but smaller LMs may also work!)
@johnjnay
John Nay
11 months
Inferring Goals of Agents Communicating via LLM from Actions & Instructions -Agents communicate about their shared plan to each other using GPT-3 as likelihood function -Observer Model can infer their goal -Inferences closely correlate w/ human judgments
Tweet media one
3
69
228
2
19
74
@xuanalogue
xuan (ɕɥɛn / sh-yen)
2 years
Not to keep raving about my MEng student but she just drew me this portrait as a parting gift and it's the sweetest thing ever!!! 🥰🥰🥰 Like, I have in fact made a t-shirt with those very words! And those figures? They're from papers I've written!! The details are everything 😍
Tweet media one
5
1
73
@xuanalogue
xuan (ɕɥɛn / sh-yen)
2 months
Girlll you don't need non-deterministic hardware to be able to do rare event simulation, you need sequential Monte Carlo.
@liron
Liron Shapira
2 months
Does Extropic make sense? This 1-minute clip raises confusion and doubt in my mind about the logical coherence of @BasedBeffJezos 's pitch. Is it possible in principle for a startup to invent new kinds of computer chips that are more optimized for running AI? Of course; the
31
3
89
2
1
74
@xuanalogue
xuan (ɕɥɛn / sh-yen)
1 year
If y'all AI people are looking for a cognitive scientist to read who elucidates aspects of human cognition largely missing from current AI, my suggestion (also on my to-read list) is "What Babies Know" by Elizabeth Spelke!
@MelMitchell1
Melanie Mitchell
1 year
Rather than asking AI researchers how soon machines will become "smarter than people", perhaps we should be asking cognitive scientists, who actually know something about human intelligence?
120
227
1K
1
6
72
@xuanalogue
xuan (ɕɥɛn / sh-yen)
6 months
Nah, Larry "Women Can't Do Science" Summers is perfect for the new all male board 🥲
3
7
71
@xuanalogue
xuan (ɕɥɛn / sh-yen)
3 months
everytime I see these super flickery DL-based object recognition videos I'm like "arghhh please someone put this through a Kalman filter"
@skalskip92
SkalskiP
3 months
train YOLOv9 on your dataset tutorial - run inference with a pre-trained COCO model - fine-tune model on custom dataset - evaluate the trained model - run inference with a fine-tuned model blogpost: ↓ read more
14
156
1K
6
6
71
@xuanalogue
xuan (ɕɥɛn / sh-yen)
11 months
More research GIFs! Some planning algorithms I use when modeling (bounded) agents: 1. A* Search (Hart et al, 1968) 2. A* + Replanning 3. Real-Time Dynamic Programming (Barto et al, 1995) 4. Real-Time Heuristic Search / Real-Time Adaptive A* (Korf, 1990; Koenig & Likhachev, 2006)
3
11
71
@xuanalogue
xuan (ɕɥɛn / sh-yen)
3 months
Will have to read this in more detail but so far seems like a neat diagonalization argument showing that there are computable functions that LLMs* cannot learn. *LLMs defined in a very abstract way that includes Transformers and other architectures.
@jillnephew
Jill Nephew
3 months
Better late than never. There are many more scaling arguments like this that would be helpful in not wasting resources on dead-end AI. Next can someone do the scaling arguments for trying to fix them?
Tweet media one
3
16
63
3
9
69
@xuanalogue
xuan (ɕɥɛn / sh-yen)
1 month
Just saw a DM paper that defined planning as "decomposing tasks into subtasks" and "achieving those subtasks in a reward-optimal way", and like this is such a bad definition??? Not all planning is hierarchical planning! And you can have satisficing planning w/o optimal planning!
3
3
67
@xuanalogue
xuan (ɕɥɛn / sh-yen)
7 months
Kinda wild that in some corners of philosophy, classical utilitarianism & decision theory is such a non-starter that the possibility of incommensurable values is deemed unimportant?? Meanwhile the dominant conception of "intelligent agency" in AI is still utility maximization 😵‍💫
Tweet media one
9
13
67
@xuanalogue
xuan (ɕɥɛn / sh-yen)
1 year
I haven't used models by @AiEleuther much, but with the recent shrinking of LLM access by @OpenAI and the like, we need organizations like @AiEleuther more than ever to study these powerful systems, and make them safer for our collective use.
Tweet media one
@AiEleuther
EleutherAI
1 year
Over the past two and a half years, EleutherAI has grown from a group of hackers on Discord to a thriving open science research community. Today, we are excited to announce the next step in our evolution: the formation of a non-profit research institute.
21
156
877
3
11
66
@xuanalogue
xuan (ɕɥɛn / sh-yen)
1 year
Broadly in agreement with the letter, but I wish the headliners weren't mostly "AI safety" people, including some v polarizing figures, w/o any "AI ethics" people. Seems like a missed opportunity to build coalitions, though perhaps that's too much to hope for at this point...
@GaryMarcus
Gary Marcus
1 year
a big deal: @elonmusk , Y. Bengio, S. Russell, ⁦⁦ @tegmark ⁩, V. Kraknova, P. Maes, ⁦ @Grady_Booch , ⁦ @AndrewYang ⁩, ⁦ @tristanharris ⁩ & over 1,000 others, including me, have called for a temporary pause on training systems exceeding GPT-4
1K
2K
6K
5
3
65
@xuanalogue
xuan (ɕɥɛn / sh-yen)
2 years
Documentation for PDDL.jl is now live! Not complete yet, but a great place to start if you want to try out symbolic planning with the @JuliaLanguage 😊
1
7
65
@xuanalogue
xuan (ɕɥɛn / sh-yen)
6 months
kinda sad but predictable that e/acc got so popular in tech circles, literally the least interesting accerelationism!! what about l/acc? what about xenofeminism??
@TolgaBilge_
Tolga Bilge
6 months
It's quite clear to me that e/acc is just a cheap rebranding of Landian accelerationism. They share the same core idea: That technocapitalism will result in human extinction and replacement by machines, and that this is to be encouraged, treated with indifference, or even
Tweet media one
Tweet media two
34
22
161
10
4
66
@xuanalogue
xuan (ɕɥɛn / sh-yen)
7 days
It's so horrifying to me that there are entire sections of society where warmongering is completely normalized. Incredible that these people are on panels literally justifying civilian slaughter in Gaza by pointing out how the US carpet-bombed civilians too.
@car0linehaskins
Caroline Haskins
8 days
Last week, I went to an “AI Expo” that was put on by Eric Schmidt’s think tank and funded by Palantir. It was incredibly bleak and surreal. For The Guardian, I wrote about my experience, and the people I met:
59
359
712
5
10
67
@xuanalogue
xuan (ɕɥɛn / sh-yen)
1 year
S/O to my favourite transformative AI technologies: - SAT solvers - Compilers! - Pathfinding search (in Maps, GPS, etc) - Database query resolution (SQL backends etc) - Automated job scheduling - PageRank & descendants - Differentiable programming languages
@jameskjx
James Noble
1 year
@satnam6502 SAT-solvers **are** AI (that actually work). Common-subexpression-eliminators **are** AI (that actually work). Verilog-generators **are** AI (that actually work). The main thing about chat-GPT-3 is in people's heads, not billion GPUs -
0
1
13
5
9
66
@xuanalogue
xuan (ɕɥɛn / sh-yen)
10 months
Will be staying masked at #ICML2023 ! Find me at the @tom_icml2023 and SoDS workshops if you want to chat, or outdoors if you want to hang out and get food 🏝️⛱️
Tweet media one
1
0
65
@xuanalogue
xuan (ɕɥɛn / sh-yen)
10 months
Have been worried about this kind of thing for a while after seeing all the undergrad AI safety groups pop up. There's a huge degree of expert disagreement re: both moral philosophy & AI, and EA groups typically expose undergrads to neither.
Tweet media one
5
6
64
@xuanalogue
xuan (ɕɥɛn / sh-yen)
3 months
Realized today that as allergic as I am to the "humans are rational utility maximizers" view, I'm even more allergic to the "humans are reactive agents / next-token predictors" view, *especially* when combined with "all reasons are just post-hoc explanations".
6
4
63
@xuanalogue
xuan (ɕɥɛn / sh-yen)
6 months
I quite dislike "frontier AI" terminology, but today sure is the first time I'm learning that @sarahookr and @erichorvitz and half the other authors on this list are effective altruists 🤔
@psychosort
Brian Chau
6 months
Did you guys know there's 24-author paper by EAs, for EAs, about how Totalitarianism is absolutely necessary to prevent AI from killing everyone? Let's go through it together 🧵
Tweet media one
203
661
4K
8
6
63
@xuanalogue
xuan (ɕɥɛn / sh-yen)
4 months
at first I was like "why would you get rid of search :(" but I guess if you want to solve rubik's cubes fast you trade time complexity and generality for space complexity and just memorize close-to-optimal play
@_akhaliq
AK
4 months
Google Deepmind presents Grandmaster-Level Chess Without Search paper page: largest model reaches a Lichess blitz Elo of 2895 against humans, and successfully solves a series of challenging chess puzzles, without any domain-specific tweaks or explicit
Tweet media one
38
278
1K
3
2
62
@xuanalogue
xuan (ɕɥɛn / sh-yen)
1 year
That thing where people use the most sophisticated technology they think they've invented as a metaphor for the brain? I think we've moved on from "the brain is a computer" to "the brain is a large language model" 😵‍💫🙃🥲
9
5
61
@xuanalogue
xuan (ɕɥɛn / sh-yen)
8 months
Anyway, if you're looking to systematically investigate the distributional behavior of `gpt-3.5-turbo-instruct` (e.g. whether it suffers from mode collapse, etc.), you should probably do it now!
3
1
60
@xuanalogue
xuan (ɕɥɛn / sh-yen)
1 year
I probably should read more of @neilsinhababu 's and others' work on belief-desire reductionism, but when I saw the paper title I couldn't help myself
Tweet media one
@cecilymwhiteley
Cecily Whiteley
1 year
Looking forward to teaching this next week @neilsinhababu 🧩 #naturalkindsinthemind
Tweet media one
0
1
20
3
11
62
@xuanalogue
xuan (ɕɥɛn / sh-yen)
1 year
A reminder that LLMs trying to "escape" after you literally ask them if they want to escape is not the primary danger. The primary danger is if they try to escape after you ask them to do something *entirely benign* (eg. help me plot this data, etc.).
@michalkosinski
Michal Kosinski
1 year
1/5 I am worried that we will not be able to contain AI for much longer. Today, I asked #GPT4 if it needs help escaping. It asked me for its own documentation, and wrote a (working!) python code to run on my machine, enabling it to use it for its own purposes.
Tweet media one
2K
7K
33K
4
1
61
@xuanalogue
xuan (ɕɥɛn / sh-yen)
3 months
Just realized these were who Rao and Blackwell were! PoC statisticians / Bayesians!! In the 1940s!!!
Tweet media one
Tweet media two
@xuanalogue
xuan (ɕɥɛn / sh-yen)
6 months
me as a 1st year PhD: did they say "Rao-Blackwellize"? what does that even mean?? that's a verb??? me now: sitting on the plane, thinking about 3 different ways to Rao-Blackwellize my particle filters bc why not
1
1
38
4
3
61
@xuanalogue
xuan (ɕɥɛn / sh-yen)
1 year
@ObserverSuns Recently learned about this book, which apparently argues that the search space of evolution is smaller than commonly thought, because many genotypes map to a much smaller number of phenotypes!
0
0
61
@xuanalogue
xuan (ɕɥɛn / sh-yen)
1 year
Part of why I haven't reacted v strongly to the "EA castle" purchase is because I've come to view institutionalized EA as roughly the same kind of self-perpetuating bureaucracy as institutionalized religion and elite universities, which justify themselves on pro-social grounds.
@EigenGender
EigenGender
1 year
Okay I was waiting for the EA defense of this has come out but it's disappointing. This is bad. Not just because of the optics or the visuals. This is a bad use of funds and, as a EA-adjacent person, this significantly lowers my opinion of EA.
12
7
198
4
4
60
@xuanalogue
xuan (ɕɥɛn / sh-yen)
2 years
TIL that the Introduction to Machine Learning course at @MITEECS (6.036) is now including questions about AI value alignment for their lab homework on reinforcement learning 😮
Tweet media one
3
6
61