
Kawin Ethayarajh
@ethayarajh
Followers
4K
Following
2K
Media
76
Statuses
1K
Assistant Professor @UChicago @ChicagoBooth. PhD @StanfordAILab @stanfordnlp.
Palo Alto, California
Joined March 2019
š¢ Belated update, but I'm thrilled to share that I've joined @UChicago @ChicagoBooth as an Assistant Professor in the newly created Applied AI group!.I'll continue to work on behavior-bound machine learning: understanding how AI shapes, is shaped, and should be shaped by the
56
27
610
RT @YejinChoinka: Honored to be back on TIME100 AI for 2025 ā alongside my longtime heroes @drfeifei and @BarzilayRegina! š. The recognitioā¦.
0
35
0
RT @PeterHndrsn: I'm starting to get emails about PhDs for next year. I'm always looking for great people to join!. For next year, I'm lookā¦.
0
28
0
RT @Diyi_Yang: Introducing āØGenerative Interfaces where LLMs respond to users by proactively generating UIs that enable adaptive interactioā¦.
0
39
0
RT @Muennighoff: Can AI solve open problems in math, physics, coding, medical sciences & beyond?. We collected unsolved questions (UQ) & teā¦.
0
244
0
2019 - 2024 was such a bizarre time to be doing a PhD in an NLP group, having first entered the field when word embeddings were just becoming a thing; it felt like I aged 20 years in 5. Iām immensely grateful for having had the support of my advisor @jurafsky ,.
1
0
14
Some questions I'll be working on:.- When we think about personalizing LLMs, we only think about changing the data, even though the objective itself encodes a utility function and different people have different utility functions (. What does.
arxiv.org
Kahneman & Tversky's $\textit{prospect theory}$ tells us that humans perceive random variables in a biased but well-defined manner (1992); for example, humans are famously loss-averse. We show...
1
0
15
RT @JeffDean: AI efficiency is important. Today, Google is sharing a technical paper detailing our comprehensive methodology for measuringā¦.
0
840
0
> Right now, there is no established RL algorithm to give a model verbal feedback and have it update it's weights. There is also no established algorithm for a model to reflect on a previous failed execution and update it's own weights. I think there are? This doesn't seem like.
Imagine you're trying to teach a human how to do a task, say install Windows XP in a virtual machine. The human walks into a room and sees a document (prompt) that you have written, that describes exactly what they are supposed to do. There is also a computer ready for their
2
1
18
RT @oshaikh13: If you thought referencing past chats was cool, we built an MCP that lets Claude use *anything you see or do on your computeā¦.
0
32
0
The problem is that the token economics arenāt sustainable:. 1. If you have too many models (e.g., 4o, o3, _and_ all the GPT-5 variants), then the cost of inference explodes: amortizing over very large batches is necessary to make any one model worth hosting. But a large minority.
ChatGPT users are canceling subscriptions. GPT-5 as the main source of intelligence does not work. Intelligence isnāt one-size-fits-all; user needs and preferences vary wildly. We want more models. We want more control.
2
2
26
RT @sayashk: How does GPT-5 compare against Claude Opus 4.1 on agentic tasks? . Since their release, we have been evaluating these models oā¦.
0
70
0
RT @nabeelqu: The 'vibe shift' on here is everyone realizing they will still have jobs in 2030.
0
439
0
RT @CAAI_Booth: Joining @ChicagoBooth's efforts to invest in #AI as it shapes business, markets, and institutions are three new Applied AIā¦.
chicagobooth.edu
Booth welcomes three new professors working at the intersection of AI, technology, business, and society.
0
3
0
RT @charles_irl: In a new blog post for @modal_labs, I argue against the prevailing denomination of LLM services in terms of dollars per toā¦.
0
4
0