manoelribeiro Profile Banner
Manoel Profile
Manoel

@manoelribeiro

Followers
5K
Following
7K
Media
622
Statuses
4K

Computational Social Science, Platforms, GenAI, Moderation

Joined January 2009
Don't wanna be here? Send us removal request.
@timalthoff
Tim Althoff
3 days
(please reshare) I'm recruiting multiple PhD students and Postdocs @uwcse @uwnlp ( https://t.co/I5wQsFnCLL). Focus areas incl. psychosocial AI simulation and safety, Human-AI collaboration. PhD: https://t.co/ku40wCrpYh Postdocs: https://t.co/K9HUIPJ5h6
8
104
381
@tiancheng_hu
Tiancheng Hu
3 days
Can AI simulate human behavior? 🧠 The promise is revolutionary for science & policy. But there’s a huge "IF": Do these simulations actually reflect reality? To find out, we introduce SimBench: The first large-scale benchmark for group-level social simulation. (1/9)
3
19
49
@cervisiarius
Bob West
3 days
📄✨Excited to share our new paper accepted to #EMNLP ’25: Combining Constrained and Unconstrained Decoding via Boosting: BoostCD and Its Application to Information Extraction https://t.co/ljsWULBHEA (led by #EPFL PhD student Marija Šakota -- soon on the job market, hire her!!)
1
7
14
@SmithaMilli
smitha milli
4 days
can we finally use natural language to optimize for deeper notions of what users want from their recommender systems?
2
12
48
@manoelribeiro
Manoel
6 days
I argue that if we consider these three points, we find that labeling with LLMs is neither trick nor treat. Treated as measurement instruments, their value lies in forcing us to confront uncertainty we once ignored, not in completely eliminating it.
0
0
1
@manoelribeiro
Manoel
6 days
Where do we go from here? I argue we should do three things: 1) clarify what problem we’re solving; 2) clarify the threat model of malicious actors who may engage in bad science; 3) we need to embrace the fact that human labels can also suck.
1
0
1
@manoelribeiro
Manoel
6 days
I cluster work in this area broadly into three waves: the “wow” phase (e.g., Gillardi’s PNAS paper), the “how do we do this right?” phase (e.g., Egami’s DSL), and the “the boat is on fire” wave (e.g., Baumann’s LM hacking).
1
0
1
@manoelribeiro
Manoel
6 days
Large language models are quietly transforming how social scientists label data. In dozens of new studies, undergrad coders and Turkers have been replaced by GPT-5 or Gemini 2.5 (or whatever new model just arrived). What began as a convenience is becoming a methodological shift.
1
0
1
@manoelribeiro
Manoel
6 days
The debate over “LLMs as annotators” feels familiar: excitement, backlash, and anxiety about bad science. My take in a new blogpost is that LLMs don’t break measurement; they expose how fragile it already was. https://t.co/6CweDPv5wG
1
8
20
@gvrkiran
Kiran Garimella
6 days
We're over-indexing on the negatives of AI companions. While the risks are real, the panic-driven narrative ignores the massive, stigmatized crisis of loneliness they're trying to solve. I wrote a short post on the positive use cases we shouldn't ignore https://t.co/ZvkxBkJswy
0
2
6
@icwsm
ICWSM
8 days
Ready for #ICWSM2026 in LA next May? 🌴 📅 Deadlines: • Jan 15, 2026 – Papers, posters, demos, datasets, tutorials (note: posters/demos/datasets submitted in Sep '25 are reviewed only after this deadline) • Jan 30, 2026 – Workshop proposals 🔗 Details: https://t.co/KTnOPzuutb
0
7
13
@ben_golub
Ben Golub
8 days
I am super excited to share a new AI tool, Refine. Refine thoroughly studies research papers like a referee and finds issues with correctness, clarity, and consistency. In my own papers, it regularly catches problems that my coauthors and I missed. 1/
37
204
1K
@jonathanstray
Jonathan Stray
9 days
Look, I'm afraid I have to tap the sign again. Last time someone looked at this, TikTok was found to "boost" Republican content. This time, the opposite. But what does it mean for an algorithm to "boost" content? What about what the users are doing? https://t.co/HbKqZjwVvU
Tweet card summary image
doomscrollingbabel.manoel.xyz
"There and back again"
@YADodeles
Yehonatan Dodeles
9 days
Our new research suggests TikTok’s algorithm systematically boosts pro-Mamdani and anti-Cuomo content in NYC’s mayoral race. This analysis was done by Spring AI. the company uncovers the hidden mechanics of social media algorithms to defend democracies from weaponized AI. 🧵
2
2
4
@jonathanstray
Jonathan Stray
8 days
GreenEarth is creating open source AI-driven recommender infrastructure for the ATProto ecosystem. Type a prompt, see your feed change. We are here for the users, the builders, the dreamers. Join us. https://t.co/5wme8hcdJv
Tweet card summary image
greenearth.social
We're building advanced open source algorithms for social media
0
3
12
@lianegalanti
Liane Galanti
9 days
Feels like a dream! I’ve recently started my Ph.D. in Computer Science @Princeton! Working on exciting research with Professors @HazanPrinceton and @tri_dao 🤩
110
96
2K
@PrincetonCS
Princeton Computer Science
9 days
Adji Bousso Dieng, assistant professor of computer science and leader of the @Vertaix_ lab, has won the Prix Galien Africa Special Prize from @galienafrique. She will receive the award in Dakar, Senegal on October 31. https://t.co/FMgpee4IZM
1
3
13
@manoelribeiro
Manoel
10 days
I'd also expect that agreement between humans is not better than 0.6 for most of these papers... The authors note that a benchmark suggests an agreement of 0.73, but there is considerable selection bias in who would publish such agreement metrics in the first place.
0
0
2
@manoelribeiro
Manoel
10 days
On that note, I am not sure this follows from the actual content of the paper. Both the recent LLM p-hacking paper and this newest paper could simply reflect that the data annotation for the original papers is bad? https://t.co/iS9Ea55cFp
@joabaum
Joachim Baumann
10 days
Cool paper by @ey_985, confirming our LLM hacking findings ( https://t.co/24Fyb4IRT3): âś“ LLMs are brittle data annotators âś“ Downstream conclusions often flip: *LLM hacking risk* is real! âś“ Bias correction methods can help but have tradeoffs âś“ Use human expert whenever possible
2
0
4
@manoelribeiro
Manoel
10 days
It is worth noting that many papers replicated in Yang et al. used the previous generation of automated methods. Most were not purely human annotations. This paper would be so cool if the authors actually reannotated some of the data by themselves...
@ey_985
Eddie Yang
11 days
New paper: LLMs are increasingly used to label data in political science. But how reliable are these annotations, and what are the consequences for scientific findings? What are best practices? Some new findings from a large empirical evaluation. Paper: https://t.co/F8FlrsLbzM
2
0
3
@amyxzh
Amy Zhang
10 days
Congrats to @galenweld for winning yet another paper award for his great work on measuring perceptions of community moderation at Reddit scale! #cscw2025
@galenweld
Galen Weld @CSCW
11 days
If you're at @ACM_CSCW 🇳🇴, come check out our 🏆honorable mention paper today on Reddit community governance (w/@amyxzh @timalthoff). 4pm in Dovregubben-2! More in thread... 🧵
1
1
25