
Diego Calanzone
@diegocalanzone
Followers
247
Following
4K
Media
69
Statuses
660
« artificia docuit fames » // phd at @Mila_Quebec, intelligence by agency + deep learning for science // AI grad @UniTrento
127.0.0.1
Joined April 2015
RT @RL_Conference: Ending with our last RLC oral, @RichardSSutton with "The Oak Architecture: A Vision of SuperIntelligence from Experienc….
0
8
0
RT @giffmana: Amazing! Truly open review, through which we all gained more insights, i love it!. Result: in multi epoch setting, making AR….
0
36
0
RT @sporadicalia: just remembered that time Noam Shazeer dropped the hardest line ever written in an ML paper
0
622
0
RT @lavoiems: 🧵 Everyone is chasing new diffusion models—but what about the representations they model from?.We introduce Discrete Latent C….
0
45
0
RT @PontiEdoardo: We blend imitation (SFT) and exploration (RLVR) in post-training with a simple idea:. Sample a prefix of an SFT demonstra….
0
7
0
RT @steveazzolin: This is an issue on multiple levels, and authors using those "shortcuts"👀 are equally responsible for this unethical beha….
0
2
0
A comprehensive article on ways to Hierarchical RL!.
As AI agents face increasingly long and complex tasks, decomposing them into subtasks becomes increasingly appealing. But how do we discover such temporal structure?. Hierarchical RL provides a natural formalism-yet many questions remain open. Here's our overview of the field🧵
0
0
2
RT @BlancheMinerva: A good warning lesson on using AIs to write papers: this alleged response to the (dubious) "Illusion of Thinking" paper….
arxiv.org
Shojaee et al. (2025) report that Large Reasoning Models (LRMs) exhibit "accuracy collapse" on planning puzzles beyond certain complexity thresholds. We demonstrate that their findings primarily...
0
47
0
RT @BlancheMinerva: Two years in the making, we finally have 8 TB of openly licensed data with document-level metadata for authorship attri….
0
69
0