timos_m Profile Banner
Timoleon (Timos) Moraitis Profile
Timoleon (Timos) Moraitis

@timos_m

Followers
1K
Following
4K
Media
176
Statuses
2K

Building brain-like AI @noemon_ai Previously @Huawei @IBMResearch @ETH_en @UZH_en @ntua

Zurich, Switzerland
Joined January 2009
Don't wanna be here? Send us removal request.
@timos_m
Timoleon (Timos) Moraitis
7 days
@timos_m
Timoleon (Timos) Moraitis
13 days
All the "ARC-AGI is toast" people forget that there are two axes. This is a benchmark for skill acquisition efficiency. @fchollet @GregKamradt @mikeknoop @arcprize
0
0
0
@timos_m
Timoleon (Timos) Moraitis
7 days
Yes. People begin realizing the real-world implications of brute-force scaling. Being "frontier" will soon be considered meaningless, unless it's the *Pareto frontier* of cost vs capability. And that needs new foundational improvements.
@theallinpod
The All-In Podcast
8 days
What Happens When AI Tokens Cost More Than Your Employees? @Jason: “We, with our agents, hit $300/day per agent using the Claude API, like instantly. And that was doing, maybe, 10 or 20%. That's $100k/year per agent.” @chamath: “We're getting to a place where we have to
2
0
1
@timos_m
Timoleon (Timos) Moraitis
13 days
All the "ARC-AGI is toast" people forget that there are two axes. This is a benchmark for skill acquisition efficiency. @fchollet @GregKamradt @mikeknoop @arcprize
1
0
12
@alexolegimas
Alex Imas
17 days
I think the best mental model for today's agents is Guy Pearce's character in one of Nolan's first films, Memento. He's got extreme amnesia, and needs to look up instructions for every single action from notes (on his body). Learning still happens, but there's no updating of
@jasoncrawford
Jason Crawford
17 days
Having Claude Code write its own skills is not far from having a highly trainable employee: you give it some feedback and it learns. Still unclear to me just how reliable this is, I have seen it ignore applicable skills… but if we're not there yet the path to it is clear
62
104
1K
@gtzi
George Tziralis
20 days
A few things to keep in mind amidst the market's downturn...
1
1
3
@timos_m
Timoleon (Timos) Moraitis
20 days
.@ylecun definitely did not just call @elonmusk a gros con (f-ing idiot in French). He only meant to encourage him to grok on.
@ylecun
Yann LeCun
20 days
@elonmusk Grokon
0
0
7
@timos_m
Timoleon (Timos) Moraitis
22 days
The METR chart clearly needs a third axis for cost. The guys @arcprize are onto something.
@scaling01
Lisan al Gaib
22 days
GPT-5.2-high took 26 TIMES LONGER than Claude 4.5 Opus to complete the METR benchmark suite
0
0
1
@timos_m
Timoleon (Timos) Moraitis
23 days
While impressive as an absolute score, this doesn't improve the scaling curve. The team @noemon_ai is feeling cute -- might announce something while in stealth. Should we?
@arcprize
ARC Prize
23 days
New SOTA public submission to ARC-AGI: - V1: 94.5%, $11.4/task - V2: 72.9%, $38.9/task Based on GPT 5.2, this bespoke refinement submission by @LandJohan ensembles many approaches together
1
0
6
@timos_m
Timoleon (Timos) Moraitis
23 days
"From my perspective, one of the biggest shortcomings is at the moment that I can't specify a problem and tell it to come back when it's found a solution, even if that takes a week. At best, it will try for an hour, and then inform me it didn't manage to solve the problem." Such
@skdh
Sabine Hossenfelder
23 days
An update on my maths/physics work with ChatGPT (Pro 5.2) I regret to inform you that ChatGPT still has not solved the Navier Stokes Millenium problem... though it has several times claimed implicitly that it either proved or disproved it. Like Gemini, ChatGPT has a peculiar
0
0
2
@leothecurious
davinci
26 days
the big world blog post by @kjaved_ is a must-read for anyone in robotics. against the complexity of real world dynamics, continual learning is a prerequisite, not an afterthought. so long as foundation models are a mere crystallization of a static model of the world,
2
3
36
@timos_m
Timoleon (Timos) Moraitis
29 days
Global structure emerging from local rules. A trillion-dollar recipe for continual learning, long-horizon agents, and post-Moore's law non-Von Neumann computing. https://t.co/vqwMr9wLJs
1
0
7
@QuanquanGu
Quanquan Gu
1 month
There’s a big difference between “we claim it works” and “we deployed it at frontier scale”.
1
6
71
@timos_m
Timoleon (Timos) Moraitis
1 month
Blessed are the poor in spirit.
1
0
2
@GreekAnalyst
The Greek Analyst
1 month
Water polo goal of the century by Stelios Argyropoulos-Kanakakis.
15
341
3K
@chrisbarber
Chris Barber
1 month
updated neolabs list. what's the right categorization? llms vs other models? agent-first vs model first @swyx? shipping to users vs not yet shipping (hype to ship ratio)? enterprise vs consumer? large general models vs specialized use cases? - @arcee_ai - @PrimeIntellect -
10
4
136
@timos_m
Timoleon (Timos) Moraitis
1 month
7/ @DBahdanau's attention, Alex Graves' neural Turing machines, Parikh's attention model, Miller's Key-Value Memory Networks, residual connections, lstms, highway networks, layer norm and fast weight programmers, etc @SchmidhuberAI's pic above is not there just for meme value.
0
0
3
@timos_m
Timoleon (Timos) Moraitis
1 month
6/ Now, why doesn't Google enforce the transformer patent? Is Google doing it for the love of the game? The more probable answer is that they just couldn't enforce it if they tried. A competitor would argue in court that the patent is merely an obvious combination of prior art.
1
0
2
@timos_m
Timoleon (Timos) Moraitis
1 month
5/ Notably, that change was only possible without changing the priority date because it's *narrowing* the protection. BTW you may notice the patent was filled in 2018, after the June 2017 paper. That's possible because they first filed a provisional in May 2017.
1
0
2
@timos_m
Timoleon (Timos) Moraitis
1 month
4/ To pin the transformer to a physical implementation that is not just math, Google added this part to the end of claim 1 of the patent. So, in Google's European transformer the encoder layers must operate in parallel!
1
0
2
@timos_m
Timoleon (Timos) Moraitis
1 month
3/ The European patent office did raise this point ("the transformer is just math"). And Google had to narrow down the European transformer.
1
0
2