Timoleon (Timos) Moraitis
@timos_m
Followers
1K
Following
4K
Media
176
Statuses
2K
Building brain-like AI @noemon_ai Previously @Huawei @IBMResearch @ETH_en @UZH_en @ntua
Zurich, Switzerland
Joined January 2009
All the "ARC-AGI is toast" people forget that there are two axes. This is a benchmark for skill acquisition efficiency. @fchollet @GregKamradt @mikeknoop @arcprize
0
0
0
Yes. People begin realizing the real-world implications of brute-force scaling. Being "frontier" will soon be considered meaningless, unless it's the *Pareto frontier* of cost vs capability. And that needs new foundational improvements.
What Happens When AI Tokens Cost More Than Your Employees? @Jason: “We, with our agents, hit $300/day per agent using the Claude API, like instantly. And that was doing, maybe, 10 or 20%. That's $100k/year per agent.” @chamath: “We're getting to a place where we have to
2
0
1
All the "ARC-AGI is toast" people forget that there are two axes. This is a benchmark for skill acquisition efficiency. @fchollet @GregKamradt @mikeknoop @arcprize
1
0
12
I think the best mental model for today's agents is Guy Pearce's character in one of Nolan's first films, Memento. He's got extreme amnesia, and needs to look up instructions for every single action from notes (on his body). Learning still happens, but there's no updating of
Having Claude Code write its own skills is not far from having a highly trainable employee: you give it some feedback and it learns. Still unclear to me just how reliable this is, I have seen it ignore applicable skills… but if we're not there yet the path to it is clear
62
104
1K
The METR chart clearly needs a third axis for cost. The guys @arcprize are onto something.
0
0
1
While impressive as an absolute score, this doesn't improve the scaling curve. The team @noemon_ai is feeling cute -- might announce something while in stealth. Should we?
New SOTA public submission to ARC-AGI: - V1: 94.5%, $11.4/task - V2: 72.9%, $38.9/task Based on GPT 5.2, this bespoke refinement submission by @LandJohan ensembles many approaches together
1
0
6
"From my perspective, one of the biggest shortcomings is at the moment that I can't specify a problem and tell it to come back when it's found a solution, even if that takes a week. At best, it will try for an hour, and then inform me it didn't manage to solve the problem." Such
An update on my maths/physics work with ChatGPT (Pro 5.2) I regret to inform you that ChatGPT still has not solved the Navier Stokes Millenium problem... though it has several times claimed implicitly that it either proved or disproved it. Like Gemini, ChatGPT has a peculiar
0
0
2
the big world blog post by @kjaved_ is a must-read for anyone in robotics. against the complexity of real world dynamics, continual learning is a prerequisite, not an afterthought. so long as foundation models are a mere crystallization of a static model of the world,
2
3
36
Global structure emerging from local rules. A trillion-dollar recipe for continual learning, long-horizon agents, and post-Moore's law non-Von Neumann computing. https://t.co/vqwMr9wLJs
1
0
7
There’s a big difference between “we claim it works” and “we deployed it at frontier scale”.
1
6
71
Water polo goal of the century by Stelios Argyropoulos-Kanakakis.
15
341
3K
updated neolabs list. what's the right categorization? llms vs other models? agent-first vs model first @swyx? shipping to users vs not yet shipping (hype to ship ratio)? enterprise vs consumer? large general models vs specialized use cases? - @arcee_ai - @PrimeIntellect -
10
4
136
7/ @DBahdanau's attention, Alex Graves' neural Turing machines, Parikh's attention model, Miller's Key-Value Memory Networks, residual connections, lstms, highway networks, layer norm and fast weight programmers, etc @SchmidhuberAI's pic above is not there just for meme value.
0
0
3
6/ Now, why doesn't Google enforce the transformer patent? Is Google doing it for the love of the game? The more probable answer is that they just couldn't enforce it if they tried. A competitor would argue in court that the patent is merely an obvious combination of prior art.
1
0
2
5/ Notably, that change was only possible without changing the priority date because it's *narrowing* the protection. BTW you may notice the patent was filled in 2018, after the June 2017 paper. That's possible because they first filed a provisional in May 2017.
1
0
2
4/ To pin the transformer to a physical implementation that is not just math, Google added this part to the end of claim 1 of the patent. So, in Google's European transformer the encoder layers must operate in parallel!
1
0
2
3/ The European patent office did raise this point ("the transformer is just math"). And Google had to narrow down the European transformer.
1
0
2