Konstantin Schürholt Profile
Konstantin Schürholt

@k_schuerholt

Followers
776
Following
7K
Media
18
Statuses
788

AI Researcher at @ndea. Previously postdoc on weight space learning @ University of St.Gallen, Switzerland.

Joined July 2019
Don't wanna be here? Send us removal request.
@k_schuerholt
Konstantin Schürholt
3 years
#NeurIPS2022 HyperNetworks have seen a revival in #stablediffusion these days (@hardmaru). Turns out you can now generate weights even unsupervised :) using generative #hyper_representations trained on populations of neural networks - with @damianborth @DocXavi @BorisAKnyazev
Tweet media one
4
39
297
@andrewgwils
Andrew Gordon Wilson
6 months
Good research is mostly about knowing what questions to ask, not about answering questions that other people are asking.
20
59
628
@arcprize
ARC Prize
20 days
We've recapped all our 30-day learnings in a full post https://t.co/t76IYmIlWT
Tweet card summary image
arcprize.org
One Month of Learnings Building Interactive Reasoning Benchmarks
1
3
43
@GregKamradt
Greg Kamradt
21 days
.@k_schuerholt was the driver behind last weeks Hierarchical Reasoning Model blog post I saw him give an internal presentation to the team about his work and said "this has to be recorded for the public" He re-recorded it and released it Great watch
@ndea
Ndea
21 days
The Surprising Performance Drivers of HRM. A paper talk from Ndea AI researcher @k_schuerholt. We ran a series of ablation studies to determine what factors had the biggest impact on the Hierarchical Reasoning Model's performance on ARC-AGI.
4
2
28
@ndea
Ndea
21 days
The Surprising Performance Drivers of HRM. A paper talk from Ndea AI researcher @k_schuerholt. We ran a series of ablation studies to determine what factors had the biggest impact on the Hierarchical Reasoning Model's performance on ARC-AGI.
11
22
101
@makingAGI
Guan Wang
22 days
Thanks to @arcprize for reproducing and verifying the results! ARC-AGI-1: public 41% pass@2 - semi private 32% pass@2 ARC-AGI-2: public 4% pass@2 - semi private 2% pass@2 Due to differences in testing environments, a certain amount of variance in results is acceptable.
Tweet media one
Tweet media two
19
33
334
@k_schuerholt
Konstantin Schürholt
23 days
Shout out to the authors of the HRM paper @makingAGI for engaging in the discussion with us and sharing their previous experiments. There’ll be a longer video explaining our findings early next week.
0
0
10
@k_schuerholt
Konstantin Schürholt
23 days
In their experiments, few-shot examples work great on large datasets, but don't work well on small scale datasets like ARC. Puzzle embeddings present obvious limitations for generalization, but seem like a very interesting starting point for future work.
1
0
9
@k_schuerholt
Konstantin Schürholt
23 days
5. Task embeddings seem critical for HRM’s learning efficiency. HRM uses embedding layers for unique task-ids rather than GPT-3 style few-shot example contexts. We did not run these experiments ourselves, but the authors kindly shared their experience.
1
0
6
@k_schuerholt
Konstantin Schürholt
23 days
4. Data augmentations matter, but significantly less than the 1000 per task proposed in the paper are sufficient. As with the outer loop, training with augmentations has a bigger impact than using augmentations at inference time to get a more predictions for the same task.
Tweet media one
1
0
7
@k_schuerholt
Konstantin Schürholt
23 days
3: The outer loop makes a dramatic difference. From 1 outer loop to 2 boosts by 13pp, from 1 to 8 doubles performance. Interestingly, refinement loops seem more important during training than during inference. Models trained with refinement have better one-shot performance.
Tweet media one
Tweet media two
1
0
7
@k_schuerholt
Konstantin Schürholt
23 days
2: The hierarchical component of the model had only minimal impact on overall performance. Swapping the HRM model with a regular transformer got within a few percentage points on arc v1 public eval. The hierarchical inner loops help, but don’t explain the overall performance.
Tweet media one
1
0
9
@k_schuerholt
Konstantin Schürholt
23 days
1. Scores on ARC (pass@2): * ARC-AGI-v1: public: 41%, semi-private: 32% * ARC-AGI-v2: public 4%, semi-private: 2% The public scores are in line with what the authors had reported. v1 scores are good for a small model trained only on arc-data. Drops on semi-private are expected.
Tweet media one
1
0
6
@k_schuerholt
Konstantin Schürholt
23 days
We took a closer look at Hierarchical Reasoning Models by @makingAGI. We verified scores and pulled it apart to understand what makes it work. Main take-aways below, full story here:
Tweet card summary image
arcprize.org
We scored on hidden tasks, ran ablations, and found that performance from the Hierarchical Reasoning Model comes from an unexpected source
4
14
157
@jm_alexia
Alexia Jolicoeur-Martineau
24 days
Summary of the findings for the Hierarchical Reasoning Model on ARC
Tweet media one
@arcprize
ARC Prize
24 days
Analyzing the Hierarchical Reasoning Model by @makingAGI We verified scores on hidden tasks, ran ablations, and found that performance comes from an unexpected source ARC-AGI Semi Private Scores: * ARC-AGI-1: 32% * ARC-AGI-2: 2% Our 4 findings:
Tweet media one
13
96
814
@arcprize
ARC Prize
24 days
Analyzing the Hierarchical Reasoning Model by @makingAGI We verified scores on hidden tasks, ran ablations, and found that performance comes from an unexpected source ARC-AGI Semi Private Scores: * ARC-AGI-1: 32% * ARC-AGI-2: 2% Our 4 findings:
Tweet media one
39
203
1K
@GregKamradt
Greg Kamradt
24 days
What makes the HRM model work so well for its size on @arcprize? We ran ablation experiments to find out what made it work Our findings show that you could replace the "hierarchical" architecture with a normal transformer with only a small performance drop We found that an
Tweet media one
14
87
842
@hardmaru
hardmaru
2 months
Blaise Agüera explains interrelated paradigm shifts which he believes are core to the future development in AI. I like his take on collective intelligence, the future of artificial life research, and the (somewhat philosophical) discussions about consciousness and theory of mind.
Tweet media one
0
5
37
@hardmaru
hardmaru
2 months
New Essay by @BlaiseAguera (@Google): “AI Is Evolving — And Changing Our Understanding Of Intelligence” Advances in AI are making us reconsider what intelligence is and giving us clues to unlocking AI’s full potential. https://t.co/PEzHQmoB9z
Tweet card summary image
noemamag.com
Advances in AI are making us reconsider what intelligence is and giving us clues to unlocking AI’s full potential.
9
34
178