Explore tweets tagged as #overparameterization
@konstmish
Konstantin Mishchenko
4 years
What happens to Adam if we turn off momentum and epsilon? If we set β₁=0, we get RMSprop. If we set β₁=β₂=0, we get signSGD. How well does signSGD with constant batch size converge? It doesn't. Not even with tiny stepsizes and overparameterization.
Tweet media one
4
27
174
@konstmish
Konstantin Mishchenko
2 years
Overparameterization empowers deep learning, but in federated learning the data is heterogeneous across clients, which makes it very difficult. We just proved that the solution to this is 𝑜𝑣𝑒𝑟𝑝𝑒𝑟𝑠𝑜𝑛𝑎𝑙𝑖𝑧𝑎𝑡𝑖𝑜𝑛, and it's always achievable:.
Tweet media one
0
21
78
@Dorialexander
Alexander Doria
4 months
A contrarian result I like a lot: smaller language models perform better on knowledge graphs than larger ones, as "overparameterization can impair reasoning due to excessive memorization".
Tweet media one
25
83
823
@preferred_jp
PFN Tech
6 years
[#NeurIPS2019] This evening poster (ID:216) on research done by our tech advisor Prof. Kenji Fukumizu at Preferred Networks. “Semi-flat minima and saddle points by embedding neural networks to overparameterization”. Paper: Slides:
Tweet media one
0
6
13
@andrewgwils
Andrew Gordon Wilson
5 years
Overparameterization isn't mysterious, if we stop parameter counting as a proxy for complexity. Our new paper "Rethinking Parameter Counting in Deep Models: Effective Dimensionality Revisited": With Wesley Maddox, @g_benton_. 1/7
Tweet media one
3
117
528
@bsimsek13
Berfin Simsek
4 years
Excited to share our paper on neural net overparameterization to appear at #ICML2021 💃🏻We asked why can’t training find a minimum in mildly overparameterized nets. Below, a 4-4-4 net can achieve a zero-loss, but any of 5-5-5 nets trained with GD can not🤨
Tweet media one
2
40
203
@BachFrancis
Francis Bach
3 years
Make some noise for SGD’s proper entrance to the blog! First part: how overparameterization might change the rules of the game.
4
116
628
@btreetaiji
Taiji Suzuki
6 years
一方で,深層学習はそのeffective dimensionalityにあたるものが何なのか,self regularizationとか最適化方法も絡んできていまいちよくわかっていない.なのでoverparameterization関連の理論研究が盛り上がっているという訳ですね..
0
26
99
@FeiziSoheil
Soheil Feizi
4 years
Check out our #ICLR2021 papers:. - Perceptual Adv. Robustness: - Influence Functions are Fragile: - Provable Defense against Poisoning: - Overparameterization in GANs: One more 👇
Tweet media one
Tweet media two
Tweet media three
Tweet media four
1
20
75
@papers_anon
PapersAnon
11 months
Stuffed Mamba: State Collapse and State Capacity of RNN-Based Long-Context Modeling. Study into the poor performance of RNNs at long context tasks. Found SC arises from state overparameterization relative to the training length. Gives 4 mitigations some training-free. Links below
Tweet media one
1
1
12
@tweetnakasho
キャルちゃん
4 years
#キャルちゃんのquantphチェック.MaxCut問題において古典的な最適化ランドスケープと、目的関数の評価に使われる量子回路との関係を調査。多量子ビット演算が必要となる箇所を特定、overparameterizationにより有利なランドスケープを得ることができることを示した。.
Tweet media one
Tweet media two
Tweet media three
Tweet media four
0
5
14
@MvsCerezo
Marco Cerezo
4 years
Did you know:.Studying the 👉Lie algebra👈 associated with a quantum neural network or parametrized quantum circuit can help you detect barren plateaus 🏜️ or study phenomena as overparameterization 📈. Tune in to my talk next week to learn more!
Tweet media one
@PennyLaneAI
PennyLane
4 years
#QHack2022 Speaker 🎤. Join Marco Cerezo (@MvsCerezo) for a talk on:. "Barren plateaus and overparametrization in quantum neural networks". Wednesday Feb 16 at 12:00 Noon EST.
Tweet media one
2
6
70
@jhhalverson
Jim Halverson
5 years
Tune in next Wednesday at 12:00 EDT for the next Physics ∩ ML seminar. You'll hear Ard Louis' insights on generalization, overparameterization, and connections to evolution. Sign up at
Tweet media one
0
7
24
@JmlrOrg
Journal of Machine Learning Research
1 month
'DRM Revisited: A Complete Error Analysis', by Yuling Jiao, Ruoxuan Li, Peiying Wu, Jerry Zhijian Yang, Pingwen Zhang. . #overparameterization #overparameterized #deep.
4
5
15
@arxiv_org
arxiv
6 years
Training Compact Neural Networks via Auxiliary Overparameterization.
Tweet media one
Tweet media two
Tweet media three
Tweet media four
0
2
3
@ducha_aiki
Dmytro Mishkin 🇺🇦
6 years
Overparameterization must go on!."Why Having 10,000 Parameters in Your Camera Model is Better Than Twelve. Thomas Schöps, Viktor Larsson, @mapo1 , Torsten Sattler".#ComputerVision #Robotics
Tweet media one
0
6
30
@andrey_kurenkov
Andrey Kurenkov
7 years
Neat paper alert!. Identity Crisis: Memorization and Generalization under Extreme Overparameterization Chiyuan Zhang Samy Bengio Moritz Hardt Yoram Singer . Interesting results. as usual, nice that it sheds more light on how the things we have work.
Tweet media one
Tweet media two
Tweet media three
1
6
38
@ajzellmer
Salamanda
7 years
Congrats to Maddie Gillman on completing her honors thesis seminar today! The impacts of overparameterization and sampling bias on modeling the future distribution of California Valley Oak trees #undergraduateresearch #speciesdistributionmodeling
Tweet media one
0
0
17
@aminkarbasi
Amin Karbasi
4 years
Our ICML workshop "Overparameterization: Pitfalls & Opportunities" will take place this Saturday. The updated schedule .and accepted papers can be found at . @QuanquanGu @HanieSedghi @yasamanbb
Tweet media one
1
8
53
@burny_tech
Burny - Effective Curiosity
2 years
Spectral bias and task-model alignment explain generalization in kernel regression and infinitely wide neural networks."A theoretical understanding of generalization remains an open problem for many machine learning models, including deep networks where overparameterization leads
Tweet media one
1
0
0