Cyrille Combettes Profile
Cyrille Combettes

@CyrilleCmbt

Followers
75
Following
163
Media
2
Statuses
23

Quantitative researcher @CFM_AM Ph.D. @MLatGT Optimization & machine learning

Joined March 2020
Don't wanna be here? Send us removal request.
@CyrilleCmbt
Cyrille Combettes
5 years
Is it possible to find better descent directions for Frank-Wolfe while remaining projection-free? We propose an intuitive and generic boosting procedure to speed up FW and variants, accepted at ICML! Check out the presentation of our paper
0
2
10
@energybants
Mark Nelson
2 years
The greatest man you've never heard of died this week on Wednesday, September 6th. Marcel Boiteux built the French nuclear fleet as head of national utility EdF, making superb, far-sighted decisions against powerful entrenched interests. Decisions such as abandoning the
167
3K
12K
@DimitrisPapail
Dimitris Papailiopoulos
2 years
Carlini got fed up breaking every adversarial defense that gets published (rightfully so), and is now using GPT4 to automate it. This is a legit statement paper, with transparency on methods, that should be read as a criticism of the related literature https://t.co/FtjMNBg5uc
4
48
260
@jerome_bolte
Jérôme Bolte
2 years
Looking for a post-doctoral candidate in optimization at Toulouse School of Economics (Toulouse, France). Funded by U.S. Air Force, renewable once, supervised by E. Pauwels and myself. Topic: deep learning, large scale optimization
1
22
32
@culturaltutor
The Cultural Tutor
3 years
Why Beauty Matters (and how it has been destroyed by "usability") A short thread...
1K
27K
139K
@e_d_andersen
Erling Andersen
4 years
@gabrielpeyre That self-concordant barrier based interior-point methods cannot be strongly polynomial has been shown recently. https://t.co/HQb6DsTW0T Nevertheless fantastic work of NN. Thanks to @HAFriberg for the link.
1
3
18
@tomgoldsteincs
Tom Goldstein
4 years
There's no evidence that SGD plays a fundamental role in generalization. With totally deterministic full-batch gradient descent, Resnet18 still gets >95% accuracy on CIFAR10. With data augmentation, full-batch Resnet152 gets 96.76%. https://t.co/iwIqQd7U1O
29
166
859
@MIT_CSAIL
MIT CSAIL
5 years
ML model saves 90% on bandwidth during video-calls by using just an image of your face & some basic motion data. Paper: https://t.co/AK5zxCRl9L (v/@nvidia) Video: https://t.co/kuTiC4QoNQ (v/@twominutepapers) #CVPR2021 #Compression #DeepFakes #KindOf
6
181
716
@nhigham
Nick Higham
5 years
Matrix Rank Relations https://t.co/BGhD1ZjA1K
3
32
153
@CyrilleCmbt
Cyrille Combettes
5 years
A review of the complexity bounds of linear minimizations and projections on several sets: https://t.co/yl5R7N9oGD w/ @spokutta The Frank-Wolfe audience was clearly in mind, hopefully it is also of interest to the broader optimization community!
1
3
10
@daniela_witten
Daniela Witten
5 years
FOUR CARDINAL RULES OF STATISTICS 📈📊📉 A thread 1/🧵
69
2K
8K
@spokutta
Sebastian Pokutta
5 years
Short informal summary of our recent paper "Projection-Free Adaptive Gradients for Large-Scale Optimization" with @CyrilleCmbt and Christoph Spiegel whether one can combine Frank-Wolfe with adaptive gradients: yes you can. https://t.co/xWppgZ6dcB #ml #opt #frankwolfe
0
1
11
@RobertTLange
Robert Lange
5 years
The lottery ticket hypothesis 🎲 states that sparse nets can be trained given the right initialisation 🧬. Since the original paper (@jefrankle & @mcarbin) a lot has happened. Checkout my blog post for an overview of recent developments & open Qs. ✍️: https://t.co/R46SCEYf8p
6
114
471
@ericjang11
Eric Jang
6 years
Every once in awhile a paper comes out that makes you breathe a sigh of relief that you don't publish in that field... https://t.co/56heAufhGA "Our results show that when hyperparameters are properly tuned via cross-validation, most methods perform similarly to one another"
31
423
2K
@CyrilleCmbt
Cyrille Combettes
6 years
Interesting paper showing that separating the architecture of a neural network into individual networks for each feature can still be very competitive while, most importantly, offering interpretability
@nickfrosst
Nick Frosst
6 years
my new paper with @rishabh_467 @geoffreyhinton, Rich Caruana and Xuezhou Zhang is out on arxiv today! It’s about interpretability and neural additive models. Don't have time to read the paper? Read this tweet thread instead :) https://t.co/eGPsL55v8G 1/7
0
6
9
@CyrilleCmbt
Cyrille Combettes
6 years
A fun read studying the impact of *fine-tuning random seeds* in BERT https://t.co/TWbxspJbq8 A familiar hyperparameter some would argue
@roydanroy
Dan Roy
6 years
Your SOTA code may only be SOTA for some random seeds. Nonsense or new reality? I suppose there are trivial ways to close the gap using restarts and validation data. https://t.co/nhuJrOlqgs
0
0
5
@edwardhkennedy
Edward Kennedy
6 years
"Behind every great theorem lies a great inequality" This inequalities cheat sheet has been really helpful to keep on hand while working - passing it along in case others find it useful https://t.co/MzWhElyksV https://t.co/VxTE4sUzKC
3
157
638
@CyrilleCmbt
Cyrille Combettes
6 years
Glad to share our new paper "Boosting Frank-Wolfe by Chasing Gradients" with @spokutta We propose to speed-up FW by moving in directions better aligned with the gradients. Turns out the progress obtained overcomes the cost of multiple oracle calls!
0
1
7
@nntaleb
Nassim Nicholas Taleb
6 years
About everybody seems to be working under the hypothesis that those who recover from the virus are permanently in the SAME condition as those who never got it.
98
745
3K
@spokutta
Sebastian Pokutta
6 years
First Lightning Talk by Cyrille W Combettes @FieldsInstitute talking about approximate Carathéodory
0
1
7