
Konstantin Mishchenko
@konstmish
Followers
7K
Following
4K
Media
169
Statuses
691
Research Scientist @AIatMeta Previously Researcher @ Samsung AI Outstanding Paper Award @icmlconf 2023 Action Editor @TmlrOrg I tweet about ML papers and math
Paris, France
Joined June 2020
I believe successful neural network training represents cases of "near convexity": the optimization landscape, while technically non-convex, behaves enough like a convex problem that standard convex optimization is often applicable. At the same time, *in general* neural nets.
The neural network objective function is a very complicated objective function. It's very non convex, and there are no mathematical guarantees whatsoever about its success. And so if you were to speak to somebody who studies optimization from a theoretical point of view, they
20
63
681
Stochastic prox is also of interest in federated learning, where each worker minimizes a regularized objective. I think this paper by @akhaledv2 and @chijinML is the best source to learn more:.3/
1
0
11
RT @wazizian: ❓ How long does SGD take to reach the global minimum on non-convex functions?. With @FranckIutzeler, J. Malick, P. Mertikopou….
0
71
0