Brandon Amos Profile
Brandon Amos

@brandondamos

Followers
19K
Following
10K
Media
446
Statuses
4K

research scientist @MetaAI (FAIR) | optimization, machine learning, control, transport | PhD from @SCSatCMU

Joined January 2014
Don't wanna be here? Send us removal request.
@brandondamos
Brandon Amos
3 years
Many standard probability distributions can be obtained by solving a maximum-entropy optimization problem over distributions. For example, the Gaussian maximizes the entropy subject to mean and covariance constraints. Jax source code: More resources: 🧵
5
137
826
@brandondamos
Brandon Amos
3 years
I just posted a new tutorial on amortized optimization! It covers foundations for learning to optimize and how they are key ingredients for VAEs, RL, meta-learning, and sparse coding with budding applications in DEQs, convex optimization, and beyond. 🧵
Tweet media one
Tweet media two
Tweet media three
9
144
751
@brandondamos
Brandon Amos
9 months
šŸ“¢ Today's my first day at Cornell Tech :). I will continue my full-time position as a scientist at Meta, and am teaching ML here on the side for the semester with @volokuleshov. The course is open source and you can follow everything in this repo:.
Tweet media one
35
46
662
@brandondamos
Brandon Amos
3 years
I just open-sourced 200+ slides from the major presentations I've given over the past 5 years:.
3
70
586
@brandondamos
Brandon Amos
2 years
šŸ“š My mini-book on amortized optimization is officially published! Via the Foundations and TrendsĀ® in Machine Learning journal (. Buy a physical copy: Free online version: Source code:
Tweet media one
13
108
589
@brandondamos
Brandon Amos
4 years
We've released differentiable convex optimization layers for JAX!. Code: Tutorial: NeurIPS paper: Blog post: With @akshaykagrawal, @ShaneBarratt, S. Boyd, S. Diamond, and @zicokolter
Tweet media one
Tweet media two
Tweet media three
5
126
561
@brandondamos
Brandon Amos
8 months
šŸ“¢ My team at Meta is hiring PhD research interns! We study core machine learning, optimization, amortization, flows, and control for modeling and interacting with complex systems (. and we use basic physics. šŸ™ƒ). Please apply here and message me:. 🧵
Tweet media one
Tweet media two
Tweet media three
Tweet media four
15
92
557
@brandondamos
Brandon Amos
6 years
A life update. I've dodged all of the attacks and have successfully defended my PhD thesis. Thanks everybody!. My thesis document and slides are available on GitHub.
19
41
518
@brandondamos
Brandon Amos
6 years
Tuning Hyperparameters without Grad Students: Scalable and Robust Bayesian Optimisation with Dragonfly.K. Kandasamy et al. Python Library: Docs:
Tweet media one
6
128
415
@brandondamos
Brandon Amos
10 months
:). (context for the reader: this is a slide from my fundamental differentiable optimization layers talk, which I've been giving since ~2017. @alfcnz has been inspirational on the "soft-argmax" naming for years!!)
Tweet media one
@AlexShtf
Alex Shtoff
11 months
I am the only one on #ML twitter who doesn't like the name 'softmax' for exp(x) / sum(exp(x)) , and instead prefers the name 'soft-argmax'? .Softmax is LogSumExp!.
5
52
421
@brandondamos
Brandon Amos
3 years
Our new paper on Meta Optimal Transport uses amortized optimization to predict initializations and improve optimal transport solver runtimes by 100x and more. With @CohenSamuel13 @GiuliaLuise1 @IevgenRedko . Paper: JAX source code:
6
58
404
@brandondamos
Brandon Amos
11 months
šŸ“¢ In our new @UncertaintyInAI paper, we do neural optimal transport with costs defined by a Lagrangian (e.g., for physical knowledge, constraints, and geodesics). Paper: JAX Code: (w/ A. Pooladian, C. Domingo-Enrich, @RickyTQChen)
6
63
410
@brandondamos
Brandon Amos
7 months
Flows and transport methods are widely used to connect one distribution to another. What about going up one level to transporting between distributions over distributions?. My new talk overviews some methods in this space:. & a quick 🧵 below
1
74
400
@brandondamos
Brandon Amos
6 years
A draft of my thesis is now on GitHub. Please let me know if you want to attack me as I'll be defending it next week.
Tweet media one
Tweet media two
19
43
382
@brandondamos
Brandon Amos
7 months
Here's a new lecture I made on learning embeddings with multidimensional scaling and TSNE for our ML course at Cornell Tech (with @volokuleshov). It covers the formulation and goes through some code for a deeper look. Notebook: PDF:
5
56
383
@brandondamos
Brandon Amos
3 years
šŸ‘€. "Amos can improve generalization and accelerate convergence"."Amos consistently outperforms the state-of-the-art"."we analyze the behavior of Amos in an intuitive and heuristic manner".
@Robin_Tian
Ran TIAN
3 years
Amos is a new optimizer that we propose to pre-train large language models. It is more efficient and converges faster than AdamW: ≤ 51% memory for slot variables, and better valid loss within ≤ 70% training time!. New from Google Research. Preprint:
Tweet media one
15
16
314
@brandondamos
Brandon Amos
7 years
Attention Solves Your TSP by W. Kool and M. Welling. This paper has a Kool intro. Paper: .@PyTorch Code:
Tweet media one
1
80
312
@brandondamos
Brandon Amos
2 months
& here are some of my favorite papers on the unification of flows and diffusion. What a decade!!. (From my presentation here:
Tweet media one
@dasayan05
Ayan Das
3 months
How Diffusion unification went:. > score based model.> then DDPM came along.> we have two formalism, DDPM & SBM.> SDE came to unify them.> now we have Score, DDPM & SDE.> Then came flow matching to unify them.> now we have Score, DDPM, SDE & Flow Models.> Then consistency models.
4
42
317
@brandondamos
Brandon Amos
6 years
Another life update. I've joined @facebookai (FAIR) in NYC as a research scientist!.
20
4
306
@brandondamos
Brandon Amos
4 years
Stoked to release our milestone #ICML2021 paper on Riemannian Convex Potential Maps! With @CohenSamuel13 and @lipmanya. Paper: JAX Code: Slides: 🧵
Tweet media one
Tweet media two
3
60
296
@brandondamos
Brandon Amos
6 years
Stoked to share a milestone project for all of us! #NeurIPS2019 paper with @akshaykagrawal, @ShaneBarratt, S. Boyd, S. Diamond, @zicokolter:. Differentiable Convex Optimization Layers. Paper: Blog Post: Repo:
@akshaykagrawal
Akshay Agrawal
6 years
CVXPY is now differentiable. Try our PyTorch and TensorFlow layers using our package, cvxpylayers: (& see our NeurIPS paper for details .
4
83
302
@brandondamos
Brandon Amos
1 year
My core ML team (@AIatMeta) is hiring research interns! Our projects span optimization, optimal transport, optimal control, generative modeling, complex systems, and geometry. Please apply here and reach out (bda@meta.com) if you're interested:.
4
36
288
@brandondamos
Brandon Amos
6 years
Excited to share my first paper with @facebookai:. The Differentiable Cross-Entropy Method.with @denisyarats. Paper: Videos:
Tweet media one
2
52
295
@brandondamos
Brandon Amos
3 years
[2017] The Consciousness Prior: [2018] Learning Awareness Models: [2021] A Theory of Consciousness from Theoretical Computer Science & the Conscious Turing Machine: [2022] Learning Consciousness Models
4
43
285
@brandondamos
Brandon Amos
8 years
Our new paper - OptNet: Differentiable Optimization as a Layer in Neural Nets. Paper: Code:
Tweet media one
4
93
275
@brandondamos
Brandon Amos
7 years
Coloring math in your papers, not just in your presentations and posters, is a great idea and helps readability if done correctly. Like in the 2011 PILCO paper:
Tweet media one
6
61
272
@brandondamos
Brandon Amos
5 months
šŸ“¢ My team at Meta is hiring visiting PhD students from CMU, UW, Berkeley, and NYU! We study core ML, optimization, amortization, transport, flows, and control for modeling and interacting with complex systems. Please apply here and message me:.
9
41
276
@brandondamos
Brandon Amos
7 years
I'm excited to announce our new @PyTorch model predictive control library! We use this in our (forthcoming) NIPS 2018 paper. This is joint work with I. Jimenez, J. Sacks, B. Boots, and @zicokolter. Project Website: GiHub Page:
Tweet media one
3
63
268
@brandondamos
Brandon Amos
7 years
@PyTorch geometric by @rusty1s et al.
Tweet media one
2
97
266
@brandondamos
Brandon Amos
5 years
One more interesting interpretation of the softmax is that it is a projection onto the simplex. The usual solution comes from the KKT conditions. Details in Ch 2 of my thesis: And you can implement it this way in PyTorch/TF with:
Tweet media one
7
50
262
@brandondamos
Brandon Amos
5 years
In our new paper we scale model-based reinforcement learning to the gym humanoid by using short-horizon model rollouts followed by a learned model-free value estimate. Paper: Videos: With @sam_d_stanton @denisyarats @andrewgwils
2
62
249
@brandondamos
Brandon Amos
4 years
We've been studying differentiable combinatorial optimization layers and just posted our #ICML2021 paper on CombOptNet! With @AnselmPaulus @MichalRolinek @vit_musil @GMartius. Paper: @PyTorch Code:
Tweet media one
3
38
222
@brandondamos
Brandon Amos
2 years
We're hiring research scientist interns in FAIR's Core ML team at Meta! Please send an application here and get in touch if you're interested.
5
31
222
@brandondamos
Brandon Amos
5 years
After 6 years of review (! longer than my PhD) one of my first papers has been published in the ACM Transactions of Mathematical Software:. Alg. 1007: QNSTOP—Quasi-Newton Algorithm for Stochastic Optimization. Paper: FORTRAN Source:
Tweet media one
8
8
217
@brandondamos
Brandon Amos
7 years
Excited to share our #nips2018 paper on differentiable MPC and our standalone @PyTorch control package!. Joint work with Jimenez, Sacks, Boots, and @zicokolter. Paper: @PyTorch MPC Solver: Experiments:
Tweet media one
Tweet media two
4
74
204
@brandondamos
Brandon Amos
5 months
šŸ¤” How to extract knowledge from LLMs to train better RL agents?. šŸ“š Our new paper (with @qqyuzu @HenaffMikael @yayitsamyzhang @adityagrover_ ) studies LLM-driven rewards for NetHack!. Paper: Code:
Tweet media one
Tweet media two
Tweet media three
2
41
206
@brandondamos
Brandon Amos
8 years
I just finished my @PyTorch implementation of DenseNet-BC, which gets 4.77% error on CIFAR-10+.
Tweet media one
6
50
187
@brandondamos
Brandon Amos
4 years
We just posted a longer version of our slides on convex optimization, flows, input-convex neural networks, and (Riemannian) optimal transportation:
@brandondamos
Brandon Amos
4 years
Stoked to release our milestone #ICML2021 paper on Riemannian Convex Potential Maps! With @CohenSamuel13 and @lipmanya. Paper: JAX Code: Slides: 🧵
Tweet media one
Tweet media two
0
26
192
@brandondamos
Brandon Amos
6 years
Hydra is an awesome new lightweight experiment manager (and beyond) from @facebookai -- I've been using it with all of my @PyTorch experiments for config/cluster scheduling/hyper-param sweeps. Code:
@Hydra_Framework
Hydra
6 years
Hydra is live at .Facebook engineering post:. More info:.
8
36
189
@brandondamos
Brandon Amos
8 years
block: My [short] new Python library for intelligent block matrices in numpy, @PyTorch, and beyond.
Tweet media one
0
86
181
@brandondamos
Brandon Amos
7 years
I am on the job market!. Website and CV: Google Scholar: GitHub:
Tweet media one
9
24
188
@brandondamos
Brandon Amos
3 years
I've posted a new paper on amortizing convex conjugates for continuous optimal transport along with new JAX software for (Euclidean) Wasserstein-2 OT!. 🧵
7
28
178
@brandondamos
Brandon Amos
5 months
šŸ“¢ My team at Meta (including @lipmanya and @RickyTQChen) is hiring a postdoctoral researcher to help us build the next generation of flow, transport, and diffusion models! Please apply here and message me:.
5
34
179
@brandondamos
Brandon Amos
9 years
Just released a new blog post and code! Image Completion with Deep Learning in TensorFlow
6
123
167
@brandondamos
Brandon Amos
6 years
Excited to share my new tech report from my @IntelAI internship on the Limited Multi-Label projection layer! Joint work with Vladlen Koltun and @zicokolter. Paper: @PyTorch Code:
Tweet media one
Tweet media two
Tweet media three
1
36
165
@brandondamos
Brandon Amos
3 years
Our Meta OT paper was rejected from @NeurIPS despite having WA/A/SA throughout the discussion period. How did it happen? The AC based the decision on a review that came a month late that we couldn't even respond to. We've made the full discussion public:.
Tweet media one
5
7
157
@brandondamos
Brandon Amos
7 months
Lately I've been interested in LLM prompt optimization and amortization. Here's my new Dagstuhl talk highlighting some of my favorite papers in the space:. šŸ“š & our AdvPrompter paper amortizes prompt optimization for jailbreaking:
4
18
156
@brandondamos
Brandon Amos
1 year
Our new #NeurIPS2023 paper shows how to learn a task-driven metric on the prediction space! Come chat at our poster (#1128) on Weds at 10:45am. Paper: Code: Talk: with @theshank9 @RickyTQChen @mukadammh
Tweet media one
2
25
146
@brandondamos
Brandon Amos
6 years
Come intern at @facebookai and do awesome ML/AI/RL/NLP/vision/optimization/game theory research with us. && write beautiful @PyTorch code while doing so. Please get in touch!. More details:
1
16
147
@brandondamos
Brandon Amos
8 years
TensorFlow has a great new K-FAC library in tf.contrib by the authors and collaborators:.
Tweet media one
Tweet media two
2
60
138
@brandondamos
Brandon Amos
7 years
GraphRNN: Generating Realistic Graphs with Deep Auto-regressive Model by @youjiaxuan, R. Ying, @xiangrenUSC, @williamleif, and @jure. #ICML2018 Paper: @PyTorch Code: Unofficial Slides:
Tweet media one
Tweet media two
0
44
140
@brandondamos
Brandon Amos
8 years
Also on debugging layers/tools, I'd love to see more "unit tests" for research, like in this ICLR 2014 paper:.
Tweet media one
0
40
140
@brandondamos
Brandon Amos
7 years
I'm in SF for the @PyTorch developer conference. Come and say hi! Here's a poster with some of my contributions to the thriving PyTorch ecosystem.
Tweet media one
2
15
137
@brandondamos
Brandon Amos
9 months
šŸŽ‰ Our new paper generalizes Flow Matching to the /Meta/ setting, where we solve (and amortize šŸ˜…) a family of FM problems simultaneously. We do this via a GNN-based conditioning and use Meta FM between cell populations. See @lazar_atan's thread for more info!.
@lazar_atan
Lazar Atanackovic
9 months
šŸš€Introducing — Meta Flow Matching (MFM) šŸš€. Imagine predicting patient-specific treatment responses for unseen cases or building generative models that adapt across different measures. MFM makes this a reality. šŸ“°Paper: šŸ’»Code:
Tweet media one
Tweet media two
6
12
136
@brandondamos
Brandon Amos
5 years
With @denisyarats, we've released the @PyTorch code and camera-ready version of our #ICML2020 paper on the differentiable cross-entropy method. Paper: Code: Videos: More details in our original thread:.
@brandondamos
Brandon Amos
6 years
Excited to share my first paper with @facebookai:. The Differentiable Cross-Entropy Method.with @denisyarats. Paper: Videos:
Tweet media one
1
41
132
@brandondamos
Brandon Amos
8 years
Why PyTorch's layer creation is powerful: Here's my layer that solves an optimization problem with a primal-dual interior point method.
Tweet media one
6
56
129
@brandondamos
Brandon Amos
7 years
Our new #iclr2018 paper on Learning Awareness Models, with @laurent_dinh @serkancabi @notmisha @NandoDF and many others. Paper: Videos:
Tweet media one
Tweet media two
Tweet media three
2
48
132
@brandondamos
Brandon Amos
7 years
Learning Latent Dynamics for Planning from Pixels.
Tweet media one
0
22
132
@brandondamos
Brandon Amos
7 years
SparseMAP: Differentiable Sparse Structured Inference by @vnfrombucharest et al. #icml2018 Paper: @PyTorch Code:
Tweet media one
Tweet media two
0
44
132
@brandondamos
Brandon Amos
5 years
Here's my #CVPR2020 deep declarative network workshop talk on differentiable optimization:. Video: Slides:
Tweet media one
2
31
129
@brandondamos
Brandon Amos
7 years
Neural Motifs: Scene Graph Parsing with Global Context by @rown et al. at #cvpr2018. Paper: @PyTorch Code: Awesome interactive web demo:
Tweet media one
Tweet media two
1
42
124
@brandondamos
Brandon Amos
6 months
Here's my take on an LLM intro for our ML course at Cornell Tech (with @volokuleshov). It starts with the foundations then goes through code to use pretrained models for 1) generation, 2) chat, and 3) code infilling. pdf: ipynb:
2
19
127
@brandondamos
Brandon Amos
4 years
The Wasserstein distance works well for imitation learning on a single system: Our new paper explores the cross-domain setting and uses the Gromov-Wasserstein distance to imitate between /different/ systems:
@arnaudfickinger
Arnaud Fickinger
4 years
We provably achieve cross-domain transfer in non-trivial continuous control domain by minimizing the Gromov-Wasserstein distance with deep RL. Paper: Site: Joint work with @CohenSamuel13 Stuart Russell @brandondamos. A thread:
Tweet media one
2
17
122
@brandondamos
Brandon Amos
7 years
Smooth Loss Functions for Deep Top-k Classification.Leonard Berrada, Andrew Zisserman, M. Pawan Kumar.ICLR 2018. @PyTorch code:
Tweet media one
0
32
120
@brandondamos
Brandon Amos
5 years
This is a great post! The argmin operations that come up here can be prototyped in PyTorch/TF/(+Jax soon) in a few lines of code with cvxpylayers: Our blog post has working code examples for the sparsemax/csparsemax/LML and others:
@srush_nlp
Sasha Rush
5 years
Notebook foršŸ¤— Reading Group this week on understanding softmax, entropy, and `Adaptively Sparse Transformers` from Correia, Niculae, and Martins.
Tweet media one
Tweet media two
Tweet media three
1
24
116
@brandondamos
Brandon Amos
5 years
Excited to share our NeurIPS workshop on Learning Meets Combinatorial Optimization (LMCA)! Consider submitting a 4-page paper if you're working in the area (due Oct 5):. With @vlastelicap, J. Song, A. Ferber, @GMartius, @BDilkina, and @yisongyue
Tweet media one
2
30
111
@brandondamos
Brandon Amos
6 years
These ICML 2019 papers on exploring via disagreement are awesome:. Model-Based Active Exploration.@recurseparadox, @WJaskowski, F. Gomez. Self-Supervised Exploration via Disagreement.@pathak2206, D. Gandi, A. Gupta. 1/5.
2
29
115
@brandondamos
Brandon Amos
10 months
Some related papers for our recent Lagrangian OT:. 0. On amortizing convex conjugates for OT.1. Neural Lagrangian Schrƶdinger Bridge.2a. Deep Generalized Schrƶdinger Bridge.2b. DGSB Matching.3. Wasserstein Lagrangian Flows.4. Metric learning via OT. A 🧵 summarizing these ā¤ļø.
@brandondamos
Brandon Amos
11 months
šŸ“¢ In our new @UncertaintyInAI paper, we do neural optimal transport with costs defined by a Lagrangian (e.g., for physical knowledge, constraints, and geodesics). Paper: JAX Code: (w/ A. Pooladian, C. Domingo-Enrich, @RickyTQChen)
4
11
110
@brandondamos
Brandon Amos
9 years
Our new paper 'Input Convex Neural Networks': The code is also online at:
Tweet media one
Tweet media two
3
44
100
@brandondamos
Brandon Amos
4 years
Excited to finally share our #ICLR2021 papers! With @RickyTQChen and @mnick. Neural Event Functions for ODEs. Neural Spatio-Temporal Point Processes. 🧵 following with virtual posters
Tweet media one
Tweet media two
1
16
97
@brandondamos
Brandon Amos
10 months
I'm giving an #ISMP2024 talk today (Tue) at 5:20pm in 514C on amortized optimization for two seemingly separate topics: optimal transport and LLMs :). If you're not around, I just uploaded the slides to and here's a gif going through them. & a short 🧵.
@brandondamos
Brandon Amos
10 months
If you're at #ISMP2024, we're having a session today on learning and verification for continuous optimizers!!. w/ @RSambharya V. Ranjan, and @b_stellato
Tweet media one
2
9
94
@brandondamos
Brandon Amos
6 years
The satire in one of my ICML reviews is pretty good. The reviewer rejected my paper in a few egregiously wrong sentences and then sprinkled in that I defined the |.| operator only for set cardinality but used it in another context for the absolute value. 😬.
4
9
96
@brandondamos
Brandon Amos
9 years
Densely Connected Convolutional Networks
Tweet media one
Tweet media two
2
65
91
@brandondamos
Brandon Amos
3 years
JAX's Optimal Transport Tools (OTT) package by @CuturiMarco et al. was crucial for developing our new Meta OT paper. I highly recommend looking into it if you are developing new OT ideas!. Code: Docs: Paper:
@brandondamos
Brandon Amos
3 years
Our new paper on Meta Optimal Transport uses amortized optimization to predict initializations and improve optimal transport solver runtimes by 100x and more. With @CohenSamuel13 @GiuliaLuise1 @IevgenRedko . Paper: JAX source code:
2
14
89
@brandondamos
Brandon Amos
5 years
I'm at #NeurIPS2019! Get in touch if you want to talk about. derivatives. And check out our paper and poster on Weds evening: (I'm also open to talking about things that are not derivatives).
@brandondamos
Brandon Amos
6 years
Stoked to share a milestone project for all of us! #NeurIPS2019 paper with @akshaykagrawal, @ShaneBarratt, S. Boyd, S. Diamond, @zicokolter:. Differentiable Convex Optimization Layers. Paper: Blog Post: Repo:
1
5
89
@brandondamos
Brandon Amos
1 year
I'm giving a talk at 1:30pm today (Sat) at the #NeurIPS2023 OTML workshop (room 220-222) on amortized optimization for optimal transport. Come by to learn more!. Workshop: Slides:
Tweet media one
Tweet media two
Tweet media three
1
18
89
@brandondamos
Brandon Amos
8 years
We also just released a fast and differentiable QP solver for @PyTorch.
Tweet media one
1
33
86
@brandondamos
Brandon Amos
5 years
Differentiating through the FrƩchet Mean.A Lou, @isaykatsman, Q. Jiang, @SergeBelongie, S. Lim, @chrismdesa.
Tweet media one
Tweet media two
Tweet media three
1
17
84
@brandondamos
Brandon Amos
9 years
Python script for illustrating CNNs:
Tweet media one
0
29
79
@brandondamos
Brandon Amos
7 years
Variational Inference in Probabilistic Submodular Models.Doctoral Dissertation, ETH Zurich.Josip Djolonga.
Tweet media one
Tweet media two
Tweet media three
0
21
78
@brandondamos
Brandon Amos
8 years
Our new paper (with @priyald17): Task-based End-to-end Model Learning. Paper: Code:
Tweet media one
0
29
79
@brandondamos
Brandon Amos
4 months
good times šŸ˜… *checks date* . over one decade ago~
Tweet media one
@KevinNaughtonJr
Kevin Naughton Jr.
4 months
please don't apply for a senior dev position if your github looks like this.
Tweet media one
2
1
81
@brandondamos
Brandon Amos
8 years
"Beware of math books that don't use pictures to convey ideas.". - Last sentence of Pugh's real analysis book. (Found by @gabrfarina)
Tweet media one
Tweet media two
1
23
75
@brandondamos
Brandon Amos
4 years
In our finalized L4DC paper, we study the stochastic value gradient for continuous model-based RL. Paper: @PyTorch Code: Slides: Talk: With @samscub @denisyarats @andrewgwils
1
12
77
@brandondamos
Brandon Amos
6 years
I'm at #NeurIPS2018! If we run out of preprints and ICLR subs to talk about, here are my papers:. Differentiable MPC (with @zicokolter et al., poster Thu 5pm): Imperfect-Info Game Solving (mostly by @polynoamial, poster Wed 5pm):
@brandondamos
Brandon Amos
7 years
Excited to share our #nips2018 paper on differentiable MPC and our standalone @PyTorch control package!. Joint work with Jimenez, Sacks, Boots, and @zicokolter. Paper: @PyTorch MPC Solver: Experiments:
Tweet media one
Tweet media two
0
17
76
@brandondamos
Brandon Amos
4 years
In our #automl workshop at @icmlconf, @trailsofwater and I look into learning better convex optimization solvers. Paper: @PyTorch Code:
Tweet media one
4
14
76
@brandondamos
Brandon Amos
6 years
Here are some of my quick thoughts on seven awesome and super-promising applications areas of differentiable optimization-based modeling. From the last chapter of my thesis:
Tweet media one
Tweet media two
Tweet media three
0
15
75
@brandondamos
Brandon Amos
1 year
Here's our new paper on jailbreaking LLMs! To do this, we learn /another/ LLM called AdvPrompter to generate a semantically meaningful suffix conditional on a malicious instruction. with @AnselmPaulus @ArmanZharmagam1 C. Guo @tydsh
Tweet media one
Tweet media two
2
13
74
@brandondamos
Brandon Amos
7 years
Deep Frank-Wolfe For Neural Network Optimization.Leonard Berrada, Andrew Zisserman, M. Pawan Kumar. @PyTorch code:
Tweet media one
1
20
69
@brandondamos
Brandon Amos
4 years
Excited to share our new @aistats_conf paper on time series alignment on incomparable spaces! With @CohenSamuel13, G. Luise, @avt_im, and @mpd37. Paper: @PyTorch Code: Blog post:
@CohenSamuel13
Samuel Cohen
4 years
Pleased this joint work on time series alignment with amazing collaborators will be at AISTATS!! @GiuliaLuise1 @avt_im @brandondamos @mpd37.
1
12
72
@brandondamos
Brandon Amos
6 years
MetaOptNet: Meta-Learning with Differentiable Convex Optimization.CVPR 2019 Oral.
Tweet media one
Tweet media two
Tweet media three
0
14
72
@brandondamos
Brandon Amos
3 years
We've added a Meta OT initializer to the Optimal Transport Tools (OTT) JAX package!. Tutorial: Docs: With @CuturiMarco, @JamesTThorn, and Michal Klein
Tweet media one
@brandondamos
Brandon Amos
3 years
Our new paper on Meta Optimal Transport uses amortized optimization to predict initializations and improve optimal transport solver runtimes by 100x and more. With @CohenSamuel13 @GiuliaLuise1 @IevgenRedko . Paper: JAX source code:
1
8
71
@brandondamos
Brandon Amos
9 years
Great slides and video series on deep learning and Torch by @AlfredoCanziani -
Tweet media one
1
29
69
@brandondamos
Brandon Amos
6 years
Stochastic Beams and Where to Find Them: The Gumbel-Top-k Trick for Sampling Sequences Without Replacement.W. Kool, H. van Hoof, M. Welling.
Tweet media one
1
9
69
@brandondamos
Brandon Amos
6 years
Deep Layers as Stochastic Solvers.
Tweet media one
Tweet media two
1
12
65
@brandondamos
Brandon Amos
7 years
There Is No Free Lunch In Adversarial Robustness.(But There Are Unexpected Benefits).
Tweet media one
Tweet media two
Tweet media three
Tweet media four
1
22
64
@brandondamos
Brandon Amos
9 years
draw_convnet - a Python script for illustrating neural nets by Gavin Weiguang Ding
Tweet media one
1
30
64
@brandondamos
Brandon Amos
7 years
šŸŽ‰ @PyTorch code:
@shaojieb
Shaojie Bai
7 years
Our new paper, with @zicokolter and Vladlen Koltun, on extensively evaluating convolutional vs. recurrent approaches to sequence tasks is on arXiv now!
2
24
62
@brandondamos
Brandon Amos
8 years
The Wasserstein GAN code is another great PyTorch example:. Paper: Code:
Tweet media one
0
21
62
@brandondamos
Brandon Amos
8 years
qpth (by me and @zicokolter) has a new website and works with the @PyTorch master branch instead of our fork.
Tweet media one
1
18
61