Taco Cohen Profile Banner
Taco Cohen Profile
Taco Cohen

@TacoCohen

Followers
21,457
Following
2,776
Media
39
Statuses
1,206

Deep learner at FAIR. Into codegen, equivariance, generative models. Spent time at Qualcomm, Scyfer (acquired), UvA, Deepmind, OpenAI.

Joined March 2013
Don't wanna be here? Send us removal request.
Explore trending content on Musk Viewer
@TacoCohen
Taco Cohen
1 year
Surprisingly little AI progress in 2023 so far. What’s going on??
128
305
5K
@TacoCohen
Taco Cohen
6 months
An interesting aspect of this discussion is the fact that LLMs will soon start affecting our thoughts, beliefs, mental & linguistic habits, and culture. The idea that we could select a handful of "trustworthy" institutions with the "correct" set of values and beliefs to shape LLM…
@karpathy
Andrej Karpathy
6 months
Thinking a lot about centralization and decentralization these few days.
826
1K
12K
66
194
1K
@TacoCohen
Taco Cohen
3 years
Rumor has it that I don't even have a PhD yet. This is in fact true... 😏 BUT! I am happy to report that I will be graduating before any of the PhD students I'm advising. The thesis is now online and I will be defending Jun 9th, 16.00 CET! Check it out:
Tweet media one
23
68
851
@TacoCohen
Taco Cohen
2 years
8 years of progress in generative modelling. What a time to be alive
Tweet media one
Tweet media two
10
47
771
@TacoCohen
Taco Cohen
4 months
Two weeks ago I joined Meta / FAIR, and I couldn't be more excited about this new chapter. Meta is indeed the only place left that supports highly ambitious long-term oriented & fundamental research projects and has a strong commitment to open science and open source. (and has…
@ylecun
Yann LeCun
4 months
There is literally no other company doing this today: - open research towards human-level AI - open source AI platform enabling a huge AI ecosystem - wearable device to interact with always-on AI assistants
120
212
2K
81
27
735
@TacoCohen
Taco Cohen
6 years
Best paper award for our ICLR paper, "Spherical CNNs"! Read it while it's hot 🔥 🔥
12
258
678
@TacoCohen
Taco Cohen
3 years
So these "Multi-Headed Vision Transformers", are they in the room with us right now?
Tweet media one
8
31
550
@TacoCohen
Taco Cohen
5 years
Interested in geometric and equivariant deep learning? Check out our latest paper on Gauge Equivariant CNNs, where we show how gauge theory makes it possible to build CNNs on general manifolds:
Tweet media one
6
155
546
@TacoCohen
Taco Cohen
2 years
After almost a decade in ML/AI, I still don't really know what a symbol is 😢
47
19
549
@TacoCohen
Taco Cohen
5 years
This. Don't waste time on domain specific tricks. Do work on abstract & general inductive biases like smoothness, relational structure, compositionality, in/equivariance, locality, stationarity, hierarchy, causality. Do think carefully & deeply about what is lacking in AI today.
@seth_stafford
Seth Stafford
5 years
The contrast btw Rich Sutton and Shimon Whiteson re the value of injecting human knowledge into models is a good definition of the word “principled”. Sutton‘s Bitter Lesson is that ad hoc tricks don’t hold up. @shimon8282 ‘s Sweet Lesson us that deeper (more principled) ideas do.
2
14
101
6
83
368
@TacoCohen
Taco Cohen
6 years
PCam: the CIFAR10 of medical imaging.
Tweet media one
2
123
340
@TacoCohen
Taco Cohen
2 years
After LLMs, the next big thing will be LCPs: Large Control Policies. Very general pretrained goal-conditioned policies for embodied agents. If you provide it with a goal vector / example / text, it can do a large number of tasks in a large number of environments. Then we retire🤖
16
34
341
@TacoCohen
Taco Cohen
2 years
👉 The first law of DL architectures 👈 "Whatever" is all you need 🤯 Any problem that can be solved by transformer / ViT can be solved by MLP / CNN, and vice versa [provided you do exhaustive tuning, and use the right inductive bias] Same for RNNs:
@_akhaliq
AK
2 years
A ConvNet for the 2020s abs: github: Constructed entirely from standard ConvNet modules, achieving 87.8% ImageNet top-1 accuracy and outperforming Swin Transformers on COCO detection and ADE20K segmentation
Tweet media one
11
210
976
6
38
335
@TacoCohen
Taco Cohen
3 years
Better late than never! A sincere thanks to my doctoral committee @geoffreyhinton @erikverlinde @mmbronstein @risi_kondor @erikjbekkers Leo Dorst & Joris Mooij, and my amazing supervisor @wellingmax . Next milestone: update my profile pic from 2010.
@wellingmax
Max Welling
3 years
Congratulations to superstar Dr. Taco Cohen for graduating Cum Laude yesterday.
Tweet media one
13
10
466
30
0
323
@TacoCohen
Taco Cohen
2 years
A very clear explanation of an idea that is at the heart of modern mathematics, and geometric deep learning as well: Klein's Erlangen Program and its generalization, here called the isomorphism philosophy. A short thread on why this matters for AI: 1/
@JDHamkins
Joel David Hamkins
2 years
The isomorphism philosophy: in any mathematical context, the genuinely mathematical ideas and properties are precisely those preserved by isomorphism.
Tweet media one
Tweet media two
Tweet media three
28
78
521
7
49
317
@TacoCohen
Taco Cohen
4 years
Angela Merkel, former quantum chemist, explaining the subtleties of exponential growth in plain language👌
@BenjAlvarez1
Benjamin Alvarez
4 years
This is how Angela Merkel explained the effect of a higher #covid19 infection rate on the country's health system. This part of today's press conf was great, so I just added English subtitels for all non-German speakers. #flattenthecurve
3K
37K
91K
7
52
290
@TacoCohen
Taco Cohen
4 years
IMO this is the most insightful way to introduce and understand convolution. Interestingly, group conv and steerable conv on homogeneous spaces can also be derived from symmetry principles. Convolution is all you need!
@mmbronstein
Michael Bronstein
4 years
Have you ever wondered what is so special about convolution? In a new blog post, I show how to derive #convolution from translational symmetry principles: This is key to extending #DeepLearning to #graphs
Tweet media one
14
324
1K
2
62
288
@TacoCohen
Taco Cohen
2 years
It's psychologically helpful to note that by duality, rejection is just acceptance into the collection of rejected papers #NeurIPS2022
5
9
276
@TacoCohen
Taco Cohen
2 years
The culmination of years of research: our neural video codec running in real time on a mobile device 😮
@aukejw
Auke Wiggers
2 years
Exciting work from our team towards making neural video compression a reality: running a neural video decoder on a mobile phone in real time. Check out the demo video at
Tweet media one
3
26
137
3
27
271
@TacoCohen
Taco Cohen
4 years
Next week I will be kicking off the virtual Physics ⋂ ML series with a talk about *Natural* Graph Networks, a new and fundamentally more flexible class of graph networks. Without a doubt the most exciting thing since Gauge CNNs 🔥 Project led by @pimdehaan with @wellingmax
@jhhalverson
Jim Halverson
4 years
Physics ∩ ML is going virtual! If you're interested in the interface of theoretical physics and ML, come hear talks by @TacoCohen , Phiala Shanahan, Ard Louis, and @hashimotostring . More info at .
Tweet media one
3
46
211
4
42
258
@TacoCohen
Taco Cohen
4 years
Short but sweet paper on recurrent autoencoder architectures for speech compression. We systematically explore the space of RNN-AEs and show that the best method, dubbed FRAE, outperforms classical codecs by a large margin. Check it out!
Tweet media one
@gsautiere
Guillaume Sautière
4 years
I am thrilled to announce our paper “Feedback Recurrent AutoEncoder” was accepted at #ICASSP2020 ! collaboration with Yang Yang, @TacoCohen and Jon Ryu. . A quick thread.
2
28
116
1
65
249
@TacoCohen
Taco Cohen
2 years
We're looking for summer interns at Qualcomm AI Research in Amsterdam! Interested in working on causal rep. learning & RL (my team), compression/generative models, combinatorial opt., model efficiency, federated learning, wireless, perception? Apply now!
8
46
242
@TacoCohen
Taco Cohen
5 years
Experts debunking AI hype
4
22
240
@TacoCohen
Taco Cohen
2 years
If we solve all benchmarks with ~current tools + large scale systems engineering, we will have learned that intelligence is a mirage; a bunch of domain-specific tricks. Imo this'd be profound, on par with "earth is just another planet" & "humans are just another kind of animal"
@Thom_Wolf
Thomas Wolf
2 years
there is a scary possibility that we may solve all the benchmarks we come up for AI... without understanding anything fundamentally deep about what intelligence is about a bummer for those like me who are see AI as a fantastic way to unlock deeper insights on human intelligence
35
44
452
16
28
230
@TacoCohen
Taco Cohen
3 years
Interested in generative modelling and image/video/audio compression? Qualcomm AI Research is hiring researchers in this exciting area in Amsterdam and San Diego!
3
44
213
@TacoCohen
Taco Cohen
2 years
It’s official, folks
@geoffreyhinton
Geoffrey Hinton
2 years
Equivariance rules!
17
145
968
6
21
214
@TacoCohen
Taco Cohen
10 months
Harm's Law of Smol Models (HLSM) tells us how much we need to scale up the data size (k_D) as we scale down the model size (k_N), if we wish to preserve the loss of a Chinchilla-optimal model.
Tweet media one
3
42
211
@TacoCohen
Taco Cohen
8 months
Super excited to present our latest work in GDL: The Geometric Algebra Transformer (AKA GATr 🐊) Combines the scalability of a transformer with general-purpose GA features & full E(3) equivariance. Check out the thread below! ⬇️
@johannbrehmer
Johann Brehmer
8 months
Are you dealing with geometric data, be it from molecules or robots? Would you like inductive biases *and* scalability? Our Geometric Algebra Transformer (GATr 🐊) may be for you. New work w/ @pimdehaan , Sönke Behrends, and @TacoCohen : 1/9
Tweet media one
4
81
353
1
31
214
@TacoCohen
Taco Cohen
2 months
A lot of people are skeptical that self-training can work. But the story of Ramanujan shows that once a certain threshold of intelligence is crossed, pure self-training in mathematics is possible even without an external reward signal provided by a proof checker.
15
17
211
@TacoCohen
Taco Cohen
3 years
Really looking forward to this! We plan to release the lecture videos & course materials online
@mmbronstein
Michael Bronstein
3 years
Together with @joanbruna @PetarV_93 @TacoCohen we will be teaching a course on #geometricdeeplearning in the @AIMS_Next #AMMI program. The course is based on our protobook Thanks @Moustapha_6C Teta Bahunde @panfordkobby for this opportunity
9
83
454
9
16
199
@TacoCohen
Taco Cohen
4 years
Dear ML twitter, Not to fear monger, but the mathematicians are closing in on us. They just reached 1999 and reduced Roweis & Ghahramani's epic paper to a slick 2-pager:
@davidad
davidad 🎇
4 years
I talked to my friend Peter (who is, among other things, stronger than I am in analysis) for some key ideas, then sat down and proved this today.
Tweet media one
2
4
54
6
30
198
@TacoCohen
Taco Cohen
2 years
Anyone know a good recent-ish review of self-supervised learning methods?
15
15
202
@TacoCohen
Taco Cohen
2 years
Very exciting result: equivariance changes the exponent of the scaling law! Equivariant nets really do *learn faster* [provided the problem has the relevant symmetries]
3
25
200
@TacoCohen
Taco Cohen
2 years
unreal
@Innov_Medicine
The Innovation | Medicine
2 years
DNA to RNA real-time speed. Gene Transcription at real-time speed. Transcription is the first step in gene expression.
82
3K
10K
2
26
193
@TacoCohen
Taco Cohen
2 years
After a few months of intensive study I'm still not 100% sure what it is that makes a variable "causal" or if all variables are causal 😢
@TacoCohen
Taco Cohen
2 years
After almost a decade in ML/AI, I still don't really know what a symbol is 😢
47
19
549
28
13
188
@TacoCohen
Taco Cohen
4 years
Rotation equivariant Steerable G-CNNs are now state of the art on tumor classification, nuclear segmentation and gland segmentation. Very exciting to see G-CNNs being used more and more in medical imaging, and working so well!
@simongraham73
Simon Graham
4 years
[1/6] We are pleased to announce our paper ‘Dense Steerable Filter CNNs for Exploiting Rotational Symmetry in Histology Image Analysis’ paper: code: @nmrajpoot @TIAwarwick
2
40
133
1
36
188
@TacoCohen
Taco Cohen
7 months
The Good Regulator Theorem states that a maximally simple regulator of a system must contain a model of that system. A regulator is kind of like a policy that controls the system to keep its outputs in some desired range. To be a model means that there exists a homomorphism from…
@wesg52
Wes Gurnee
7 months
Do language models have an internal world model? A sense of time? At multiple spatiotemporal scales? In a new paper with @tegmark we provide evidence that they do by finding a literal map of the world inside the activations of Llama-2!
183
1K
6K
13
29
182
@TacoCohen
Taco Cohen
5 years
With this Deep 1-pager, @jaschasd has reached the global minimum of paper writing
@jaschasd
Jascha Sohl-Dickstein
5 years
Eliminating All Bad Local Minima from Loss Landscapes Without Even Adding an Extra Unit It's less than one page. It may be deep. It may be trivial. It will definitely help you understand how some claims in recent theory papers could possibly be true.
Tweet media one
6
178
703
2
18
180
@TacoCohen
Taco Cohen
3 years
🥱 Tired: teach kids trigonometry 😲 Wired: teach kids statistics 🤩 Inspired: teach kids causal inference
4
13
171
@TacoCohen
Taco Cohen
1 year
"Identifiability proofs", which are conspicuously absent for all modern AI methods that actually work, are considered indispensable in the causal inference & causal representation learning communities. Without a proof, the method is not "truly causal".
10
24
167
@TacoCohen
Taco Cohen
4 years
More evidence that roto-translation equivariant G-CNNs outperform conventional CNNs by a large margin on medical imaging problems with rotation symmetry. G-CNN on 25%-50% of data outperforms CNN on 100% (+data augmentation). Great paper with lots of details & careful experiments.
@erikjbekkers
Erik Bekkers
4 years
Great work by Maxime Lafarge indeed! It shows that group CNNs again consistently outperform regular CNNs and it shows the power of G-convs with a fine rotation resolution (finer than standard 90 degree rotations). Includes a careful analysis of obtained equivariance of the nets.
0
15
71
3
40
163
@TacoCohen
Taco Cohen
5 years
A beautiful demonstration of the mathematical fact that it is not possible to map a non-trivial orbit of SO(3) [the rotating car] to a Euclidean latent space in a continuous and invertible manner. More research needed!
@SyntopiaDK
Mikael H Christensen
5 years
GAN's may be evaluated based on how smooth (disentangled) the latent space interpolations are. It is impressive how #StyleGAN can interpolate between different orientations - even with no concept of 3D.
3
91
431
2
30
161
@TacoCohen
Taco Cohen
3 years
If true, this would be a big vindication for equivariant nets. ...Dreaming of a day I will give a talk and nobody asks why we don’t just do data augmentation... 😌
@FabianFuchsML
Fabian Fuchs
3 years
There is still quite a bit of mystery around the details of @DeepMind 's AlphaFold 2, but equivariance & symmetries may have played a significant role in their success. This is @JustasDauparas 's and my take 🧐:
7
83
329
10
7
159
@TacoCohen
Taco Cohen
5 years
The world may finally see a safe and beneficial version of Clippy
@ilyasut
Ilya Sutskever
5 years
Super exciting news: Microsoft is investing $1B into OpenAI, and we're partnering to build giant NN computers within Azure that will train giant NNs!
20
215
1K
3
5
156
@TacoCohen
Taco Cohen
1 year
Tomorrow at 4pm GMT: a new talk on grounding causal models in dynamical systems and MDPs for the cats4ai series:
5
24
142
@TacoCohen
Taco Cohen
2 years
👉👉👉Applications are now open for internships at Qualcomm AI Research! 👈👈👈 Apply now to work with our amazing team on topics ranging from model compression to RL, federated learning, generative models, causality and more.
2
31
151
@TacoCohen
Taco Cohen
4 years
e2cnn: A comprehensive library for easy construction of rotation-reflection-translation equivariant CNNs in @PyTorch + thorough a experimental study of equivariant network architectures. By @_gabrielecesa_ and @maurice_weiler .
@maurice_weiler
Maurice Weiler
4 years
Check out our poster #143 on general E(2)-Steerable CNNs tomorrow, Thu 10:45AM. Our work solves for the most general isometry-equivariant convolutional mappings and implements a wide range of related work in a unified framework. With @_gabrielecesa_ #NeurIPS2019 #NeurIPS
1
33
136
1
28
144
@TacoCohen
Taco Cohen
3 months
Hardly anyone believes that LLMs learn or think the way humans do, but if you are instead looking for the essence of intelligence, compression (what LLMs are trained for) is a decent starting point.
11
10
129
@TacoCohen
Taco Cohen
3 years
Had some more printed, so still have a few copies! DM your address if you want one, I only charge for shipping and even that is free if you can’t afford it.
5
3
126
@TacoCohen
Taco Cohen
2 years
Nice example of theory informing practice: Tune hyperparams on a small model using muParameterization, transfer them to a large model without further tuning. Big deal if it works as advertised.
@arankomatsuzaki
Aran Komatsuzaki
2 years
Tensor Programs V: Tuning Large Neural Networks via Zero-Shot Hyperparameter Transfer By transferring from 40M parameters, µTransfer outperforms the 6.7B GPT-3, with tuning cost only 7% of total pretraining cost. abs: repo:
Tweet media one
5
43
248
1
24
125
@TacoCohen
Taco Cohen
4 years
Dear Twitterverse: it is 2020 and the GAN literature is huge. What are some of the best methods to stabilize training and prevent mode dropping?
10
17
125
@TacoCohen
Taco Cohen
3 years
Look mom, I'm on twitter!
@XEng
Engineering
3 years
Geometric Deep Learning, a new proto-book on deep learning, co-authored by @mmbronstein Head of Twitter Graph Learning Research with @PetarV_93 , @joanbruna , @TacoCohen .
2
55
292
0
2
125
@TacoCohen
Taco Cohen
4 years
John Schulman's opinionated guide to ML research
0
27
117
@TacoCohen
Taco Cohen
6 years
Chainer continues to amaze me. With a tiny team, they built a DL framework that is competitive with or superior to the major (well-funded) DL frameworks in terms of speed, ease of use, and features (e.g. Chainer pioneered dynamic computation graphs).
@CuPy_Team
CuPy
6 years
Released Chainer/CuPy v4.0.0! #Chainer : Major performance improvements including TensorCore support and iDeep backend, NCCL2 support, Caffe export. #CuPy : CUDA 9.1 support, wheel package, FFT support, etc. More in the blog post and release notes.
0
62
129
1
35
111
@TacoCohen
Taco Cohen
4 years
Quanta Magazine covers Geometric DL & G-CNNs. For technical details, check out some of the original papers mentioned in the article:
@wellingmax
Max Welling
4 years
Nice piece in Quanta about gauge CNNs and geometric deep learning
1
67
208
0
28
110
@TacoCohen
Taco Cohen
4 years
Some people are sometimes able to correctly predict some things about the distant future. It's remarkable!
Tweet media one
2
13
108
@TacoCohen
Taco Cohen
4 years
There has been some discussion on ML twitter about the meaning of the word compositionality. It is a word that, like "disentangling", has many meanings. But there is a mathematical framework that captures all of them: category theory.
@aggielaz
Angeliki Lazaridou
4 years
On the topic of compositionality: I was recently tasked with giving a talk on the topic (what do people mean, how do they measure, how to achieve it etc) ->
3
48
308
7
17
103
@TacoCohen
Taco Cohen
6 months
If AI actually gets good at finding security vulnerabilities, software will quickly become much more secure.
10
9
104
@TacoCohen
Taco Cohen
3 years
Of course, attention and symmetries are all you need for protein structure prediction 😁
2
13
102
@TacoCohen
Taco Cohen
6 years
First steps towards learning representations that respect the topology of the data manifold: "Explorations in Homeomorphic Variational Auto-Encoding" by @lcfalors @pimdehaan @im_td @nicola_decao M. Weiler, P. Forre, yours truly. Check out poster 19 at #TADGM #ICML2018
Tweet media one
1
30
95
@TacoCohen
Taco Cohen
2 months
Now that machine learning is kind of working, it's time to focus on machine teaching.
5
9
94
@TacoCohen
Taco Cohen
5 years
Finding applications of inapplicable math is one of my favorite things. So I would like to take this opportunity to apologize to all the group representation theorists, non-commutative harmonic analysts, differential geometers, and fiber bundlists whose work I have made use of.
@KeithEPeterson_
Keith E Peterson
5 years
Tweet media one
0
1
30
1
7
96
@TacoCohen
Taco Cohen
5 years
In deep learning, it is acceptable to add an inductive bias to your model, but only if you don't understand why it works. Understanding things via mathematics was already tried by the SVM folks and it didn't work.
@dileeplearning
Dileep George
5 years
"Artificial General Generality": a new cartoon in the AGI series + some old ones in the thread.
Tweet media one
1
9
39
4
16
93
@TacoCohen
Taco Cohen
2 years
Yuval Harari ( @harari_yuval ) noted in Sapiens that humans may be unique in their ability to imagine non-existing things like a person with a lion's head. Interestingly, generative models already appear to be quite good at this.
Tweet media one
Tweet media two
6
4
91
@TacoCohen
Taco Cohen
3 years
Our neural video codec running realtime on a mobile device! Super proud of the team.
@QCOMResearch
Qualcomm Research & Technologies
3 years
Check out @Qualcomm #AI Research's latest breakthrough: the world’s first software-based neural video decoder running HD format in real-time on a commercial smartphone. Learn more:
0
25
83
3
6
89
@TacoCohen
Taco Cohen
2 months
🚨 Hiring Alert🚨 The FAIR CodeGen team in Paris is looking for research engineers! Come join this super talented team, help release open models to the world, and push the frontiers of code generation research!
@syhw
Gabriel Synnaeve
2 months
The CodeGen team at FAIR *in Paris* is recruiting junior and senior research engineers! Come work with us @jadecopet @b_roziere @qcar_ @FabianGloeckle @KunhaoZ et al., and folks in EMEA @jnsgehring @TacoCohen @adiyossLC @FelixKreuk et al.
1
27
105
1
11
82
@TacoCohen
Taco Cohen
5 years
A principled way to deal with scale variation in convolutional nets. Neat!
@danielewworrall
Daniel Worrall
5 years
Check out my new work with @wellingmax on Deep Scale Spaces (link: ). We develop a new kind of 'semigroup convolution', generalizing the group conv of @TacoCohen , and present the connection with classical scale-spaces from CV
Tweet media one
Tweet media two
Tweet media three
Tweet media four
4
41
162
0
23
81
@TacoCohen
Taco Cohen
6 years
Very clear introduction to equivariant convolutional networks by one of the experts in the field. Highly recommended.
@danielewworrall
Daniel Worrall
6 years
So I gave a talk on my research on equivariant CNNs at a London #machinelearning #meetup . Watch here:
0
23
76
0
21
79
@TacoCohen
Taco Cohen
5 years
A fairly realistic depiction of academic debates on twitter
0
8
77
@TacoCohen
Taco Cohen
4 years
To me, the current phase is even more exciting than the last. To make progress, we need to rethink foundations: causality and explanation, learning without rewards, common sense reasoning, etc.. Not easy, but certainly tractable.
@Sam_L_Shead
Sam Shead
4 years
As the new decade gets underway, AI appears to be transitioning to a new phase. But what does it look like? I spoke to academics and researchers at companies like Facebook, DeepMind, and Microsoft to try and find out
2
6
26
1
4
74
@TacoCohen
Taco Cohen
6 years
Any theory that explains how or why neural nets work so well should be consistent with the fact that NNs that don't throw away any information until the very last layer work just fine.
@RogerGrosse
Roger Grosse
6 years
Fully reversible extension of RevNets by Jörn Jacobsen and colleagues, plus a neat connection to Sweldens' lifting scheme for wavelets.
2
27
75
3
16
75
@TacoCohen
Taco Cohen
5 years
Green AI: "[Deep Learning] computations have a surprisingly large carbon footprint. [...] This position paper advocates a practical solution by making efficiency an evaluation criterion for research along-side accuracy and related measures"
3
16
74
@TacoCohen
Taco Cohen
1 year
Looking forward to an in-person NeurIPS! I will be at the Qualcomm booth Tue & Wed from 9-11 and 13-15. Stop by anytime or send me a DM if you want to chat!
1
2
72
@TacoCohen
Taco Cohen
4 years
Happening today! OmniCV workshop @ CVPR. I’ll be giving a (pre-recorded) talk on Spherical CNNs, Icosahedral CNNs, Gauge CNNs, Mesh CNNs and all that, and doing a live Q&A
0
12
74
@TacoCohen
Taco Cohen
2 years
I highly recommend this course on Equivariant DL by Erik Bekkers. It does a great job covering the fundamentals as well as recent developments. Check it out!
@erikjbekkers
Erik Bekkers
2 years
Dear GDL friends! Here's a🧵on our mini-course ✨Group Equivariant Deep Learning✨ See for YT playlist (21 vids), colabs, slides, lecture notes. Topics: 1⃣regular & 2⃣steerable g-convs 3⃣equivariant graph NNs 4⃣geometric latent space models 1/14
4
72
356
1
12
74
@TacoCohen
Taco Cohen
5 years
WIRED: the AlphaStar Transformer-LSTM-AutoRegressive-PointerNet cognitive architecture TIRED: deep learning is just curve fitting EXPIRED: arguments about symbolic AI
@OriolVinyalsML
Oriol Vinyals
5 years
Happy that we could share #AlphaStar progress with you all! Good Games @LiquidTLO and @Liquid_MaNa , and @Artosis and @RotterdaM08 for a great show! You can see all the details in the blog.
47
450
1K
0
7
72
@TacoCohen
Taco Cohen
4 years
Self attention, group convolution, and a figure that could pass for a Picasso. What more do you want?
Tweet media one
@davidwromero
David W. Romero
4 years
We present the attentive group convolution, a generalization of the group convolution that uses attention during the group convolution to focus on relevant symmetry combinations. It generates equivariant attention maps as well. @erikjbekkers @jmtomczak
Tweet media one
Tweet media two
Tweet media three
3
54
185
0
18
70
@TacoCohen
Taco Cohen
1 year
Guess the paper!
Tweet media one
2
0
66
@TacoCohen
Taco Cohen
6 years
Max Welling @wellingmax is doing an AMA today! Just a few hours left to ask questions:
1
21
67
@TacoCohen
Taco Cohen
3 years
Join us tomorrow at the ICLR workshop "Neural Compression: From Information Theory to Applications"! With a wonderful list of speakers & panelists:
3
9
62
@TacoCohen
Taco Cohen
11 months
The two best Euclidean-equivariant CNN libraries now both have Jaxx as well as Pytorch implementations. Using equivariant nets has never been easier.
@MathieuEmile
Emile Mathieu
11 months
I've written a Jax version of the great _escnn_ () python library for training equivariant neural networks by @_gabrielecesa_ It's over there! Hope you'll find it useful 🙌
5
12
92
0
8
63
@TacoCohen
Taco Cohen
6 years
I've been saying this for a while now. Having a prior belief about the value of a meaningless parameter makes no sense. Important corollary: number of parameters is not a great measure of model complexity.
@dustinvtran
Dustin Tran
6 years
Think in function space, not parameter space. @yeewhye 's talk on Bayesian deep learning at #NIPS2017
Tweet media one
3
90
322
0
11
60
@TacoCohen
Taco Cohen
6 years
Another exciting workshop coming up: "Towards learning with limited labels: Equivariance, Invariance, and Beyond". With talks by Bengio, Poggio, Soatto, Gupta, Pathak & yours truly. Submissions due May 20th! (2 days after the NIPS deadline)
Tweet media one
0
14
60
@TacoCohen
Taco Cohen
3 years
Very excited about this project and the future possibilities for instance-adaptive compression. Great work by joint first authors @tivaro & @IamHuijben !
@tivaro
Ties van Rozendaal @[email protected]
3 years
In our new paper with @IamHuijben and @TacoCohen (accepted at #ICLR2021 ), we improve neural I-frame compression with 1 dB by overfitting the full compression model on the data instance that we want to transmit! (1/3)
2
13
72
1
6
60
@TacoCohen
Taco Cohen
2 years
The bitter-sweet lesson: methods that can efficiently leverage compute & data work best, but you still need to respect the symmetries. #geometricdeeplearning #compchem
1
4
60
@TacoCohen
Taco Cohen
3 months
Come work with us!
@syhw
Gabriel Synnaeve
3 months
We’re hiring PhD interns to work on code generation research at FAIR in EMEA! Please apply at if you’re interested by research in Code Llama, LLMs, code generation, compilers, reinforcement learning.
5
26
124
1
5
59
@TacoCohen
Taco Cohen
6 years
Latest news from Equivariland: "Clebsch-Gordan Networks: a Fully Fourier Space Spherical Convolutional Neural Network", by @risi_kondor , Zhen Lin & @_onionesque . Easy to implement and numerically stable 3D rotation-equivariant networks.
@_onionesque
Shubhendu Trivedi
6 years
Our new paper: "Clebsch-Gordan Networks: a Fully Fourier Space Spherical Convolutional Neural Network" The architecture here avoids forward and backward Fourier transforms needed in prior art by making use of the C-G transform as the non-linearity.
3
24
76
1
11
59
@TacoCohen
Taco Cohen
2 years
New PhD project on geometric DL for spatiotemporal data in Amsterdam by @egavves ! (I will serve as industry co-supervisor) The project is quite open ended, so lots of room for your input. Great opportunity to work in an exciting area with top-notch colleagues in the QUVA lab.
@egavves
Efstratios Gavves
2 years
Interested in 'Geometric Deep Learning of Space and Time'?The portal is now online! Apply *now* for our ELLIS PhD program for a PhD position at the QUVA Lab of the University of Amsterdam, with @TacoCohen ! #ECCV2022 #NeurIPS2022
1
18
92
0
4
57
@TacoCohen
Taco Cohen
2 months
@NandoDF Another issue is paper length. Many of the tech reports on LLMs and code models are necessarily very long and won’t fit into 8 pages. Maybe there should be a special venue for such engineering heavy research?
5
2
57
@TacoCohen
Taco Cohen
3 years
@fhuszar @wellingmax PhD defense in 2021 🤷‍♂️
Tweet media one
2
0
53
@TacoCohen
Taco Cohen
2 years
Regular reminder that Qualcomm AI Research is hiring DL researchers and software engineers!
0
10
52
@TacoCohen
Taco Cohen
1 year
deeply concerned that after all these years we haven't even solved the alignment problem for Ikea furniture
4
2
53
@TacoCohen
Taco Cohen
4 years
Constrained optimization has several practical advantages over the standard beta-VAE (rate/distortion) loss for training compression models. Check out the paper! 👇
@tivaro
Ties van Rozendaal @[email protected]
4 years
Still training β-VAEs for lossy compression? Why not use constrained optimization? Have a look at our CLIC CVPR paper: Lossy Compression with Distortion Constrained Optimization Joint work with @TacoCohen and @gsautiere
Tweet media one
0
27
77
0
10
53
@TacoCohen
Taco Cohen
4 years
Natural Graph Networks: Tomorrow, Wednesday 6th at 12:00 EDT!
@jhhalverson
Jim Halverson
4 years
Physics ∩ ML is now listed on , with easy calendar sync. Come hear @TacoCohen tomorrow @ 12:00 EDT on "Natural Graph Networks." Info sent via mailing list, register at .
0
18
54
3
13
51