New work on how to construct neural network layers on composite objects of scalars, vectors, bivectors, … --> multivectors! Via Clifford algebras, we generalize convolution and Fourier transforms to multivectors, especially relevant for PDE modeling:
Work done as a close collaboration with
@vdbergrianne
and
@wellingmax
, at our new AI4Science team at Microsoft Research
@MSFTResearch
Amsterdam, and...
@rejuvyesh
at Microsoft Autonomous Systems and Robotics Research Group.
2/n
In sciences, PDEs describe physical processes as scalar and vector fields interacting and coevolving over time. We want neural PDE surrogates to take into account the relationship between different fields and their internal components, via so-called multivectors.
3/n
In Clifford algebras, different spatial primitives can be summarized into such multivectors.
A multivector does not only have scalar, and vector components. But also has plane (bivector) and volume (trivector) primitives.
4/n
An example of a bivector is the cross-product of two vectors in 3d, which is a plane segment spanned by these two vectors, and is often represented as a vector itself. But the cross-product has a sign flip under reflection which a true vector does not. It is a bivector.
5/n
We introduce Clifford neural layers, both for convolution and for Fourier operations, that handle scalar (e.g. density), vector (e.g. wind/electric field), bivector (e.g. magnetic field) and higher order fields as proper geometric objects termed as multivector fields.
6/n
And that is the cool thing: we can build neural networks on top of this concept of multivectors😊If there are physical connections between fields, e.g. between electric and magnetic fields, Clifford layers give you a handle on how to encode these.
7/n
We introduce Clifford convolution: instead of convolving over channels of feature maps, we convolve over channels of multivector feature maps using multivector kernel.
8/n
We therefore evaluate on fluid dynamics (Navier-Stokes equations) where scalar smoke and vector velocity fields form multivectors, coupled over force terms.
10/n
We replace each convolution with a Clifford convolution in ResNet architectures, and each Fourier Neural Operator layer with a Clifford Fourier layer in Fourier Neural Operators (FNOs):
12/n
We finally evaluate on Maxwell’s equations in 3 dimensions, which beautifully illustrate the multivector nature of electromagnetism. This is because the magnetic field is the dual (bivector) of the electric field (vector).
13/n
For all these datasets, we experiment with different number of training trajectories. Clifford neural layers consistently improve generalization capabilities of the tested architectures.
But… in the end, it seems we have just touched the tip of the iceberg when it comes to marrying geometric algebra and deep learning. There is much more to harness :)
@andrewwhite01
@mpspells
Thank you very much for forwarding this paper. We did our best to cite as much as possible, seems we overlooked this important one. This is a nice piece of work!
@mpspells
builds multivectors by applying vector-vector operations, e.g. uses the cross-product to obtain
1/2
@jo_brandstetter
I look forward to Clifford Algebras being a hot topic of research in ML. I've been looking for the past year but from a basic data modeling lens.
A big limitation is the exponential growth of dimension in degree. Do you restrict yourself up to 3-vectors?
Linear systems are the linchpin of many problems in science & engineering
What if I told you they are solved error-free by extending their geometry, in terms of oriented
• lines (vectors)
• areas (bivectors)
• volumes (trivectors)
• etc.
Case in point: an N×2 linear system
@alkalait
We need less channels if we use multivector objects. In 2D and 3D experiments, Clifford networks performed better for the same nr of params. For 4D and higher, you are right, there might be a tradeoff which needs to be explored :)