Liu Liu
@liuliu
Followers
2K
Following
822
Media
52
Statuses
1K
Maintains https://t.co/VoCwlJ9Eq0 / https://t.co/bMI9arVwcR / https://t.co/2agmCPOZ2t. Wrote iOS app Snapchat (2014-2020). Founded Facebook Videos with others (2013). Sometimes writes at https://t.co/Gyt4J9Z9Tv
San Francisco, CA
Joined April 2010
QR: Gemini 3 knowledge seems fresher (can reference things from September without RAG). It now answers correctly a few questions previously failed on 2.5 Pro (but succeeded on GPT-5). Haven't yet found a problem that GPT-5 fails but Gemini 3 succeeds. Will take a few days.
0
0
3
Once we started to talk about deployment-focused frameworks, there are too many to be enumerated. Many of them still active today (look no further than ncnn, still going strong a decade later! Also onnx..). They brought a lot of sweat and tears to the table, but not modernity.
1
0
0
llama.cpp: similar design backbone to Caffe2 (highly likely just a coincidence). Brought quantization techniques and one-click deployment to mass.
1
0
0
mxnet: started similarly to Caffe, mid-way changed to be similar to TensorFlow. Ran out of steam when trying to be more like PyTorch.
1
0
0
Caffe2: similar design as TensorFlow, but tries to be a bit more developer-friendly.
1
0
0
Here are some side-quests: Chainer: more performant Theano, I honestly don't know why it didn't succeed but PyTorch did.
1
0
0
There are many more frameworks today, that brings nothing new beyond PyTorch offers (will list these "side-quest" later). The only novelty to me since PyTorch is the single-coordinator pattern to launch large-scale single training run.
1
0
0
PyTorch: eager execution, but no memory-leaks and implicit asynchronous kernel launches gave good-enough performance with good debugging experience.
1
0
0
Keras: modularize your complex neural networks with side-effect-free functions.
1
0
0
TensorFlow: popularize the idea you can use DAG to represent your neural networks, gave people hope that frameworks can do crazy optimizations (they cannot).
1
0
0
LuaTorch: eager execution. Ran your neural network with a script s.t. you can debug internals.
1
0
0
Caffe: popularize im2col op to do CNN such that your shining GTX 980 can use cuBLAS ran CNN without writing your own kernel. Also popularized modular layer-based neural network construction, which is slightly easier than cuda-convnet.
1
0
0
cuda-convnet: show-case it is possible to run convolutional neural networks at 10x~20x speed as CPU at that time with GTX 780.
1
0
1
Modernity and deep learning frameworks since 2012. There are many deep learning frameworks since the renaissance, but only some of them brought new ideas (or brought back old ideas in significant way, as someone might say there is no new ideas in DL). Here is my list:
1
0
4
(The matmul (we care about) are 4096x3072x3072 / 4096x12288x3072 / 4096x3072x12288, and on M5 (iPad Pro), it tops ~12.5 TFLOPs TFLOPs w/ NA).
2
1
6