ImageNet is the new CIFAR! My students made FFCV (), a drop-in data loading library for training models *fast* (e.g., ImageNet in half an hour on 1 GPU, CIFAR in half a minute).
FFCV speeds up ~any existing training code (no training tricks needed) (1/3)
FFCV is easy to use, minimally invasive, fast, and flexible: . We're really excited to both release FFCV today, and start unveiling (soon!) some of the large-scale empirical work it has enabled us to perform on an academic budget. (2/3)
PS A few examples: in 30 min, we can train ResNet-18 to 67% ImageNet acc on *one A100*. In 20 mins, ResNet-50 to 75.6% on a p4d AWS machine (<$5!). CIFAR costs 2 cents/model. My students tell me this is fast ;) [More seriously, we haven't seen anything in PyTorch that compares.]
@aleks_madry
Impressive. Sometimes I should compile a list of libraries and techniques that demonstrate that computers are significantly faster than we think.
@aleks_madry
@aleks_madry
(and team) this is really cool! nice job.
A few points on the image i posted. Excited to allow PL to work with ffcv so you can get both benefits (in addition to things like deepspeed, etc...)