Pruna AI
@PrunaAI
Followers
1K
Following
277
Media
170
Statuses
307
High-performance AI models that combine speed, quality, and specialization for key use cases. ๐ฑ Make AI accessible and sustainable for everyone!
Munich & Paris
Joined April 2023
See you next Tuesday at @NeurIPSConf. Weโll be at ๐ธ๐ถ๐ผ๐๐ธ ๐ณ! Make sure to stop by, discover our latest research, and pick up a sticker!
12
3
23
To celebrate ๐๐น๐๐
๐ฎ ๐ฟ๐๐ป๐ป๐ถ๐ป๐ด ๐ฎ๐ ๐ณ๐๐น๐น ๐๐ฝ๐ฒ๐ฒ๐ฑ after our collaboration with @bfl_ml and @replicate, weโre launching the Flux 2 x Pruna Contest on X! ๐จ Create an image using any Flux 2 model on BFL or Replicate ๐ Prizes: ๐ญ๐ฑ๐ฌโฌ ๐ฝ๐ฒ๐ฟ ๐ฐ๐ฎ๐๐ฒ๐ด๐ผ๐ฟ๐ +
37
34
113
Z-model from @AlibabaGroup is now running is less than 2 seconds on @replicate optimized by @PrunaAI ! We've accelerated inference so you can run Z-model faster, cheaper, and greener. And this is just the beginning, even more speed-ups are coming soon! ๐ Try it on Replicate:
9
14
106
Pruna is built โฆ โข with Pretzels & Croissants ๐ฅจ ๐ฅ โข in Paris & Munich & in the metaverse ๐ฉ๐ชย ๐ซ๐ทย ๐พ So, we moved our offices online to @gather_town and things have been great! Cool things weโve seen happen: โข more engaging and fund global meetings โข great excuses for
4
0
14
Flux 2 from @bfl_ml is out at maximum inference speed on @replicate with @PrunaAI optimization! - The novelty: Stunning 4MP quality with new state-of-the-art realism and high-precision control with up to 8 reference images. This is a huge milestone for Black Forest Labs! - The
4
12
37
โฐ ๐ฆ๐๐ฎ๐ฟ๐๐ถ๐ป๐ด ๐ฎ๐ ๐ฒ:๐ฌ๐ฌ ๐ฝ๐บ ๐๐๐ง! Learn from @chrismdesa from Cornell University how simple linear error feedback unifies two key challenges in modern ML: efficient training and efficient inference. In todayโs webinar, Chris will present: โข ๐๐ฟ๐ฎ๐
0
0
4
๐ Excited to see you at our ๐๐ ๐๐ณ๐ณ๐ถ๐ฐ๐ถ๐ฒ๐ป๐ฐ๐ ๐ ๐ฒ๐ฒ๐๐๐ฝ ๐ถ๐ป ๐ ๐๐ป๐ถ๐ฐ๐ต on ๐๐ฒ๐ฐ๐ฒ๐บ๐ฏ๐ฒ๐ฟ ๐ญ๐ฌ๐๐ต ๐ฎ๐ ๐ฒ:๐ฏ๐ฌ ๐ฃ๐ ๐๐๐ฆ๐ง! ๐
Agenda โข 6:30 PM: Doors open โข 7:00โ8:00 PM: Talks featuring, โข 8:00โ9:00 PM: Hang out with likeminded developers & Food ๐ ๐
0
1
2
๐๏ธโโ๏ธย A proper warmup is important, especially for compiled models, but Pruna can now do without it. torch.compile can accelerate your model even with LoRA. But the warmup and hot swapping can be rather slow. ๐ย Our improvements: ๐ง Portable Compilation: Save compiled artefacts
0
0
3
We kicked off the @dotaiconf with a burst of creativity, designing new ๐ฃ๐ฟ๐๐ป๐ฎ ๐๐ ๐๐๐ถ๐ฐ๐ธ๐ฒ๐ฟ๐. Developers, researchers, and ML engineers teamed up to create merch design that captures what we care about: ๐ฒ๐ณ๐ณ๐ถ๐ฐ๐ถ๐ฒ๐ป๐, ๐๐ฐ๐ฎ๐น๐ฎ๐ฏ๐น๐ฒ, ๐ฎ๐ป๐ฑ ๐ณ๐๐ป ๐๐. Some
3
0
3
๐ Learn from @chrismdesa how simple linear error feedback unifies two key challenges in modern ML: efficient training and efficient inference. In this webinar, on Nov. 24th, Chris will present: โข ๐๐ฟ๐ฎ๐ (๐๐ฟ๐ฎ๐ฑ๐ถ๐ฒ๐ป๐ ๐๐ฎ๐น๐ฎ๐ป๐ฐ๐ถ๐ป๐ด) โ a new way to select training
0
1
1
โก๏ธ AI is slow, expensiveโฆ and bad for the environment! At @dotaiconf, I shared how energy utilization interact with AI models โ and how we can make it much more efficient. ๐กWhat drives AI progress? Energy utilization drives AI breakthroughs like transformer architectures,
0
2
5
2 seconds for 4MP is wild. Great work from the @PrunaAI team!
While everyone is waiting for Flux 2 from @bfl_ml, we built an ultra-fast 4 MP model based on Flux-Schnell. It only takes 2 seconds per generation. You can try it now on @replicate.
2
3
10
While everyone is waiting for Flux 2 from @bfl_ml, we built an ultra-fast 4 MP model based on Flux-Schnell. It only takes 2 seconds per generation. You can try it now on @replicate.
3
4
24
๐ Learn from @chrismdesa how simple linear error feedback unifies two key challenges in modern ML: efficient training and efficient inference. In this webinar, on Nov. 24th, Chris will present: โข ๐๐ฟ๐ฎ๐ (๐๐ฟ๐ฎ๐ฑ๐ถ๐ฒ๐ป๐ ๐๐ฎ๐น๐ฎ๐ป๐ฐ๐ถ๐ป๐ด) โ a new way to select training
0
1
3
Do you want to know what we can do for models on your costly H100s?! From March to May, we did some pretty cool things on @replicate โข FLUX runs in less than a second (0.5s) โข Hidream models series run 1.3x to 2.5x faster โข Wan 2.2 runs on 8 GPUs instead of 1, making it
0
1
18
Huge thank you to everyone who joined us yesterday to create a new merch design for @PrunaAI ! The design will be launched soon, stay tuned ๐ If youโre at the @dotaiconf today, come say hi. And donโt miss @Bertrand_Charp talk at 2:50 PM, see you there! ๐
0
3
2
We want to congratulate @bria_ai_ on releasing FIBO, their frontier image generation model, which was trained on structured JSON for precise, controllable generation. Some cool things - x5 faster and powered by our optimisation expertise - JSON prompts? Yes, it gives you more
2
3
12
๐ย Nothing scarier than slow and costly models. Joking but not joking. Happy Halloween! Pruna is ready for it. Boooo. Spooky!
0
1
6
Get your laptops ready and put on your optimisation gloves for a hands-on session in Paris on Nov 5 (18:30โ21:00) with @dotaiconf! We're running a live session on model optimisation, including quantisation, compilation, LoRA tricks, and more. Get ready to get hands-on! ๐ฉโ๐ป Ideal
0
2
5