Kfir Goldberg
@kfir99
Followers
163
Following
2K
Media
10
Statuses
245
Happy to share something cool we made at @bria_ai_ - Zero-to-Wan: a minimalistic, easy-to-modify, finetuning repo for @Alibaba_Wan 1.3B video generation model - their most accessible one ๐น Check out the code here: https://t.co/42y9i3mpnp
3
5
31
๐ฌ Introducing ๐๐ซ๐ข๐ ๐๐ข๐๐๐จ ๐๐ซ๐๐ฌ๐๐ซ Erase objects in 3 ways: โ๏ธ ๐ป๐๐๐-๐๐๐๐๐
(e.g. โremove the red carโ) ๐ฏ ๐ท๐๐๐๐ & ๐๐๐๐๐ auto-detection ๐ญ ๐ช๐๐๐๐๐ ๐๐๐๐๐ for precise control Try it in the on @bria_ai_ lab: https://t.co/dsc8olaFWg Or on
0
3
7
๐งต New paper: "Simple Context Compression" - we show that mean-pooling beats the widely-used compression-tokens method for compressing contexts in LLMs, while being simpler and more efficient! with @yoavartzi (1/7)
3
13
41
If you missing publishing your research and contributing to the open community - my research group is hiring and provide competitive offering See our latest paper release in the comments Additional models and papers are already in the oven - great time to join us ๐
1
2
12
๐ฅFIBO News: Paper & Training Code are out!๐ฅ Making the model fully open for the community to explore, reproduce and build upon is one of our core values. Eager to hear your thoughts ๐ Paper: https://t.co/KwuIKdbV8K ๐ Repo:
github.com
FIBO is a SOTA, first open-source, JSON-native text-to-image model built for controllable, predictable, and legally safe image generation. - Bria-AI/FIBO
3
12
49
FIBO isnโt just another text-to-image model. It achieves SOTA benchmarks in aesthetics + alignment, is fast, and cost-effective. It generates structured JSON alongside each image, so the results are controllable and easy to integrate into agentic and enterprise workflows.
๐ New open-source model drop: FIBO by @bria_ai_ FIBO transforms intent into structured JSON prompts for high-quality, repeatable image generation.
1
9
96
๐ New open-source model drop: FIBO by @bria_ai_ FIBO transforms intent into structured JSON prompts for high-quality, repeatable image generation.
4
5
48
@bria_ai_ just dropped FIBO ๐ฅ โจ 8B text-to-image model โจ first open model to use json prompts natively! I tested it with one of the most complex prompts of all time and it did sooo well ๐คฏ >prompt: a man, a woman and a dog are standing against a backdrop, the backdrop is
5
9
47
Welcome ๐
๐๐๐!! The worldโs first JSON-native, Open Source T2I model! Winning both aesthetics and alignment in the benchmarks by being built ground up for ๐๐๐๐๐๐๐, ๐๐๐๐๐๐๐๐๐, and ๐๐๐๐๐๐
๐๐๐๐๐๐๐๐๐ https://t.co/LMewUlpjzg The is big!
6
19
85
Fibo is a very cool model - you know exactly what it "sees" through a very detailed json prompt. You can easily refine and control your generations in an easy iterative flow ๐จ
Generating an image from 1,000 words. Very excited to release Fibo ๐, the first ever open-source model trained exclusively on long, structured captions. Fibo sets a new standard for controllability and disentanglement in image generation [1/6] ๐งต
0
0
7
๐จ FIBO is now live on fal! ๐ง JSON-native text-to-image model for structured, controllable generation ๐๏ธ Edit lighting, camera, and color without breaking the scene ๐งฉ Trained exclusively on licensed datasets
5
13
113
Incredible demo by @MokadyRon, announcing @bria_ai_'s new model, Fibo! The most controllable image model we've ever seen, and open-source ๐
4
12
88
T2I models excel at realism, but true creativity means generating what doesn't exist yet. How do you prompt for something you can't describe? ๐จ We introduce VLM-Guided Adaptive Negative Prompting: inference time method that promotes creative image generation. 1/6
4
44
165
three years ago, DiT replaced the legacy unet with a transformer-based denoising backbone. we knew the bulky VAEs would be the next to go -- we just waited until we could do it right. today, we introduce Representation Autoencoders (RAE). >> Retire VAEs. Use RAEs. ๐(1/n)
57
332
2K
I finally solved player recognition - player and number detection with RF-DETR - player tracking with SAM2 - team clustering with SigLIP, UMAP and KMeans - number recognition with SmolVLM2 stay tuned for YT tutorial: https://t.co/6Iq051w9Tn โ full breakdown + code
129
406
4K
๐ก๐ฒ๐ฒ๐ฑ ๐๐ผ ๐๐ฝ๐๐ฐ๐ฎ๐น๐ฒ ๐ฎ ๐๐ถ๐ฑ๐ฒ๐ผ ๐๐ฝ ๐๐ผ ๐ด๐? Bria Upscaler on fal = sharper, cleaner, flawless.
11
21
175
1/ What if you could animate a face directly from text? ๐ญ Meet Express4D - a dataset of expressive 4D facial motions captured from natural language prompts, designed for generative models and animation pipelines. ๐ https://t.co/okK0JgZMhV ๐น๐
2
17
21
From my meetup talk a month ago: The biggest problem in text-to-image isnโt scaling or architectures. Itโs evaluation. Today, we evaluate with โarenasโ: user preference on short prompts. But humans struggle to follow long instructions, so to win you just optimize for average
4
3
34
Classic background removal uses binary masks, and thatโs the problem. Real-world edges aren't just black or white. Soft alpha matting assigns per-pixel transparency and changes everything. Read my latest blog to see how we built it @bria_ai_:
blog.bria.ai
Soft alpha matting vs binary masks: How 8-bit transparency revolutionizes background removal for glass, hair, and product photography
0
2
10
Very excited about our ICCV work ๐โโ๏ธ My favorite part: instead of using inversion, we just train the model to perform the identity function and used the emerged internal features at inference. Check the thread below for more details ๐
Very excited to introduce our ๐บ#ICCV2025 highlight๐บ paper โBlendedPCโ! This work sets a new standard in localised semantic editing of point clouds, using purely text as guidance. Project page: https://t.co/7waIFnDhVM Wanna hear more? ๐๐งต
3
2
21