
Gargi Ghosh
@gargighosh
Followers
716
Following
78
Media
14
Statuses
68
Researcher at FAIR (Meta AI)
Bellevue, WA
Joined December 2009
We released new research - Byte Latent Transformer(BLT).BLT encodes bytes into dynamic patches using light-weight local models and processes them with a large latent transformer. Think of it as a transformer sandwich!
New from Meta FAIR โ Byte Latent Transformer: Patches Scale Better Than Tokens introduces BLT, which for the first time, matches tokenization-based LLM performance at scale with significant improvements in inference efficiency & robustness. Paper โก๏ธ
11
83
667
RT @liang_weixin: ๐ Excited to share: "๐๐ข๐ฑ๐ญ๐ฎ๐ซ๐-๐จ๐-๐๐ซ๐๐ง๐ฌ๐๐จ๐ซ๐ฆ๐๐ซ๐ฌ (๐๐จ๐)" has been officially accepted to TMLR (March 2025) and the code is nowโฆ.
0
84
0
RT @ylecun: Rob Fergus is the new head of Meta-FAIR!.FAIR is refocusing on Advanced Machine Intelligence: what others would call human-leveโฆ.
0
43
0
RT @fb_engineering: Meta and NVIDIA have teamed up to supercharge vector search on GPUs by integrating NVIDIA cuVS into Faiss v1.10, Metaโsโฆ.
0
128
0
We are releasing Collaborative Reasoner, a self improving social agent that achieves stronger performance through collaboration. This research leverages social skills such as effective communication, providing feedback, having empathy, and theory-of-mind.
๐ Meta FAIR is releasing several new research artifacts on our road to advanced machine intelligence (AMI). These latest advancements are transforming our understanding of perception. 1๏ธโฃ Meta Perception Encoder: A large-scale vision encoder that excels across several image &
1
0
2
Excited to share that we are open sourcing BLT model weights by popular demand(Code was open sourced already): paper:
ai.meta.com
Meta FAIR is releasing several new research artifacts that advance our understanding of perception and support our goal of achieving advanced machine intelligence (AMI).
2
6
26
RT @InceptionAILabs: We are excited to introduce Mercury, the first commercial-grade diffusion large language model (dLLM)! dLLMs push theโฆ.
0
986
0
RT @ClementDelangue: Our science team has started working on fully reproducing and open-sourcing R1 including training data, training scripโฆ.
0
533
0
RT @ykilcher: ๐ฅNew Video๐ฅ.I delve (ha!) into Byte Latent Transformer: Patches Scale Better Than Tokens where the authors do away with tokenโฆ.
0
106
0
RT @nrehiew_: Wrote about some of my favourite papers over the past year or so and some research directions that I am excited about in 2025โฆ.
0
77
0
RT @AIatMeta: New research from Meta FAIR โ Meta Memory Layers at Scale. This work takes memory layers beyond proof-of-concept, proving theโฆ.
0
178
0
Joint wrk with @mingdachen @LukeZettlemoyer @scottyih , Alicia Sun, Yang Li, Karthik Padthe, @RulinShao.
0
0
3