Gargi Ghosh Profile
Gargi Ghosh

@gargighosh

Followers
716
Following
78
Media
14
Statuses
68

Researcher at FAIR (Meta AI)

Bellevue, WA
Joined December 2009
Don't wanna be here? Send us removal request.
@gargighosh
Gargi Ghosh
7 months
We released new research - Byte Latent Transformer(BLT).BLT encodes bytes into dynamic patches using light-weight local models and processes them with a large latent transformer. Think of it as a transformer sandwich!
Tweet media one
@AIatMeta
AI at Meta
7 months
New from Meta FAIR โ€” Byte Latent Transformer: Patches Scale Better Than Tokens introduces BLT, which for the first time, matches tokenization-based LLM performance at scale with significant improvements in inference efficiency & robustness. Paper โžก๏ธ
Tweet media one
11
83
667
@gargighosh
Gargi Ghosh
6 hours
0
0
2
@gargighosh
Gargi Ghosh
7 hours
Outstanding paper award! @aclmeeting - BLT:
Tweet media one
5
5
60
@gargighosh
Gargi Ghosh
3 months
RT @liang_weixin: ๐ŸŽ‰ Excited to share: "๐Œ๐ข๐ฑ๐ญ๐ฎ๐ซ๐ž-๐จ๐Ÿ-๐“๐ซ๐š๐ง๐ฌ๐Ÿ๐จ๐ซ๐ฆ๐ž๐ซ๐ฌ (๐Œ๐จ๐“)" has been officially accepted to TMLR (March 2025) and the code is nowโ€ฆ.
0
84
0
@gargighosh
Gargi Ghosh
3 months
RT @ylecun: Rob Fergus is the new head of Meta-FAIR!.FAIR is refocusing on Advanced Machine Intelligence: what others would call human-leveโ€ฆ.
0
43
0
@gargighosh
Gargi Ghosh
3 months
RT @fb_engineering: Meta and NVIDIA have teamed up to supercharge vector search on GPUs by integrating NVIDIA cuVS into Faiss v1.10, Metaโ€™sโ€ฆ.
0
128
0
@gargighosh
Gargi Ghosh
3 months
Thanks @AnsongNi @real_asli , Ruta.
0
0
2
@gargighosh
Gargi Ghosh
3 months
We are releasing Collaborative Reasoner, a self improving social agent that achieves stronger performance through collaboration. This research leverages social skills such as effective communication, providing feedback, having empathy, and theory-of-mind.
@AIatMeta
AI at Meta
3 months
๐Ÿš€ Meta FAIR is releasing several new research artifacts on our road to advanced machine intelligence (AMI). These latest advancements are transforming our understanding of perception. 1๏ธโƒฃ Meta Perception Encoder: A large-scale vision encoder that excels across several image &
1
0
2
@gargighosh
Gargi Ghosh
3 months
This enables you all to explore efficient post training strategies with byte level models and much more. Stay tuned for next set of research from BLT team.
0
0
0
@gargighosh
Gargi Ghosh
3 months
0
0
1
@gargighosh
Gargi Ghosh
3 months
Excited to share that we are open sourcing BLT model weights by popular demand(Code was open sourced already): paper:
Tweet card summary image
ai.meta.com
Meta FAIR is releasing several new research artifacts that advance our understanding of perception and support our goal of achieving advanced machine intelligence (AMI).
2
6
26
@gargighosh
Gargi Ghosh
5 months
RT @InceptionAILabs: We are excited to introduce Mercury, the first commercial-grade diffusion large language model (dLLM)! dLLMs push theโ€ฆ.
0
986
0
@gargighosh
Gargi Ghosh
6 months
RT @ClementDelangue: Our science team has started working on fully reproducing and open-sourcing R1 including training data, training scripโ€ฆ.
0
533
0
@gargighosh
Gargi Ghosh
7 months
RT @elonmusk: A friend in LA just took this video
0
82K
0
@gargighosh
Gargi Ghosh
7 months
RT @ykilcher: ๐Ÿ”ฅNew Video๐Ÿ”ฅ.I delve (ha!) into Byte Latent Transformer: Patches Scale Better Than Tokens where the authors do away with tokenโ€ฆ.
0
106
0
@gargighosh
Gargi Ghosh
7 months
RT @nrehiew_: Wrote about some of my favourite papers over the past year or so and some research directions that I am excited about in 2025โ€ฆ.
0
77
0
@gargighosh
Gargi Ghosh
7 months
RT @AIatMeta: New research from Meta FAIR โ€” Meta Memory Layers at Scale. This work takes memory layers beyond proof-of-concept, proving theโ€ฆ.
0
178
0
@gargighosh
Gargi Ghosh
7 months
Joint wrk with @mingdachen @LukeZettlemoyer @scottyih , Alicia Sun, Yang Li, Karthik Padthe, @RulinShao.
0
0
3
@gargighosh
Gargi Ghosh
7 months
Ewe outperforms strong baselines on four fact-seeking long-form generation datasets, increasing the factuality metric, VeriScore, by 2 to 10 points absolute without sacrificing the helpfulness of the responses. Ewe can be easily adapted to various large language models.
Tweet media one
1
0
3
@gargighosh
Gargi Ghosh
7 months
The working memory consists of k memory units, each unit stores the representations of each feedback message of M tokens.The memory is refreshed based on online fact-checking and retrieval feedback, allowing Ewe to rectify false claims.
Tweet media one
1
0
1
@gargighosh
Gargi Ghosh
7 months
Last one of the year - EWE: Ewe (Explicit Working Memory), enhances factuality in long-form text generation by integrating a working memory that receives real-time feedback from external resources.
Tweet media one
2
22
103