Sahil1V Profile Banner
Sahil Verma Profile
Sahil Verma

@Sahil1V

Followers
586
Following
6K
Media
17
Statuses
583

PhD student @uwcse. Robustness and Interpretability. Currently at @MSFTResearch. Former intern at @amazon, @itsArthurAI. Undergrad @IITKanpur

Seattle, WA
Joined September 2013
Don't wanna be here? Send us removal request.
@Sahil1V
Sahil Verma
11 days
Glad to share that our paper was accepted the main EMNLP 2025 Conference!.
@Sahil1V
Sahil Verma
3 months
๐Ÿšจ New Paper! ๐Ÿšจ.Guard models slow, language-specific, and modality-limited?. Meet OmniGuard that detects harmful prompts across multiple languages & modalities all using one approach with SOTA performance in all 3 modalities!! while being 120X faster ๐Ÿš€.
Tweet media one
4
7
69
@Sahil1V
Sahil Verma
1 month
RT @soumyesinghal: Llama Nemotron model just got Super-Charged โšก๏ธWe released Llama-Nemotron-Super-v1.5 today! The best open model that canโ€ฆ.
Tweet card summary image
huggingface.co
0
7
0
@grok
Grok
4 days
Join millions who have switched to Grok.
194
397
3K
@Sahil1V
Sahil Verma
1 month
RT @_shruti_joshi_: I will be at the Actionable Interpretability Workshop (@ActInterp, #ICML) presenting *SSAEs* in the East Ballroom A froโ€ฆ.
0
6
0
@Sahil1V
Sahil Verma
1 month
RT @zvez11: Transformers struggle with length generalization and long context. What can we do about it?. Our new #TMLR paper with @rolandalโ€ฆ.
0
7
0
@Sahil1V
Sahil Verma
2 months
RT @zvez11: Are you compositionally curious ๐Ÿค“. Want to know how to learn embeddings using๐ŸŒฒ?. In our new #ICML2025 paper, we present Banyan:โ€ฆ.
0
13
0
@Sahil1V
Sahil Verma
2 months
RT @fengyao1909: ๐Ÿ˜ตโ€๐Ÿ’ซ Struggling with ๐Ÿ๐ข๐ง๐ž-๐ญ๐ฎ๐ง๐ข๐ง๐  ๐Œ๐จ๐„?. Meet ๐ƒ๐ž๐ง๐ฌ๐ž๐Œ๐ข๐ฑ๐ž๐ซ โ€” an MoE post-training method that offers more ๐ฉ๐ซ๐ž๐œ๐ข๐ฌ๐ž ๐ซ๐จ๐ฎ๐ญ๐ž๐ซ ๐ ๐ซ๐š๐๐ข๐žโ€ฆ.
0
60
0
@Sahil1V
Sahil Verma
2 months
RT @avibose22: ๐Ÿšจ Code is live! Check out LoRe โ€“ a modular, lightweight codebase for personalized reward modeling from user preferences. ๐Ÿ“ฆ Fโ€ฆ.
Tweet card summary image
github.com
Code to reproduce results of our experiments using LoRe - facebookresearch/LoRe
0
6
0
@Sahil1V
Sahil Verma
3 months
Using retrieval? --> check out this work by my awesome collaborator on how to increase diversity when retrieving!.
@arnaved
Arnav Das
3 months
1/8 ๐Ÿš€ How can retrieval augmentation be made both relevant and non-redundant for few-shot adaptation? I'm excited to introduce COBRA. Catch our poster at #CVPR25 (ExHall D, Poster #450) on Sat 14 Jun, 5โ€“7 p.m. CDT:
Tweet media one
0
2
7
@Sahil1V
Sahil Verma
3 months
RT @fengyao1909: ๐Ÿ”ฅ "Vibe coding" is everywhereโ€”but is it really care-free?. We introduce ๐‘๐ž๐š๐‹, an RL framework that trains LLMs with automaโ€ฆ.
0
41
0
@Sahil1V
Sahil Verma
3 months
Joint work with awesome collaborators:.@keeghin, @jbilmes, @_siskac, @LukeZettlemoyer,.@hila_gonen, and @csinva. Also huge thanks to @arnaved, @butcher_jasper, @BhattGantavya, @zvez11, @faeze_brh, @MakeshNarsimhan, @soumyesinghal, @_shruti_joshi_ for helpful discussions.
0
0
7
@Sahil1V
Sahil Verma
3 months
OmniGuard is also the fastest and most accurate at rapid adaptation to new settings using few-shot examples (one of the purported benefits of separate Guard models built using LLMs).
Tweet media one
1
0
3
@Sahil1V
Sahil Verma
3 months
Using the internal representations for safety classification bypasses the need for a separate Guard model while making OmniGuard 120X faster than the fastest baseline Guard model.
Tweet media one
1
0
3
@Sahil1V
Sahil Verma
3 months
How does OmniGuard detect harmful prompts across languagesโ€”even ciphersโ€”and modalities?. 1๏ธโƒฃ Identifies universal internal representations from models (LLMs/MLLMs). 2๏ธโƒฃ Builds powerful classifiers using these shared representations. A single unified guard ! ๐ŸŒ๐Ÿ”’
Tweet media one
1
0
3
@Sahil1V
Sahil Verma
3 months
๐Ÿšจ New Paper! ๐Ÿšจ.Guard models slow, language-specific, and modality-limited?. Meet OmniGuard that detects harmful prompts across multiple languages & modalities all using one approach with SOTA performance in all 3 modalities!! while being 120X faster ๐Ÿš€.
Tweet media one
1
39
82
@Sahil1V
Sahil Verma
3 months
RT @jinaycodes: Introducing soarXiv โœˆ๏ธ, the most beautiful way to explore human knowledge. Take any paper's URL and replace arxiv with soarโ€ฆ.
0
1K
0
@Sahil1V
Sahil Verma
5 months
RT @soumyesinghal: โšกโšก Llama-Nemotron-Ultra-253B just dropped: our most advanced open reasoning model.๐Ÿงต๐Ÿ‘‡
Tweet media one
0
13
0
@Sahil1V
Sahil Verma
5 months
RT @soumyesinghal: โšก๏ธ Llama-Nemotron-Ultra is fully open โ€” weights and post-training data. Achieves 76.0% on GPQA via FP8 RL training withโ€ฆ.
0
4
0
@Sahil1V
Sahil Verma
6 months
RT @soumyesinghal: ๐Ÿš€ Meet Llama-Nemotron-Super-49B, our teamโ€™s new reasoning model released at #GTC25! Proud to have contributed ๐Ÿง . Optimizโ€ฆ.
Tweet card summary image
huggingface.co
0
4
0
@Sahil1V
Sahil Verma
6 months
RT @snehaark: Come for the ridiculous 30 column spreadsheet created at @sweetgreen, stay for a critical discussion of how you *actually* scโ€ฆ.
0
3
0
@Sahil1V
Sahil Verma
6 months
RT @_shruti_joshi_: 1\ Hi, can I get an unsupervised sparse autoencoder for steering, please? I only have unlabeled data varying across mulโ€ฆ.
0
10
0