Akshara Prabhakar Profile
Akshara Prabhakar

@aksh_555

Followers
401
Following
1K
Media
7
Statuses
70

applied scientist @SFResearch | prev @princeton_nlp, @surathkal_nitk

bay area
Joined October 2019
Don't wanna be here? Send us removal request.
@aksh_555
Akshara Prabhakar
9 months
🤖 NEW PAPER 🤖. Chain-of-thought reasoning (CoT) can dramatically improve LLM performance. Q: But what *type* of reasoning do LLMs use when performing CoT? Is it genuine reasoning, or is it driven by shallow heuristics like memorization?. A: Both!. 🔗 1/n
Tweet media one
7
48
315
@aksh_555
Akshara Prabhakar
1 month
RT @quantumbytz: Salesforce AI Introduces CRMArena-Pro: The First Multi-Turn and Enterprise-Grade Benchmark for LLM Agents.#AI #MachineLear….
0
2
0
@aksh_555
Akshara Prabhakar
1 month
RT @SFResearch: 🚨 Introducing CRMArena-Pro: The first multi-turn, enterprise-grade benchmark for LLM agents. ✍️Blog: .
0
26
0
@aksh_555
Akshara Prabhakar
2 months
RT @salesforce: .@SFResearch’s new series “AI Research Lab - Explained” just dropped!. First up? See how we fine-tune specialized models to….
0
5
0
@aksh_555
Akshara Prabhakar
2 months
RT @silviocinguetta: Enterprise General Intelligence (EGI) won't require bigger models—it will demand better data! Our recent research demo….
0
4
0
@aksh_555
Akshara Prabhakar
2 months
🚀 Just dropped APIGen-MT-5k — 5K high-quality multi-turn agent interactions, generated with our APIGen-MT framework! Built for training & evaluating AI agents.
@SFResearch
Salesforce AI Research
2 months
Introducing APIGen-MT: Our agentic pipeline for multi-turn synthetic data generation that produces high-quality training data for tuning AI agents! Try our open-sourced dataset today!. 📊 Paper: .🤗 Dataset: We used APIGen-MT to
Tweet media one
1
2
14
@aksh_555
Akshara Prabhakar
4 months
RT @AndrewLampinen: New preprint! In “Naturalistic Computational Cognitive Science: Towards generalizable models and theories that capture….
0
81
0
@aksh_555
Akshara Prabhakar
4 months
It was a great experience interacting with the @buZZrobot community 😃, thanks for the invite @sopharicks!. Talk link:
@buZZrobot
BuzzRobot
5 months
Join us on Feb 20 for a talk on Generalization vs. Memorization in LLMs. @aksh_555 from @SFResearch will dive into how Chain-of-Thought prompting impacts reasoning—does it truly enhance logic, or is it just smart memorization?.
0
0
11
@aksh_555
Akshara Prabhakar
7 months
RT @SFResearch: 🤖 Fresh from #NeurIPS2024: Our AI research scientist Akshara Prabhakar @aksh_555 discusses our demo of xLAM's specialized a….
0
4
0
@aksh_555
Akshara Prabhakar
7 months
RT @LiuZuxin: 🚀 Introducing our #NeurIPS'24 (D&B track) paper, APIGen - an Automated PIpeline for Generating high-quality agentic data. Wh….
0
7
0
@aksh_555
Akshara Prabhakar
7 months
RT @SFResearch: 🇨🇦🇨🇦🇨🇦 Welcome to Vancouver! 🇨🇦🇨🇦🇨🇦.13 Paper links below! 👇. The @Salesforce AI Research team brought a baker's dozen AI Re….
0
7
0
@aksh_555
Akshara Prabhakar
7 months
RT @taoyds: Text-to-SQL has been my passion since Yale Spider 1.0! But as LLMs master it, real-world complexity demands more. 🚀After a yea….
0
56
0
@aksh_555
Akshara Prabhakar
7 months
RT @iscreamnearby: It actually reminds me of the multi-source domain adaptation work - where knowing domain index during training makes the….
0
1
0
@aksh_555
Akshara Prabhakar
7 months
RT @murefil: Super enjoyable read: promising results that model mixing via a small, learnable router on top of independently trained "skill….
0
3
0
@aksh_555
Akshara Prabhakar
7 months
RT @LiuZuxin: Super interesting work & definitely check it out if you are attending NeurIPS! It reminds me the paper we published at ICLR t….
0
7
0
@aksh_555
Akshara Prabhakar
7 months
RT @aksh_555: Have a task that can be decomposed into two tasks requiring different skills? BUT.- it is difficult to generate expert-curate….
0
11
0
@aksh_555
Akshara Prabhakar
7 months
Work done with the amazing team: Yuanzhi Li, @karthik_r_n, @ShamKakade6, and guided by Samy Jelassi and @EranMalach. To appear in COLING 2025 Industry track!. 6/6.
0
0
1
@aksh_555
Akshara Prabhakar
7 months
🤔 Limitations.- Scaling to more than two tasks is still a challenge.- Initial experiments show scope for further improvements!. 5/n.
1
0
1
@aksh_555
Akshara Prabhakar
7 months
🌟 We explore 4 practical settings.- solving hard math problems with code.- domain specific question-answering bots.- comprehension on technical documents.- robustness to prompt formats. 4/n
Tweet media one
1
0
1
@aksh_555
Akshara Prabhakar
7 months
🔄 The traditional approach? Training over a mixture of the two datasets. However, concatenating LoRAs performs better than this and other merging methods (TIES, DARE, LoRAHub) 🎉. 3/n
Tweet media one
1
0
0
@aksh_555
Akshara Prabhakar
7 months
🤖 Imagine a question-answering bot for a medical textbook without RAG. You can train a model for next-token prediction on the textbook, and use an instruction-following model (trained on Alpaca for eg.) and compose them!. 2/n.
1
0
0