
Samaya AI
@samaya_AI
Followers
3K
Following
26
Media
11
Statuses
50
An AI-powered Knowledge Discovery Platform
Joined April 2022
Big news at Samaya: our Founder and CEO, @maithra_raghu, has been named to the @TIME 100 AI 2025 list! ๐ Through her AI research at Google Brain, working with the "Godfather of AI" Geoffrey Hinton, to building @samaya_AI, Maithra has been driven by the vision of Expert AIs that
Surprised and delighted to be on the @TIME 100 AI list! (And in very good company!) Itโs especially meaningful to have this recognition this year, which has been one of incredible growth and milestones for Samaya. We closed our Series A led by NEA, and with other stellar
0
0
10
Live testing GPT-5 on challenging financial questions and still seeing hallucinations and other errors crop up. The automatic routing to thinking is also frustrating at times --- GPT-5 thinks for a long time only to provide a subpar answer. (See below for an example on Figma's
4
3
20
Under the hood: it draws from reasoning frameworks like Causal World Models ๐ ( https://t.co/Mvaj60xqQZ) So it doesnโt just summarize headlines, it simulates ripple effects across markets, sectors, and systems.
samaya.ai
At Samaya, we are pushing the boundaries of reasoning systems by building causal world models that are designed to predict and explain economic outcomes resulting from unseen events.
1
1
3
โ
Analyzes world events using second-order reasoning to surface market-moving insights โ
Eliminates hallucinations using primary financial sources โ
Cites exact sentences in the source docs, no vague summaries โ
Built specifically for financial workflows (not generalist
1
1
1
Most AI โdeep researchโ agents hallucinate ๐ค๐ญ And almost none can connect second-order effects. After 4 months of iteration with users, weโre launching Samayaโs Deep Research Agent, built to solve both problems. (๐ธ example below โฌ๏ธ)
1
3
16
Congratulations to @figma ๐ Use Samaya to dig into Figma's S-1 and research breaking market updates in real time.
0
1
5
๐ ๐๐น๐ผ๐ฏ๐ฎ๐น ๐๐ฟ๐ถ๐๐ถ๐ ๐๐ฒ๐บ๐ฎ๐ป๐ฑ๐ ๐๐น๐ฎ๐ฟ๐ถ๐๐ โ ๐ฆ๐ฎ๐บ๐ฎ๐๐ฎโ๐ ๐๐ ๐๐ด๐ฒ๐ป๐๐ ๐๐ฒ๐น๐ถ๐๐ฒ๐ฟ ๐๐ป๐๐ถ๐ด๐ต๐๐ ๐ถ๐ป ๐ฅ๐ฒ๐ฎ๐น ๐ง๐ถ๐บ๐ฒ. As the Israel-Iran conflict unfolds, markets are shifting fast. Our latest agent-generated report breaks down the potential
0
0
1
This brings us closer to our goal: An assistant that doesnโt just respondโit thinks ahead. By detecting and filling gaps automatically, Samaya delivers answers experts can rely on. Full story โ https://t.co/9IOnCI9mCA
samaya.ai
By automatically spotting and filling information gaps in real time, Samaya's self-correction engine helps users trust their AI research tool for high-stakes decisions.
0
0
1
Impact? ๐ Up to 60% more facts recovered on complex queries ๐ 10.5% boost on our toughest benchmark (Criteria-Eval)
1
0
1
Hereโs how it works: As Samaya generates an answer, a second process checks for gaps. If anythingโs off, it launches targeted background searchesโquietly and in real time. It then uses the additional evidence to improve the answer, without noticeable additional delay.
1
0
1
Even great answers can have subtle gapsโmissing numbers, overlooked facts. In high-stakes workflows, those gaps cost time and trust. At Samaya, we built a real-time self-correction engine that quality-checks our summaries and fills in missing info before the user notices. ๐งต๐
1
2
7
With over 3,000 carefully annotated queries and 8,000+ hours of expert annotations, Criteria-Eval has transformed how we monitor, debug, and continuously improve our platform. Read more details in our blog post:
samaya.ai
A checklist-based evaluation framework that directly aligns with how expert users judge quality of long-form LLM-generated answers.
0
0
1
Why Criteria-Eval - Reproducible: Clear binary criteria enable objective scoring - Flexible: Checklists allow multiple valid answers - Expert-aligned: Directly encodes expectations of experts - End-to-end: Evaluates full pipeline interactions, not isolated components
1
0
1
How Criteria-Eval works 1. Given a query, domain experts (finance) provide an exhaustive list of yes/no criteria each answer must meet, e.g., "Appleโs fiscal 2023 revenue was $383.3 billion". 2. LLM judges score each criterion individually 3. Score = % of criteria satisfied
1
0
0
Evaluating long-form answers to complex technical questions is very challenging. Existing methods fall short in this setting. At Samaya, we built Criteria-Eval, a checklist-based evaluation that aligns with how domain experts judge answers. ๐งต https://t.co/BxidFqA7hc โ๏ธ
1
8
31
๐Big news: Weโre announcing our Series A led by NEA and $43.5m in funding to build the future of expert intelligence in finance. Led by @NEA with support from visionaries like @ericschmidt, @ylecun, David Siegel, and Marty Chavez. Trusted by premier financial institutions like
12
7
70
Everyone's talking about #AgenticAI, but what does it actually look like in finance? @Cortandr from @samaya_AI shares how we build AI agents that domain experts can trust and use every day. Real workflows. Real stakes. ๐ง https://t.co/sV9sfCeN27
#FinTechAI #RAG #AIAgents
0
2
7
๐งย Real-world queries can be long and complex in RAG applications - but todayโs retrieval models often fail to understand the subtleties in them. ๐ย In our latest blog, we introduce โจ Promptriever โจ, our #ICLR2025 work that supercharges retrieval models with instruction
0
1
7