Daniel
@danielmadii
Followers
261
Following
2K
Media
113
Statuses
2K
Still replying to customers manually in 2026? 🤖 I build AI that replies, follows up & closes for you ⚡ Founder @TryHalaAI
Hala AI HQ
Joined July 2013
You’re not losing customers because of competition. You’re losing them because you reply too late. We fixed that. Meet Hala AI. @TryHalaAI
0
0
15
In the next wave, the winners won’t be the smartest models. They’ll be the ones that: 👉 know their limits 👉 expose uncertainty 👉 design around it
0
0
3
Because automation isn’t about replacing humans. It’s about: 👉 knowing when NOT to act That’s where most systems break.
1
0
2
This is exactly how we think about it at Hala AI. Not just: “Respond fast” But: 👉 “Respond with calibrated confidence” 👉 “Know when to escalate”
1
0
1
Imagine: Instead of: “Book this surgery.” You get: “Top diagnosis: 42% Second: 38% Uncertainty: high → Recommend doctor review” Now that’s usable.
1
0
1
Real opportunity? Not “better models.” 👉 Better exposure of model uncertainty That’s where the edge is.
1
0
0
If you’re building in AI and you’re not handling uncertainty, your product is: 👉 dangerous in high-stakes domains 👉 weak in competitive ones
1
0
0
This is where most AI startups fail. They build: “Ask anything → get answer” But ignore: “Should you trust this answer?”
1
0
0
So the real issue isn’t intelligence. It’s the interface. We wrapped probabilistic systems in a clean chat UI and expected truth to come out.
1
0
0
Let that sink in: The model knows when it’s unsure. It just doesn’t tell you properly.
1
0
0
But internal probabilities? 👉 Consistently predictive 👉 Actually useful Switching to them turned garbage signals into real ones.
1
0
0
The paper measured this using AUC-ROC: • 100% = perfect predictor • 50% = random guessing Self-reported confidence? 👉 Barely above random. Useless.
1
0
0
And humans are terrible at expressing uncertainty. We round numbers. We overstate confidence. We avoid saying “I don’t know.” So the model does the same.
1
0
0
When you ask: “How confident are you?” The model doesn’t break character. It doesn’t become a scientist. It becomes… 👉 a human pretending to be confident
1
0
0
Self-reported confidence = performance. It’s not truth. It’s not calibration. It’s roleplay.
1
0
0
Internal probability = math. It reflects how strongly the model believes an answer based on training data patterns. When uncertain → probabilities spread. When certain → probabilities spike. This is usable.
1
0
0
Let’s break the illusion. There are two types of confidence inside an LLM: 1.Internal probability (real signal) 2.Self-reported confidence (what it tells you) These are NOT the same thing.
1
0
0
Every single model was: 👉 Overconfident when wrong 👉 Just as confident when right No signal. No calibration. Just confidence theater.
1
0
0
A paper tested 9 major LLMs on medical licensing exams. ~12,000 questions Multiple countries Multiple languages Results? Up to 89% accuracy → near expert level. Sounds impressive. It is. But here’s the catch 👇
1
0
0
This isn’t a model problem. This is a product design failure. And 99% of AI founders are building on top of it blindly.
1
0
0