Lior Alexander
@LiorOnAI
Followers
107K
Following
6K
Media
619
Statuses
4K
Helping devs stay up to date with AI. CEO @AlphaSignalAI → 250k users • MLE since 2017
Joined November 2012
Deep research mode goes further. It pulls arXiv papers, summarizes them, compares your method, and generates citation-ready tables, all without leaving Overleaf. It feels like a review committee wired into your document. If this scales, Overleaf stops being just an editor and
1
1
3
How it works: → Highlight a rough paragraph → it launches a critique + rewrite pipeline. → Reviewer, Enhancer, Scoring, and Researcher agents run in parallel. → You get before/after diffs and instant patches. → Multi-agent scheduling handles orchestration. → An MCP
1
0
2
Why this matters → Most LLM tools can’t read fine-grained context. → PaperDebugger adds bidirectional sync → It patches your source with clean diff-based updates. → All of it runs through a Chrome extension and a Kubernetes-native backend.
1
0
0
Writing papers just got 100x easier. This tool is a must for every researcher. A team just released PaperDebugger, an in-editor, multi-agent system that plugs directly into Overleaf. Not an external assistant. Not a sidebar chatbot. It lives inside the editor and works with
1
4
12
If you train large models: You either use SDPA output gating… or you keep debugging runaway activations, broken heads, and unstable runs. Small gate. Massive impact. Congrats to the @Alibaba_Qwen team.
2
2
4
Why It Works: Attention used to behave like a giant unprotected global variable. One bad value? Entire system destabilizes. The new gate acts like: → A non-linear firewall between attention and the next layer → A per-head circuit breaker that stops activation spikes → A
1
1
1
The Solution: The Qwen team made one tiny architectural change: A head-specific, query-dependent sigmoid gate after SDPA. Think of SDPA like a noisy function that sometimes returns normal values… and sometimes returns spikes so huge they crash the entire system. You don’t
1
1
2
The Problem: Softmax attention had two chronic failures: • Massive activations: like a function call that allocates memory until your laptop wheezes. • Attention sink:one token acting like a broken while(true) loop, hijacking the whole head. Every big model hit these
1
1
1
Qwen just won Best Paper Award at NeurIPS. And it wasn’t for a flashy new architecture. It was for fixing a problem Transformers had for years. Here’s what you need to know:
8
11
46
▸ 5-min daily newsletter for developers to keep up with AI:
alphasignal.ai
The Best of Machine Learning. Summarized by AI.
0
0
18
Be Alex Krizhevsky. Born in the Soviet Union. Join Hinton’s lab. Create AlexNet. Train it on GPUs in your bedroom. Breaks every record. Spark the Deep Learning revolution. Get 181,495 citations. Disappear.
94
316
5K
Of all possible targets, they chose Charlie Kirk, a man who, despite holding strong views, was among the few political influencers who always made the effort to engage civilly and in good faith with people who disagreed with him. Perhaps this is why he was feared.
350
4K
40K
Most of your habits aren’t thoughtful decisions, they’re just the easiest ways you’ve found to avoid fear, effort, or change.
3
0
11
Reminder: The future is already here, it's just unevenly distributed.
3
1
11
▸ 5-min daily newsletter for developers to keep up with AI: https://t.co/ZJ2Iz2bdY5 ▸ Source:
github.com
Flexible and powerful framework for managing multiple AI agents and handling complex conversations - awslabs/agent-squad
0
8
47
AWS released an open-source framework that lets you orchestrate multiple AI agents and handle complex conversations. Can be deployed locally on your computer.
25
137
1K
Always start strict, then ease up. Works with employees, relationships, laws, and deals. Doing the opposite rarely works.
5
1
16
I think I just found the best prompt... "Review this code as a senior dev" It’s improving my entire codebase. 10x cleaner code 10x fewer bugs 10x impostor syndrome
6
1
19