neptune_ai Profile Banner
neptune.ai Profile
neptune.ai

@neptune_ai

Followers
7K
Following
1K
Media
3K
Statuses
13K

Experiment tracker purpose-built for foundation model training. We tweet about #LLM best practices & other cool stuff. Read our blog at https://t.co/4eACuib1QI

Warsaw, Poland
Joined January 2018
Don't wanna be here? Send us removal request.
@neptune_ai
neptune.ai
10 months
We built Neptune Scale to let you monitor such training and debug any issues quickly. Now available in beta: https://t.co/7lWqtFSm7g Coming soon for everyone.
12
1
17
@neptune_ai
neptune.ai
2 days
Track down gradient explosions to the exact layer and training step. — Watch the full product demo: https://t.co/xGvTQ78QwB
1
0
0
@neptune_ai
neptune.ai
4 days
When you’re tracking experiments, benchmarks matter A LOT. With Neptune, you can now overlay baselines, thresholds, or success criteria directly on your charts. No more mental math or custom expression workarounds. Just clear, reliable context right in the plots.
1
0
0
@neptune_ai
neptune.ai
8 days
[Editor's Pick] Learnings From Teams Training Large-Scale Models: Challenges and Solutions For Monitoring at Hyperscale Author: Siddhant Sadangi Reading time: 5 min — Full article: https://t.co/W6VoKxsJmf
0
1
1
@neptune_ai
neptune.ai
11 days
Mixture-of-Experts can save compute or burn it. At Qualcomm, research intern Maciej Pióro ran a grid of #MOE experiments that spiraled into 50,000 A100 hours of re-training. - Why: The load-balancing loss was averaged across layers, so deeper experts ignored it. Tokens
Tweet card summary image
arxiv.org
Mixture of Experts (MoE) architectures have significantly increased computational efficiency in both research and real-world applications of large-scale machine learning models. However, their...
0
0
0
@neptune_ai
neptune.ai
12 days
The Neptune chart legend has come a long way to become a real tool for inspecting and navigating run results: → Started as a simple floating label → Now: resizable, draggable, pinnable → Add params, tags, or any other attributes → Search through it easily → Use it to
1
0
1
@neptune_ai
neptune.ai
13 days
[Editor's Pick] Hyperparameter Optimization For LLMs: Advanced Strategies Authors: Gabriel Souto Augusto Dutra, Kilian Kluge Reading time: 14 min — Full article: https://t.co/S8Im20OCGa
0
0
0
@neptune_ai
neptune.ai
14 days
Spot anomalies at 600K (or any other) steps without zooming in or digging through logs. — Watch the full product demo: https://t.co/xGvTQ78QwB
0
0
0
@neptune_ai
neptune.ai
16 days
Not all layers learn at the same rate. Some layers in deep models "go silent" during training—gradients shrinking to near-zero. Others behave erratically, with values jumping unpredictably. Both cases harm convergence, and neither is obvious from loss curves alone. That’s why,
0
0
0
@neptune_ai
neptune.ai
19 days
Now in Neptune: Shared zoom. Shared cursor. Shared highlights. These should help you stay oriented when you’re debugging and the metrics volume gets real.
0
0
0
@neptune_ai
neptune.ai
19 days
[Editor's Pick] Open LLMs are Necessary For Current Private Adaptations and Outperform Their Closed Alternatives Author: Olatunji Iyiola Emmanuel Reading time: 5 min — Full article: https://t.co/Qez4mcpmWO
0
0
0
@neptune_ai
neptune.ai
21 days
Scatter plots make it easy to spot trends. In this example, a look at BLEU scores and a config parameter shows that the best models used the lowest training floor. — Discover more ways to analyze your experiments with Neptune: https://t.co/xGvTQ78QwB
0
0
0
@neptune_ai
neptune.ai
22 days
Every researcher has a story that starts with: “We thought everything was fine… until it wasn't.” At ICML 2025, we turned those moments into a series, TrainFM, where top researchers share their biggest training failures and how they fixed them. It’s raw. It’s messy. And
0
0
0
@neptune_ai
neptune.ai
23 days
When you’re tracking experiments, benchmarks matter A LOT. With Neptune, you can now overlay baselines, thresholds, or success criteria directly on your charts. No more mental math or custom expression workarounds. Just clear, reliable context right in the plots.
0
0
0
@neptune_ai
neptune.ai
25 days
Monitor → Spot spikes → Debug → Fork In this short demo, you can see how to do the forking with Neptune. Update training config, and continue from a stable checkpoint. The forked run inherits metrics up to the split, giving you a complete view of both runs side by side. —
0
0
0
@neptune_ai
neptune.ai
27 days
[New on our blog] How to Optimize LLM Inference Author: Alek Pikl Reading time: 12 min — Full article: https://t.co/CiPTdvUmmv
0
0
0
@neptune_ai
neptune.ai
28 days
A slow tracker kills research momentum. It made our day to hear that Alex Immer, a Senior Research Scientist at Bioptimus, was “very pleasantly surprised” by Neptune’s performance. Lightning-fast UI, real-time logging, no overhead, even at large scale. — Watch how Bioptimus
1
0
1
@neptune_ai
neptune.ai
29 days
Metric blindness can cost you hundreds of GPU days. That’s the painful lesson Saaketh Narayan from Meta's Llama pre-training team shared with us. During FP8 training, unnoticed gNorm anomalies led to catastrophic loss spikes, wasting ~512 GPU days. The fix? Architectural
0
0
0
@neptune_ai
neptune.ai
1 month
We’ve made a lot of small improvements to @neptune_ai’s chart legend over time. Together, they’ve added up to something that’s quietly powerful. → Floating or attached to the bottom of the chart → Resizable, draggable, pinnable → Search inside the legend → Add custom
1
0
1
@neptune_ai
neptune.ai
1 month
“Usually, the first idea doesn't check out. That's just the nature of #AIresearch.” For Derek Li, Senior Researcher at Noah's Ark Lab, that first idea was training all multitask #reinforcementlearning objectives together. What went wrong: Runs underperformed baselines, with
0
0
0
@neptune_ai
neptune.ai
1 month
Before: Shaded regions in @neptune_ai = min–max range. Now: You decide. Custom Error Bands let you bring your own error metrics like standard error, confidence intervals, or any bounds you care about.
0
0
0