Pritamstudyai Profile Banner
Pritam Profile
Pritam

@Pritamstudyai

Followers
255
Following
12K
Media
97
Statuses
1K

kernel
Joined October 2024
Don't wanna be here? Send us removal request.
@Pritamstudyai
Pritam
14 hours
add turbodiffusion
@fujikanaeda
Eric W. Tramel @ Home
1 day
what a week for flopheads 🚀: - BLASST: 1.62x prefill & 1.48x decode speed that folds naturally into FA - SonicMoE: 1.86x faster MoE - Nemotron 3: Mamba2 SSM hybrid transformers going almost 2x on tok/sec over non-hybrid. - Nanomoe: RDEP & nvfp4 on blackwell
1
0
1
@Pritamstudyai
Pritam
1 day
Another resource stacked up
@MatejKripner
Matěj Kripner
1 day
I'm releasing nanoproof - a minimal open-source implementation of AlphaProof (@TZahavy). It uses nanochat (@karpathy) for pretrain+midtrain+sft, and the official AlphaProof pseudocode for MCTS+the RL loop. First version has 32.8% on minif2f with ~0.001% of the compute. (link to
0
0
1
@Jeyffre
Jeffrey Scholz
1 day
If you want to do ZK in 2026, here are the courses I'd take: 1 - A linear algebra course. This is the foundation of almost all non-trivial fields of programming. 2 - A discrete math course (especially one that includes elementary number theory) 3 - A proofs course (as a
@summit_defi
Defi Security Summit
12 days
How do you actually audit ZK circuits? At DSS 101, @Jeyffre, founder of @rareskills_io, breaks down the core building blocks of ZK auditing: ✅ What ZK proofs really verify ✅ How constraints encode real-world logic ✅ Common pitfalls (missing constraints, compilers, memory
28
78
532
@_xjdr
xjdr
2 days
today we’re open-sourcing nmoe: https://t.co/iq6HliUqpq i started this because training deepseek-shaped ultra-sparse moes should be straightforward at research scale, but in practice it’s painful: - expert flops get stranded (router shatters your batch → tiny per-expert
Tweet card summary image
github.com
MoE training for Me and You and maybe other people - GitHub - Noumena-Network/nmoe: MoE training for Me and You and maybe other people
24
71
584
@Pritamstudyai
Pritam
2 days
if you wanna learn inference optimisation:
@lmsysorg
LMSYS Org
2 days
How long have you been "planning to understand" how modern LLM inference works? We just gave you a readable version of SGLang you can finish over the weekend. Introducing mini-SGLang ⚡ We distilled SGLang from 300K into 5,000 lines. Kept the core design, cut the complexity.
0
0
3
@Pritamstudyai
Pritam
2 days
they did it🥳, congrats team
@Winterice10
Jintao Zhang
4 days
TurboDiffusion: 100–205× faster video generation on a single RTX 5090 🚀 Only takes 1.8s to generate a high-quality 5-second video. The key to both high speed and high quality? 😍SageAttention + Sparse-Linear Attention (SLA) + rCM Github: https://t.co/ybbNBjgHFP Technical
0
0
3
@mkurman88
Mariusz Kurman
3 days
This is it: A single-person project; Trained from scratch on TPUs (@Google TRC) on the one and only SYTNH dataset by @pleiasfr; Neuroblast-v3 architecture running on my local vLLM instance Just wow (I'm amazed by how good it looks; speed is incredible, here slightly slowed by
31
16
270
@jiqizhixin
机器之心 JIQIZHIXIN
3 days
What if you could speed up a cutting-edge AI image generator without retraining it from scratch? Apple presents a new method that unifies two major AI image models (diffusion & flow matching) under a simple statistical rule. This lets them directly apply a powerful
3
24
168
@mjsMLP
Mahmoud Soliman
4 days
My team is looking for interns! Please reach out with your resume if you want to contribute to the state of art deep learning frameworks and compilers, powering frontier models on pre/post training and inference!
58
41
621
@Pritamstudyai
Pritam
3 days
so if that's the case one research can alone train and release sota models
@natolambert
Nathan Lambert
3 days
I estimate that each frontier lab could have more research compute than all academic institutions in the US combined. Even if it’s not quite true yet, it’s telling.
0
0
0
@0xDevShah
Dev Shah
4 days
This is the DeepSeek moment for Voice AI. Today we’re releasing Chatterbox Turbo — our state-of-the-art MIT licensed voice model that beats ElevenLabs Turbo and Cartesia Sonic 3! We’re finally removing the trade-offs that have held voice AI back. Fast models sound robotic.
161
381
4K
@HuggingPapers
DailyPapers
5 days
Microsoft just dropped VITRA-VLA, a new Vision-Language-Action model for robotics on Hugging Face. It learns dexterous manipulation from over 1 million real-life human hand activity videos.
4
52
304
@Pritamstudyai
Pritam
5 days
Insecurity is profit. Human insecurity is the prime psychological driver of consumerism and a systemic driver of modern capitalism. [in current economics and attention economics]
0
0
1
@Pritamstudyai
Pritam
5 days
Are my thoughts mine? or are they simply influenced by everyone around me?
1
0
2
@Pritamstudyai
Pritam
5 days
if you wanna learn Pytorch in 2025 , Andrew ng's @DeepLearningAI dropped a brand new course.
0
0
2
@Pritamstudyai
Pritam
6 days
Are you @karpathy willing a drop a video on Nanochat?
0
0
3
@Pritamstudyai
Pritam
8 days
this is really cool way to learn about LLMs and Rust.
@brk0v
Viacheslav Biriukov
8 days
🦀 Building an LLM From Scratch in Rust Want to really understand how LLMs work under the hood? Start at the foundation: a complete GPT-2 style transformer. https://t.co/qpsi9QN6RV #rust #rustlang
0
0
6
@AnthropicAI
Anthropic
8 days
We’re opening applications for the next two rounds of the Anthropic Fellows Program, beginning in May and July 2026. We provide funding, compute, and direct mentorship to researchers and engineers to work on real safety and security projects for four months.
90
283
3K
@CSProfKGD
Kosta Derpanis
10 days
#KostasKeynoteLessons: Curious about the "Keynote magic" behind my slides? I’m releasing the full Keynote source file for my recent Gaussian Splatting lecture, all 10 GIGAbytes of it! Grab the files in the thread and feel free to remix.
16
43
387
@Pritamstudyai
Pritam
9 days
thanks @thinkymachines , ig 1st one from india! it's time to cook.
44
7
498