Shivam Singh
@er_shivamsingh0
Followers
772
Following
6K
Media
251
Statuses
2K
Engineer| koinophobic | 22 | AI | GPU POOR | Building Neo clouds https://t.co/qGAknj71kz
localhost:8080/
Joined March 2023
I spent last week at Neurips (the big ML/AI conference) and had hundreds of amazing technical conversations with old friends and new. Beyond the technical, I'm deeply grateful for the love from all the folks who stopped me in the hallways and in the street to say kind words and
140
100
2K
Andrej Karpathy, Eureka Labs founder & former Director of AI at Tesla, breaks down how LLMs like ChatGPT "download the Internet." Watch the full beginner-friendly breakdown here: https://t.co/doph3alGnE
18
418
4K
never think before spending money on books. also, if you want one and dont want to buy, drop the link below and i will gift it to you
1K
137
4K
I am feeling sick ๐คง๐คข
Turned 22 today. AMA !! Quick look at the past year: > landed an ML internship at @Apple, then joined @MSFTResearch to work on agentic memory > secured a six figure USD research grant from @Meta to build SoTA AI models at @cognitivelab_ai > crossed 10k GitHub stars across my
0
0
1
Turned 22 today. AMA !! Quick look at the past year: > landed an ML internship at @Apple, then joined @MSFTResearch to work on agentic memory > secured a six figure USD research grant from @Meta to build SoTA AI models at @cognitivelab_ai > crossed 10k GitHub stars across my
105
45
1K
No worries ๐คฃ @rrishabk please try to learn AI from scratch Data analytics-> ML, DL ,CV,NLP, THEN MOVE TO LEARNING ADV LLMs -> gen ai ( how it works exactly) the move to fine tuning and other adv concepts Kid !! ๐ค๐ค
1
0
1
tldr: 2 guys with a laptop removed ~70% of the worldโs compute bill for free, cuz why not
You can now train LLMs 3ร faster with no accuracy loss, via our new RoPE and MLP kernels. Our Triton kernels plus smart auto packing delivers ~3ร faster training & 30% less VRAM vs optimized FA3 setups. Train Qwen3-4B 3x faster on just 3.9GB VRAM. Blog: https://t.co/j3HpsmzrCw
37
236
5K
Build this Markdown render website where you can share the preview via url ( i am not finding any so i created this ) https://t.co/ZtjgYzOwdL
2
0
2
i won the @xai hackathon by making ads for X Videos introducing Halftime. targeted ad generation using AI that feels like a part of your movies and shows built with @yuviecodes @lohanipravin
503
156
4K
I have been fine-tuning LLMs for over 2 years now! Here are the top 5 LLM fine-tuning techniques, explained with visuals: First of all, what's so different about LLM finetuning? Traditional fineโtuning is impractical for LLMs (billions of params; 100s GB). Since this kind of
44
327
2K
Also so many are down again Obviously if we are dependent on a product which is that much big This is possible Kind a monopoly
0
0
1
> my top 8 local LLMs > to run at home > for December 2025 > MiniMax-M2 at #1 > great for agentic workflows thanks to interleaved thinking > very solid at coding & good at UI/design as well > fits on 2x RTX PRO 6000 > or 8x RTX 3090s w/ full context > (specifically
38
56
458