er_shivamsingh0 Profile Banner
Shivam Singh Profile
Shivam Singh

@er_shivamsingh0

Followers
772
Following
6K
Media
251
Statuses
2K

Engineer| koinophobic | 22 | AI | GPU POOR | Building Neo clouds https://t.co/qGAknj71kz

localhost:8080/
Joined March 2023
Don't wanna be here? Send us removal request.
@er_shivamsingh0
Shivam Singh
5 months
It's now me vs me
1
0
12
@lexfridman
Lex Fridman
2 days
I spent last week at Neurips (the big ML/AI conference) and had hundreds of amazing technical conversations with old friends and new. Beyond the technical, I'm deeply grateful for the love from all the folks who stopped me in the hallways and in the street to say kind words and
140
100
2K
@MIT_CSAIL
MIT CSAIL
2 days
Andrej Karpathy, Eureka Labs founder & former Director of AI at Tesla, breaks down how LLMs like ChatGPT "download the Internet." Watch the full beginner-friendly breakdown here: https://t.co/doph3alGnE
18
418
4K
@TheIshanGoswami
Ishan Goswami
2 days
never think before spending money on books. also, if you want one and dont want to buy, drop the link below and i will gift it to you
1K
137
4K
@er_shivamsingh0
Shivam Singh
2 days
๐Ÿ˜‹๐Ÿ˜‹๐Ÿ˜‹๐Ÿ™‚๐ŸคŒ
1
0
3
@er_shivamsingh0
Shivam Singh
5 days
I am feeling sick ๐Ÿคง๐Ÿคข
@adithya_s_k
Adithya S K
5 days
Turned 22 today. AMA !! Quick look at the past year: > landed an ML internship at @Apple, then joined @MSFTResearch to work on agentic memory > secured a six figure USD research grant from @Meta to build SoTA AI models at @cognitivelab_ai > crossed 10k GitHub stars across my
0
0
1
@adithya_s_k
Adithya S K
5 days
Turned 22 today. AMA !! Quick look at the past year: > landed an ML internship at @Apple, then joined @MSFTResearch to work on agentic memory > secured a six figure USD research grant from @Meta to build SoTA AI models at @cognitivelab_ai > crossed 10k GitHub stars across my
105
45
1K
@er_shivamsingh0
Shivam Singh
5 days
No worries ๐Ÿคฃ @rrishabk please try to learn AI from scratch Data analytics-> ML, DL ,CV,NLP, THEN MOVE TO LEARNING ADV LLMs -> gen ai ( how it works exactly) the move to fine tuning and other adv concepts Kid !! ๐ŸคŒ๐ŸคŒ
1
0
1
@Yampeleg
Yam Peleg
6 days
tldr: 2 guys with a laptop removed ~70% of the worldโ€™s compute bill for free, cuz why not
@UnslothAI
Unsloth AI
6 days
You can now train LLMs 3ร— faster with no accuracy loss, via our new RoPE and MLP kernels. Our Triton kernels plus smart auto packing delivers ~3ร— faster training & 30% less VRAM vs optimized FA3 setups. Train Qwen3-4B 3x faster on just 3.9GB VRAM. Blog: https://t.co/j3HpsmzrCw
37
236
5K
@er_shivamsingh0
Shivam Singh
5 days
A new era will begin now
@karpathy
Andrej Karpathy
6 days
nanoGPT - the first LLM to train and inference in space ๐Ÿฅน. It begins.
0
0
0
@AdiOltean
Adi Oltean
6 days
We have just used the @Nvidia H100 onboard Starcloud-1 to train the first LLM in space! We trained the nano-GPT model from Andrej @Karpathy on the complete works of Shakespeare and successfully ran inference on it. We have also run inference on a preloaded Gemma model, and we
191
570
4K
@er_shivamsingh0
Shivam Singh
7 days
Build this Markdown render website where you can share the preview via url ( i am not finding any so i created this ) https://t.co/ZtjgYzOwdL
2
0
2
@krishgarg09
Krish
8 days
i won the @xai hackathon by making ads for X Videos introducing Halftime. targeted ad generation using AI that feels like a part of your movies and shows built with @yuviecodes @lohanipravin
503
156
4K
@_avichawla
Avi Chawla
12 days
I have been fine-tuning LLMs for over 2 years now! Here are the top 5 LLM fine-tuning techniques, explained with visuals: First of all, what's so different about LLM finetuning? Traditional fineโ€‘tuning is impractical for LLMs (billions of params; 100s GB). Since this kind of
44
327
2K
@er_shivamsingh0
Shivam Singh
10 days
Someone please write AWS @awscloud docs again using @mintlify please I am tired of reading boring docs
0
0
3
@er_shivamsingh0
Shivam Singh
11 days
Also so many are down again Obviously if we are dependent on a product which is that much big This is possible Kind a monopoly
0
0
1
@er_shivamsingh0
Shivam Singh
11 days
Again @Cloudflare ๐Ÿ˜ญ๐Ÿ™
0
0
1
@TheAhmadOsman
Ahmad
11 days
> my top 8 local LLMs > to run at home > for December 2025 > MiniMax-M2 at #1 > great for agentic workflows thanks to interleaved thinking > very solid at coding & good at UI/design as well > fits on 2x RTX PRO 6000 > or 8x RTX 3090s w/ full context > (specifically
38
56
458
@archiexzzz
Archie Sengupta
12 days
In this article, I have tried to cover the breadth of distributed GPU training, which includes: > arithmetic intensity via shared memory data reuse > BF16 precision for memory and training stability > CUDA kernel programming for GPU execution > data, pipeline, and Tensor
12
46
516