sasank51 Profile Banner
Sasank Chilamkurthy Profile
Sasank Chilamkurthy

@sasank51

Followers
2K
Following
19K
Media
340
Statuses
3K

Building JOHNAIC, your personal AI server! Read more: https://t.co/yHTE5CHIci

Bengaluru, India
Joined July 2014
Don't wanna be here? Send us removal request.
@PyTorch
PyTorch
4 days
Large Language Models (#LLMs) are optimized for Intel GPUs labeled as xpu in #PyTorch. Learn how to speed up local inference on Intel Arc discrete, built-in, and Arc Pro GPUs, bringing advanced AI to laptops and desktops. 🔗 https://t.co/D36g0nGBPB #PyTorch #LLM #OpenSourceAI
Tweet media one
8
24
108
@prajdabre
Raj Dabre
11 days
Wow! Prof Mitesh Khapra's name is in Times 100/AI in the same list as Fei-Fei Li, @sama, @elonmusk, @JeffDean and so on. I bless the day I was invited to work alongside him in AI4Bharat! Truly a visionary! https://t.co/grN3LcPE2u Reply 🐐 to show respect!
Tweet media one
34
49
732
@dharmeshba
Dharmesh Ba
26 days
In India, AI won’t replace humans to cut costs. It’ll replace middle management to save dignity. We’re building algorithms that watch workers - not because labor is expensive, but because trust is.
2
4
48
@tarundua81
Tarun Dua
1 month
Agree with the sentiment. How about we start with repatriation of Zomato compute workloads back to @e2enetworks Public Cloud or on prem Sovereign Cloud Platform ? @deepigoyal
@deepigoyal
Deepinder Goyal
1 month
Every few years, the world reminds us of our place. A threat here, a tariff there. But the message is the same: stay in your lane, India. Global powers will always bully us, unless we take our destiny in our own hands. And the only way to do that is if we collectively decide to
10
29
157
@MuraliSrinivasa
Murali Srinivasa
1 month
Speed of PCB fab Execution in 🇮🇳 that made us get a FREE machine from Japan. Couple of years back when we were building a new PCB fab. A delegation including CEO,from a multibillion$ cos in 🇯🇵 visited our construction site. We told this plant will be functional in 3 months.
Tweet media one
203
1K
6K
@sasank51
Sasank Chilamkurthy
1 month
When you're on cloud, your data is no longer yours. Big tech cloud providers can suspend services at their whim if you're considered US enemy. This is the reason why our customers decide to go onprem. They can't afford to have this sort of risk.
@cvkrishnan
Krishnan
1 month
Microsoft is restricting Nayara’s “access to its own data, proprietary tools, and products — despite these being acquired under fully paid-up licenses,”. Nayara Energy has filed a lawsuit in the High Court of Delhi against Microsoft
2
2
16
@huybery
Binyuan Hui
1 month
Our crazy week keeps going! We've updated a smaller Qwen3, called Qwen3-30B-A3B-Instruct-2507! Hope you all have fun playing with it locally! Guess what's next? 👨🏻‍💻
@Alibaba_Qwen
Qwen
1 month
🚀 Qwen3-30B-A3B Small Update: Smarter, faster, and local deployment-friendly. ✨ Key Enhancements: ✅ Enhanced reasoning, coding, and math skills ✅ Broader multilingual knowledge ✅ Improved long-context understanding (up to 256K tokens) ✅ Better alignment with user intent
Tweet media one
59
65
724
@sasank51
Sasank Chilamkurthy
1 month
Me: Move to onprem, it cuts your cloud costs by 5-10x Developer: We don't care about cloud costs Me: OK sure Developer, now back to his stand-up call: Team, let's spend next 2 months optimizing cloud configuration to save 10% of cloud costs Me: 🤦
0
3
6
@sasank51
Sasank Chilamkurthy
1 month
MoE architecture is great for apple style unified memory systems. You can scale memory relatively cheaply. Perf won't be bad because only few of the weights are activated.
@ivanfioravanti
Ivan Fioravanti ᯅ
1 month
Qwen3-235B-A22B-Instruct-2507-3bit-DWQ is here! You can run this on 128GB machine! Model uses ~103GB of RAM On M3 Ultra 512GB: 33 tokens/sec On M4 Max 128GB: 31.1 tokens/sec Have fun! 🔥
Tweet media one
1
0
3
@sasank51
Sasank Chilamkurthy
2 months
Had fun doing viva for our JOHNAIC trainees on introduction to operating systems! Interns who pass our training programs, am sure, will kick ass! https://t.co/3XMXmEgNYe
Tweet media one
1
1
7
@_iunravel
Ashwin Sankar
2 months
I'm happy today because I got to know that someone is building on top of indic-parler-tts. The outputs also sound very natural. Rooting for them to succeed.
0
2
14
@sasank51
Sasank Chilamkurthy
2 months
People pay for radiology because it provides insane value for money. Seeing AI as a way to decrease this 'cost' has been proven a wrong perspective. Successful AI deployments see radiology + AI as a lever to create new care pathways. Radiology 'cost' becomes investment.
@dvasishtha
Dhruv Vasishtha
2 months
Radiology is one of the most unsustainable cost drivers in healthcare right now. So many trainees were worried that AI would replace radiologists that far fewer entered the field relative to current demand. At the same time AI isn't delivering the efficiency gains expected.
1
0
6
@sasank51
Sasank Chilamkurthy
2 months
Or consider JOHNAIC and move onprem!
@nikhilkumarks
Nikhil Kumar
2 months
It'll be nice if you guys take a look at AWS prices too... thank you
0
0
0
@sasank51
Sasank Chilamkurthy
2 months
Looks like code of an interpretor.
@burkov
BURKOV
2 months
This is what they call "an agent."
Tweet media one
0
0
1
@amit_sethi
Amit Sethi
2 months
@sasank51 Those without strong fundamentals will make mistakes by over-relying on LLMs for coding. We need to ban LLM-based coding for those below a certain level of proficiency. After that, LLM use should be rather encouraged. Scared for the next generation. https://t.co/Ix2OlHnX1l
@sasank51
Sasank Chilamkurthy
2 months
I hate it when I have to work with vibe coders who don't even bother to read the code cursor generates. Me - Why did you put this code here? Vibe coder - cursor did it Me - but what does it do? Vibe coder - no clue So frustrating.
0
2
7
@sasank51
Sasank Chilamkurthy
2 months
I hate it when I have to work with vibe coders who don't even bother to read the code cursor generates. Me - Why did you put this code here? Vibe coder - cursor did it Me - but what does it do? Vibe coder - no clue So frustrating.
3
1
18
@sasank51
Sasank Chilamkurthy
2 months
Announcing JOHNAIC 120: 120 GB GPU RAM in a complete cloud-in-a-box system! Read more: https://t.co/nzZ7A1AHpN
Tweet media one
Tweet media two
Tweet media three
7
3
39
@sasank51
Sasank Chilamkurthy
3 months
Nvidia makes you believe AI software works fine on their gpus. In reality, VLLM, one of the best inferencing frameworks doesn't work on Blackwell. Blackwell generation is a significant shake up compared to Hooper.
0
0
5
@sasank51
Sasank Chilamkurthy
4 months
Amazing to see how narrative changed from LLMs to retrieval models yet again. LLMs are generative while retrieval models are discriminative. You need to have both. Read about GANs. Don't fall for the hype that generative models aka genai is everything.
@NirantK
Nirant
4 months
vector search isn't a monolith; it's a pipeline. treating retrieval as "embed → retrieve → LLM" can waste money, increase latency, and lead to poor results. i wrote a blog on designing retrieval systems for cost, latency, and clarity—not just vibes. learn about BM25, mixing
Tweet media one
0
2
5
@aryanagxl
Aryan Agal
5 months
The weird part is - the loss actually didn't go down that much - it feels like some sort of noisy variation with barely any trend
Tweet media one
1
1
2