
Raghu Ganti
@RaghuGanti
Followers
297
Following
857
Media
6
Statuses
262
Researcher, A.I., dancer
White Plains, NY
Joined January 2012
RT @PyTorch: The Kubeflow Trainer project has been integrated into the PyTorch Ecosystem! This integration ensures that Kubeflow Trainer al….
0
34
0
RT @PyTorch: PyTorch and vLLM are both critical to the AI ecosystem and are increasingly being used together for cutting edge generative AI….
0
65
0
RT @LysandreJik: The Transformers library is undergoing it's largest pivot to date 🙌. It now cements its role as the central model definiti….
0
59
0
RT @PyTorch: PyTorch Foundation has expanded into an umbrella foundation. @vllm_project and @DeepSpeedAI have been accepted as hosted proje….
0
47
0
RT @tri_dao: Very strong 8B and 56B Mamba hybrid models trained to 20T tokens, on 6K H100s, with FP8!.This answers many of the open questio….
0
57
0
Job alert: . Red Hat team is expanding folks in PyTorch! You will be working side-by-side with the inference experts @neuralmagic !.
linkedin.com
I'm looking to hire PyTorch community members and contributors to help us build Red Hat's investment in PyTorch. We believe the future of AI is Open, and that PyTorch is at the centre of it. If you...
0
0
2
RT @danielhanchen: Excited to share that @UnslothAI now supports:. • Full fine-tuning + 8bit.• Nearly any model like Mixtral, Cohere, Grani….
0
53
0
RT @PyTorch: Optimize your model training with smarter memory management!. Check out our latest blog post to learn how PyTorch’s new activa….
0
22
0
Super excited about the release of GneissWeb dataset from IBM:
research.ibm.com
At IBM Research, we’re inventing what’s next in AI, quantum computing, and hybrid cloud to shape the world ahead.
1
6
39
For the MoE experts, I am curious as to how many experts are triggered in a batch of sequences? If I have a batch of 16, would it be possible to trigger (16 x num active parameters) in one forward pass?. Could not find any study either 😢.@deepseek_ai @MistralAI.
0
0
0
RT @PyTorch: Presenting HadaCore: Tensor Core Accelerated Hadamard Transform Kernel 🔍 Take a look at how we achieve state-of-the-art perfor….
0
31
0
Super kicked by this work from @PyTorch team at @AIatMeta and @IBMResearch team. @TheZachMueller , now we just need to get this into @huggingface accelerate and enable the community!. Let’s go and make training even faster!!! 🔥🔥🔥.
Supercharging Training using float8 and FSDP2 ⚡ .Read our latest blog to find out how we achieve up to 50% throughput speedup while achieving loss & evaluation benchmark parity in training:
0
0
3
Great work from @IBM and @PyTorch ! This will allow for native tensor parallel using @huggingface transformers library! Long sequences are now being tamed ;).
Native tensor parallel has landed in transformers!!! thanks a lot to the torch team for their support! . Contributions are welcome to support more models! 🔥.
0
0
3
RT @aashkaa_: I’m presenting our poster on INDUS: Effective & Efficient Language Models for Scientific Applications at #EMNLP2024 tomorrow….
0
2
0
RT @PyTorch: Learn the inner workings of Triton, the hardware agnostic language for GPU programming and powering TorchInductor: https://t.c….
0
35
0