Explore tweets tagged as #TensorRTLLM
@ManuAGI01
ManuAGI 🤖 - ( ManuIn )
19 days
🚀TensorRTLLM🔥. 'TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and support state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes
Tweet media one
1
0
0
@ManuAGI01
ManuAGI 🤖 - ( ManuIn )
4 months
🚀Project Number 5 - TensorRT-LLM🔥. ' Turbocharging Your Large Language Model Inference!'. #opensource #github #trending #AI #coding #search #softwaredevelopment #technology #innovation #programming #AItools #TensorRTLLM
1
0
0
@moficodes
Mofi Rahman
1 year
3. TensorRTLLM with GPUs -
1
1
3
@0xblacklight
Kyle Mistele 🏴‍☠️
1 year
Time to rip out vllm for triton+tensorrtllm immediately after ripping out llama.cpp for vllm
Tweet media one
0
1
1
@Synapsi_Off
Synapsi Community
1 year
E poiché tutto si svolge localmente sul tuo PC o workstation Windows RTX, avrai risultati veloci e sicuri. #ChatRTX #GPT #RAG #TensorRTLLM #RTX #Personalizzazione #Chatbot.
Tweet media one
0
0
0
@Aratako_LM
Aratako
7 months
このvLLM vs TensorRT-LLMシリーズがかなり良かった.LLMの推論周りの諸々について分かりやすい説明から入ってその後パフォーマンス比較されるので、vLLMやTensorRT-LLM自体に興味ない人にも勉強になると思う. [vLLM vs TensorRT-LLM] #1. An Overall Evaluation - SqueezeBits
1
14
101
@akashians_
Akashians
1 year
@NVIDIAAIDev pushing @akashnet_ $akt #akt . Check it out ⬇️⬇️. Optimize your Meta Llama 3 #inference with TensorRT-LLM. …and how! ✨ in only 4 steps ✨. 📗➡️. @AIatMeta
0
6
13
@GenAiAlien
Kain Jares - The GenAI Alien
7 months
@Apple working with @nvidia to improve the speed of #TensorRTLLM by almost a favor of 3x was not on my bingo card for 2024.
Tweet media one
0
0
0
@NVIDIAAIDev
NVIDIA AI Developer
1 year
🦙🦙🦙 Optimize your Meta Llama 3 #inference with TensorRT-LLM. …and how! ✨ in only 4 steps ✨. 📗➡️. @AIatMeta
Tweet media one
4
23
102
@dpinto64
David Pinto
1 year
Optimize your Meta Llama 3 #inference with TensorRT-LLM in only 4 steps. @AIatMeta @nvidiaaidev
Tweet media one
0
0
1
@andrey_cheptsov
Andrey Cheptsov
6 months
Benchmarking @vllm_project vs @nvidia TensorRT LLM (VLMs) by SqueezeBits team
0
1
7
@KMatiDev1
Mati
2 years
I ran a benchmark with int4 and achieved an inference speed of about 100 tokens per second. 🚀. Here is the script : 👇👇👇 . #buildinpublic #indiehackers #AI #TensorRTLLM.
0
0
3
@jessebenisrael
Jesse Ben Israel
1 year
Exciting news: Gemma 7B has been deployed with TensorRT-LLM, achieving over 500 tokens per second, a significant advancement in language processing! #Gemma7B #TensorRTLLM.
0
0
1
@genainewstop
GenAINews.co
8 months
Accelerate time to first token with NVIDIA TensorRT-LLM KV cache early reuse techniques! Learn how to optimize KV cache for faster response times. #TensorRTLLM #KVCacheReuse #NVIDIA #AI #Efficiency".
0
0
0
@JigarHalani3
Jigar Halani
1 year
Optimize your Meta Llama 3 #inference with TensorRT-LLM in only 4 steps. @AIatMeta @nvidiaaidev
Tweet media one
0
0
1
@NVIDIAAP
NVIDIA Asia Pacific
2 years
NVIDIA has just released TensorRT-LLM, a game-changer for optimizing large language models in AI inference. Don't worry, you can catch up by reading our technical blog. Dive into the details now! #NVIDIA #TensorRTLLM #AIinference
0
0
1
@aioaipl
AI o AI
2 years
Prędkość spotyka wydajność! 🚀 NVIDIA TensorRT-LLM, potężne narzędzie optymalizacji, teraz dostępne za darmo! Nie przegap szansy na przyspieszenie swoich projektów #AI. Sprawdź darmową bibliotekę #NVIDIA #TensorRTLLM #OptymalizacjaAI #Technologia.
0
1
1
@nauzetjesus_
nauzetjesus.com
1 year
Chat con RTX: tu chatbot personal de IA en tu PC .#ChatConRTX #NVIDIA #IAGenerativa #Chatbot #LLM #GPT #RAG #TensorRTLLM #AceleraciónRTX #GPU.
0
1
0
@sagar_desai_x
sagar desai
1 year
Optimize your Meta Llama 3 #inference with TensorRT-LLM in only 4 steps. @AIatMeta @nvidiaaidev
Tweet media one
0
0
0