Explore tweets tagged as #WebLLM
@GitHub_Daily
GitHubDaily
23 days
GitHub 上一款完全本地化运行的 AI 助手浏览器插件:NativeMind,. 支持 Ollama 和 WebLLM 两种本地模型调用,所有数据处理都在设备本地完成,无需担心隐私泄露。. GitHub: 主要功能:. - 智能对话助手,支持多标签页上下文理解.- 网页内容分析,可快速总结页面和文档内容
10
32
155
@ldjconfirmed
LDJ
2 years
Running locally on iphone 13 mini without cellular, without wifi, without bluetooth. Real-time, not sped-up. Insightful in-depth response to a question involving 2 relatively obscure and complex topics. (I plan to make it even more effecient soon 😉) ty to WebLLM & MLCChat.
37
43
325
@samwillis
Sam Willis
1 year
What if psql had an \ai command?.This is PGlite (WASM Postgres) and Llama3 (via WebLLM & WebGPU) running completely in the browser - no cloud, no server🤯.Local AI is coming sooner than you think!
7
35
192
@npaka123
布留川英一 / Hidekazu Furukawa
10 months
WebLLM を使って Qwen2.5-1.5B のブラウザ上でのローカル実行をお試し中。
0
4
20
@yi_xin_dong
Yixin Dong
8 months
🚀✨Introducing XGrammar: a fast, flexible, and portable engine for structured generation!. 🤖Accurate JSON/grammar generation.⚡️3-10x speedup in latency.🤝Easy LLM engine integration.✅ Now in MLC-LLM, SGLang, WebLLM; vLLM & HuggingFace coming soon!.
Tweet media one
Tweet media two
6
65
257
@Tahabsn
Taha Bouhsine
4 months
we just added webllm support to @bashnota .now you can still vibe code in the woods
1
1
6
@calebfahlgren
Caleb
11 months
Running Phi-3.5-Mini in the browser at 67 tokens per second just feels right 🤗. 🔸 Powered by MLC WebLLM + WebGPU.🔸 Fully Private / Running on Device
2
3
14
@nash_su
nash_su - e/acc
2 years
手机上跑LLM大模型.继WebLLM,作者又发布了MLC-LLM.在任意支持 CUDA、Vulkan、Metal的平台上,当然包括手机,就可以运行大模型. Demo:.
Tweet media one
2
59
191
@reach_vb
Vaibhav (VB) Srivastav
9 months
Fully open source code interpreter running 100% local and in your browser! 🔥. Powered by Qwen Code 2.5 & WebLLM! ⚡
12
83
585
@calebfahlgren
Caleb
10 months
Qwen-2.5 on WebGPU 🏎️. • 42 tok/sec for Qwen2.5-Coder-1.5B on Mac ⚡.• Powered by MLC WebLLM and WebGPU 🔥. Watch Qwen2.5-Coder-1.5B build a website entirely in the browser!
12
50
390
@reach_vb
Vaibhav (VB) Srivastav
8 months
LFG!! XGrammar: a lightning fast, flexible, and portable engine for structured generation! 🔥. > Accurate JSON/grammar generation.> 3-10x speedup in latency.> 14x faster JSON-schema generation and up to 80x CFG-guided generation.> Now in MLC-LLM, SGLang, WebLLM; vLLM &
Tweet media one
2
29
114
@kalyan_kpl
Kalyan KS
10 months
Run Llama 3.2 LLM in Web Browser Locally. Llama 3.2 is the new series of LLMs in Llama family. Now you can run Llama 3.2 (1B and 3B) in your browser locally with the help of WebLLM and WebGPU. WebLLM is a high-performance in-browser LLM inference engine. WebGPU is a new,
0
1
7
@jtpio
Jeremy Tuloup
5 months
What if we could use AI models like Llama 3.2 or Mistral 7B in the browser with JupyterLite? 🤯. Still at a very early stage of course, but making some good progress!. Thanks to WebLLM, which brings hardware accelerated language model inference onto web browsers, via WebGPU 🚀
0
2
5
@reach_vb
Vaibhav (VB) Srivastav
1 year
Gemma 2 2B running in a browser, powered by WebLLM & WebGPU! 🔥. 100% local & on-device. In less than 24 hours, we've already got the model to the edge! ⚡. Try it out on an HF space below:
10
113
484
@calebfahlgren
Caleb
10 months
Run Llama 3.2 in the browser on WebGPU 🔥. • Llama 3.2 (1B + 3B)🤏 🦙.• Running 100% locally in the browser at 62 tok/sec 🏎️.• Powered by MLC WebLLM + WebGPU ⚡
4
9
57
@calebfahlgren
Caleb
10 months
Try out the new WebLLM Playground with the new model picker 🏎️. Try LLama 3.2, Qwen 2.5, Mistral, Gemma and more entirely in the browser.
5
8
54
@hokazuya
ホーダチ-Hodatsu | LLM Researcher × AI Engineer
11 months
Phi-3.5-miniが、WebLLMとして利用可能に. インストール、セットアップ不要でLLMをブラウザから自分のPCのGPUで。. ほかの人と共有しないので、いつでも基本的に同じ性能で利用できるところが良さ。. VRAM6GBくらいあれば動くと思うが、
1
13
63
@Aut4rk
Autark
5 days
Just adding real time streaming batched inference to WebLLM. No big deal.
Tweet media one
1
0
2
@josedonato__
José Donato
9 months
wip: allowing local models via WebLLM for OpenBB Copilot 🤯🤯🤯. all data stays in your control without leaving the browser. small models are becoming powerful. WebLLM project from @charlie_ruan is just mind blowing, took me a few minutes to integrate. even supports streaming
2
7
26