MadeWithML Profile Banner
Made With ML Profile
Made With ML

@MadeWithML

Followers
10K
Following
598
Media
78
Statuses
649

Learn how to responsibly develop, deploy & manage machine learning. Maintained by @GokuMohandas

Learn machine learning β†’
Joined May 2019
Don't wanna be here? Send us removal request.
@GokuMohandas
Goku Mohandas
2 years
Excited to share our end-to-end LLM workflows guide that we’ve used to help our industry customers fine-tune and serve OSS LLMs that outperform closed-source models in quality, performance and cost. https://t.co/u9hvVj7E24 1/🧡
Tweet card summary image
anyscale.com
Execute end-to-end LLM workflows to develop & productionize LLMs at scale.
1
45
238
@rauchg
Guillermo Rauch
2 years
An AI-generated clone of HN built with @nextjs App Router β—† Uses PPR and streaming Node.js SSR β—† Fully dynamic, fresh data from Postgres β—† All the UIs bootstrapped with @v0 β—† Content via @mistralai 8x7B and @anyscalecompute Tools What I've learned 🧡 https://t.co/HSbl34jzXY
Tweet card summary image
next-ai-news.vercel.app
A version of Hacker News where the stories and comments are 100% generated by LLMs
31
104
903
@anyscalecompute
Anyscale
2 years
@rauchg Glad you're finding it useful! Check out our accompanying blog post and the evaluation experiments we ran comparing across a suite of open-source and proprietary LLMs:
anyscale.com
Powered by Ray, Anyscale empowers AI builders to run and scale all ML and AI workloads on any cloud and on-prem.
0
7
31
@rauchg
Guillermo Rauch
2 years
Very impressed with @anyscalecompute's endpoints, which support tools / function calling. 2LOC to play with Mixtral as a replacement for GPT 🀯
10
28
388
@GokuMohandas
Goku Mohandas
2 years
It's been nice to see small jumps in output quality in our RAG applications from chunking experiments, contextual preprocessing, prompt engineering, fine-tuned embeddings, lexical search, reranking, etc. but we just added Mixtral-8x7B-Instruct to the mix and we're seeing a 🀯
12
68
443
@robertnishihara
Robert Nishihara
2 years
The Llama Guard model is now available on Anyscale Endpoints. Get started here: https://t.co/SBYL7T5NQO Example:
@AIatMeta
AI at Meta
2 years
At release, Purple Llama includes: - CyberSecEval - Llama Guard model - Tools for insecure code detection & testing for cyber attack compliance We're also publishing two new whitepapers outlining this work. Get Purple Llama ➑️ https://t.co/YSSBNXUiZm
2
21
56
@robertnishihara
Robert Nishihara
2 years
One of the most common asks we get is for public (and reproducible) performance benchmarks. LLM inference performance benchmarks are subtle, and this is a rapidly evolving space, so numbers quickly become stale. But to make comparisons, we need to be talking about the same
4
30
120
@bhutanisanyam1
Sanyam Bhutani
2 years
The definitive guide to RAG in production! πŸ™ @GokuMohandas walks us through implementing RAG from scratch, building a scalable app It now has updated discussion on embedding fine-tuning, re-ranking and effectively routing requests I think this is easily the most complete
12
90
559
@robertnishihara
Robert Nishihara
2 years
We updated our production RAG application guide with a number of new sections: β˜‘οΈ When to fine-tune embeddings β˜‘οΈ When to augment vector-based retrieval with traditional lexical search β˜‘οΈ When to rerank retrieved context β˜‘οΈ How to update & reindex as data changes Importantly,
@GokuMohandas
Goku Mohandas
2 years
Added some new components (fine-tuning embeddings, lexical search, reranking, etc.) to our production guide for building RAG-based LLM applications. Combination of these yielded significant retrieval and quality score boosts (evals included). Blog: https://t.co/6LUe8Z6DMm
0
12
48
@GokuMohandas
Goku Mohandas
2 years
Added some new components (fine-tuning embeddings, lexical search, reranking, etc.) to our production guide for building RAG-based LLM applications. Combination of these yielded significant retrieval and quality score boosts (evals included). Blog: https://t.co/6LUe8Z6DMm
@GokuMohandas
Goku Mohandas
2 years
Excited to share our production guide for building RAG-based LLM applications where we bridge the gap between OSS and closed-source LLMs. - πŸ’» Develop a retrieval augmented generation (RAG) based LLM application from scratch. - πŸš€ Scale the major workloads (load, chunk, embed,
7
49
206
@adithyan_ai
Adithyan
2 years
I burned inπŸ”₯2000$ in finetuning so you don't have to. I fine-tuned models with @OpenAI and @anyscalecompute API endpoints with 50million tokens. Here are the results I wish I knew before getting into finetuning. If you just want a quick snapshot, look at the figure. A longer
30
74
674
@AIatMeta
AI at Meta
2 years
Anyscale Endpoints enables AI application developers to easily swap closed models for the Llama 2 models β€” or use open models along with closed models in the same application.
@raydistributed
ray
2 years
The team @MetaAI has done a tremendous amount to move the field forward with the Llama models. We're thrilled to collaborate to help grow the Llama ecosystem. https://t.co/1bB64cs1sf
10
28
168
@raydistributed
ray
2 years
The team @MetaAI has done a tremendous amount to move the field forward with the Llama models. We're thrilled to collaborate to help grow the Llama ecosystem. https://t.co/1bB64cs1sf
Tweet card summary image
anyscale.com
We are excited to announce collaboration between Meta and Anyscale to bolster the Llama ecosystem.
2
17
81
@TheTuringPost
Ksenia_TuringPost
2 years
3 free MLOps courses you should know about: β–ͺ️ MLOps Course, @GokuMohandas β–ͺ️ CS 329S: Machine Learning Systems Design @Stanford β–ͺ️ MLOps Zoomcamp @Al_Grigor 🧡
1
74
283
@llama_index
LlamaIndex πŸ¦™
2 years
New LLM integration πŸ”₯: @anyscalecompute endpoints allows any developer to easily run + finetune open-source LLMs through an API. Best of all you get the full power of Ray Serve/Train for scalable/efficient training and inference ⚑️ Big s/o to kylehh: https://t.co/5NK1zy35T3
2
10
31
@NianticEng
Niantic Engineering
2 years
Later this month, Niantic will present at Ray Summit 23 and our own @dreamingleo89 wrote about how we are using Ray to improve multiple aspects of our scanning and mapping infrastructures, and we're just getting started.
Tweet card summary image
nianticlabs.com
Niantic uses Ray for scaling complex distributed workloads
2
5
23
@CyrusHakha
kourosh hakhamaneshi
2 years
πŸ€” Fine-tuning LLMs: LoRA or Full-Parameter? Which should you choose? Uncover the insights in our latest technical blog. πŸ”— Link: https://t.co/3pvQ9TAksF 🧡 Thread (1/N) πŸ‘‡
Tweet card summary image
anyscale.com
In this blog, we compare full-parameter fine-tuning with LoRA and answer questions around the strengths and weaknesses of the two techniques.
4
43
204
@bhutanisanyam1
Sanyam Bhutani
2 years
High signal ML for developers guide! πŸ™ Building Machine Learning Applications in real world involves a lot of moving parts and ideas. This series covers all of them really well Made with ML by @GokuMohandas is the best no-nonsense collection of guides with every module
2
22
120
@anyscalecompute
Anyscale
2 years
Save cloud costs while keeping quality high with your open source LLM - > Llama 2 is about as factually accurate as GPT-4 for summaries and is 30 times cheaper https://t.co/b3C7gkOajx via @anyscalecompute #ML #AI #ArtificialIntelligence #LLM
Tweet card summary image
anyscale.com
Using Anyscale Endpoints, we compared Llama 2 7b, 13b and 70b vs. OpenAI's GPT-3.5-turbo and GPT-4 for accuracy and cost.
1
4
19
@GokuMohandas
Goku Mohandas
2 years
A very comprehensive case study on fine-tuning Llama-2 across three different tasksπŸ‘‡ - code for distributed fine-tuning w/ @raydistributed + @huggingface Accelerate + @MSFTDeepSpeed - data prep + eval + baselines - when to & not to fine-tune - using perplexity for checkpointing
@CyrusHakha
kourosh hakhamaneshi
2 years
πŸš€ Exploring Llama-2’s Quality: Can we replace generalist GPT-4 endpoints with specialized OSS models? Dive deep with our technical blogpost to understand the nuances and insights of fine-tuning OSS models. πŸ”— https://t.co/zVStDCoG4y 🧡 Thread 1/NπŸ‘‡
1
20
98