
Luke Metz
@Luke_Metz
Followers
17K
Following
4K
Media
51
Statuses
326
Thinking Machines Previously: OpenAI, Google Brain
San Francisco, CA
Joined October 2012
RT @NandoDF: Thinking Machines’ @Luke_Metz giving a clear, beautifully simple, but extremely clever and informative tutorial on post trai….
0
4
0
Super excited to share what we've been working on. It's a privilege to be working with such an amazing team!.
Today, we are excited to announce Thinking Machines Lab (, an artificial intelligence research and product company. We are scientists, engineers, and builders behind some of the most widely used AI products and libraries, including ChatGPT,.
7
8
243
RT @chipro: It’s done! 150,000 words, 200+ illustrations, 250 footnotes, and over 1200 reference links. My editor just told me the manuscr….
0
626
0
I'm leaving OpenAI after over 2 years of wild ride. Alongside @barret_zoph , @LiamFedus , @johnschulman2 , and many others I got to build a “low key research preview” product that became ChatGPT. While we were all excited to work on it, none of us expected it to be where it is.
129
126
3K
It has been a pleasure working with you John. I am extremely sad to see you go. Best of luck with your new adventures!.
I shared the following note with my OpenAI colleagues today:. I've made the difficult decision to leave OpenAI. This choice stems from my desire to deepen my focus on AI alignment, and to start a new chapter of my career where I can return to hands-on technical work. I've decided.
1
0
53
RT @chipro: A challenge of building AI applications is choosing which model to use. What if we don’t have to? What if we can predict the be….
0
52
0
RT @jaschasd: Have you ever done a dense grid search over neural network hyperparameters? Like a *really dense* grid search? It looks like….
0
2K
0
New gradient estimation technique lead by the fantastic @OscarLi101!. It provides low variance estimates gradients of unrolled, or iterative computation graphs such as those found in rl, learned optimizers, meta optimization. If you’re at NeuRIPS go check out the poster!.
Introducing: Noise-Reuse Evolution Strategies, an unbiased, online, memory efficient, variance-reduced gradient estimator that can outperform many other methods (including Backprop) on some particularly challenging unrolled computation graph problems. A Thread🧵.
0
1
18
RT @OscarLi101: 📝Quiz time: when you have an unrolled computation graph (see figure below), how would you compute the unrolling parameters'….
0
13
0
RT @chipro: New blog post: Multimodality and Large Multimodal Models (LMMs). Being able to work with data of different modalities -- e.g. t….
0
192
0
RT @jmes_harrison: Want to learn about learned optimization? I gave a tutorial at @CoLLAs_Conf which is now public!
0
9
0
RT @giffmana: What makes CLIP work?.The contrast with negatives via softmax?.The more negatives, the better -> large batch-size?. We'll ans….
0
292
0
RT @chipro: Open challenges in LLM research. The first two challenges, hallucinations and context learning, are probably the most talked ab….
0
398
0
RT @__ishaan: New paper with @tatsu_hashimoto! Likelihood-Based Diffusion Language Models: Likelihood-based traini….
0
35
0
RT @chipro: New post: RLHF - Reinforcement Learning from Human Feedback. Discussing 3 phases of ChatGPT development, where RLHF fits in, ho….
0
215
0
RT @chipro: New post: bringing LLM applications to production!. 1. Challenges of LLM engineering & the solutions that I’ve seen. 2. How to….
0
222
0