Arvind Neelakantan Profile
Arvind Neelakantan

@arvind_io

Followers
6K
Following
6K
Media
22
Statuses
105

Research Scientist, @GoogleDeepMind Past: @AIatMeta , @OpenAI, @Google Brain PhD @UMassAmherst

Joined January 2012
Don't wanna be here? Send us removal request.
@arvind_io
Arvind Neelakantan
2 months
thrilled to be back @Google in the @GoogleDeepMind team! The technical breadth and expertise across the whole stack (hardware->infra->deep learning->products) is truly mind-blowing. Great to see a lot of familiar faces and meet new friends. Look forward to learning a lot!.
34
30
1K
@arvind_io
Arvind Neelakantan
9 months
Excited to join @AIatMeta! The past 4.5 years at @OpenAI,working on embeddings, GPT-3 & 4,API and ChatGPT, have been career highlights. Now, I'm thrilled to work on the next generations of Llama and contribute to its impact on the developer ecosystem and billions of users!🚀 1/2.
44
26
1K
@arvind_io
Arvind Neelakantan
6 years
We explore a simple approach to task-oriented dialog. A single neural network consumes conversation history and external knowledge as input and generates the next turn text response along with the action (when necessary) as output. Paper: 1/4
Tweet media one
3
57
229
@arvind_io
Arvind Neelakantan
7 years
We develop a non-autoregressive machine translation model whose accuracy almost matches a strong greedy autoregressive baseline Transformer, while being 3.3 times faster at inference. Joint work with @ashVaswani @nikiparmar09 Aurko Roy
1
50
204
@arvind_io
Arvind Neelakantan
3 years
A thread on how we evaluate our embedding models in OpenAI’s API. We achieve state-of-the-art results in linear probe classification, text search and code search. It’s not fine-tuned, so it works great in the real world — and our customers love it. 1/7.
6
29
157
@arvind_io
Arvind Neelakantan
2 years
@tszzl imagine being told you are wrong million times a second, for a few months.
3
0
87
@arvind_io
Arvind Neelakantan
3 years
Zero-shot results of OpenAI API’s embeddings on the FIQA search dataset. Evaluation script: We zero-shot evaluated on 14 text search datasets, our embeddings outperform keyword search and previous dense embedding methods on 11 of them!
Tweet media one
@arvind_io
Arvind Neelakantan
3 years
In text search tasks, we obtain best zero-shot results in msmarco, triviaQA, and NQ and also the best transfer results on the BEIR benchmark. 5/7
Tweet media one
0
15
68
@arvind_io
Arvind Neelakantan
9 months
look forward to working with @manohar_paluri, @Ahmad_Al_Dahle, @edunov and many others in the excellent @AIatMeta team! 2/2.
4
0
54
@arvind_io
Arvind Neelakantan
3 years
Thanks for a balanced take! Couple of comments that are also added to the video description now: 1/4.
@ykilcher
Yannic Kilcher 🇸🇨
3 years
🔥New Video🔥.OpenAI now offers embeddings for text similarity and search, but are they holding up? We look at the release, the paper, the criticism, and most important: the price! Are the embeddings worth it? Watch here to find out:.
Tweet media one
4
9
53
@arvind_io
Arvind Neelakantan
3 years
Small models specifically fine-tuned on a dataset can do well on a narrow benchmark, but they far underperform in real-world settings, as many of our customers are discovering. This study from @FineTuneLearn shows our API performance. 7/7
Tweet media one
2
14
48
@arvind_io
Arvind Neelakantan
3 years
OpenAI embeddings work on a very broad set of use cases. Here, Viable gets a 7.7% absolute improvement in clustering quality using OpenAI embeddings when compared to previous methods!.
@askviable
Viable 🎯
3 years
We tested different embedding models and show the data behind why GPT-3 was the clear winner for our clustering needs.
0
3
43
@arvind_io
Arvind Neelakantan
3 years
The cost to run this experiment with text-search-ada, embedding both documents and queries, is ~$80. text-search-ada achieves a 62% relative improvement over keyword search here!.
@arvind_io
Arvind Neelakantan
3 years
Zero-shot results of OpenAI API’s embeddings on the FIQA search dataset. Evaluation script: We zero-shot evaluated on 14 text search datasets, our embeddings outperform keyword search and previous dense embedding methods on 11 of them!
Tweet media one
1
7
39
@arvind_io
Arvind Neelakantan
1 year
@OpenAI embeddings api over time
Tweet media one
1
2
36
@arvind_io
Arvind Neelakantan
6 years
We describe a simple technique to parallelize Scheduled Sampling across time that allows us to apply Scheduled Sampling for problems that involve generating very long sequences. We get better sample quality and train almost as fast as teacher-forcing.
Tweet media one
2
5
35
@arvind_io
Arvind Neelakantan
3 years
@ylecun For the same reason a kind of unsupervised learning that people were always doing was branded as self-supervised learning 😉.
0
1
30
@arvind_io
Arvind Neelakantan
2 years
@OpenAI embeddings achieve better retrieval performance and are also lot cheaper!.Results taken from:
Tweet media one
Tweet media two
4
3
23
@arvind_io
Arvind Neelakantan
3 years
My team and I trained the model. We look at 33 datasets across four different categories: linear probe classification, sentence similarity, text search, and code search. All these results and figures were in our paper, released this week. 2/7.
2
0
23
@arvind_io
Arvind Neelakantan
3 years
In text search tasks, we obtain best zero-shot results in msmarco, triviaQA, and NQ and also the best transfer results on the BEIR benchmark. 5/7
Tweet media one
3
0
21
@arvind_io
Arvind Neelakantan
3 years
OpenAI Embeddings helps you go beyond keyword search!
Tweet media one
Tweet media two
Tweet media three
@lilianweng
Lilian Weng
3 years
The code is actually extremely simple for a cool app like this - open sourced here:
0
0
19
@arvind_io
Arvind Neelakantan
21 days
TPU -> XLA -> JAX -> Transformer, MoE, Chinchilla, AlphaGo, . -> Gemini, Veo, . -> Search, YouTube, Waymo, . -> Chrome, Android, . 🤯🤯🤯.
0
0
19
@arvind_io
Arvind Neelakantan
3 years
We also achieve new state-of-the-art results on code search. 6/7
Tweet media one
2
0
16
@arvind_io
Arvind Neelakantan
9 months
0
0
4
@arvind_io
Arvind Neelakantan
5 years
Check out our spotlight talk and poster describing the Neural Assistant work in the ConvAI workshop tomorrow @NeurIPSConf #neurips19 .
Tweet media one
@arvind_io
Arvind Neelakantan
6 years
We explore a simple approach to task-oriented dialog. A single neural network consumes conversation history and external knowledge as input and generates the next turn text response along with the action (when necessary) as output. Paper: 1/4
Tweet media one
0
1
14
@arvind_io
Arvind Neelakantan
3 years
In sentence similarity tasks, we perform worse than previous work. This was explained in our paper as well. 4/7
Tweet media one
2
0
12
@arvind_io
Arvind Neelakantan
2 years
@WilliamWangNLP Thanks for having me, I had a fun time visiting @ucsbNLP !.
0
0
10
@arvind_io
Arvind Neelakantan
4 years
@ilyasut Belief is all you need!.
1
0
9
@arvind_io
Arvind Neelakantan
8 years
Our paper () on neural program induction accepted to #ICLR2017 !.Code: #DeepLearning #NLProc.
0
2
8
@arvind_io
Arvind Neelakantan
3 years
in case people are counting, I forgot to share the results for text search from 3 more datasets (apart from the 11 text search results already reported) 🙂
Tweet media one
@arvind_io
Arvind Neelakantan
3 years
My team and I trained the model. We look at 33 datasets across four different categories: linear probe classification, sentence similarity, text search, and code search. All these results and figures were in our paper, released this week. 2/7.
0
0
7
@arvind_io
Arvind Neelakantan
3 years
More details in the paper:
Tweet media one
0
0
7
@arvind_io
Arvind Neelakantan
9 years
We get good results on real-world question answering with neural semantic parsing/program induction. Code is here:
@StatMLPapers
Stat.ML Papers
9 years
Learning a Natural Language Interface with Neural Programmer. (arXiv:1611.08945v1 [cs.CL])
0
2
6
@arvind_io
Arvind Neelakantan
7 years
@GaryMarcus Things are changing : and multiple other recent work in nlp.
0
0
5
@arvind_io
Arvind Neelakantan
6 years
In our experiments we find that: 1) our model was able to incorporate external knowledge and generate factual text response with weak supervision signal. 2) our model can incorporate medium-size knowledge bases with only 8K training examples over multiple verticals.
Tweet media one
1
1
4
@arvind_io
Arvind Neelakantan
3 years
@bobvanluijt @SeMI_tech @CShorten30 @OpenAI This was fun, thanks for having me!.
1
0
5
@arvind_io
Arvind Neelakantan
2 years
@sdand Any feedback for us ? :).
0
0
5
@arvind_io
Arvind Neelakantan
6 years
@GoogleAI @Google Implementation of Neural Assistant: Joint Action Prediction, Response Generation, and Latent Knowledge Reasoning:
0
1
5
@arvind_io
Arvind Neelakantan
3 years
@jobergum our method actually zero-shot transfers better than bm25 to 11 search tasks on average as shown in the entire table. even our smallest models are better than bm25. while it is not the only way to exploit training data with bm25, we perform better than one such method docT5 query
Tweet media one
1
0
4
@arvind_io
Arvind Neelakantan
6 years
@quocleix Agree! But, I think once widely used brown clusters (e.g., : should also be given credit. They use language model pre-training objective on unlabeled data and transfer the word clusters to supervised tasks. They are not "contextual" though.
0
0
4
@arvind_io
Arvind Neelakantan
6 years
Work done with awesome intern Semih Yavuz and many awesome colleagues @GoogleAI @Google.
1
0
3
@arvind_io
Arvind Neelakantan
3 years
We leave out 6 not 7 BEIR datasets.Results on MSMARCO, NQ, TriviaQA are in a separate table (Table 5 in the paper).NQ is part of BEIR too and we didn't want to repeat it.The 6 datasets we leave out are not readily available and it is common to leave them out in prior work too.3/4.
1
0
3
@arvind_io
Arvind Neelakantan
5 years
@egrefen @pfau what are the drawbacks of the benchmark/metric and any suggestions on how they can be improved ?.
0
0
4
@arvind_io
Arvind Neelakantan
3 years
The code for FIQA experiments to reproduce the results in the paper using the API: . There's no discrepancy AFAIK. 2/4.
@arvind_io
Arvind Neelakantan
3 years
Zero-shot results of OpenAI API’s embeddings on the FIQA search dataset. Evaluation script: We zero-shot evaluated on 14 text search datasets, our embeddings outperform keyword search and previous dense embedding methods on 11 of them!
Tweet media one
1
0
4
@arvind_io
Arvind Neelakantan
3 years
For example, see SPLADE v2 ( also evaluates on the same 12 BEIR datasets. Discussion from their paper: 4/4
Tweet media one
1
0
4
@arvind_io
Arvind Neelakantan
6 years
@emnlp2019 Data: Work done with many awesome colleagues at Google Assistant team and.@GoogleAI along with student researcher Chinnadhurai Shankar.
0
0
3
@arvind_io
Arvind Neelakantan
2 months
@melvinjohnsonp @Google @GoogleDeepMind thank you, Melvin! look forward to working with you as well :).
0
0
2
@arvind_io
Arvind Neelakantan
3 years
and also impressive performance on text classification and search!
Tweet media one
Tweet media two
1
0
3
@arvind_io
Arvind Neelakantan
7 years
@earnmyturns @yoavgo Also, Inductive bias of Transformer makes it easier to skip words and learn long-range dependencies compared to RNNs . This paper has some supporting experiments.
0
0
3
@arvind_io
Arvind Neelakantan
3 years
@AndrewMayne @rushbhatia Awesome, congratulations!!!.
0
0
2
@arvind_io
Arvind Neelakantan
5 years
@quocleix @xpearhead @lmthang Awesome work! 🙂.
0
0
2
@arvind_io
Arvind Neelakantan
3 years
we see massive improvement in code search using our models!
Tweet media one
2
0
2
@arvind_io
Arvind Neelakantan
3 years
@doomie @poolio Noe Cafe!.
2
0
2
@arvind_io
Arvind Neelakantan
1 year
@ZhuyunDai @OpenAI 11 beir datasets used in the embeddings v1 paper:
0
1
2
@arvind_io
Arvind Neelakantan
7 years
@strubell @emnlp2018 Congratulations!!!.
0
0
1
@arvind_io
Arvind Neelakantan
2 months
@quocleix @Google @GoogleDeepMind thank you, Quoc! it was a great chat, felt like I never left :).
0
0
1
@arvind_io
Arvind Neelakantan
6 years
Joint work with Daniel Duckworth, Ben Goodrich, @lukaszkaiser and Samy Bengio.
0
0
1
@arvind_io
Arvind Neelakantan
6 years
@julianharris The conversation is annotated with accept/reject. At test time we would want the third-party business to implement a boolean function that returns whether transaction can be completed.Neural Assistant will learn to work with the response as it has been annotated at training time.
1
0
1
@arvind_io
Arvind Neelakantan
6 years
@julianharris hope it answers your question!.
0
0
1
@arvind_io
Arvind Neelakantan
6 years
@dmimno congratulations!!!.
0
0
1
@arvind_io
Arvind Neelakantan
6 years
0
0
0
@arvind_io
Arvind Neelakantan
2 months
0
0
1
@arvind_io
Arvind Neelakantan
5 years
@DBahdanau Nice work! 🙂.
0
0
1
@arvind_io
Arvind Neelakantan
6 years
@colinraffel @unccs Congratulations!!!.
1
0
1
@arvind_io
Arvind Neelakantan
2 months
@JeffDean @Google @GoogleDeepMind thank you, Jeff! so happy to be back :).
0
0
1