daibond_alpha Profile Banner
Bo Dai Profile
Bo Dai

@daibond_alpha

Followers
3K
Following
3K
Media
9
Statuses
223

Assistant Professor at @gtcse, Research Scientist at @GoogleDeepMind | ex @googlebrain

California, USA
Joined October 2012
Don't wanna be here? Send us removal request.
@daibond_alpha
Bo Dai
1 year
RL is so back, as David Silver predicted. https://t.co/fa0RJEI1je
2
10
181
@daibond_alpha
Bo Dai
11 months
RL is sparkling again.
@daibond_alpha
Bo Dai
1 year
RL is so back, as David Silver predicted. https://t.co/fa0RJEI1je
0
0
54
@daibond_alpha
Bo Dai
1 year
We scale up the diffusion transformer for SoTA performance with fast inference!
@haotiansun014
Haotian Sun
1 year
(3/N) EC-DIT outperforms dense models while maintaining competitive inference speed. Our largest model (64 experts) hits a GenEval score of 71.68%🔝, with around 23% additional overhead to the dense model.
0
0
35
@daibond_alpha
Bo Dai
1 year
I did not even have 10 submissions…. There are two different “Bo Dai”.
@peter_richtarik
Peter Richtarik
1 year
11
3
139
@daibond_alpha
Bo Dai
1 year
Please consider joining us to explore the frontier on generative foundation model for decision making, planning, and reasoning.
@hanjundai
Hanjun Dai
1 year
Our team (w/Dale, @daibond_alpha, @mengjiao_yang + others) at Google DeepMind is looking to hire. If you are interested in foundation models+decision making, and making real-world impact through Gemini and cloud solutions, please consider applying through https://t.co/KfhYZuohIY
0
1
24
@daibond_alpha
Bo Dai
1 year
My Twitter Interaction Circle ➡️ https://t.co/lM7gVhpC2t
0
0
12
@daibond_alpha
Bo Dai
1 year
Our black-box adaptation for LLMs has been accepted to #ICML2024. We provide offline and online learning strategy for a value function to pivot the LLMs decoding procedure, only with the access to the output sentences of LLMs.
@daibond_alpha
Bo Dai
2 years
We make local private adaptation of GPT possible!
0
1
18
@daibond_alpha
Bo Dai
1 year
Thanks for sharing our work! We present this principled optimistic/pessimistic policy optimization without uncertainty estimation, and test it with LLMs with my great collaborators @tsen9731, Jincheng @Kgoshvadi @hanjundai Tong Yang @mengjiao_yang, Dale, @yuejiec
@arankomatsuzaki
Aran Komatsuzaki
1 year
Google presents Value-Incentivized Preference Optimization: A Unified Approach to Online and Offline RLHF https://t.co/nMIzXVKDv4
5
8
33
@daibond_alpha
Bo Dai
2 years
This is quite aligned with what we have completed: energy-based black box adaptation https://t.co/sfVJfpzoNP. More interesting part is this adaptor can be trained and used only with sampled sentences from API, agnostic to the logins. And can be used in a plug-and-play way!
Tweet card summary image
arxiv.org
Adapting state-of-the-art Large Language Models (LLMs) like GPT-4 and Gemini for specific tasks is challenging. Due to the opacity in their parameters, embeddings, and even output probabilities,...
@ai_for_success
AshutoshShrivastava
2 years
Q* Leaked info: Source: An unspecified PasteBin.(L-I-N-K in next post) Can't confirm the authenticity as it's from unknown source but you can have a look. Q* is a dialog system conceptualized by OpenAI, designed to enhance the traditional dialog generation approach through the
0
0
13
@daibond_alpha
Bo Dai
2 years
We make local private adaptation of GPT possible!
@haotiansun014
Haotian Sun
2 years
Having troubles with blind domain adaptation for GPTs through OpenAI or Azure 🤔? We are excited to introduce BBox-Adapter 🔌— Lightweight Adapting for Black-Box #LLMs📦. BBox-Adapter offers a transparent, privacy-conscious, and cost-effective solution for customizing
0
2
18
@daibond_alpha
Bo Dai
2 years
Please come to our poster to see the control closed-loop LLM agent.
@chaozhangcs
Chao Zhang
2 years
Want smarter LLM agents? 🤖 Join Haotian's @haotiansun014 poster on AdaPlanner tomorrow! 📅 It enables LLMs to think ahead & plan adaptively based on feedback. #NeurIPS2023 #LLMs #LLMagent https://t.co/byl5Stx2uD
0
4
37
@JackK
Jack Krawczyk
2 years
👀
@neilpatel
Neil Patel
2 years
Do people prefer content written by ChatGPT or Bard?  We had ChatGPT create 1000 articles on various topics. We then gave Bard the same instructions on the same topics. In total, we created 2000 pieces of content with AI. 1000 from ChatGPT. 1000 from Bard. When then paid
12
8
86
@daibond_alpha
Bo Dai
2 years
Great work of using video model in RL!
@sherryyangML
Sherry Yang
2 years
Introducing Universal Simulator (UniSim), an interactive simulator of the real world. Interactive website: https://t.co/c3aQazNYXq Paper: https://t.co/1IdxKQAHsd
0
0
13
@daibond_alpha
Bo Dai
2 years
0
0
4
@daibond_alpha
Bo Dai
2 years
Do not miss the opportunity! I really appreciate the MLSS 2011 Singapore, which led me to this amazing area :)
@myamada0
myamada0
2 years
The submission page for MLSS 2024 in Okinawa is now OPEN! Don't miss this incredible opportunity to expand your research network, enhance your knowledge in machine learning, and connect with experts in the field. Submission page: https://t.co/9lonTZ3yeo Deadline: Sep/30/2023
0
1
13
@daibond_alpha
Bo Dai
2 years
Aloha! Just arrive Honolulu for #ICML23.
0
0
37
@daibond_alpha
Bo Dai
2 years
Glad to see the constrastive representation works in robotics. There is a connection between the constrastive representation and linear MDP as we investigated in our icml2022 paper
@svlevine
Sergey Levine
2 years
Contrastive RL provides a way to use contrastive learning methods to learn general-purpose goal-conditioned policies, uniting representation learning and RL. We recently got this working at scale with real robots! You can read more here: https://t.co/PZxgiViMw8 A short 🧵👇
0
1
11
@daibond_alpha
Bo Dai
2 years
As a concurrent work of Voyager with similar components, we echo that these techniques are generally useful beyond Minecraft @DrJimFan.
8
0
3