OpenChat
@OpenChatDev
Followers
2K
Following
143
Media
10
Statuses
60
Advancing Open Source LLMs with Mixed Quality Data through offline RL-inspired C-RLFT. ⠀⠀⠀⠀⠀⠀⠀⠀⠀ ⠀⠀⠀⠀⠀⠀⠀⠀⠀ ⠀⠀𝗣𝗿𝗼𝗷𝗲𝗰𝘁 𝗟𝗲𝗮𝗱: Guan Wang, @AlpayAriyak
Joined July 2023
🚀Introducing OpenChat 3.6 🌟Surpassed official Llama3-Instruct—with 1-2M synthetic data compared to ~10M human labels 🤫GPTs are close to limits—excel at generation but fall short at complex tasks 🎯We are training next gen—capable of deterministic reasoning and planning 🔗
9
68
291
Will Sudoku become the MNIST for reasoning? Simple rules, clear structure, unique solutions—yet surprisingly challenging for modern LLMs, often requiring explicit trial-and-error to solve. https://t.co/BmqUaW8Qno
huggingface.co
5
10
71
🚀Introducing Hierarchical Reasoning Model🧠🤖 Inspired by brain's hierarchical processing, HRM delivers unprecedented reasoning power on complex tasks like ARC-AGI and expert-level Sudoku using just 1k examples, no pretraining or CoT! Unlock next AI breakthrough with
227
652
4K
Thrilled to see RSP featured at AAAI'25! This pioneering concept was a key inspiration for developing OpenChat! 🚀 #AI #AAAI25
🚨Recursive Skip-Step Planning (RSP) Relying on larger, expressive models for sequential decision-making has recently become a popular choice, but are they truly necessary? Can we replace these heavy models? Yes—RSP empowers shallow MLPs to excel in long-horizon tasks!🧵(1/n)
1
0
6
skronge bones in that one 🔍 excellent job! a 7b model out cracking gpt4 turbo and gpt4o and claude 3 sonnet!
🚀Excited to share our Storm-7B🌪️. This model achieves a 50.5% length-controlled win rate against GPT-4 Preview, making it the first open-source model to match GPT-4 Preview on AlpacaEval 2.0. 📄 https://t.co/WRsxvdVD9x 🤗 https://t.co/GdSdE8lvew
0
3
8
4)This is why I am embarking on a journey to explore new frontiers in AI, specifically targeting the current limitations of GPTs in Planning and Reasoning.
2
0
23
3) However, while training these new models, I can't help but realize the upper limit of what autoregressive models can do. They struggle to solve complex tasks such as software engineering, advanced mathematics, and creating super assistants. It is mathematically challenging for
2
0
11
2) In Openchat 3.6, we pushed Llama3 8B to a new level of performance while retaining the flexibility for further SFT, so developers can better tailor our model for each unique use-cases
1
0
9
1)We developed a new continous pre-training method, Meta-Alignment, for LLMs which achieves similar results to extensive RLHF training that Meta did with Llama3 Instruct. This process is both data and compute efficient using primarily synthetic data at less than 10% of data set
1
0
12
🚀Introducing OpenChat 3.6 🌟Surpassed official Llama3-Instruct—with 1-2M synthetic data compared to ~10M human labels 🤫GPTs are close to limits—excel at generation but fall short at complex tasks 🎯We are training next gen—capable of deterministic reasoning and planning 🔗
9
68
291
🚀 The World's First Gemma fine-tune based on openchat-3.5-0106 data and method (C-RLFT). Almost the same performance as the Mistral-based version. 6T tokens = secret recipe? HuggingFace:
huggingface.co
11
29
178
🚀Kudos to @huggingface ! OpenChat-3.5 Update 0106 has landed on HuggingChat & Spaces! Explore now! Experience open-source AI at ChatGPT & Grok level! 🤗 HuggingChat: https://t.co/ic5jOEYlUm 🌌 Spaces: https://t.co/ov8AdxWSgD 🖥️ OpenChat UI:
huggingface.co
Making the community's best AI chat models available to everyone
🚀Announcing OpenChat-3.5 Update 0106: 𝗪𝗼𝗿𝗹𝗱’𝘀 𝗕𝗲𝘀𝘁 𝗢𝗽𝗲𝗻 𝗦𝗼𝘂𝗿𝗰𝗲 𝟳𝗕 𝗟𝗟𝗠! Experience ChatGPT & Grok-level AI locally 💿! Surpassing Grok-0 (33B) across all 4 benchmarks and Grok-1 (???B) on average and 3/4 benchmarks 🔥. 🎯 This update mainly enhanced
2
10
57
🚀Announcing OpenChat-3.5 Update 0106: 𝗪𝗼𝗿𝗹𝗱’𝘀 𝗕𝗲𝘀𝘁 𝗢𝗽𝗲𝗻 𝗦𝗼𝘂𝗿𝗰𝗲 𝟳𝗕 𝗟𝗟𝗠! Experience ChatGPT & Grok-level AI locally 💿! Surpassing Grok-0 (33B) across all 4 benchmarks and Grok-1 (???B) on average and 3/4 benchmarks 🔥. 🎯 This update mainly enhanced
31
128
625
(1/3) Additional benchmark results, including ChatGPT and other open-source models
2
0
22
🚀Announcing OpenChat-3.5 Update 0106: 𝗪𝗼𝗿𝗹𝗱’𝘀 𝗕𝗲𝘀𝘁 𝗢𝗽𝗲𝗻 𝗦𝗼𝘂𝗿𝗰𝗲 𝟳𝗕 𝗟𝗟𝗠! Experience ChatGPT & Grok-level AI locally 💿! Surpassing Grok-0 (33B) across all 4 benchmarks and Grok-1 (???B) on average and 3/4 benchmarks 🔥. 🎯 This update mainly enhanced
31
128
625
We achieved almost ideal MoE fine-tuning performance (equivalent to a dense model with the same active parameters). <10% overhead
3
0
61