yichern_tan Profile Banner
Yi-Chern Profile
Yi-Chern

@yichern_tan

Followers
218
Following
220
Media
2
Statuses
23

lead, post-training @cohere. previously @Waymo @Facebook @Yale. 🇸🇬

San Francisco, CA
Joined September 2022
Don't wanna be here? Send us removal request.
@yichern_tan
Yi-Chern
9 days
deep research like the best on just 1 gpu.
@cohere
cohere
9 days
Command A Reasoning powers end-to-end systems involving chained and hierarchical agents and leveraging the most relevant tools to accomplish tasks – for example our Deep Research system, which is coming soon to North.
Tweet media one
1
3
49
@yichern_tan
Yi-Chern
16 days
RT @cohere: We’re excited to announce $500M in new funding to accelerate our global expansion and build the next generation of enterprise A….
0
49
0
@grok
Grok
3 days
Join millions who have switched to Grok.
154
290
2K
@yichern_tan
Yi-Chern
30 days
eyes for enterprise.
@cohere
cohere
30 days
Introducing Command A Vision, a state-of-the-art generative model that excels across multimodal image capabilities that matter for enterprises!
Tweet media one
0
3
27
@yichern_tan
Yi-Chern
3 months
check out how we use system prompt learning to reverse engineer human preferences!.
@LisaAlazraki
Lisa Alazraki
3 months
Thrilled to share our new preprint on Reinforcement Learning for Reverse Engineering (RLRE) 🚀. We demonstrate that human preferences can be reverse engineered effectively by pipelining LLMs to optimise upstream preambles via reinforcement learning 🧵⬇️
Tweet media one
0
1
17
@yichern_tan
Yi-Chern
6 months
1313 in 13th position after launching on 13th mar.
@lmarena_ai
lmarena.ai
6 months
🚀 Big news @cohere's latest Command A now climbs to #13 on Arena!. Another organization joining the top-15 club - congrats to the Cohere team!. Highlights:.- open-weight model (111B).- 256K context window.- $2.5/$10 input/output MTok. More analysis👇
Tweet media one
1
6
49
@yichern_tan
Yi-Chern
6 months
i’m so proud of the village we assembled to do this. we used a scalable post-training recipe which we’ll share more about in a technical report soon.
0
0
10
@yichern_tan
Yi-Chern
6 months
gpt-4o perf on enterprise and stem tasks, >deepseek-v3 on many languages including chinese human eval, >gpt-4o on enterprise rag human eval. 2 gpus 256k context length, 156 tops at 1k context, 73 tops at 100k context. this is your workhorse.
@aidangomez
Aidan Gomez
6 months
Today @cohere is very excited to introduce Command A, our new model succeeding Command R+. Command A is an open-weights 111B parameter model with a 256k context window focused on delivering great performance across agentic, multilingual, and coding usecases. 🧵
Tweet media one
1
10
54
@yichern_tan
Yi-Chern
6 months
eyes for aya, on a frontier
Tweet media one
@Cohere_Labs
Cohere Labs
6 months
Introducing ✨ Aya Vision ✨ - an open-weights model to connect our world through language and vision. Aya Vision adds breakthrough multimodal capabilities to our state-of-the-art multilingual 8B and 32B models. 🌿
0
5
20
@yichern_tan
Yi-Chern
6 months
RT @cohere: We’re excited to release Command R7B Arabic – a compact open-weights AI model optimized to deliver state-of-the-art Arabic lang….
0
53
0
@yichern_tan
Yi-Chern
7 months
Let's rethink how LLMs learn from mistakes 📚 . Look out for this new work led by @LisaAlazraki while @cohere:.
Tweet card summary image
arxiv.org
Showing incorrect answers to Large Language Models (LLMs) is a popular strategy to improve their performance in reasoning-intensive tasks. It is widely assumed that, in order to be helpful, the...
@LisaAlazraki
Lisa Alazraki
7 months
Do LLMs need rationales for learning from mistakes? 🤔 When LLMs learn from previous incorrect answers, they typically observe corrective rationales explaining each mistake. In our new preprint, we find these rationales do not help, in fact they hurt performance!. 🧵
1
3
10
@yichern_tan
Yi-Chern
9 months
Very happy to share Command R7B, our smallest and final model in the R series. It's an all-rounder and great for agentic workflows in a small package. I'm at #NeurIPS2024, happy to chat post-training generally capable agents, evals, synthetic data, and what we do at cohere!.
@cohere
cohere
9 months
Introducing Command R7B: the smallest, fastest, and final model in our R series of enterprise-focused LLMs!. It delivers a powerful combination of state-of-the-art performance in its class and efficiency to lower the cost of building AI applications.
1
12
65
@yichern_tan
Yi-Chern
1 year
🏟Command R+ is 6th in the arena leaderboard, as the first open-weights model to surpass earlier versions of GPT-4. 🤔No RAG in the arena yet!. Download at or try via @cohere's API with the @CohereForAI Research Grant Program
Tweet media one
1
6
45
@yichern_tan
Yi-Chern
1 year
➕more weights here, released on 🤗. For the fullest Command R+ experience with RAG, grounded generations, and citations, go to
@aidangomez
Aidan Gomez
1 year
⌘R+. Welcoming Command R+, our latest model focused on scalability, RAG, and Tool Use. Like last time, we're releasing the weights for research use, we hope they're useful to everyone!
0
3
29
@yichern_tan
Yi-Chern
1 year
We’re not done yet,➕more soon.
@lmarena_ai
lmarena.ai
1 year
[Arena Update]. @cohere's Command R is now top-10 in Arena leaderboard🔥. It's now one of the best open models reaching the level of top proprietary models. We find the model great at handling longer context, which we plan to separate as a new category in Arena very soon.
Tweet media one
1
1
26
@yichern_tan
Yi-Chern
1 year
RT @aidangomez: ⌘-R. Introducing Command-R, a model focused on scalability, RAG, and Tool Use. We've also released the weights for research….
Tweet card summary image
cohere.com
Command R is a scalable generative model targeting RAG and Tool Use to enable production-scale AI for enterprise.
0
182
0
@yichern_tan
Yi-Chern
3 years
Excited to share what we've been working on - create a chatbot on top of LLMs with any persona you want using conversant!.
@eddotman
Edward Kim
3 years
(1/2) Large language models' (LLM) generative capabilities can be used to create immersive dialogue agents (i.e., chatbots) for a wide variety of applications. However, you'll also need to manage chat logs, write LLM prompts, and keep track of facts. ↓
0
2
29