yoavartzi Profile Banner
Yoav Artzi Profile
Yoav Artzi

@yoavartzi

Followers
17K
Following
19
Media
13
Statuses
105

Research/prof @cs_cornell + @cornell_tech🚡 / https://t.co/9YnWry7yHs / asso. faculty director @arxiv / building https://t.co/f9QkzO5kaC and @COLM_conf

New York, NY
Joined June 2011
Don't wanna be here? Send us removal request.
@yoavartzi
Yoav Artzi
5 months
It's now public! My postdoc call is for the inaugural postdoc as part of this $10.5M gift for a new AI fellows program at Cornell. There's a lot more in this program, so more exciting things to happen here real soon!. Application:
Tweet media one
@yoavartzi
Yoav Artzi
5 months
I am looking for a postdoc. A serious-looking call coming soon, but this is to get it going. Topics include (but not limited to): LLMs (🫢!), multimodal LLMs, interaction+learning, RL, intersection with cogsci, . see our work to get an idea:..Plz RT 🙏.
3
23
108
@yoavartzi
Yoav Artzi
9 days
RT @COLM_conf: The list of accepted papers for COLM 2025 is now available here:. The papers will be made available….
0
3
0
@yoavartzi
Yoav Artzi
16 days
#COLM-ing
Tweet media one
3
1
48
@yoavartzi
Yoav Artzi
20 days
RT @Ar_Douillard: @yoavartzi if you do pre-training, you could try that @cranialxix . it w….
0
4
0
@yoavartzi
Yoav Artzi
21 days
What work/software is out there about training models on heterogeneous clusters? Let's say I have access to various machines, some with more contemporary GPUs, some with much older.
5
1
8
@yoavartzi
Yoav Artzi
25 days
Check out our LMLM, our take on what is now being called a "cognitive core" (as far as branding go, this one is not bad) can look like, how it behaves, and how you train for it.
Tweet card summary image
arxiv.org
Neural language models are black-boxes -- both linguistic patterns and factual knowledge are distributed across billions of opaque parameters. This entangled encoding makes it difficult to...
@karpathy
Andrej Karpathy
26 days
The race for LLM "cognitive core" - a few billion param model that maximally sacrifices encyclopedic knowledge for capability. It lives always-on and by default on every computer as the kernel of LLM personal computing. Its features are slowly crystalizing:. - Natively multimodal.
2
7
34
@yoavartzi
Yoav Artzi
1 month
RT @tallinzen: I'm hiring at least one post-doc! We're interested in creating language models that process language more like humans than m….
0
52
0
@yoavartzi
Yoav Artzi
1 month
Hearing a lot of opposition to RAG, largely sounding like "it (will) never works". Trying to reconcile this with almost every query I put into ChatGPT doing a web search and retrieving content. So, definitely seems to work. What am I missing? Is it a definition discreprency?.
3
3
30
@yoavartzi
Yoav Artzi
1 month
RT @COLM_conf: COLM 2025 will include 6 plenary sessions. The details about the format (panel vs. keynote) and topic will come up soon. We….
0
6
0
@yoavartzi
Yoav Artzi
1 month
RT @COLM_conf: 📈
Tweet media one
Tweet media two
0
1
0
@yoavartzi
Yoav Artzi
1 month
RT @COLM_conf: We are making progress on discussions, but also running out of time. Discussion ends tomorrow. Reviewers and ACs, please get….
0
2
0
@yoavartzi
Yoav Artzi
2 months
❤️.
@AiEleuther
EleutherAI
2 months
@yoavartzi We've always been strongly indebted to arXiv. Thank you for your amazing work.
0
0
3
@yoavartzi
Yoav Artzi
2 months
👏
Tweet media one
@AiEleuther
EleutherAI
2 months
Can you train a performant language models without using unlicensed text?. We are thrilled to announce the Common Pile v0.1, an 8TB dataset of openly licensed and public domain text. We train 7B models for 1T and 2T tokens and match the performance similar models like LLaMA 1&2
Tweet media one
1
0
13
@yoavartzi
Yoav Artzi
2 months
RT @COLM_conf: We are doing our best to encourage engagement during the discussion period. It's moving, even if we wish folks would engage….
0
5
0
@yoavartzi
Yoav Artzi
2 months
RT @COLM_conf: Always interesting to see how deadlines impact engagement trends
Tweet media one
0
1
0
@yoavartzi
Yoav Artzi
2 months
RT @COLM_conf: The 2nd stage of the discussion period has now started. The intermediate response deadline was very effective, so now we hav….
0
3
0
@yoavartzi
Yoav Artzi
2 months
RT @COLM_conf: Here are some graphs
Tweet media one
Tweet media two
Tweet media three
0
2
0
@yoavartzi
Yoav Artzi
2 months
RT @COLM_conf: Our discussion period just started. Authors, please read our instructions carefully. We require responses by June 2. But,….
0
3
0
@yoavartzi
Yoav Artzi
2 months
Really happy about this new work. Trying to think a lot about disentangling knowledge and reasoning/ling skill in LLMs, and this is a promising method in this direction. There is a lot of exciting things happening here, and a lot to build on.
@linxizhao4
Linxi Zhao
2 months
🚀Excited to share our latest work:. LLMs entangle language and knowledge, making it hard to verify or update facts. We introduce LMLM 🐑🧠 — a new class of models that externalize factual knowledge into a database and learn during pretraining when and how to retrieve facts
0
3
28
@yoavartzi
Yoav Artzi
2 months
RT @COLM_conf: The full list of COLM 2025 workshops is now online!. Most deadlines are June 23, but check the specific CFP of each workshop….
0
7
0