aicoding_ Profile Banner
Andrea | πŸ‡ΈπŸ‡ͺπŸ‡ͺπŸ‡ΈπŸ‡»πŸ‡ͺ Profile
Andrea | πŸ‡ΈπŸ‡ͺπŸ‡ͺπŸ‡ΈπŸ‡»πŸ‡ͺ

@aicoding_

Followers
229
Following
7
Media
0
Statuses
209

Computer Vision Engineer currently working as a Machine Learning Engineer. https://t.co/xLVKLO30rv https://t.co/DiKrU5Eya5

Joined July 2016
Don't wanna be here? Send us removal request.
@xiangyue96
Xiang Yue
1 year
Introducing Critique Fine-Tuning (CFT): a more effective SFT method for enhancing LLMs' reasoning abilities. πŸ“„ Paper: https://t.co/BmotfUjBWP CFT is simple: instead of training models to directly answer questions, we train them to critique noisy answers. What's fascinating is
11
69
306
@UnslothAI
Unsloth AI
1 year
Run DeepSeek-R1 (671B) locally on @OpenWebUI - Full Guide No GPU required. Using our 1.58-bit Dynamic GGUF and llama.cpp. Tutorial: https://t.co/p5WCA3olgJ
16
179
845
@illyism
ILIAS ISM
1 year
You don't need a reasoning model like R1 or o3, just use this .cursorrules with Claude Sonnet to add a thinking step, works 100x better.
81
274
5K
@ivanfioravanti
Ivan Fioravanti α―…
1 year
πŸ”₯ o3-mini-high beats deepseek r1 and o1-pro! in a p5.js challenge! 03-mini result is so good that deserves a video on its own. deepseek r1 (bad result) and o1-pro (better) in comments below. Prompt in last comment. 1/4
71
133
1K
@DimitrisPapail
Dimitris Papailiopoulos
1 year
Transformers can overcome easy-to-hard and length generalization challenges through recursive self-improvement. Paper on arxiv coming on Monday. Link to a talk I gave on this below πŸ‘‡ Super excited about this work!
19
145
1K
@sama
Sam Altman
1 year
o3-mini is out! smart, fast model. available in ChatGPT and API. it can search the web, and it shows its thinking. available to free-tier users! click the "reason" button. with ChatGPT plus, you can select "o3-mini-high", which thinks harder and gives better answers.
2K
2K
26K
@SeunghyunSEO7
Seunghyun Seo
1 year
what up guys, I made a one-page comparison of MHA and MLA from @deepseek_ai for those who skipped the DS-V2 paper. pls correct me if I'm wrong.
4
51
363
@LangChain
LangChain
1 year
πŸ“šπŸ€– Advanced RAG + Agents Cookbook A comprehensive open-source guide delivering production-ready implementations of cutting-edge RAG techniques with AI agents. Built with LangChain and LangGraph, it features advanced implementations like Hybrid, Self, and ReAct RAG. Learn
5
159
706
@andimarafioti
Andi Marafioti
1 year
Fuck it, today we're open-sourcing the codebase used to train SmolVLM from scratch on 256 H100sπŸ”₯ Inspired by our team's effort to open-source DeepSeek's R1 training, we are releasing the training and evaluation code on top of the weights 🫑 Now you can train any of our
34
215
1K
@_akhaliq
AK
1 year
OpenAI o3-mini System Card
11
70
364
@hxiao
Han Xiao
1 year
Letter-dropping physics comparison: o3-mini vs. deepseek-r1 vs. claude-3.5 in one-shot - which is the best? Prompt: Create a JavaScript animation of falling letters with realistic physics. The letters should: * Appear randomly at the top of the screen with varying sizes * Fall
155
260
3K
@omarsar0
elvis
1 year
AI Agents for Computer Use This report provides a comprehensive overview of the emerging field of instruction-based computer control, examining available agents – their taxonomy, development, and resources.
15
142
659
@G4brym
Gabriel Massadas
1 year
Gemini 2.0 doesn’t get nearly enough credit. I just dumped all my workers-qb source code into it, hit it with a simple, humble prompt, and boom => it one-shotted the docs. Not just good docs, way better than what I had before, packed with examples. Kinda insane.
30
62
720
@_akhaliq
AK
1 year
OpenAI o3-mini just one shotted this prompt: write a script for 100 bouncing yellow balls within a sphere, make sure to handle collision detection properly. make the sphere slowly rotate. make sure balls stays within the sphere. implement it in p5.js
137
410
4K
@abacaj
anton
1 year
Finished a run (R1 style) GRPO on Qwen-2.5-0.5B (base model) yield +10 accuracy points on GSM8K. Literally just works. Base model scores 41.6% as reported on qwen paper vs 51%~ GRPO
41
109
1K
@doesdatmaksense
Antaripa Saha
1 year
for people learning gpu programming and especially triton should check out liger kernel by linkedin it was released last year and built on top of triton to provide pre-optimized, ready-to-use implementations gpu optimization techniques specifically tailored for llm training
9
64
625
@CalebPeffer
Caleb Peffer (Hiring!)
1 year
Excited to announce https://t.co/azlzx4Rrah A website that turns any website into a get API with @firecrawl /extract endpoint. Data on the web has never been more accessible! Thanks to @devdigest, for starting this fabulous trend. Check out his GitHub repo below!
38
199
2K
@lexfridman
Lex Fridman
1 year
OpenAI o3-mini is a good model, but DeepSeek r1 is similar performance, still cheaper, and reveals its reasoning. Better models will come (can't wait for o3pro), but the "DeepSeek moment" is real. I think it will still be remembered 5 years from now as a pivotal event in tech
981
1K
13K
@ArtificialAnlys
Artificial Analysis
1 year
OpenAI’s o3-mini is here - a significant jump forward from o1-mini Initial results (full benchmarking coming soon): ➀ Artificial Analysis Quality Index of 89, matching DeepSeek R1 and just below o1 ➀ Cheaper - $1.1/$4.4 input/output pricing per million tokens, lower than many
24
63
406
@IntuitMachine
Carlos E. Perez
1 year
When working with o1/o3 models, I always have this feeling that I'm leaving a lot on the table with my prompting. Creating a long sequence of prompts for regular LLMs is good practice. This is because you don't want to overload what an LLM can process (or it'll lead to
21
79
526