Amalik Amriou
@amalik
Followers
2K
Following
2K
Media
3K
Statuses
24K
Data & AI Product Lead @OCTOTechnology - part of @Accenture - DeFi enthusiast - Irregular cyclist - Inconsistent runner
Fontainebleau, France
Joined November 2007
Traditional RAG sucks because it promises "relevant chunks" but in fact returns "similar chunks". Relevancy requires reasoning. Introducing ReAG - Reasoning Augmented Generation
150
291
4K
How to Scale Your Model Google DeepMind just released an awesome book on scaling language models on TPUs. This is gold! Worth checking you are an LLM developer.
4
72
315
Language Models Use Trigonometry to Do Addition "We first discover that numbers are represented in these LLMs as a generalized helix, which is strongly causally implicated for the tasks of addition and subtraction, and is also causally relevant for integer division,
57
349
2K
I wrote a short overview of DeepSeek R1's training process: https://t.co/OLsz2u1fwJ Will follow-up with the knowledge distillation later this week.
15
144
768
Woow a fully open source reasoning model on par with OpenAI o1 just released Deepseek R1 even outperforms Claude 3.5 Sonnet and o1-mini in almost all benchmarks. You can already use it for free (see below)
103
485
5K
⚠️ Developers, please be careful when installing Homebrew. Google is serving sponsored links to a Homebrew site clone that has a cURL command to malware. The URL for this site is one letter different than the official site.
255
3K
11K
🌻kotaemon An open-source clean & customizable RAG UI for chatting with your documents Built with both end users and developers in mind. You can use off the shelf, or easily customize https://t.co/nKSe4sBYHM
11
114
539
Optimizing LLM-based Multi-Agent Systems Presents a novel framework, OPTIMA, to enhance both communication efficiency and task effectiveness in LLM-based multi-agent systems through LLM training. Proposes an iterative generate, rank, select, and train paradigm with a reward
5
55
270
Useful Prompting technique. Simply ask the LLM to re-read the question - significantly boosts LLM reasoning across diverse tasks and model types. 💡 Repeats question input twice in prompt, unlocks latent reasoning potential **Problem** 🤔: Decoder-only LLMs with
38
166
1K
MemLong utilizes an external retriever for retrieving historical information which enhances the capabilities of long-context LLMs. It consistently outperforms other SoTA LLMs on long-context benchmarks and can extend the context length on a single 3090 GPU from 4k up to 80k.
8
78
315
RAGEval Proposes a simple framework to automatically generate evaluation datasets to assess knowledge usage of different LLM under different scenarios. It defines a schema from seed documents and then generates diverse documents which leads to question-answering pairs. The QA
4
108
500
Direct Preference Optimization (DPO) has become one of the go-to methods to align large language models (LLMs) more closely with user preferences. If you want to learn how it works, I coded it from scratch: https://t.co/VioT1zVn68
19
310
2K
Few #Bitcoin traders understand this: If you want to long BTC with leverage. DON'T buy futures, buy spot BTC with margin. EXPLAINER (1) Buying futures can be fulfilled by any counter trader who has USD collateral, together you are minting new synthetic BTC to the supply which
157
520
4K
Attention Instruction: Amplifying Attention in the Middle via Prompting Key findings: 1. LLMs lack relative position awareness 2. We can guide LLM to a specific region with position-based indexing Paper: https://t.co/G7u2lr2QOv Thanks to: @mengzaiqiao and @nigelhcollier
4
35
120
Cool paper proposing a graph-based agent system to enhance the long-context abilities of LLMs. It first structures long text into a graph (elements and facts) and employs an agent to explore the graph using predefined functions guided by a step-by-step rational plan. The agent
4
82
353
Large Language Models Must Be Taught to Know What They Don't Know abs: https://t.co/ccKTs39nu7 Prompting is not enough for LLMs to produce accurate estimates of its uncertainty of its responses, but can be finetuned with as little as 1000 examples and outperform baselines for
9
130
732
My amalik@twitter Twitter handle can now receive crypto. #CRYPTO Get your own $FIO Crypto Handle for free now! https://t.co/1mmCEW9EL4 via @joinfio
0
0
2
I just voted in @SweatEconomy ’s governance vote in the #SweatWallet app! Join me here https://t.co/yfBkCaQ8Ir via @SweatEconomy
0
0
0
0
7
45