labmlai Profile Banner
labml.ai Profile
labml.ai

@labmlai

Followers
13K
Following
604
Media
204
Statuses
654

๐Ÿ“ Annotated paper implementations https://t.co/qeO4UTbrJ3

Joined December 2020
Don't wanna be here? Send us removal request.
@nothiingf4
Himanshu Singh
20 days
As an ML engineer, implementation >>>>everything. Knowing is theory. Implementation is understanding. Few outstanding topics it has: 1. Reinforcement Learning - ppo, dqn 2. Transformer - classical to Retro, switch, gpt models 3. Diffusion models - stable, DDPM, DDIM, UNET 4.
16
58
612
@notbadai
NOTBAD AI
2 months
We've open-sourced our internal AI coding IDE. We built this IDE to help with coding and to experiment with custom AI workflows. It's based on a flexible extension system, making it easy to develop, test, and tweak new ideas quickly. Each extension is a Python script that runs
2
5
11
@vpj
vpj
3 months
GEPA appears to be an effective method for enhancing LLM performance, requiring significantly fewer rollouts than reinforcement learning (RL). It maintains a pool of system prompts. It uses an the LLM to improve them by reflecting on the generated answers and the scores/feedback
2
4
17
@vpj
vpj
4 months
Wrote an annotated Triton implementation of Flash Attention 2. (Links in reply) This is based on the flash attention implementation by the Triton team. Changed it to support GQA and cleaned up a little bit. Check it out to read the code for forward and backward passes along
3
9
48
@vpj
vpj
4 months
Added the JAX transformer model to annotated paper implementations project. https://t.co/PBFbvaeVY3 Link ๐Ÿ‘‡
@vpj
vpj
4 years
Coded a transformer model in JAX from scratch. This was my first time with JAX so it might have mistakes. https://t.co/J0AE8RH7zn This doesn't using any high-level frameworks such as Flax. ๐Ÿงต๐Ÿ‘‡
1
4
24
@vpj
vpj
7 months
The following scripts stalls and times out on B200 x 8. Seems like we are having problems with NCCL. Anyone else experiencing this? @PyTorch
1
1
4
@vpj
vpj
8 months
The new training also improved GPQA from 64.2% to 67.3% and MMLU Pro from 64.2% to 67.3%. This model was also trained with the same reasoning datasets we used to train the v1.0 model. We mixed more general instruction data with answers sampled from the
@notbadai
NOTBAD AI
8 months
We are releasing an updated reasoning model with improvements on IFEval scores (77.9%) than our previous model (only 51.4%). ๐Ÿ‘‡ Links to try the model and to download weights below
1
6
7
@notbadai
NOTBAD AI
8 months
We are releasing an updated reasoning model with improvements on IFEval scores (77.9%) than our previous model (only 51.4%). ๐Ÿ‘‡ Links to try the model and to download weights below
1
6
10
@notbadai
NOTBAD AI
8 months
We just released a Python coding reasoning dataset with 200k samples on @huggingface This was generated by our RL-based self-improved Mistral 24B 2501 model. This dataset was used to train train Notbad v1.0 Mistral 24B. ๐Ÿค— Links in replies ๐Ÿ‘‡
2
7
19
@vpj
vpj
8 months
Uploaded the dataset of 270k math reasoning samples that we used to finetune Notbad v1.0 Mistral 24B (MATH-500=77.52% GSM8k Platinum=97.55%) to @huggingface (link in reply) Follow @notbadai for updates
9
13
62
@notbadai
NOTBAD AI
8 months
We're open-sourcing a math reasoning dataset with 270k samples, generated by our RL-based self-improved Mistral 24B 2501 model and used to train Notbad v1.0 Mistral 24B. Available on Hugging Face:
Tweet card summary image
huggingface.co
0
4
9
@notbadai
NOTBAD AI
9 months
๐Ÿ“ข We are excited to announce Notbad v1.0 Mistral 24B, a new reasoning model trained in math and Python coding. This model is built upon the @MistralAI Small 24B 2501 and has been further trained with reinforcement learning on math and coding.
1
8
23
@labmlai
labml.ai
9 months
Try it on
Tweet card summary image
chat.labml.ai
NotBadAI models generate shorter, cleaner reasoning outputs through self-improved capabilities, independently developed without distillation from other models.
@notbadai
NOTBAD AI
9 months
๐Ÿ“ข We are excited to announce Notbad v1.0 Mistral 24B, a new reasoning model trained in math and Python coding. This model is built upon the @MistralAI Small 24B 2501 and has been further trained with reinforcement learning on math and coding.
0
0
6
@labmlai
labml.ai
1 year
@luck_not_shit
hehehehe
1 year
Now Our visualization library Inspectus can visualize values related to tokens in LLM outputs. This demo shows some outputs from using entropyx (by @_xjdr) on Llama 3. Had fun making this. (jk I didnโ€™t) ๐Ÿ”—๐Ÿ‘‡
0
0
6
@labmlai
labml.ai
1 year
We added token visualization to Inspectus. It lets you visualize metrics associated with tokens such as loss, entropy, KL div, etc. It works on notebooks and pretty easy to use. ๐Ÿ‘‡ https://t.co/uWLPkorrtN
1
4
27
@labmlai
labml.ai
1 year
Our open source deep learning experiment monitoring library now has 2000 stars! Thank you
1
3
18
@notbadai
NOTBAD AI
1 year
Weโ€™ve been training @nvidia Mistral-NeMo-Minitron-8B-Base for math reasoning on the GSM8K-Aug dataset, and we have a version with a 70.2% gsm8k score, up from a 58.5% cot score (reported in the paper LLM Pruning and distillation). ๐Ÿ‘‡
1
9
19