labmlai Profile Banner
labml.ai Profile
labml.ai

@labmlai

Followers
13K
Following
593
Media
204
Statuses
649

📝 Annotated paper implementations https://t.co/qeO4UTbrJ3

Joined December 2020
Don't wanna be here? Send us removal request.
@labmlai
labml.ai
2 months
RT @vpj: The following scripts stalls and times out on B200 x 8. Seems like we are having problems with NCCL. Anyone else experiencing this….
0
1
0
@labmlai
labml.ai
4 months
RT @vpj: The new training also improved GPQA from 64.2% to 67.3% and MMLU Pro from 64.2% to 67.3%. This model was also trained with the sa….
0
6
0
@labmlai
labml.ai
4 months
RT @notbadai: We are releasing an updated reasoning model with improvements on IFEval scores (77.9%) than our previous model (only 51.4%).….
0
6
0
@labmlai
labml.ai
4 months
RT @notbadai: We just released a Python coding reasoning dataset with 200k samples on @huggingface. This was generated by our RL-based self….
0
7
0
@labmlai
labml.ai
4 months
RT @vpj: Uploaded the dataset of 270k math reasoning samples that we used to finetune Notbad v1.0 Mistral 24B (MATH-500=77.52% GSM8k Platin….
0
13
0
@labmlai
labml.ai
4 months
RT @notbadai: We're open-sourcing a math reasoning dataset with 270k samples, generated by our RL-based self-improved Mistral 24B 2501 mode….
Tweet card summary image
huggingface.co
0
4
0
@labmlai
labml.ai
4 months
RT @notbadai: 📢 We are excited to announce Notbad v1.0 Mistral 24B, a new reasoning model trained in math and Python coding. This model is….
0
8
0
@labmlai
labml.ai
4 months
Try it on
Tweet card summary image
chat.labml.ai
NotBadAI models generate shorter, cleaner reasoning outputs through self-improved capabilities, independently developed without distillation from other models.
@notbadai
NOTBAD AI
4 months
📢 We are excited to announce Notbad v1.0 Mistral 24B, a new reasoning model trained in math and Python coding. This model is built upon the @MistralAI Small 24B 2501 and has been further trained with reinforcement learning on math and coding.
Tweet media one
Tweet media two
Tweet media three
0
0
6
@labmlai
labml.ai
10 months
@luck_not_shit
hehehehe
10 months
Now Our visualization library Inspectus can visualize values related to tokens in LLM outputs. This demo shows some outputs from using entropyx (by @_xjdr) on Llama 3. Had fun making this. (jk I didn’t). 🔗👇
0
0
6
@labmlai
labml.ai
10 months
We added token visualization to Inspectus. It lets you visualize metrics associated with tokens such as loss, entropy, KL div, etc. It works on notebooks and pretty easy to use. 👇.
2
4
28
@labmlai
labml.ai
10 months
Our open source deep learning experiment monitoring library now has 2000 stars! Thank you
Tweet media one
1
3
18
@labmlai
labml.ai
1 year
RT @notbadai: We’ve been training @nvidia Mistral-NeMo-Minitron-8B-Base for math reasoning on the GSM8K-Aug dataset, and we have a version….
0
9
0
@labmlai
labml.ai
1 year
2/2) We have included sample script to fine-tune GPT-2 model using LoRA. Code to finetune GPT-2 with LoRA: Colab:
Tweet card summary image
colab.research.google.com
Run, share, and edit Python notebooks
0
2
9
@labmlai
labml.ai
1 year
Annotated @PyTorch implementation of of LoRA (Low Lank Adaptation of LLMs). 📝 Code + Notes: 📎 Paper: LoRA freezes the pre-trained model and trains smaller injected weights, enabling faster and memory efficient fine-tuning. 👇
Tweet media one
2
22
99
@labmlai
labml.ai
1 year
We should be able to release an update of labml experiment monitoring library very soon 😂. It has a bunch of cool new features.
@luck_not_shit
hehehehe
1 year
🙂‍↕️.When you work from home: when you work from office:
Tweet media one
Tweet media two
0
3
4
@labmlai
labml.ai
1 year
RT @luck_not_shit: I tried this with MNIST. In the loss curve the mean (dashed line) is much higher than the median. After plotting with bo….
0
1
0
@labmlai
labml.ai
1 year
RT @vpj: I first found plotting the distribution useful when I was trying RL algorithms on Atari around 2018/19. I used Tensorboard back th….
0
2
0