
labml.ai
@labmlai
Followers
13K
Following
593
Media
204
Statuses
649
📝 Annotated paper implementations https://t.co/qeO4UTbrJ3
Joined December 2020
You can now download the Notbad v1.0 Mistral 24B model from @huggingface . Try it on
chat.labml.ai
NotBadAI models generate shorter, cleaner reasoning outputs through self-improved capabilities, independently developed without distillation from other models.
0
1
9
RT @notbadai: We just released a Python coding reasoning dataset with 200k samples on @huggingface. This was generated by our RL-based self….
0
7
0
RT @notbadai: We're open-sourcing a math reasoning dataset with 270k samples, generated by our RL-based self-improved Mistral 24B 2501 mode….
huggingface.co
0
4
0
Try it on
chat.labml.ai
NotBadAI models generate shorter, cleaner reasoning outputs through self-improved capabilities, independently developed without distillation from other models.
📢 We are excited to announce Notbad v1.0 Mistral 24B, a new reasoning model trained in math and Python coding. This model is built upon the @MistralAI Small 24B 2501 and has been further trained with reinforcement learning on math and coding.
0
0
6
Now Our visualization library Inspectus can visualize values related to tokens in LLM outputs. This demo shows some outputs from using entropyx (by @_xjdr) on Llama 3. Had fun making this. (jk I didn’t). 🔗👇
0
0
6
Quick star: GitHub:
labml.ai
Organize machine learning experiments and monitor training progress from mobile.
0
0
4
2/2) We have included sample script to fine-tune GPT-2 model using LoRA. Code to finetune GPT-2 with LoRA: Colab:
colab.research.google.com
Run, share, and edit Python notebooks
0
2
9
RT @luck_not_shit: I tried this with MNIST. In the loss curve the mean (dashed line) is much higher than the median. After plotting with bo….
0
1
0