synthetic GPT-3.5 data works amazing for fine-tuning smaller models!
generated a 13,000 Q&A dataset on the Aztec Empire in ten minutes. answers were designed to be sarcastic, upper-case, concise, and have an [END] token.
#llama
-2-7b (base) was able to learn all these guidelines…