Llama 3.1 8B Instruct trained on 9 000 000 Claude Opus/Sonnet tokens. Trained on 2 epoch, for 6 hours on 8x H100 NVL. Dataset used : - Norquinal/claude_multiround_chat_30k - kalomaze/Opus_Instruct_3k - mahiatlinux/Claude3-Opus-Instruct-ShareGPT-14k - kalomaze/Opus_Instruct_25k - meseca/opus-instruct-9k - Gryphe/Sonnet3.5-SlimOrcaDedupCleaned - Gryphe/Opus-WritingPrompts ![image/png](https://cdn-uploads.huggingface.co/production/uploads/63ab1241ad514ca8d1430003/gCAp_3Fdsq6xojlInnoQr.png)