Llama 3.1 8B Instruct trained on 9 000 000 Claude Opus/Sonnet tokens.
Trained on 2 epoch, for 6 hours on 8x H100 NVL.
Dataset used :
- Norquinal/claude_multiround_chat_30k
- kalomaze/Opus_Instruct_3k
- mahiatlinux/Claude3-Opus-Instruct-ShareGPT-14k
- kalomaze/Opus_Instruct_25k
- meseca/opus-instruct-9k
- Gryphe/Sonnet3.5-SlimOrcaDedupCleaned
- Gryphe/Opus-WritingPrompts