Llama 3.1 trained on 9 000 000 of Claude Opus/Sonnet. Trained on 2 epoch, for 6 hours on 8x H100 NVL. Dataset used : - Norquinal/claude_multiround_chat_30k - kalomaze/Opus_Instruct_3k - mahiatlinux/Claude3-Opus-Instruct-ShareGPT-14k - kalomaze/Opus_Instruct_25k - meseca/opus-instruct-9k - Gryphe/Sonnet3.5-SlimOrcaDedupCleaned - Gryphe/Opus-WritingPrompts