Papers
arxiv:2408.11796

LLM Pruning and Distillation in Practice: The Minitron Approach

Published on Aug 21
ยท Submitted by pmolchanov on Aug 22
#2 Paper of the day
Authors:
,
,

Abstract

We present a comprehensive report on compressing the Llama 3.1 8B and Mistral NeMo 12B models to 4B and 8B parameters, respectively, using pruning and distillation. We explore two distinct pruning strategies: (1) depth pruning and (2) joint hidden/attention/MLP (width) pruning, and evaluate the results on common benchmarks from the LM Evaluation Harness. The models are then aligned with NeMo Aligner and tested in instruct-tuned versions. This approach produces a compelling 4B model from Llama 3.1 8B and a state-of-the-art Mistral-NeMo-Minitron-8B (MN-Minitron-8B for brevity) model from Mistral NeMo 12B. We found that with no access to the original data, it is beneficial to slightly fine-tune teacher models on the distillation dataset. We open-source our base model weights on Hugging Face with a permissive license.

Community

Paper submitter

Models are at https://huggingface.co/collections/nvidia/minitron-669ac727dc9c86e6ab7f0f3e
Quick demo of Base models (not Instruct): https://huggingface.co/spaces/nvidia/minitron
Examples and tutorial (WIP): https://github.com/NVlabs/Minitron
Authors are happy to answer your questions.

This is an automated message from the Librarian Bot. I found the following papers similar to this paper.

The following papers were recommended by the Semantic Scholar API

Please give a thumbs up to this comment if you found it helpful!

If you want recommendations for any Paper on Hugging Face checkout this Space

You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend

We wrote a summary about this paper and a few other papers here. Please do give it a read.
Contents:

  1. Nvidia MiniTron
  2. 1.5 Pints
  3. Jamba-1.5
  4. FocusLLM

https://datta0.substack.com/p/ai-unplugged-18-minitron-and-llama

Sign up or log in to comment

Models citing this paper 7

Browse 7 models citing this paper

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2408.11796 in a dataset README.md to link it from this page.

Spaces citing this paper 24

Collections including this paper 12