jamba-small-v1 / README.md
OxxoCodes's picture
Update README.md
12155e1 verified
|
raw
history blame
1.28 kB
metadata
library_name: transformers
tags: []

Jamba-Small v1

This is a pruned version of AI21 Labs' Jamba-v0.1 model that is ~25% the size of Jamba-v0.1.

Model Details

Whereas Jamba-v0.1 contains 4 Jamba blocks, Jamba-Small contains only 1 Jamba block. Jamba-Small's Jamba blocks follow the same structure seen in Jamba-v0.1, with a 1:7 ratio of attention-to-Mamba layers and MoE applied every 2 layers.

Jamba-Small's weights are initialized from various layers in the original Jamba-v0.1 model. For v1, the layer weights are mapped as follows (left is Jamba-Small layer number, right is Jamba-v0.1 layer number):

0: 0
1: 1
2: 2
3: 3
4: 4
5: 5
6: 30
7: 31

Note that no additional fine-tuning has been performed on this model. As such, its performance is exceptionally poor. This should not be used in production without additional training.

Model Description

  • Developed by: Nathan Brown (OxxoCodes)
  • Compute provided by [optional]: Clemson Palmetto Cluster
  • Model type: Joint Attention and Mamba (Jamba)
  • Language(s) (NLP): English
  • License: Apache 2.0
  • Original model: Jamba-v0.1
  • Jamba paper: https://arxiv.org/pdf/2403.19887.pdf