what's the dtype (precision) used during training?

#33
by keunwoochoi - opened

hi, thanks for the nice model.
when we get the model through transformers, the default data format is float32.

https://huggingface.co/nomic-ai/nomic-embed-text-v1.5/blob/main/config.json#L49

but the technical report says the model was done on bfloat16.

i'd like to confirm what is the correct dtype. asking because it is critical as the size doubles with float32.

any insight would be great. thanks!

Nomic AI org
edited Aug 26

hey @keunwoochoi the correct dtype is bf16 !

zpn changed discussion status to closed
Nomic AI org

To clarify, we use torch.autocast with bfloat16 precision during training

Nomic AI org

I've also run inference of the model in bfloat16 and found pretty similar results but YMMV @keunwoochoi

Sign up or log in to comment