OPT RM
Collection
OPT reward models
•
3 items
•
Updated
This is a fine tuned OPT 1.3b model for reward modelling. The finetuning has been done on top of the full SLF5K dataset following the method presented in the paper Training Language Models with Language Feedback at Scale. The main results can be seen in the following table:
Model | # Params | Validation Accuracy (in %) |
---|---|---|
OPT LM Loss | 13B | 73.4 +/- 1.9 |
OPT LM Loss | 1.3B | 69.6 +/- 2.0 |
OPT RM Loss | 13B | 71.8 +/- 2.0 |
If using this model, please cite the following paper:
@article{scheurer2023training,
title={Training Language Models with Language Feedback at Scale},
author={Scheurer, J{\'e}r{\'e}my and Campos, Jon Ander and Korbak, Tomasz and Chan, Jun Shern and Chen, Angelica and Cho, Kyunghyun and Perez, Ethan},
journal={arXiv preprint arXiv:2303.16755},
year={2023}
}