Text Generation
Transformers
PyTorch
English
llama
conversational
text-generation-inference
Inference Endpoints
hamishivi commited on
Commit
f1a3088
1 Parent(s): f05fb21

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -6
README.md CHANGED
@@ -20,6 +20,8 @@ Tulu is a series of language models that are trained to act as helpful assistant
20
  Tulu V2 DPO 7B is a fine-tuned version of Llama 2 that was trained on on a mix of publicly available, synthetic and human datasets using [Direct Preference Optimization (DPO)](https://arxiv.org/abs/2305.18290).
21
  This model is a strong alternative to Llama 2 7b Chat.
22
 
 
 
23
 
24
  ## Model description
25
 
@@ -134,12 +136,13 @@ The following hyperparameters were used during DPO training:
134
  If you find Tulu 2 is useful in your work, please cite it with:
135
 
136
  ```
137
- @misc{ivison2023changing,
138
- title={Camels in a Changing Climate: Enhancing LM Adaptation with Tulu 2},
139
- author={Hamish Ivison and Yizhong Wang and Valentina Pyatkin and Nathan Lambert and Matthew Peters and Pradeep Dasigi and Joel Jang and David Wadden and Noah A. Smith and Iz Beltagy and Hannaneh Hajishirzi},
140
- year={2023},
141
- archivePrefix={arXiv},
142
- primaryClass={cs.CL}
 
143
  }
144
  ```
145
 
 
20
  Tulu V2 DPO 7B is a fine-tuned version of Llama 2 that was trained on on a mix of publicly available, synthetic and human datasets using [Direct Preference Optimization (DPO)](https://arxiv.org/abs/2305.18290).
21
  This model is a strong alternative to Llama 2 7b Chat.
22
 
23
+ For more details, read the paper: [Camels in a Changing Climate: Enhancing LM Adaptation with Tulu 2
24
+ ](https://arxiv.org/abs/2311.10702).
25
 
26
  ## Model description
27
 
 
136
  If you find Tulu 2 is useful in your work, please cite it with:
137
 
138
  ```
139
+ @misc{ivison2023camels,
140
+ title={Camels in a Changing Climate: Enhancing LM Adaptation with Tulu 2},
141
+ author={Hamish Ivison and Yizhong Wang and Valentina Pyatkin and Nathan Lambert and Matthew Peters and Pradeep Dasigi and Joel Jang and David Wadden and Noah A. Smith and Iz Beltagy and Hannaneh Hajishirzi},
142
+ year={2023},
143
+ eprint={2311.10702},
144
+ archivePrefix={arXiv},
145
+ primaryClass={cs.CL}
146
  }
147
  ```
148