Update model card
Browse files
README.md
CHANGED
@@ -14,8 +14,8 @@ Proc-RoBERTa is a pre-trained language model for procedural text. It was built b
|
|
14 |
@inproceedings{bai-etal-2021-pre,
|
15 |
title = "Pre-train or Annotate? Domain Adaptation with a Constrained Budget",
|
16 |
author = "Bai, Fan and
|
17 |
-
|
18 |
-
|
19 |
booktitle = "Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing",
|
20 |
month = nov,
|
21 |
year = "2021",
|
|
|
14 |
@inproceedings{bai-etal-2021-pre,
|
15 |
title = "Pre-train or Annotate? Domain Adaptation with a Constrained Budget",
|
16 |
author = "Bai, Fan and
|
17 |
+
Ritter, Alan and
|
18 |
+
Xu, Wei",
|
19 |
booktitle = "Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing",
|
20 |
month = nov,
|
21 |
year = "2021",
|