jaspercatapang commited on
Commit
a1664de
1 Parent(s): 8c6efff

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +50 -0
README.md ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ datasets:
3
+ - SirNeural/flan_v2
4
+ ---
5
+ # OPT-FLAN-IML-6.7B
6
+
7
+ This model is patterned after the methodology presented in the original [OPT-IML (OPT + Instruction Meta-Learning)](https://arxiv.org/abs/2212.12017) paper, made available for the 6.7 billion parameter size variant of OPT. However, this model is fine-tuned on FLAN v2.
8
+
9
+ ### How to use
10
+ According to Meta's model card for OPT-IML-30B, it is recommended to directly call the [`generate`](https://huggingface.co/docs/transformers/main/en/main_classes/text_generation#transformers.generation_utils.GenerationMixin.generate)
11
+ method as follows:
12
+
13
+ ```python
14
+ >>> from transformers import AutoModelForCausalLM, AutoTokenizer
15
+ >>> import torch
16
+
17
+ >>> model = AutoModelForCausalLM.from_pretrained("MayaPH/OPT-FLAN-IML-6.7B", torch_dtype=torch.float16).cuda()
18
+
19
+ >>> # the fast tokenizer currently does not work correctly
20
+ >>> tokenizer = AutoTokenizer.from_pretrained("MayaPH/OPT-FLAN-IML-6.7B", use_fast=False)
21
+
22
+ >>> prompt = "What is the color of the sea?\nA:"
23
+
24
+ >>> input_ids = tokenizer(prompt, return_tensors="pt").input_ids.cuda()
25
+
26
+ >>> generated_ids = model.generate(input_ids)
27
+
28
+ >>> tokenizer.batch_decode(generated_ids, skip_special_tokens=True)
29
+ ```
30
+
31
+ ## Further Information
32
+ For additional information or inquiries about OPT-FLAN-IML-6.7B, please contact the Maya Philippines iOps Team via [email protected].
33
+
34
+ ## Disclaimer
35
+ OPT-FLAN-IML-6.7B is an AI language model trained by Maya Philippines. It is provided "as is" without warranty of any kind, express or implied. The model developers and Maya Philippines shall not be liable for any direct or indirect damages arising from the use of this model.
36
+
37
+ ## Acknowledgments
38
+ The development of OPT-FLAN-IML-6.7B was made possible by Maya Philippines and the curation and creation of the instruction-following dataset, FLAN v2, was made possible by Google.
39
+
40
+ ### BibTeX entry and citation info
41
+ ```bibtex
42
+ @misc{iyer2022opt,
43
+ title={OPT-IML: Scaling Language Model Instruction Meta Learning through the Lens of Generalization},
44
+ author={Iyer, Srinivasan and Lin, Xi Victoria and Pasunuru, Ramakanth and Mihaylov, Todor and Simig, D{\'a}niel and Yu, Ping and Shuster, Kurt and Wang, Tianlu and Liu, Qing and Koura, Punit Singh and others},
45
+ year={2022},
46
+ eprint={2212.12017},
47
+ archivePrefix={arXiv},
48
+ primaryClass={cs.CL}
49
+ }
50
+ ```