omitakahiro
commited on
Commit
•
c8f1288
1
Parent(s):
e72c788
Update README.md
Browse files
README.md
CHANGED
@@ -43,6 +43,10 @@ output = tokenizer.decode(tokens[0], skip_special_tokens=True)
|
|
43 |
print(output)
|
44 |
```
|
45 |
|
|
|
|
|
|
|
|
|
46 |
## Training dataset
|
47 |
- Japanese Web Corpus (ja): 8.6B tokens (This dataset will not be released.)
|
48 |
- Wikipedia (ja): 0.88B tokens
|
|
|
43 |
print(output)
|
44 |
```
|
45 |
|
46 |
+
## Example:
|
47 |
+
|
48 |
+
- LoRA tuning: https://huggingface.co/stockmark/gpt-neox-japanese-1.4b/blob/main/notebooks/LoRA.ipynb
|
49 |
+
|
50 |
## Training dataset
|
51 |
- Japanese Web Corpus (ja): 8.6B tokens (This dataset will not be released.)
|
52 |
- Wikipedia (ja): 0.88B tokens
|