aashish1904 commited on
Commit
d20e08a
1 Parent(s): 8d5e613

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +34 -0
README.md ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+
4
+ license: apache-2.0
5
+ language:
6
+ - en
7
+ - hi
8
+ metrics:
9
+ - perplexity
10
+ base_model: meta-llama/Llama-2-7b-hf
11
+ pipeline_tag: text-generation
12
+ library_name: transformers
13
+ tags:
14
+ - code
15
+ datasets:
16
+ - zicsx/mC4-Hindi-Cleaned-3.0
17
+
18
+ ---
19
+
20
+ ![](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)
21
+
22
+ # QuantFactory/Llama2-7B-Hindi-finetuned-GGUF
23
+ This is quantized version of [subhrokomol/Llama2-7B-Hindi-finetuned](https://huggingface.co/subhrokomol/Llama2-7B-Hindi-finetuned) created using llama.cpp
24
+
25
+ # Original Model Card
26
+
27
+
28
+
29
+ # Finetune Llama-2-7B-hf on Hindi dataset after transtokenization
30
+ This model was trained on 24GB of RTX A500 on zicsx/mC4-Hindi-Cleaned-3.0 dataset (1%) for 3 hours.
31
+
32
+ We used Hugging Face PEFT-LoRA PyTorch for training.
33
+
34
+ Transtokenization process in --