inetnuc commited on
Commit
2670e02
1 Parent(s): 256dd2d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -8
README.md CHANGED
@@ -19,13 +19,14 @@ tags:
19
 
20
  This LLAMA-3 model was finetuned to enhance capabilities in text generation for nuclear-related topics. The training was accelerated using [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library, achieving a 2x faster performance.
21
 
22
- Finetuning Process
 
23
  The model was finetuned using the Unsloth library, leveraging its efficient training capabilities. The process included the following steps:
24
 
25
- Data Preparation: Loaded and preprocessed nuclear-related data.
26
- Model Loading: Utilized unsloth/llama-3-8b-bnb-4bit as the base model.
27
- LoRA Patching: Applied LoRA (Low-Rank Adaptation) for efficient training.
28
- Training: Finetuned the model using Hugging Face's TRL library with optimized hyperparameters.
29
 
30
  ## Model Details
31
 
@@ -50,6 +51,3 @@ model = AutoModelForCausalLM.from_pretrained("inetnuc/llama-3-8b-chat-nuclear-lo
50
  inputs = tokenizer("what is the iaea approach for cyber security?", return_tensors="pt")
51
  outputs = model.generate(**inputs, max_new_tokens=128)
52
  print(tokenizer.decode(outputs[0], skip_special_tokens=True))
53
-
54
-
55
-
 
19
 
20
  This LLAMA-3 model was finetuned to enhance capabilities in text generation for nuclear-related topics. The training was accelerated using [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library, achieving a 2x faster performance.
21
 
22
+ ## Finetuning Process
23
+
24
  The model was finetuned using the Unsloth library, leveraging its efficient training capabilities. The process included the following steps:
25
 
26
+ 1. **Data Preparation:** Loaded and preprocessed nuclear-related data.
27
+ 2. **Model Loading:** Utilized `unsloth/llama-3-8b-bnb-4bit` as the base model.
28
+ 3. **LoRA Patching:** Applied LoRA (Low-Rank Adaptation) for efficient training.
29
+ 4. **Training:** Finetuned the model using Hugging Face's TRL library with optimized hyperparameters.
30
 
31
  ## Model Details
32
 
 
51
  inputs = tokenizer("what is the iaea approach for cyber security?", return_tensors="pt")
52
  outputs = model.generate(**inputs, max_new_tokens=128)
53
  print(tokenizer.decode(outputs[0], skip_special_tokens=True))