xianchaowu commited on
Commit
1ae6855
1 Parent(s): cd0f47e
Files changed (1) hide show
  1. usage.py +2 -2
usage.py CHANGED
@@ -13,7 +13,7 @@ import torch
13
 
14
  #import ipdb; ipdb.set_trace()
15
  cache_dir="/workspace/asr/peft/qlora"
16
- # TODO set this cache_dir to the path where you stored (or, want to store) llama2-7bhf model
17
 
18
  lazylora_dir=os.getcwd() # the path that contains 'adapter_config.json' and 'adapter_model.bin'
19
 
@@ -41,7 +41,7 @@ model = AutoModelForCausalLM.from_pretrained(
41
  )
42
  #model.print_trainable_parameters()
43
  print(sum(p.numel() for p in model.parameters()))
44
- # 6,671,979,520 -> half-size of 7B due to 4-bit loading
45
 
46
  model = PeftModel.from_pretrained(model, lazylora_dir)
47
  print('after adding lazy lora parameters:')
 
13
 
14
  #import ipdb; ipdb.set_trace()
15
  cache_dir="/workspace/asr/peft/qlora"
16
+ # TODO set this cache_dir to the path where you stored (or, want to store) llama2-13b-hf model
17
 
18
  lazylora_dir=os.getcwd() # the path that contains 'adapter_config.json' and 'adapter_model.bin'
19
 
 
41
  )
42
  #model.print_trainable_parameters()
43
  print(sum(p.numel() for p in model.parameters()))
44
+ # 6,671,979,520 -> half-size of 13B due to 4-bit loading
45
 
46
  model = PeftModel.from_pretrained(model, lazylora_dir)
47
  print('after adding lazy lora parameters:')