File size: 681 Bytes
d37fab3
 
7978a78
fa4e210
7978a78
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
launch_mode: hf   # hf or local
hf_log_path: embodied-generalist/leo_demo_log
hf_ckpt_path: [huangjy-pku/embodied-generalist, weights/leo_noact_hf.pth]
local_ckpt_path: /mnt/huangjiangyong/leo/hf_assets/weights/leo_noact_lora_yesno.pth
model:
  name: LeoAgentLLM
  # vision modules omitted
  llm:
    name: Vicuna7B
    hf_cfg_path: huangjy-pku/vicuna-7b
    local_cfg_path: /mnt/huangjiangyong/vicuna-7b
    truncation_side: right
    prompt: ""
    max_out_len: 256
    max_context_len: 256   # for prompt_after_obj
    lora:
      flag: True
      rank: 16
      alpha: 16
      target_modules: [q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj]
      dropout: 0.0