qnguyen3 commited on
Commit
4d18fd2
1 Parent(s): 917a798

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +1 -1
app.py CHANGED
@@ -14,7 +14,7 @@ subprocess.run('pip install flash-attn --no-build-isolation', env={'FLASH_ATTENT
14
  torch.set_default_device('cuda')
15
 
16
  tokenizer = AutoTokenizer.from_pretrained(
17
- 'qnguyen3/nanoLLaVA',
18
  trust_remote_code=True)
19
 
20
  model = LlavaQwen2ForCausalLM.from_pretrained(
 
14
  torch.set_default_device('cuda')
15
 
16
  tokenizer = AutoTokenizer.from_pretrained(
17
+ 'qnguyen3/nanoLLaVA-1.5',
18
  trust_remote_code=True)
19
 
20
  model = LlavaQwen2ForCausalLM.from_pretrained(