Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -34,18 +34,14 @@ class ChatBot:
|
|
34 |
user_input_ids = tokenizer.encode(input + tokenizer.eos_token, return_tensors="pt")
|
35 |
|
36 |
# Concatenate the user input with chat history
|
37 |
-
if self.history:
|
38 |
chat_history_ids = torch.cat([self.history, user_input_ids], dim=-1)
|
39 |
else:
|
40 |
chat_history_ids = user_input_ids
|
41 |
|
42 |
# Generate a response using the PEFT model
|
43 |
-
# response = peft_model.generate(chat_history_ids, max_length=512, pad_token_id=tokenizer.eos_token_id)
|
44 |
-
# response = peft_model.generate(chat_history_ids)
|
45 |
response = peft_model.generate(input_ids=chat_history_ids, max_length=512, pad_token_id=tokenizer.eos_token_id)
|
46 |
|
47 |
-
|
48 |
-
|
49 |
# Update chat history
|
50 |
self.history = chat_history_ids
|
51 |
|
|
|
34 |
user_input_ids = tokenizer.encode(input + tokenizer.eos_token, return_tensors="pt")
|
35 |
|
36 |
# Concatenate the user input with chat history
|
37 |
+
if self.history.numel() > 0:
|
38 |
chat_history_ids = torch.cat([self.history, user_input_ids], dim=-1)
|
39 |
else:
|
40 |
chat_history_ids = user_input_ids
|
41 |
|
42 |
# Generate a response using the PEFT model
|
|
|
|
|
43 |
response = peft_model.generate(input_ids=chat_history_ids, max_length=512, pad_token_id=tokenizer.eos_token_id)
|
44 |
|
|
|
|
|
45 |
# Update chat history
|
46 |
self.history = chat_history_ids
|
47 |
|