Remove full_messages
Browse files
app.py
CHANGED
|
@@ -110,11 +110,9 @@ def chat_with_model(messages, pid):
|
|
| 110 |
pad_id = current_tokenizer.pad_token_id or current_tokenizer.unk_token_id or 0
|
| 111 |
eos_id = current_tokenizer.eos_token_id
|
| 112 |
|
| 113 |
-
# FULL conversation
|
| 114 |
-
full_messages = system_messages + messages
|
| 115 |
|
| 116 |
# --- Generate from full context
|
| 117 |
-
prompt = format_prompt(
|
| 118 |
|
| 119 |
device = torch.device("cuda")
|
| 120 |
current_model.to(device).half()
|
|
|
|
| 110 |
pad_id = current_tokenizer.pad_token_id or current_tokenizer.unk_token_id or 0
|
| 111 |
eos_id = current_tokenizer.eos_token_id
|
| 112 |
|
|
|
|
|
|
|
| 113 |
|
| 114 |
# --- Generate from full context
|
| 115 |
+
prompt = format_prompt(messages)
|
| 116 |
|
| 117 |
device = torch.device("cuda")
|
| 118 |
current_model.to(device).half()
|