Update app.py
Browse files
app.py
CHANGED
|
@@ -110,7 +110,7 @@ def engine(input_embeds):
|
|
| 110 |
batch_input_embeds = input_embeds[start:start+batch_size]
|
| 111 |
outputs = m.generate(
|
| 112 |
inputs_embeds = batch_input_embeds.to("cuda"),
|
| 113 |
-
max_new_tokens =
|
| 114 |
do_sample = True,
|
| 115 |
temperature = 0.6,
|
| 116 |
top_p = 0.9,
|
|
@@ -126,7 +126,7 @@ def chat_engine(input_ids):
|
|
| 126 |
with torch.no_grad():
|
| 127 |
outputs = m.generate(
|
| 128 |
input_ids = input_ids.to("cuda"),
|
| 129 |
-
max_new_tokens =
|
| 130 |
do_sample = True,
|
| 131 |
temperature = 0.6,
|
| 132 |
top_p = 0.9,
|
|
|
|
| 110 |
batch_input_embeds = input_embeds[start:start+batch_size]
|
| 111 |
outputs = m.generate(
|
| 112 |
inputs_embeds = batch_input_embeds.to("cuda"),
|
| 113 |
+
max_new_tokens = 16,
|
| 114 |
do_sample = True,
|
| 115 |
temperature = 0.6,
|
| 116 |
top_p = 0.9,
|
|
|
|
| 126 |
with torch.no_grad():
|
| 127 |
outputs = m.generate(
|
| 128 |
input_ids = input_ids.to("cuda"),
|
| 129 |
+
max_new_tokens = 512,
|
| 130 |
do_sample = True,
|
| 131 |
temperature = 0.6,
|
| 132 |
top_p = 0.9,
|