Spaces:
Runtime error
Runtime error
Oleg Lavrovsky
commited on
Recommended params
Browse files
app.py
CHANGED
|
@@ -115,10 +115,19 @@ async def predict(q: str):
|
|
| 115 |
tokenize=False,
|
| 116 |
add_generation_prompt=True,
|
| 117 |
)
|
| 118 |
-
model_inputs = tokenizer(
|
|
|
|
|
|
|
|
|
|
|
|
|
| 119 |
|
| 120 |
# Generate the output
|
| 121 |
-
generated_ids = model.generate(
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 122 |
|
| 123 |
# Get and decode the output
|
| 124 |
output_ids = generated_ids[0][len(model_inputs.input_ids[0]) :]
|
|
|
|
| 115 |
tokenize=False,
|
| 116 |
add_generation_prompt=True,
|
| 117 |
)
|
| 118 |
+
model_inputs = tokenizer(
|
| 119 |
+
[text],
|
| 120 |
+
return_tensors="pt",
|
| 121 |
+
add_special_tokens=False
|
| 122 |
+
).to(model.device)
|
| 123 |
|
| 124 |
# Generate the output
|
| 125 |
+
generated_ids = model.generate(
|
| 126 |
+
**model_inputs,
|
| 127 |
+
top_p=0.9,
|
| 128 |
+
temperature=0.8,
|
| 129 |
+
max_new_tokens=512
|
| 130 |
+
)
|
| 131 |
|
| 132 |
# Get and decode the output
|
| 133 |
output_ids = generated_ids[0][len(model_inputs.input_ids[0]) :]
|