Oleg Lavrovsky commited on
Commit
4b1aae8
·
unverified ·
1 Parent(s): 93fa881

Recommended params

Browse files
Files changed (1) hide show
  1. app.py +11 -2
app.py CHANGED
@@ -115,10 +115,19 @@ async def predict(q: str):
115
  tokenize=False,
116
  add_generation_prompt=True,
117
  )
118
- model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
 
 
 
 
119
 
120
  # Generate the output
121
- generated_ids = model.generate(**model_inputs, max_new_tokens=32768)
 
 
 
 
 
122
 
123
  # Get and decode the output
124
  output_ids = generated_ids[0][len(model_inputs.input_ids[0]) :]
 
115
  tokenize=False,
116
  add_generation_prompt=True,
117
  )
118
+ model_inputs = tokenizer(
119
+ [text],
120
+ return_tensors="pt",
121
+ add_special_tokens=False
122
+ ).to(model.device)
123
 
124
  # Generate the output
125
+ generated_ids = model.generate(
126
+ **model_inputs,
127
+ top_p=0.9,
128
+ temperature=0.8,
129
+ max_new_tokens=512
130
+ )
131
 
132
  # Get and decode the output
133
  output_ids = generated_ids[0][len(model_inputs.input_ids[0]) :]