kz209 commited on
Commit
8e22bd4
·
1 Parent(s): 331a464
Files changed (1) hide show
  1. utils/model.py +1 -1
utils/model.py CHANGED
@@ -56,7 +56,7 @@ class Model(torch.nn.Module):
56
  sampling_params = SamplingParams(
57
  temperature=temp,
58
  max_tokens=max_length,
59
- top_p=0.95 if do_sample else 1.0,
60
  stop_token_ids=[self.tokenizer.eos_token_id]
61
  )
62
  outputs = self.llm.generate(content_list, sampling_params)
 
56
  sampling_params = SamplingParams(
57
  temperature=temp,
58
  max_tokens=max_length,
59
+ #top_p=0.95 if do_sample else 1.0,
60
  stop_token_ids=[self.tokenizer.eos_token_id]
61
  )
62
  outputs = self.llm.generate(content_list, sampling_params)