reduce repetition penalty
This commit is contained in:
parent
0f69a0c19e
commit
926291940d
|
@ -53,7 +53,7 @@ def main():
|
||||||
"temperature": 0.95,
|
"temperature": 0.95,
|
||||||
"num_beams": 1,
|
"num_beams": 1,
|
||||||
"max_new_tokens": 256,
|
"max_new_tokens": 256,
|
||||||
"repetition_penalty": 1.5,
|
"repetition_penalty": 1.0,
|
||||||
"logits_processor": get_logits_processor()
|
"logits_processor": get_logits_processor()
|
||||||
}
|
}
|
||||||
with torch.no_grad():
|
with torch.no_grad():
|
||||||
|
|
|
@ -105,7 +105,7 @@ def predict(input, chatbot, max_length, top_p, temperature, history):
|
||||||
"temperature": temperature,
|
"temperature": temperature,
|
||||||
"num_beams": 1,
|
"num_beams": 1,
|
||||||
"max_length": max_length,
|
"max_length": max_length,
|
||||||
"repetition_penalty": 1.5,
|
"repetition_penalty": 1.0,
|
||||||
"logits_processor": get_logits_processor()
|
"logits_processor": get_logits_processor()
|
||||||
}
|
}
|
||||||
with torch.no_grad():
|
with torch.no_grad():
|
||||||
|
|
Loading…
Reference in New Issue