@@ -50,7 +50,7 @@ class GptParams:
50
50
# If chat ended prematurely, append this to the conversation to fix it.
51
51
# Set to "\nUser:" etc.
52
52
# This is an alternative to input_prefix which always adds it, so it potentially duplicates "User:""
53
- fix_prefix : str = " "
53
+ fix_prefix : str = ""
54
54
output_postfix : str = ""
55
55
input_echo : bool = True ,
56
56
@@ -75,7 +75,7 @@ def gpt_params_parse(argv = None, params: Optional[GptParams] = None):
75
75
parser .add_argument ("--top_p" , type = float , default = 0.95 , help = "top-p samplin" ,dest = "top_p" )
76
76
parser .add_argument ("--top_k" , type = int , default = 40 , help = "top-k sampling" ,dest = "top_k" )
77
77
parser .add_argument ("--temp" , type = float , default = 0.80 , help = "temperature" ,dest = "temp" )
78
- parser .add_argument ("--n_predict" , type = int , default = 128 , help = "number of model parts " ,dest = "n_predict" )
78
+ parser .add_argument ("--n_predict" , type = int , default = 128 , help = "number of tokens to predict (-1 = infinity) " ,dest = "n_predict" )
79
79
parser .add_argument ("--repeat_last_n" , type = int , default = 64 , help = "last n tokens to consider for penalize " ,dest = "repeat_last_n" )
80
80
parser .add_argument ("--repeat_penalty" , type = float , default = 1.10 , help = "penalize repeat sequence of tokens" ,dest = "repeat_penalty" )
81
81
parser .add_argument ("-b" , "--batch_size" , type = int , default = 8 , help = "batch size for prompt processing" ,dest = "n_batch" )
0 commit comments