@@ -91,7 +91,7 @@ class ChatConfig:
91
91
:class:`mlc_chat.ChatModule` instance to override the default setting in
92
92
``mlc-chat-config.json`` under the model folder.
93
93
94
- Since the configuraiton is partial, everything will be ``Optional``.
94
+ Since the configuration is partial, everything will be ``Optional``.
95
95
96
96
Note that we will exploit this class to also represent ``mlc-chat-config.json``
97
97
during intermediate processing.
@@ -131,14 +131,19 @@ class ChatConfig:
131
131
For additional information on top-p sampling, please refer to this blog
132
132
post: https://huggingface.co/blog/how-to-generate#top-p-nucleus-sampling.
133
133
mean_gen_len : Optional[int]
134
+ The approximated average number of generated tokens in each round. Used
135
+ to determine whether the maximum window size would be exceeded.
134
136
max_gen_len : Optional[int]
137
+ The maximum number of tokens to be generated in each round. Would simply
138
+ stop generating after this number is exceeded.
135
139
shift_fill_factor : Optional[float]
140
+ The fraction of maximum window size to shift when it is exceeded.
136
141
tokenizer_files : Optional[List[str]]
137
142
List of tokenizer files of the model.
138
143
conv_config : Optional[ConvConfig]
139
144
The partial overriding configuration for conversation template. Will first
140
145
load the predefined template with the name specified in ``conv_template``
141
- and then override some of the configuraitons specified in ``conv_config``.
146
+ and then override some of the configurations specified in ``conv_config``.
142
147
model_category : Optional[str]
143
148
The category of the model's architecture (e.g. ``llama``, ``gpt_neox``, ``rwkv``).
144
149
model_name : Optional[str]
@@ -216,7 +221,11 @@ class GenerationConfig:
216
221
For additional information on top-p sampling, please refer to this blog
217
222
post: https://huggingface.co/blog/how-to-generate#top-p-nucleus-sampling.
218
223
mean_gen_len : Optional[int]
224
+ The approximated average number of generated tokens in each round. Used
225
+ to determine whether the maximum window size would be exceeded.
219
226
max_gen_len : Optional[int]
227
+ The maximum number of tokens to be generated in each round. Would simply
228
+ stop generating after this number is exceeded.
220
229
"""
221
230
222
231
temperature : Optional [float ] = None
0 commit comments