update configs
Browse files- config.json +1 -1
- generation_config.json +2 -1
- modeling_qwen.py +8 -2
config.json
CHANGED
|
@@ -23,7 +23,7 @@
|
|
| 23 |
"n_embd": 4096,
|
| 24 |
"n_head": 32,
|
| 25 |
"n_layer": 32,
|
| 26 |
-
"n_positions":
|
| 27 |
"no_bias": true,
|
| 28 |
"onnx_safe": null,
|
| 29 |
"padded_vocab_size": 151936,
|
|
|
|
| 23 |
"n_embd": 4096,
|
| 24 |
"n_head": 32,
|
| 25 |
"n_layer": 32,
|
| 26 |
+
"n_positions": 8192,
|
| 27 |
"no_bias": true,
|
| 28 |
"onnx_safe": null,
|
| 29 |
"padded_vocab_size": 151936,
|
generation_config.json
CHANGED
|
@@ -1,8 +1,9 @@
|
|
| 1 |
{
|
| 2 |
"chat_format": "chatml",
|
| 3 |
"eos_token_id": 151643,
|
| 4 |
-
"max_new_tokens": 512,
|
| 5 |
"pad_token_id": 151643,
|
|
|
|
|
|
|
| 6 |
"do_sample": true,
|
| 7 |
"top_k": 0,
|
| 8 |
"top_p": 0.5,
|
|
|
|
| 1 |
{
|
| 2 |
"chat_format": "chatml",
|
| 3 |
"eos_token_id": 151643,
|
|
|
|
| 4 |
"pad_token_id": 151643,
|
| 5 |
+
"max_window_size": 6144,
|
| 6 |
+
"max_new_tokens": 512,
|
| 7 |
"do_sample": true,
|
| 8 |
"top_k": 0,
|
| 9 |
"top_p": 0.5,
|
modeling_qwen.py
CHANGED
|
@@ -994,12 +994,15 @@ class QWenLMHeadModel(QWenPreTrainedModel):
|
|
| 994 |
if stop_words_ids is None:
|
| 995 |
stop_words_ids = []
|
| 996 |
|
|
|
|
|
|
|
|
|
|
| 997 |
raw_text, context_tokens = make_context(
|
| 998 |
tokenizer,
|
| 999 |
query,
|
| 1000 |
history=history,
|
| 1001 |
system=system,
|
| 1002 |
-
max_window_size=
|
| 1003 |
chat_format=self.generation_config.chat_format,
|
| 1004 |
)
|
| 1005 |
|
|
@@ -1045,12 +1048,15 @@ class QWenLMHeadModel(QWenPreTrainedModel):
|
|
| 1045 |
if stop_words_ids is None:
|
| 1046 |
stop_words_ids = []
|
| 1047 |
|
|
|
|
|
|
|
|
|
|
| 1048 |
raw_text, context_tokens = make_context(
|
| 1049 |
tokenizer,
|
| 1050 |
query,
|
| 1051 |
history=history,
|
| 1052 |
system=system,
|
| 1053 |
-
max_window_size=
|
| 1054 |
chat_format=self.generation_config.chat_format,
|
| 1055 |
)
|
| 1056 |
|
|
|
|
| 994 |
if stop_words_ids is None:
|
| 995 |
stop_words_ids = []
|
| 996 |
|
| 997 |
+
max_window_size = kwargs.get('max_window_size', None)
|
| 998 |
+
if max_window_size is None:
|
| 999 |
+
max_window_size = self.generation_config.max_window_size
|
| 1000 |
raw_text, context_tokens = make_context(
|
| 1001 |
tokenizer,
|
| 1002 |
query,
|
| 1003 |
history=history,
|
| 1004 |
system=system,
|
| 1005 |
+
max_window_size=max_window_size,
|
| 1006 |
chat_format=self.generation_config.chat_format,
|
| 1007 |
)
|
| 1008 |
|
|
|
|
| 1048 |
if stop_words_ids is None:
|
| 1049 |
stop_words_ids = []
|
| 1050 |
|
| 1051 |
+
max_window_size = kwargs.get('max_window_size', None)
|
| 1052 |
+
if max_window_size is None:
|
| 1053 |
+
max_window_size = self.generation_config.max_window_size
|
| 1054 |
raw_text, context_tokens = make_context(
|
| 1055 |
tokenizer,
|
| 1056 |
query,
|
| 1057 |
history=history,
|
| 1058 |
system=system,
|
| 1059 |
+
max_window_size=max_window_size,
|
| 1060 |
chat_format=self.generation_config.chat_format,
|
| 1061 |
)
|
| 1062 |
|