From f5244dd8748e57036650dedb5ab7e6b6732c09fc Mon Sep 17 00:00:00 2001 From: Cherrytest Date: Wed, 5 Mar 2025 16:27:05 +0000 Subject: [PATCH] Update config.json --- config.json | 10 ++++------ 1 file changed, 4 insertions(+), 6 deletions(-) diff --git a/config.json b/config.json index 4484b10..315d9c2 100644 --- a/config.json +++ b/config.json @@ -1,5 +1,4 @@ { - "_name_or_path": "/cpfs01/data/shared/Group-m6/chuanyi.yl/ckpts/sft/dense/hf-ckpts/32b.stage2--cptv9-base100w-cpt32k-0218_Ma_Co_STE_Rea_Age_Gen_dgb-32B.qwen2.5B-bf16-mp8-pp1-lr-7e-6-minlr-7e-7-bs-128-gpus-128-seqlen-32768-step8392", "architectures": [ "Qwen2ForCausalLM" ], @@ -10,20 +9,19 @@ "hidden_size": 5120, "initializer_range": 0.02, "intermediate_size": 27648, - "max_position_embeddings": 65536, + "max_position_embeddings": 131072, "max_window_layers": 64, "model_type": "qwen2", "num_attention_heads": 40, "num_hidden_layers": 64, "num_key_value_heads": 8, "rms_norm_eps": 1e-05, - "rope_scaling": null, "rope_theta": 1000000.0, - "sliding_window": null, + "sliding_window": 32768, "tie_word_embeddings": false, "torch_dtype": "bfloat16", - "transformers_version": "4.45.2", + "transformers_version": "4.43.1", "use_cache": true, "use_sliding_window": false, "vocab_size": 152064 -} +} \ No newline at end of file