From 45fd0841b89954569262e2cc55412fab474b2f39 Mon Sep 17 00:00:00 2001 From: zRzRzRzRzRzRzR <2448370773@qq.com> Date: Sat, 25 Oct 2025 21:27:31 +0800 Subject: [PATCH] update for new format config of transformers lib / sglang lib --- config.json | 68 ++++++++++++++++++++++-------------------- generation_config.json | 2 +- 2 files changed, 37 insertions(+), 33 deletions(-) diff --git a/config.json b/config.json index d301ef0..ec604c8 100644 --- a/config.json +++ b/config.json @@ -3,38 +3,50 @@ "Glm4vForConditionalGeneration" ], "model_type": "glm4v", - "attention_bias": true, - "attention_dropout": 0.0, - "pad_token_id": 151329, - "eos_token_id": [ - 151329, - 151336, - 151338, - 151348 - ], "image_start_token_id": 151339, "image_end_token_id": 151340, "video_start_token_id": 151341, "video_end_token_id": 151342, "image_token_id": 151343, "video_token_id": 151344, - "hidden_act": "silu", - "hidden_size": 4096, - "initializer_range": 0.02, - "intermediate_size": 13696, - "max_position_embeddings": 65536, - "num_attention_heads": 32, - "num_hidden_layers": 40, - "num_key_value_heads": 2, - "rms_norm_eps": 1e-05, - "rope_theta": 10000.0, "tie_word_embeddings": false, - "torch_dtype": "bfloat16", - "transformers_version": "4.53.0dev", - "use_cache": true, - "vocab_size": 151552, - "partial_rotary_factor": 0.5, + "transformers_version": "4.57.1", + "text_config": { + "model_type": "glm4v_text", + "attention_bias": true, + "attention_dropout": 0.0, + "pad_token_id": 151329, + "eos_token_id": [ + 151329, + 151336, + 151338, + 151348 + ], + "hidden_act": "silu", + "hidden_size": 4096, + "initializer_range": 0.02, + "intermediate_size": 13696, + "max_position_embeddings": 65536, + "num_attention_heads": 32, + "num_hidden_layers": 40, + "num_key_value_heads": 2, + "rms_norm_eps": 1e-05, + "dtype": "bfloat16", + "use_cache": true, + "vocab_size": 151552, + "partial_rotary_factor": 0.5, + "rope_theta": 10000, + "rope_scaling": { + "rope_type": "default", + "mrope_section": [ + 8, + 12, + 12 + ] + } + }, "vision_config": { + "model_type": "glm4v", "hidden_size": 1536, "depth": 24, "num_heads": 12, @@ -49,13 +61,5 @@ "rms_norm_eps": 1e-05, "spatial_merge_size": 2, "temporal_patch_size": 2 - }, - "rope_scaling": { - "type": "default", - "mrope_section": [ - 8, - 12, - 12 - ] } } \ No newline at end of file diff --git a/generation_config.json b/generation_config.json index 0540e78..51d9c1b 100644 --- a/generation_config.json +++ b/generation_config.json @@ -11,5 +11,5 @@ "top_p": 0.6, "temperature": 0.8, "top_k": 2, - "transformers_version": "4.53.0dev" + "transformers_version": "4.57.1" } \ No newline at end of file