Skip to content

Commit 2b201af

Browse files
authored
fix some issues of qwen2.5_vl (#631)
1 parent 33191bd commit 2b201af

2 files changed

Lines changed: 2 additions & 2 deletions

File tree

examples/qwen2_5_vl/pretrain_qwen.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -83,7 +83,7 @@ def model_provider(
8383

8484
vision_config = get_vision_model_config(args, deepcopy(config))
8585
vision_config.pipeline_model_parallel_size = 1
86-
vision_config.first_pipeline_num_layers = None
86+
vision_config.num_layers_in_first_pipeline_stage = None
8787
vision_projector_config = get_vision_projection_config(deepcopy(config), vision_config.hidden_size, args.spatial_merge_size)
8888

8989
print_rank_0("building Qwen2-5-VL model in TE...")

toolkits/model_checkpoints_convertor/utils/__init__.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -113,7 +113,7 @@ def build_layer_id_mapping(args):
113113
remained_stages = args.pipeline_model_parallel_size - 1
114114
assert remained_layers % remained_stages == 0
115115
pp_layers_per_stage = [args.target_decoder_first_pipeline_num_layers] +([remained_layers // remained_stages] * remained_stages)
116-
116+
offset = 0
117117
for pp_id, num_layers in enumerate(pp_layers_per_stage):
118118
for global_layer_id in range(offset, offset + num_layers):
119119
# NOTE: map a global transformer layer to a local pp rank

0 commit comments

Comments
 (0)