Skip to content

Commit abba29b

Browse files
[BugFix] fix mm rope (#7274)
1 parent 8f21c9c commit abba29b

2 files changed

Lines changed: 5 additions & 5 deletions

File tree

fastdeploy/model_executor/layers/rotary_embedding.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -458,12 +458,12 @@ def __call__(self, position_ids, max_len_lst, cumsum_seqlens):
458458

459459
# Build position_ids_3d: [bsz, max_position, 3]
460460
position_ids_3d = paddle.tile(
461-
paddle.arange(self.max_position, dtype="int64").unsqueeze(0).unsqueeze(-1),
461+
paddle.arange(self.max_position, dtype="float32").unsqueeze(0).unsqueeze(-1),
462462
[bsz, 1, 3],
463463
)
464464
for i in range(bsz):
465465
position_ids_cur = position_ids[cumsum_seqlens[i] : cumsum_seqlens[i + 1]]
466-
prefix_max_position_ids = paddle.max(position_ids_cur) + 1
466+
prefix_max_position_ids = paddle.max(position_ids_cur[..., 0]) + 1
467467
dec_pos_ids = paddle.tile(
468468
paddle.arange(max_len_lst[i], dtype="int64").unsqueeze(-1),
469469
[1, 3],
@@ -530,12 +530,12 @@ def __call__(self, position_ids, max_len_lst, cumsum_seqlens):
530530
bsz = len(cumsum_seqlens) - 1
531531
# position_ids_3d: [bsz, seq_len, 3]
532532
position_ids_3d = paddle.tile(
533-
paddle.arange(self.max_position, dtype="int64").unsqueeze(0).unsqueeze(-1),
533+
paddle.arange(self.max_position, dtype="float32").unsqueeze(0).unsqueeze(-1),
534534
[bsz, 1, 3],
535535
)
536536
for i in range(bsz):
537537
position_ids_cur = position_ids[cumsum_seqlens[i] : cumsum_seqlens[i + 1]]
538-
prefix_max_position_ids = paddle.max(position_ids_cur) + 1
538+
prefix_max_position_ids = paddle.max(position_ids_cur[..., 0]) + 1
539539
dec_pos_ids = paddle.tile(
540540
paddle.arange(max_len_lst[i], dtype="int64").unsqueeze(-1),
541541
[1, 3],

fastdeploy/worker/gpu_model_runner.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1013,7 +1013,7 @@ def insert_tasks_v1(self, req_dicts: List[Request], num_running_requests: int =
10131013

10141014
if len(rope_3d_position_ids["position_ids_idx"]) > 0 and self.enable_mm:
10151015
packed_position_ids = paddle.to_tensor(
1016-
np.concatenate(rope_3d_position_ids["position_ids_lst"]), dtype="int64"
1016+
np.concatenate(rope_3d_position_ids["position_ids_lst"]), dtype="float32"
10171017
)
10181018
rope_3d_lst = self.prepare_rope3d(
10191019
packed_position_ids,

0 commit comments

Comments
 (0)