Skip to content

Commit 367d37b

Browse files
authored
fix typo (#7147)
1 parent f422f83 commit 367d37b

File tree

8 files changed

+22
-22
lines changed

8 files changed

+22
-22
lines changed
Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -16,7 +16,7 @@
1616

1717
import asyncio
1818

19-
from fastdeploy.input.tokenzier_client import (
19+
from fastdeploy.input.tokenizer_client import (
2020
AsyncTokenizerClient,
2121
ImageDecodeRequest,
2222
ImageEncodeRequest,

fastdeploy/entrypoints/openai/response_processors.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,7 @@
1818
from typing import Any, Dict, List, Optional
1919

2020
from fastdeploy.entrypoints.openai.usage_calculator import count_tokens
21-
from fastdeploy.input.tokenzier_client import AsyncTokenizerClient, ImageDecodeRequest
21+
from fastdeploy.input.tokenizer_client import AsyncTokenizerClient, ImageDecodeRequest
2222

2323

2424
class ChatResponseProcessor:

fastdeploy/entrypoints/openai/v1/serving_chat.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -44,7 +44,7 @@
4444
OpenAiServingBase,
4545
ServingResponseContext,
4646
)
47-
from fastdeploy.input.tokenzier_client import AsyncTokenizerClient, ImageDecodeRequest
47+
from fastdeploy.input.tokenizer_client import AsyncTokenizerClient, ImageDecodeRequest
4848
from fastdeploy.metrics.metrics import main_process_metrics
4949
from fastdeploy.utils import api_server_logger
5050
from fastdeploy.worker.output import LogprobsLists

fastdeploy/worker/gpu_model_runner.py

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -638,12 +638,12 @@ def _process_mm_features(self, request_list: List[Request]):
638638
image_features_output is not None
639639
), f"image_features_output is None, images_lst length: {len(multi_vision_inputs['images_lst'])}"
640640
grid_thw = multi_vision_inputs["grid_thw_lst_batches"][index][thw_idx]
641-
mm_token_lenght = inputs["mm_num_token_func"](grid_thw=grid_thw)
642-
mm_feature = image_features_output[feature_idx : feature_idx + mm_token_lenght]
641+
mm_token_length = inputs["mm_num_token_func"](grid_thw=grid_thw)
642+
mm_feature = image_features_output[feature_idx : feature_idx + mm_token_length]
643643

644644
# add feature to encoder cache
645645
self.encoder_cache[mm_hash] = mm_feature.detach().cpu()
646-
feature_idx += mm_token_lenght
646+
feature_idx += mm_token_length
647647
thw_idx += 1
648648

649649
feature_start = feature_position.offset
@@ -663,13 +663,13 @@ def _process_mm_features(self, request_list: List[Request]):
663663
merge_image_features, thw_idx = [], 0
664664
for feature_position in feature_position_item:
665665
grid_thw = grid_thw_lst[thw_idx]
666-
mm_token_lenght = inputs["mm_num_token_func"](grid_thw=grid_thw)
667-
mm_feature = image_features_output[feature_idx : feature_idx + mm_token_lenght]
666+
mm_token_length = inputs["mm_num_token_func"](grid_thw=grid_thw)
667+
mm_feature = image_features_output[feature_idx : feature_idx + mm_token_length]
668668

669669
feature_start = feature_position.offset
670670
feature_end = feature_position.offset + feature_position.length
671671
merge_image_features.append(mm_feature[feature_start:feature_end])
672-
feature_idx += mm_token_lenght
672+
feature_idx += mm_token_length
673673
thw_idx += 1
674674
image_features_list.append(paddle.concat(merge_image_features, axis=0))
675675
for idx, index in req_idx_img_index_map.items():

fastdeploy/worker/metax_model_runner.py

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -571,12 +571,12 @@ def _process_mm_features(self, request_list: List[Request]):
571571
image_features_output is not None
572572
), f"image_features_output is None, images_lst length: {len(multi_vision_inputs['images_lst'])}"
573573
grid_thw = multi_vision_inputs["grid_thw_lst_batches"][index][thw_idx]
574-
mm_token_lenght = inputs["mm_num_token_func"](grid_thw=grid_thw)
575-
mm_feature = image_features_output[feature_idx : feature_idx + mm_token_lenght]
574+
mm_token_length = inputs["mm_num_token_func"](grid_thw=grid_thw)
575+
mm_feature = image_features_output[feature_idx : feature_idx + mm_token_length]
576576

577577
# add feature to encoder cache
578578
self.encoder_cache[mm_hash] = mm_feature.detach().cpu()
579-
feature_idx += mm_token_lenght
579+
feature_idx += mm_token_length
580580
thw_idx += 1
581581

582582
feature_start = feature_position.offset
@@ -596,13 +596,13 @@ def _process_mm_features(self, request_list: List[Request]):
596596
merge_image_features, thw_idx = [], 0
597597
for feature_position in feature_position_item:
598598
grid_thw = grid_thw_lst[thw_idx]
599-
mm_token_lenght = inputs["mm_num_token_func"](grid_thw=grid_thw)
600-
mm_feature = image_features_output[feature_idx : feature_idx + mm_token_lenght]
599+
mm_token_length = inputs["mm_num_token_func"](grid_thw=grid_thw)
600+
mm_feature = image_features_output[feature_idx : feature_idx + mm_token_length]
601601

602602
feature_start = feature_position.offset
603603
feature_end = feature_position.offset + feature_position.length
604604
merge_image_features.append(mm_feature[feature_start:feature_end])
605-
feature_idx += mm_token_lenght
605+
feature_idx += mm_token_length
606606
thw_idx += 1
607607
image_features_list.append(paddle.concat(merge_image_features, axis=0))
608608
for idx, index in req_idx_img_index_map.items():

fastdeploy/worker/xpu_model_runner.py

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -485,12 +485,12 @@ def _process_mm_features(self, request_list: List[Request]):
485485
image_features_output is not None
486486
), f"image_features_output is None, images_lst length: {len(multi_vision_inputs['images_lst'])}"
487487
grid_thw = multi_vision_inputs["grid_thw_lst"][thw_idx]
488-
mm_token_lenght = inputs["mm_num_token_func"](grid_thw=grid_thw)
489-
mm_feature = image_features_output[feature_idx : feature_idx + mm_token_lenght]
488+
mm_token_length = inputs["mm_num_token_func"](grid_thw=grid_thw)
489+
mm_feature = image_features_output[feature_idx : feature_idx + mm_token_length]
490490

491491
# add feature to encoder cache
492492
self.encoder_cache[mm_hash] = mm_feature.detach().cpu()
493-
feature_idx += mm_token_lenght
493+
feature_idx += mm_token_length
494494
thw_idx += 1
495495

496496
feature_start = feature_position.offset
@@ -510,13 +510,13 @@ def _process_mm_features(self, request_list: List[Request]):
510510
image_features_output = self.extract_vision_features(multi_vision_inputs)
511511
for feature_position in multi_vision_inputs["feature_position_list"]:
512512
grid_thw = multi_vision_inputs["grid_thw_lst"][thw_idx]
513-
mm_token_lenght = inputs["mm_num_token_func"](grid_thw=grid_thw)
514-
mm_feature = image_features_output[feature_idx : feature_idx + mm_token_lenght]
513+
mm_token_length = inputs["mm_num_token_func"](grid_thw=grid_thw)
514+
mm_feature = image_features_output[feature_idx : feature_idx + mm_token_length]
515515

516516
feature_start = feature_position.offset
517517
feature_end = feature_position.offset + feature_position.length
518518
merge_image_features.append(mm_feature[feature_start:feature_end])
519-
feature_idx += mm_token_lenght
519+
feature_idx += mm_token_length
520520
thw_idx += 1
521521
self.share_inputs["image_features"] = paddle.concat(merge_image_features, axis=0)
522522

tests/input/test_tokenizer_client.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,7 @@
1818
import pytest
1919
import respx
2020

21-
from fastdeploy.input.tokenzier_client import (
21+
from fastdeploy.input.tokenizer_client import (
2222
AsyncTokenizerClient,
2323
ImageEncodeRequest,
2424
VideoEncodeRequest,

0 commit comments

Comments
 (0)