Skip to content
Merged
Prev Previous commit
Next Next commit
Address comment
  • Loading branch information
chang-l committed Jul 27, 2025
commit cd7393a1b4fdcfe96f37b08a0c69a05cb79a27df
2 changes: 1 addition & 1 deletion tensorrt_llm/_torch/models/modeling_llama.py
Original file line number Diff line number Diff line change
Expand Up @@ -819,7 +819,7 @@ def __init__(self,
}).cuda()
load_sharded_checkpoint(self.encoder, model_path, strict=False)

def attch_multimodal_embeddings(
def attach_multimodal_embeddings(
self, inputs: TextPrompt, multimodal_embedding: Dict[str,
List[Dict[str,
Any]]],
Expand Down
2 changes: 1 addition & 1 deletion tensorrt_llm/_torch/models/modeling_llava_next.py
Original file line number Diff line number Diff line change
Expand Up @@ -195,7 +195,7 @@ def _postprocess(self, input_ids, mm_features):
mm_features = mm_features.view(-1, mm_features.shape[-1])
return fused_input_ids, mm_features

def attch_multimodal_embeddings(
def attach_multimodal_embeddings(
self, inputs: TextPrompt,
multimodal_embedding: Dict[str, List[torch.Tensor]],
sampling_params: SamplingParams
Expand Down
2 changes: 1 addition & 1 deletion tensorrt_llm/llmapi/llm.py
Original file line number Diff line number Diff line change
Expand Up @@ -379,7 +379,7 @@ def generate_async(
inputs, sampling_params)
elif 'multi_modal_embeddings' in inputs:
mm_embedding_info = inputs['multi_modal_embeddings']
prompt_token_ids, extra_processed_inputs = self.input_processor.attch_multimodal_embeddings(
prompt_token_ids, extra_processed_inputs = self.input_processor.attach_multimodal_embeddings(
inputs, mm_embedding_info, sampling_params)
else:
with nvtx_range_debug("input_processor"):
Expand Down