We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent 17dd222 commit 703cccbCopy full SHA for 703cccb
vllm/entrypoints/llm.py
@@ -536,6 +536,16 @@ def sort_beams_key(x: BeamSearchSequence) -> float:
536
tokenizer.eos_token_id,
537
length_penalty)
538
539
+ # TODO - fix handling of multimodal data for beam search; we pass it
540
+ # through in the async version on the abstract EngineClient, but not
541
+ # here.
542
+ if any("multi_modal_data" in prompt
543
+ and prompt["multi_modal_data"] is not None
544
+ for prompt in prompts):
545
+ logger.warning(
546
+ "Multimodal data appears to have been provided, but is not"
547
+ " currently being passed through in LLM.beam_search()!")
548
+
549
tokenizer = self.get_tokenizer()
550
# generate 2 * beam_width candidates at each step
551
# following the huggingface transformers implementation
0 commit comments