vllm.v1.engine.mm_input_cache
 
 Used by P0 to check whether multi-modal kwargs are cached in P1.
Source code in vllm/v1/engine/mm_input_cache.py
  instance-attribute  ¶
 mm_cache = get_lru_cache(
    get_mm_input_cache_gb(), MultiModalCacheItemMetadata
)
 
 __init__(
    model_config: ModelConfig,
    mm_registry: MultiModalRegistry,
) -> None
Source code in vllm/v1/engine/mm_input_cache.py
  
 get_and_update(
    mm_inputs: Sequence[MultiModalKwargs],
    mm_hashes: list[str],
) -> Sequence[Optional[MultiModalKwargs]]
Source code in vllm/v1/engine/mm_input_cache.py
  
 Used by P1 to avoid requiring past multi-modal kwargs from P0.
Source code in vllm/v1/engine/mm_input_cache.py
  
 __init__(
    model_config: ModelConfig,
    mm_registry: MultiModalRegistry,
) -> None
Source code in vllm/v1/engine/mm_input_cache.py
  
 get_and_update(
    mm_inputs: Sequence[Optional[MultiModalKwargs]],
    mm_hashes: list[str],
) -> Sequence[MultiModalKwargs]