Bases: Mistral3ForConditionalGeneration
Source code in vllm/model_executor/models/lightonocr.py
  class-attribute instance-attribute  ¶
 hf_to_vllm_mapper = WeightsMapper(
    orig_to_new_prefix={
        "model.vision_encoder.": "vision_tower.",
        "model.vision_projection.": "multi_modal_projector.",
        "lm_head.": "language_model.lm_head.",
        "model.language_model.": "language_model.model.",
    }
)
 instance-attribute  ¶
 language_model = init_vllm_registered_model(
    vllm_config=vllm_config,
    hf_config=text_config,
    prefix=maybe_prefix(prefix, "language_model"),
)
 instance-attribute  ¶
   instance-attribute  ¶
 multi_modal_projector = Mistral3MultiModalProjector(
    vision_hidden_size=hidden_size,
    text_hidden_size=hidden_size,
    projector_hidden_act=projector_hidden_act,
    spatial_merge_size=spatial_merge_size,
    patch_size=patch_size,
    multimodal_projector_bias=multimodal_projector_bias,
    quant_config=quant_config,
    prefix=maybe_prefix(prefix, "multi_modal_projector"),
)
 instance-attribute  ¶
 vision_tower = init_vision_tower_for_llava(
    config,
    quant_config,
    require_post_norm=False,
    prefix=maybe_prefix(prefix, "vision_tower"),
)
 
 __init__(
    *, vllm_config: VllmConfig, prefix: str = ""
) -> None
Source code in vllm/model_executor/models/lightonocr.py
  
    
  Bases: BaseMultiModalProcessor[Mistral3ProcessingInfo]
Source code in vllm/model_executor/models/lightonocr.py
  
 _call_hf_processor(
    prompt: str,
    mm_data: Mapping[str, object],
    mm_kwargs: Mapping[str, object],
    tok_kwargs: Mapping[str, object],
) -> BatchFeature
Source code in vllm/model_executor/models/lightonocr.py
  
 _get_mm_fields_config(
    hf_inputs: BatchFeature,
    hf_processor_mm_kwargs: Mapping[str, object],
) -> Mapping[str, MultiModalFieldConfig]
Source code in vllm/model_executor/models/lightonocr.py
   
 _get_prompt_updates(
    mm_items: MultiModalDataItems,
    hf_processor_mm_kwargs: Mapping[str, object],
    out_mm_kwargs: MultiModalKwargs,
) -> Sequence[PromptUpdate]
Source code in vllm/model_executor/models/lightonocr.py
  
 _build_LightOnOCR_processor(
    info: _I,
    dummy_inputs: BaseDummyInputsBuilder[_I],
    *,
    cache: BaseMultiModalProcessorCache | None = None,
)