Skip to content

Handle VLM weight name prefixes in QuantizedModel loader#1996

Draft
uday610 wants to merge 1 commit intomicrosoft:mainfrom
uday610:vlm_quant_model_load
Draft

Handle VLM weight name prefixes in QuantizedModel loader#1996
uday610 wants to merge 1 commit intomicrosoft:mainfrom
uday610:vlm_quant_model_load

Conversation

@uday610
Copy link
Copy Markdown

@uday610 uday610 commented Feb 28, 2026

This change makes the quantized model weight loader work with VLM checkpoints where weight names include extra prefixes like model.language_model.* and vision tower weights like model.visual.*.

  • Skip weights under model.visual.*

  • Normalize model.language_model.* → model.* before existing pattern matching and layer parsing

No functional change for pure LLM checkpoints.

This change makes the quantized model weight loader work with VLM checkpoints where weight names include extra prefixes like `model.language_model.*` and vision tower weights like `model.visual.*`.

- Skip weights under `model.visual.*`

- Normalize `model.language_model.* → model.*` before existing pattern matching and layer parsing

No functional change for pure LLM checkpoints.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant