Enable NextStepDiffusion and support multi-device tuning for diffusion#1640
Enable NextStepDiffusion and support multi-device tuning for diffusion#1640
Conversation
Signed-off-by: Xin He <xin3.he@intel.com>
There was a problem hiding this comment.
Pull request overview
Fixes model loading for the “nextstep” model type by selecting an appropriate AutoModel loader, and adjusts multimodal key detection to recognize “image”-named components.
Changes:
- Force
AutoModelformodel_type == "nextstep"during MLLM model loading. - Add
"image"toMM_KEYSto broaden multimodal component detection.
Reviewed changes
Copilot reviewed 2 out of 2 changed files in this pull request and generated 2 comments.
| File | Description |
|---|---|
auto_round/utils/model.py |
Adds a NextStep-specific loader class override to resolve loading failures. |
auto_round/utils/common.py |
Extends multimodal key matching to include "image" for downstream detection/mapping. |
Signed-off-by: Xin He <xin3.he@intel.com>
|
The |
|
better add next_step to mllm support matrix |
|
I need to upstream a model before updating the support matrix (requires model link). |
If the model’s license allows upstreaming, we can upload it. Otherwise, we can leave the link blank. |
|
The status has been reverted to "Draft," as only RTN is currently supported; upstream adaptation and optimization work is currently underway. |
…model loading for NextStep Signed-off-by: Xin He <xin3.he@intel.com>
Signed-off-by: Xin He <xin3.he@intel.com>
Signed-off-by: Xin He <xin3.he@intel.com>
… gptqmodel fix Signed-off-by: Xin He <xin3.he@intel.com>
for more information, see https://pre-commit.ci
…imports Signed-off-by: Xin He <xin3.he@intel.com>
for more information, see https://pre-commit.ci
|
/azp run Unit-Test-CUDA-AutoRound |
|
Azure Pipelines successfully started running 1 pipeline(s). |
Description
fix nextstep loading issue
example_prompt = "A REALISTIC PHOTOGRAPH OF A WALL WITH \"TOWARD AUTOREGRESSIVE IMAGE GENERATION WITH CONTINUOUS TOKENS AT SCALE\" PROMINENTLY DISPLAYED"Raw model output:
W4A16 model output with torch backend on CPU:
W4A16 model output with
gptqmodel:marlinbackend on CUDA:Type of Change
Related Issues
Fixes or relates to #
Checklist Before Submitting