MickJ's picture
Sync clean ModelOpt weights from BBuf/Qwen-Image-ModelOpt-FP8-SGLang
7b87ffd verified
metadata
language:
  - en
library_name: sglang
base_model: Qwen/Qwen-Image
tags:
  - sglang
  - diffusion
  - modelopt
  - fp8
  - transformer

qwen-image-modelopt-fp8-sglang-transformer

This repository contains the SGLang-native ModelOpt FP8 transformer override for Qwen/Qwen-Image.

It is intended to be used with SGLang Diffusion through --transformer-path while keeping the base model separate:

sglang generate \
  --model-path Qwen/Qwen-Image \
  --transformer-path lmsys/qwen-image-modelopt-fp8-sglang-transformer \
  --prompt "A cinematic scene with detailed lighting" \
  --save-output

Contents

The repository is intentionally minimal and contains only:

  • config.json
  • *.safetensors weight shard files
  • *.safetensors.index.json when the checkpoint is sharded

Validation images, benchmark outputs, profiler traces, and conversion scratch artifacts are not stored in this model repository.

Notes

  • Quantization config is stored in config.json with quant_method=modelopt and quant_algo=FP8.
  • Use this checkpoint with an SGLang version that includes diffusion ModelOpt support for the corresponding model family.
  • The original base model license and usage terms still apply.