<RETURN_TO_BASE

Qwen-Image-Edit: Alibaba’s New 20B Model for Precise Bilingual Image Editing and Novel-View Synthesis

'Alibaba's Qwen-Image-Edit extends Qwen-Image to provide precise semantic and appearance edits, bilingual text handling and easy deployment via Hugging Face.'

What Qwen-Image-Edit Does

Instruction-based image editing models are changing how creators modify visual content. Released in August 2025 by Alibaba’s Qwen Team, Qwen-Image-Edit builds on the 20B-parameter Qwen-Image foundation to provide advanced semantic and appearance editing while keeping strong multilingual text rendering in English and Chinese. It integrates with Qwen Chat and is available on Hugging Face, lowering barriers for professional content creation from IP design to correcting generated artwork.

Architecture and key innovations

Qwen-Image-Edit extends the Multimodal Diffusion Transformer (MMDiT) architecture of Qwen-Image. The system combines Qwen2.5-VL (an MLLM) for text conditioning, a VAE for image tokenization, and the MMDiT backbone for joint modeling. For editing it uses a dual encoding strategy: Qwen2.5-VL extracts high-level semantic features while the VAE captures low-level reconstruction details; both streams are concatenated into the MMDiT image stream. This balance supports semantic coherence and visual fidelity.

The model also augments Multimodal Scalable RoPE (MSRoPE) with a frame dimension to mark pre- and post-edit images, enabling tasks such as text-image-to-image (TI2I) editing. The VAE is fine-tuned on text-rich data and achieves 33.42 PSNR on general images and 36.63 on text-heavy ones, outperforming FLUX-VAE and SD-3.5-VAE — enabling high-fidelity bilingual text edits that preserve font, size and style.

Key features

  • Semantic and appearance editing: supports high-level semantic transformations (style transfer, novel view synthesis up to 180 degrees, IP generation) and precise low-level appearance edits (add/remove/modify elements while preserving unmodified regions).
  • Precise text editing: bilingual Chinese and English text editing with preservation of original font, size and style.
  • Strong benchmark performance: state-of-the-art results on multiple public editing benchmarks.

Training and data pipeline

The model leverages Qwen-Image’s curated dataset of billions of image-text pairs across Nature, Design, People and Synthetic domains. Training unifies T2I, I2I and TI2I objectives using a multi-task paradigm and a seven-stage filtering pipeline. Synthetic text rendering strategies (Pure, Compositional, Complex) help address long-tail Chinese character challenges. The training stack uses flow matching with a Producer-Consumer framework, supervised fine-tuning and reinforcement learning (DPO and GRPO) for preference alignment. Editing-specific tasks include novel view synthesis and depth estimation with DepthPro as a teacher model.

Advanced editing capabilities

Examples include creating MBTI-themed emojis from a mascot while keeping character consistency, 180-degree novel view synthesis with 15.11 PSNR on GSO, portrait style transfer (e.g., Studio Ghibli) and delicate appearance edits like adding realistic reflections or removing fine hair strands. Bilingual text edits allow changing words on posters or correcting Chinese calligraphy via bounding-box edits; chained edits enable iterative fixes until accurate.

Benchmarks and evaluations

Qwen-Image-Edit scores 7.56 on GEdit-Bench-EN and 7.52 on CN, outperforming several competitors. On ImgEdit it achieves 4.27 overall with strong task-specific scores. Depth estimation reaches 0.078 AbsRel on KITTI. Human evaluations place its base model high for text rendering and instruction-following.

Deployment and usage

Qwen-Image-Edit is deployable via Hugging Face Diffusers. Example usage:

from diffusers import QwenImageEditPipeline
import torch
from PIL import Image
 
pipeline = QwenImageEditPipeline.from_pretrained("Qwen/Qwen-Image-Edit")
pipeline.to(torch.bfloat16).to("cuda")
 
image = Image.open("input.png").convert("RGB")
prompt = "Change the rabbit's color to purple, with a flash light background."
output = pipeline(image=image, prompt=prompt, num_inference_steps=50, true_cfg_scale=4.0).images
output.save("output.png")

Alibaba Cloud Model Studio provides API access for scalable inference. The project is Apache 2.0 licensed and the GitHub repo includes training code and tutorials.

Future implications

The unified approach to understanding and generation positions Qwen-Image-Edit as a step toward richer vision-language interfaces, with potential extensions into video and 3D, enabling new AI-driven design workflows.

🇷🇺

Сменить язык

Читать эту статью на русском

Переключить на Русский