跳到主要导航 跳到搜索 跳到主要内容

Zero-Shot Image Harmonization With Generative Model Prior

  • Beihang University

科研成果: 期刊稿件文章同行评审

摘要

We propose a zero-shot approach to image harmonization, aiming to overcome the reliance on large amounts of synthetic composite images in existing methods. These methods, while showing promising results, involve significant training expenses and often struggle with generalization to unseen images. To this end, we introduce a fully modularized framework inspired by human behavior. Leveraging the reasoning capabilities of recent foundation models in language and vision, our approach comprises three main stages. Initially, we employ a pretrained vision-language model (VLM) to generate descriptions for the composite image. Subsequently, these descriptions guide the foreground harmonization direction of a text-to-image generative model (T2I). We refine text embeddings for enhanced representation of imaging conditions and employ self-attention and edge maps for structure preservation. Following each harmonization iteration, an evaluator determines whether to conclude or modify the harmonization direction. The resulting framework, mirroring human behavior, achieves harmonious results without the need for extensive training. We present compelling visual results across diverse scenes and objects, along with quantitative comparisons validating the effectiveness of our approach.

源语言英语
页(从-至)4494-4507
页数14
期刊IEEE Transactions on Multimedia
27
DOI
出版状态已出版 - 2025

指纹

探究 'Zero-Shot Image Harmonization With Generative Model Prior' 的科研主题。它们共同构成独一无二的指纹。

引用此