Skip to yearly menu bar Skip to main content


Poster

Robust-Wide: Robust Watermarking against Instruction-driven Image Editing

Runyi Hu · Jie Zhang · Ting Xu · Jiwei Li · Tianwei Zhang

[ ]
Fri 4 Oct 1:30 a.m. PDT — 3:30 a.m. PDT

Abstract:

Instruction-driven image editing allows users to quickly edit an image according to text instructions in a forward pass. Nevertheless, malicious users can easily exploit this technique to create fake images, which could cause a crisis of trust and harm the rights of the original image owners. Watermarking is a common solution to trace such malicious behavior. Unfortunately, instruction-driven image editing can significantly change the watermarked image at the se- mantic level, making it less robust and effective. We propose Robust-Wide, the first robust watermark- ing methodology against instruction-driven image editing. Specifically, we adopt the widely-used encoder-decoder framework for watermark embedding and extraction. To achieve robustness against semantic distortions, we intro- duce a novel Partial Instruction-driven Denoising Sampling Guidance (PIDSG) module, which consists of a large vari- ety of instruction injections and substantial modifications of images at different semantic levels. With PIDSG, the encoder tends to embed the watermark into more robust and semantic-aware areas, which remains in existence even after severe image editing. Experiments demonstrate that Robust-Wide can effectively extract the watermark from the edited image with a low bit error rate of nearly 2.6% for 64-bit watermark messages. Meanwhile, it only induces a neglectable influence on the visual quality and editability of the original images. Moreover, Robust-Wide holds general robustness against different sampling configura- tions and other image editing methods such as ControlNet- InstructPix2Pix, MagicBrush, Inpainting and DDIM Inver- sion.

Live content is unavailable. Log in and register to view live content