{"title":"通过水印信息混合实现潜在扩散模型的有效用户归属","authors":"Yongyang Pan, Xiaohong Liu, Siqi Luo, Yi Xin, Xiao Guo, Xiaoming Liu, Xiongkuo Min, Guangtao Zhai","doi":"arxiv-2409.10958","DOIUrl":null,"url":null,"abstract":"Rapid advancements in multimodal large language models have enabled the\ncreation of hyper-realistic images from textual descriptions. However, these\nadvancements also raise significant concerns about unauthorized use, which\nhinders their broader distribution. Traditional watermarking methods often\nrequire complex integration or degrade image quality. To address these\nchallenges, we introduce a novel framework Towards Effective user Attribution\nfor latent diffusion models via Watermark-Informed Blending (TEAWIB). TEAWIB\nincorporates a unique ready-to-use configuration approach that allows seamless\nintegration of user-specific watermarks into generative models. This approach\nensures that each user can directly apply a pre-configured set of parameters to\nthe model without altering the original model parameters or compromising image\nquality. Additionally, noise and augmentation operations are embedded at the\npixel level to further secure and stabilize watermarked images. Extensive\nexperiments validate the effectiveness of TEAWIB, showcasing the\nstate-of-the-art performance in perceptual quality and attribution accuracy.","PeriodicalId":501289,"journal":{"name":"arXiv - EE - Image and Video Processing","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Towards Effective User Attribution for Latent Diffusion Models via Watermark-Informed Blending\",\"authors\":\"Yongyang Pan, Xiaohong Liu, Siqi Luo, Yi Xin, Xiao Guo, Xiaoming Liu, Xiongkuo Min, Guangtao Zhai\",\"doi\":\"arxiv-2409.10958\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Rapid advancements in multimodal large language models have enabled the\\ncreation of hyper-realistic images from textual descriptions. However, these\\nadvancements also raise significant concerns about unauthorized use, which\\nhinders their broader distribution. Traditional watermarking methods often\\nrequire complex integration or degrade image quality. To address these\\nchallenges, we introduce a novel framework Towards Effective user Attribution\\nfor latent diffusion models via Watermark-Informed Blending (TEAWIB). TEAWIB\\nincorporates a unique ready-to-use configuration approach that allows seamless\\nintegration of user-specific watermarks into generative models. This approach\\nensures that each user can directly apply a pre-configured set of parameters to\\nthe model without altering the original model parameters or compromising image\\nquality. Additionally, noise and augmentation operations are embedded at the\\npixel level to further secure and stabilize watermarked images. Extensive\\nexperiments validate the effectiveness of TEAWIB, showcasing the\\nstate-of-the-art performance in perceptual quality and attribution accuracy.\",\"PeriodicalId\":501289,\"journal\":{\"name\":\"arXiv - EE - Image and Video Processing\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-17\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - EE - Image and Video Processing\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.10958\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - EE - Image and Video Processing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.10958","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Towards Effective User Attribution for Latent Diffusion Models via Watermark-Informed Blending
Rapid advancements in multimodal large language models have enabled the
creation of hyper-realistic images from textual descriptions. However, these
advancements also raise significant concerns about unauthorized use, which
hinders their broader distribution. Traditional watermarking methods often
require complex integration or degrade image quality. To address these
challenges, we introduce a novel framework Towards Effective user Attribution
for latent diffusion models via Watermark-Informed Blending (TEAWIB). TEAWIB
incorporates a unique ready-to-use configuration approach that allows seamless
integration of user-specific watermarks into generative models. This approach
ensures that each user can directly apply a pre-configured set of parameters to
the model without altering the original model parameters or compromising image
quality. Additionally, noise and augmentation operations are embedded at the
pixel level to further secure and stabilize watermarked images. Extensive
experiments validate the effectiveness of TEAWIB, showcasing the
state-of-the-art performance in perceptual quality and attribution accuracy.