TY - GEN
T1 - Guided Safe Diffusion
T2 - 31st International Conference on Neural Information Processing, ICONIP 2024
AU - Jiang, Sidong
AU - Zhang, Rui
AU - Yang, Xi
AU - Dong, Bin
AU - Huang, Kaizhu
N1 - Publisher Copyright:
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2026.
PY - 2025/7/19
Y1 - 2025/7/19
N2 - The increasing deployment of large generative models has heightened concerns over security and privacy, particularly regarding the generation of inappropriate content such as violent, explicit, or sensitive images, as well as the potential for creating fake images that spread misinformation and cause social problems. In this work, we propose Guided Safe Diffusion (GSD), an inference-time method specifically designed for diffusion models to prevent the generation of images with undesirable content as defined in a prohibited content list. Our method integrates safety guidance during the denoising steps of the model’s inference process, modifying the predicted noise to steer the generation process away from unwanted content. This approach allows the model to accept both an input image and a text description, facilitating controlled image generation. Unlike previous methods, our technique does not necessitate retraining or fine-tuning of the model. We conduct qualitative and quantitative experiments to assess the effectiveness of our method, demonstrating that GSD can remove the unwanted content while preserving unrelated content. The results validate our method’s ability to mitigate risks while maintaining the generative utility of diffusion models.
AB - The increasing deployment of large generative models has heightened concerns over security and privacy, particularly regarding the generation of inappropriate content such as violent, explicit, or sensitive images, as well as the potential for creating fake images that spread misinformation and cause social problems. In this work, we propose Guided Safe Diffusion (GSD), an inference-time method specifically designed for diffusion models to prevent the generation of images with undesirable content as defined in a prohibited content list. Our method integrates safety guidance during the denoising steps of the model’s inference process, modifying the predicted noise to steer the generation process away from unwanted content. This approach allows the model to accept both an input image and a text description, facilitating controlled image generation. Unlike previous methods, our technique does not necessitate retraining or fine-tuning of the model. We conduct qualitative and quantitative experiments to assess the effectiveness of our method, demonstrating that GSD can remove the unwanted content while preserving unrelated content. The results validate our method’s ability to mitigate risks while maintaining the generative utility of diffusion models.
KW - AI ethics
KW - Diffusion models
KW - Image protection
UR - https://www.scopus.com/pages/publications/105011971487
U2 - 10.1007/978-981-96-7036-9_11
DO - 10.1007/978-981-96-7036-9_11
M3 - Conference Proceeding
AN - SCOPUS:105011971487
SN - 9789819670352
T3 - Communications in Computer and Information Science
SP - 153
EP - 164
BT - Neural Information Processing - 31st International Conference, ICONIP 2024, Proceedings
A2 - Mahmud, Mufti
A2 - Doborjeh, Maryam
A2 - Doborjeh, Zohreh
A2 - Wong, Kevin
A2 - Leung, Andrew Chi Sing
A2 - Tanveer, M.
PB - Springer Science and Business Media Deutschland GmbH
Y2 - 2 December 2024 through 6 December 2024
ER -