Education, Science, Technology, Innovation and Life
Open Access
Sign In

PGGAN: Probability Guided Generative Adversarial Network for Image Inpainting

Download as PDF

DOI: 10.23977/acss.2025.090106 | Downloads: 23 | Views: 576

Author(s)

Chaoqun Dong 1, Yang Yang 1

Affiliation(s)

1 School of Information Science and Technology, Yunnan Normal University, Kunming, China

Corresponding Author

Yang Yang

ABSTRACT

Probability Guided Generative Adversarial Network (PG-GAN) aims to address key challenges in image inpainting, particularly in capturing structural information over long distances. Firstly, we design the IAModule, which provides semantic attention based on the distribution characteristics of input features, thereby enhancing semantic coherence in image inpainting. Secondly, we propose RR-SSIM Loss, a new loss function aimed at solving the problem of Structural Similarity (SSIM) that is difficult to capture long-distance structural information through sliding window calculations. Finally, we provide a new feature enhancement mechanism through channel dimension Fourier transform and design it as a HybridFFTModule. This module enhances the distinguishability of global representation through channel modeling, effectively adjusting the representation space of global information and further improving the effectiveness of image inpainting. In the experimental section, we validate the superior performance of PG-GAN on CelabA-HQ dataset. In summary, our PG-GAN provides a new and effective method for image inpainting, with broad application prospects.

KEYWORDS

Image Inpainting, Generative Adversarial Network (GAN), Deep Learning, Fourier Transform

CITE THIS PAPER

Chaoqun Dong, Yang Yang, PGGAN: Probability Guided Generative Adversarial Network for Image Inpainting. Advances in Computer, Signals and Systems (2025) Vol. 9: 33-41. DOI: http://dx.doi.org/10.23977/acss.2025.090106.

REFERENCES

[1] Criminisi A, Pérez P, Toyama K. Region filling and object removal by exemplar-based image inpainting[J]. IEEE Transactions on image processing, 2004, 13(9): 1200-1212. 
[2] Darabi S, Shechtman E, Barnes C, et al. Image melding: Combining inconsistent images using patch-based synthesis [J]. ACM Transactions on graphics (TOG), 2012, 31(4): 1-10. 
[3] Song Y, Bao L, He S, et al. Stylizing face images via multiple exemplars[J]. Computer Vision and Image Understanding, 2017, 162: 135-145. 
[4] Barnes C, Shechtman E, Finkelstein A, et al. PatchMatch: A randomized correspondence algorithm for structural image editing[J]. ACM Trans. Graph., 2009, 28(3): 24. 
[5] Xu Z, Sun J. Image inpainting by patch propagation using patch sparsity[J]. IEEE transactions on image processing, 2010, 19(5): 1153-1165. 
[6] Yu J, Lin Z, Yang J, et al. Generative image inpainting with contextual attention[C]. Proceedings of the IEEE conference on computer vision and pattern recognition. 2018: 5505-5514. 
[7] Liu G, Reda F A, Shih K J, et al. Image inpainting for irregular holes using partial convolutions[C]. Proceedings of the European conference on computer vision (ECCV). 2018: 85-100. 
[8] Yan Z, Li X, Li M, et al. Shift-net: Image inpainting via deep feature rearrangement[C]. Proceedings of the European conference on computer vision (ECCV). 2018: 1-17. 
[9] Iizuka S, Simo-Serra E, Ishikawa H. Globally and locally consistent image completion[J]. ACM Transactions on Graphics (ToG), 2017, 36(4): 1-14. 
[10] Altinel F, Ozay M, Okatani T. Deep structured energy-based image inpainting[C]. 2018 24th International Conference on Pattern Recognition (ICPR). IEEE, 2018: 423-428. [11] Zhu X, Qian Y, Zhao X, et al. A deep learning approach to patch-based image inpainting forensics[J]. Signal Processing: Image Communication, 2018, 67: 90-99. 
[12] Wang Y, Tao X, Qi X, et al. Image inpainting via generative multi-column convolutional neural networks[J]. Advances in neural information processing systems, 2018, 31. [13] Hedjazi M A, Genc Y. Efficient texture-aware multi-GAN for image inpainting[J]. Knowledge-Based Systems, 2021, 217: 106789. 
[14] Zheng C, Cham T J, Cai J. Pluralistic image completion[C]. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2019: 1438-1447. 
[15] Sola S, Gera D. Unmasking your expression: Expression-conditioned gan for masked face inpainting[C]. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2023: 5908-5916. 
[16] Li W, Lin Z, Zhou K, et al. Mat: Mask-aware transformer for large hole image inpainting[C]. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2022: 10758-10768. 
[17] Wan Z, Zhang J, Chen D, et al. High-fidelity pluralistic image completion with transformers[C]. Proceedings of the IEEE/CVF international conference on computer vision. 2021: 4692-4701. 
[18] Xu Q, Zhang R, Zhang Y, et al. A fourier-based framework for domain generalization [C]. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2021: 14383-14392. 
[19] Lee S, Bae J, Kim H Y. Decompose, adjust, compose: Effective normalization by playing with frequency for domain generalization[C]. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2023: 11776-11785. 
[20] Chi L, Jiang B, Mu Y. Fast fourier convolution[J]. Advances in Neural Information Processing Systems, 2020, 33: 4479-4488. 
[21] Rao Y, Zhao W, Zhu Z, et al. Global filter networks for image classification[J]. Advances in neural information processing systems, 2021, 34: 980-993. 
[22] Karras T, Aila T, Laine S, et al. Progressive growing of GANs for improved quality, stability, and variation International Conference on Learning Representations. 2018[EB/OL].(2018) 
[23] Zheng C, Cham T J, Cai J, et al. Bridging global context interactions for high-fidelity image completion[C]. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2022: 11512-11522. 
[24] Liu H, Wan Z, Huang W, et al. Pd-gan: Probabilistic diverse gan for image inpainting[C]. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2021: 9371-9381. 
[25] Zeng Y, Fu J, Chao H, et al. Aggregated contextual transformations for high-resolution image inpainting[J]. IEEE Transactions on Visualization and Computer Graphics, 2022, 29(7): 3266-3280.

Downloads: 38553
Visits: 697938

Sponsors, Associates, and Links


All published work is licensed under a Creative Commons Attribution 4.0 International License.

Copyright © 2016 - 2031 Clausius Scientific Press Inc. All Rights Reserved.