adoption in practical applications. In the future,
researchers can start from the directions of improving
training strategies, enhancing model robustness, and
exploring unsupervised or weakly supervised
learning methods to promote the further development
of GAN technology. By systematically combing the
research progress of GAN in image processing, this
paper aims to provide comprehensive theoretical
references and practical guidance for researchers in
related fields. At the same time, this paper also points
out the limitations of the current research and looks
forward to possible future research directions, with a
view to providing useful insights for the further
development and application of GAN technology.
REFERENCES
Choi, Y., Choi, M., Kim, M., Ha, J. W., Kim, S., & Choo,
J. (2018). StarGAN: Unified generative adversarial
networks for multi-domain image-to-image translation.
In Proceedings of the IEEE conference on computer
vision and pattern recognition (pp. 8789-8797).
Gao, C., Liu, Q., Xu, Q., Wang, L., Liu, J., & Zou, C.
(2020). SketchyCOCO: Image generation from
freehand scene sketches. In Proceedings of the
IEEE/CVF conference on computer vision and pattern
recognition (pp. 5174-5183).
Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B.,
Warde-Farley, D., Ozair, S., ... & Bengio, Y. (2014).
Generative adversarial nets. Advances in neural
information processing systems, 27.
Isola, P., Zhu, J. Y., Zhou, T., & Efros, A. A. (2017). Image-
to-image translation with conditional adversarial
networks. In Proceedings of the IEEE conference on
computer vision and pattern recognition (pp. 1125-
1134).
Johnson, J., Gupta, A., & Fei-Fei, L. (2018). Image
generation from scene graphs. In Proceedings of the
IEEE conference on computer vision and pattern
recognition (pp. 1219-1228).
Li, Y., Liu, H., Wu, Q., Mu, F., Yang, J., Gao, J., ... & Lee,
Y. J. (2023). GLIGEN: Open-set grounded text-to-
image generation. In Proceedings of the IEEE/CVF
Conference on Computer Vision and Pattern
Recognition (pp. 22511-22521).
Liao, W., Hu, K., Yang, M. Y., & Rosenhahn, B. (2022).
Text-to-image generation with semantic-spatial aware
GAN. In Proceedings of the IEEE/CVF conference on
computer vision and pattern recognition (pp. 18187-
18196).
Liu, M. Y., Breuel, T., & Kautz, J. (2017). Unsupervised
image-to-image translation networks. Advances in
neural information processing systems, 30.
Lu, Y., Wu, S., Tai, Y. W., & Tang, C. K. (2018). Image
generation from sketch constraint using contextual
GAN. In Proceedings of the European conference on
computer vision (ECCV) (pp. 205-220).
Men, Y., Mao, Y., Jiang, Y., Ma, W. Y., & Lian, Z. (2020).
Controllable person image synthesis with attribute-
decomposed GAN. In Proceedings of the IEEE/CVF
conference on computer vision and pattern recognition
(pp. 5084-5093).
Siarohin, A., Sangineto, E., Lathuiliere, S., & Sebe, N.
(2018). Deformable GANs for pose-based human
image generation. In Proceedings of the IEEE
conference on computer vision and pattern recognition
(pp. 3408-3416).
Xu, W., Long, C., Wang, R., & Wang, G. (2021). DRB-
GAN: A dynamic resblock generative adversarial
network for artistic style transfer. In Proceedings of the
IEEE/CVF international conference on computer vision
(pp. 6383-6392).
Yang, J., Kannan, A., Batra, D., & Parikh, D. (2017). LR-
GAN: Layered recursive generative adversarial
networks for image generation. arXiv preprint
arXiv:1703.01560.
Zhang, Y., Tang, F., Dong, W., Huang, H., Ma, C., Lee, T.
Y., & Xu, C. (2022, July). Domain-enhanced arbitrary
image style transfer via contrastive learning. In ACM
SIGGRAPH 2022 conference proceedings (pp. 1-8).