to the edge discriminator for evaluation. Next, they
designed a novel loss function that measures the
image discrepancy between the focusing frequency
loss and the spatial loss (Lin, Xu, Liu, 2023).
Zhao proposed a new image canonisation architecture
that extracts anime image features independently,
making it controllable and scalable in practical
applications (Zhao, Zhu, Huang, et al., 2024). The
architecture is a GAN model based on multiple
decompositions. The image features generated by the
generator network training are decomposed and
extracted, and the texture module takes the extracted
texture and trains the image using a discriminator to
create textures with more cartoon image features.
Such a structural module uses the pre-trained network
to efficiently extract image structural features and
preserve image content information (Zhao, Zhu,
Huang, et al., 2024).
Li proposed an improved GAN-based style
migration algorithm for anime images (Li, Zhu, Liu,
et al., 2023). Firstly, the improved inverted residual
block and the efficient attention mechanism are used
to form a feature transformation module, which
enriches the local feature attributes of the migrated
image and improves the expression ability of style
features; secondly, the style loss function is improved
to suppress the interference of colour and luminance
on texture learning. Experiments reveal that the
introduced algorithm excels in producing images with
superior quality and enhanced visual realism (Li, Zhu,
Liu, et al., 2023).
2.4 Oil Painting
To solve the detail loss and blurring that appears
locally in the generated oil painting images problem,
Han proposed a multi-feature extractor to perform
style transfer of shape, texture, color, and space in oil
painting (Han X, Wu Y, and Wan R, 2023). The
extractor contains U-net, MiDas, texture extraction
network, and chromatic histogram. Meanwhile, the
autoencoder structure is employed as the main
framework for feature extraction. After identifying
the input and style images, DE-GAN is trained using
the model architecture, generating a network that
shares style parameters with the feature extraction
module. Finally, implement the generation of realistic
photos in the oil painting style (Han, Wu, and Wan,
2023).
To address the challenges of generating and
reconstructing image details in oil painting, Liu
proposed a method that combines the LBP algorithm
with an improved loss function (Liu, 2021). Liu
initially applied the Wasserstein metric to the GAN's
objective function to enhance training stability(Liu,
2021). Then add gradient penalty to loss function
(WGAN-GP), it can deal with gradient vanishing
during training. By incorporating noise control in the
CycleGAN loss function, boundary details and
surface patterns are enhanced after oil painting style
transfer, along with the addition of LBP structural
features and total variation.
3 RESEARCH
3.1 Quantitative Results of Ink
Painting Experiments
Hu's experiments found that the ChipGAN model,
which incorporates a residual dense network and
introduces a multi-scale discriminator, can largely
improve the quality of the generated images (Hu,
2023). The quality of an image is assessed by
quantitatively analyzing the image using Peak Signal
Noise Ratio (PSNR), Structural Similarity (SSIM),
Frechette's Distance (FID), and Cosine Similarity
(CosSim) as metrics (Hu, 2023). PSNR is used to
measure the generation quality of an image, and the
higher its value, the closer the style-migrated image
is to the original content, maximizing the retention of
content information. FID is used to compute the
similarity between two images in terms of feature
vectors, the smaller its value, the more similar it
represents the distribution of the generated image and
the real image in space. CosSim is used to measure
the pixel-wise similarity of two images, with higher
values indicating a smaller angle between the two
vectors represented by the image features and a
higher degree of similarity. SSIM is a measure of the
structural similarity of the two images, where higher
values indicate that the reconstructed image is more
similar to the original image in terms of brightness,
contrast, and structural inverse.
Hu uses four methods, method 1 is ChipGAN,
method 2 is ChipGAN based on RDN, method 3 is
ChipGAN using a multi-scale discriminator, and
method 4 is ChipGAN based on RDN and multi-scale
discriminator. Among them, method 1 is for the
ChipGAN model without any improvement, which is
a variant of GAN. The basic principles include:
generative network generates IC layout according to
the input design parameters; discriminative network
accepts the generated layout and the real layout and
outputs the true-false discriminative results;
optimizing the generative network and discriminative
network through adversarial training, so that the
quality of the generated layout is continuously