MFFCN-GAN: Multi-Scale Feature Fusion CNN with GAN for Automated Artistic Scene Generation in Film Animation

Xianrui Liu, Hang Zhao

Abstract


In the context of the digital transformation of the film art industry, the traditional animation scene production model faces the problems of low efficiency, high cost, and difficulty in meeting the audience's demand for high-quality scenes. To overcome this dilemma, this paper utilizes convolutional neural networks for the automatic generation of film art animation scenes. By constructing a multi-scale feature fusion convolutional network (MFFCN), the multi-scale convolution kernels are integrated to extract features in parallel, and the attention mechanism is combined with the generative adversarial network for scene generation. The experiment uses Kaggle's Anime Images Dataset, which includes fantastical landscapes and futuristic cityscapes. The proposed MFFCN model, with three convolutional branches and two attention modules, is compared to four models, including a geometric rule-based model and a support vector machine. Results demonstrate that MFFCN improves PSNR by 15 dB and SSIM by over 40% over the geometric model. It also excels in scene richness and visual style. This research advances computer graphics and deep learning in art generation, providing a realistic and intelligent solution for animation scene development that improves film industry operations and stylization.


Full Text:

PDF


DOI: https://doi.org/10.31449/inf.v49i9.8903

Creative Commons License
This work is licensed under a Creative Commons Attribution 3.0 License.