Neural Network-Based Mapping Mining of Image Style Transfer in Big Data Systems

Hong An Li, Qiaoxue Zheng, Xin Qi*, Wenjing Yan, Zheng Wen, Na Li, Chu Tang

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

Image style transfer can realize the mutual transfer between different styles of images and is an essential application for big data systems. The use of neural network-based image data mining technology can effectively mine the useful information in the image and improve the utilization rate of information. However, when using the deep learning method to transform the image style, the content information is often lost. To address this problem, this paper introduces L1 loss on the basis of the VGG-19 network to reduce the difference between image style and content and adds perceptual loss to calculate the semantic information of the feature map to improve the model's perceptual ability. Experiments show that the proposal in this paper improves the ability of style transfer, while maintaining image content information. The stylization of the improved model can better meet people's requirements for stylization, and the evaluation indexes of structural similarity, cosine similarity, and mutual information value have increased by 0.323%, 0.094%, and 3.591%, respectively.

Original languageEnglish
Article number8387382
JournalComputational Intelligence and Neuroscience
Volume2021
DOIs
Publication statusPublished - 2021

ASJC Scopus subject areas

  • Computer Science(all)
  • Neuroscience(all)
  • Mathematics(all)

Fingerprint

Dive into the research topics of 'Neural Network-Based Mapping Mining of Image Style Transfer in Big Data Systems'. Together they form a unique fingerprint.

Cite this