Abstract:
With astonishing successes of GAN-based style transfer techniques, real-world photo translation to animation style images recently has attracted some interest. In particular, the transfer of a selfie to a cartoon has become quite popular as it can serve as a cartoon filter in social media. Unlike most of the Image-to-Image translation tasks, the selfie to anime task requires preserving the contour in the selfie image in the transfer process while it transforms other local characteristics into an animation style. Since the gap between the selfie domain and anime domain is quite large, as it can be imagined in the case of transforming a person into a cartoon animal like a mouse, developing an effective method remains a difficult challenge. In this paper, we propose an Adaptive Content Feature Enhancement Generative Adversarial Networks (ACFE-GAN) for a selfie to anime translation. By our model, the preservation of content features of selfie is improved. In addition to facial and hair contours, the shapes of worn items (eg{} hat and glasses) are also better preserved compared to existing methods. Our method also captures local features more accurately and selectively in translating them into the animation domain. Moreover, compared to the previous photo to anime translation models, we implement it with multimodal translation. Experiments on the selfie2anime dataset demonstrate that our method delivers superior performance in terms of selective preservation of content features.