Exocentric to egocentric image generation via parallel generative adversarial network

G Liu, H Tang, H Latapie, Y Yan - ICASSP 2020-2020 IEEE …, 2020 - ieeexplore.ieee.org
ICASSP 2020-2020 IEEE International Conference on Acoustics …, 2020ieeexplore.ieee.org
Cross-view image generation has been recently proposed to generate images of one view
from another dramatically different view. In this paper, we investigate exocentric (third-
person) view to egocentric (first-person) view image generation. This is a challenging task
since egocentric view sometimes is remarkably different from exocentric view. Thus,
transforming the appearances across the two views is a nontrivial task. To this end, we
propose a novel Parallel Generative Adversarial Network (P-GAN) with a novel cross-cycle …
Cross-view image generation has been recently proposed to generate images of one view from another dramatically different view. In this paper, we investigate exocentric (third-person) view to egocentric (first-person) view image generation. This is a challenging task since egocentric view sometimes is remarkably different from exocentric view. Thus, transforming the appearances across the two views is a nontrivial task. To this end, we propose a novel Parallel Generative Adversarial Network (P-GAN) with a novel cross-cycle loss to learn the shared information for generating egocentric images from exocentric view. We also incorporate a novel contextual feature loss in the learning procedure to capture the contextual information in images. Extensive experiments on the Exo-Ego datasets [1] show that our model outperforms the state-of-the-art approaches.
ieeexplore.ieee.org