Semantically-aware Discriminators

A Shared Representation for Photorealistic Driving Simulators

Saeed Saadatnejad, Siyuan Li, Taylor Mordan, Alexandre Alahi

(published in IEEE Transactions on Intelligent Transportation Systems, 2021)

A powerful simulator highly decreases the need for real-world tests when training and evaluating autonomous vehicles. Data-driven simulators flourished with the recent advancement of conditional Generative Adversarial Networks (cGANs), providing high-fidelity images. The main challenge is synthesizing photo-realistic images while following given constraints. In this work, we propose to improve the quality of generated images by rethinking the discriminator architecture. The focus is on the class of problems where images are generated given semantic inputs, such as scene segmentation maps or human body poses. We build on successful cGAN models to propose a new semantically-aware discriminator that better guides the generator. We aim to learn a shared latent representation that encodes enough information to jointly do semantic segmentation, content reconstruction, along with a coarse-to-fine grained adversarial reasoning. The achieved improvements are generic and simple enough to be applied to any architecture of conditional image synthesis. We demonstrate the strength of our method on the scene, building, and human synthesis tasks across three different datasets.

Code , paper and arxiv.


A Shared Representation for Photorealistic Driving Simulators

S. Saadatnejad; S. Li; T. Mordan; A. Alahi 

IEEE Transactions on Intelligent Transportation Systems. 2021-12-03.  p. 1-11. DOI : 10.1109/TITS.2021.3131303.