Synthesizing Optical and SAR Imagery From Land Cover Maps and Auxiliary Raster Data

November 23, 2020 Β· Entered Twilight Β· πŸ› IEEE Transactions on Geoscience and Remote Sensing

πŸŒ… TWILIGHT: Old Age
Predates the code-sharing era β€” a pioneer of its time

"Last commit was 5.0 years ago (β‰₯5 year threshold)"

Evidence collected by the PWNC Scanner

Repo contents: README.md, datasets, fid_comp.py, loss.py, models, options, pix_acc_iou_comp.py, synth_img.py, test.py, test_unet.py, train.py, train_unet.py, trainer.py, utils.py

Authors Gerald Baier, Antonin Deschemps, Michael Schmitt, Naoto Yokoya arXiv ID 2011.11314 Category cs.CV: Computer Vision Citations 47 Venue IEEE Transactions on Geoscience and Remote Sensing Repository https://github.com/gbaier/rs_img_synth ⭐ 33 Last Checked 1 month ago
Abstract
We synthesize both optical RGB and synthetic aperture radar (SAR) remote sensing images from land cover maps and auxiliary raster data using generative adversarial networks (GANs). In remote sensing, many types of data, such as digital elevation models (DEMs) or precipitation maps, are often not reflected in land cover maps but still influence image content or structure. Including such data in the synthesis process increases the quality of the generated images and exerts more control on their characteristics. Spatially adaptive normalization layers fuse both inputs and are applied to a full-blown generator architecture consisting of encoder and decoder to take full advantage of the information content in the auxiliary raster data. Our method successfully synthesizes medium (10 m) and high (1 m) resolution images when trained with the corresponding data set. We show the advantage of data fusion of land cover maps and auxiliary information using mean intersection over unions (mIoUs), pixel accuracy, and FrΓ©chet inception distances (FIDs) using pretrained U-Net segmentation models. Handpicked images exemplify how fusing information avoids ambiguities in the synthesized images. By slightly editing the input, our method can be used to synthesize realistic changes, i.e., raising the water levels. The source code is available at https://github.com/gbaier/rs_img_synth and we published the newly created high-resolution dataset at https://ieee-dataport.org/open-access/geonrw.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

πŸ“œ Similar Papers

In the same crypt β€” Computer Vision