Exploring Generative Adversarial Networks for Image-to-Image Translation in STEM Simulation

October 29, 2020 ยท Entered Twilight ยท ๐Ÿ› arXiv.org

๐ŸŒ… TWILIGHT: Old Age
Predates the code-sharing era โ€” a pioneer of its time

"Last commit was 5.0 years ago (โ‰ฅ5 year threshold)"

Evidence collected by the PWNC Scanner

Repo contents: .gitignore, Calculate_NRMSE.py, FNN, LICENSE, README.md, Utils, fnn.py, pix2pix

Authors Nick Lawrence, Mingren Shen, Ruiqi Yin, Cloris Feng, Dane Morgan arXiv ID 2010.15315 Category cs.CV: Computer Vision Cross-listed cond-mat.mtrl-sci, cs.LG, eess.IV Citations 2 Venue arXiv.org Repository https://github.com/uw-cmg/GAN-STEM-Conv2MultiSlice โญ 2 Last Checked 2 months ago
Abstract
The use of accurate scanning transmission electron microscopy (STEM) image simulation methods require large computation times that can make their use infeasible for the simulation of many images. Other simulation methods based on linear imaging models, such as the convolution method, are much faster but are too inaccurate to be used in application. In this paper, we explore deep learning models that attempt to translate a STEM image produced by the convolution method to a prediction of the high accuracy multislice image. We then compare our results to those of regression methods. We find that using the deep learning model Generative Adversarial Network (GAN) provides us with the best results and performs at a similar accuracy level to previous regression models on the same dataset. Codes and data for this project can be found in this GitHub repository, https://github.com/uw-cmg/GAN-STEM-Conv2MultiSlice.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computer Vision