Deep Multiple Description Coding by Learning Scalar Quantization

November 05, 2018 ยท Entered Twilight ยท ๐Ÿ› Data Compression Conference

๐ŸŒ… TWILIGHT: Old Age
Predates the code-sharing era โ€” a pioneer of its time

"Last commit was 5.0 years ago (โ‰ฅ5 year threshold)"

Evidence collected by the PWNC Scanner

Repo contents: MDQNN-supp.pdf, README.md, test-image-126

Authors Lijun Zhao, Huihui Bai, Anhong Wang, Yao Zhao arXiv ID 1811.01504 Category cs.MM: Multimedia Cross-listed cs.CV Citations 5 Venue Data Compression Conference Repository https://github.com/mdcnn/Deep-Multiple-Description-Coding โญ 1 Last Checked 2 months ago
Abstract
In this paper, we propose a deep multiple description coding framework, whose quantizers are adaptively learned via the minimization of multiple description compressive loss. Firstly, our framework is built upon auto-encoder networks, which have multiple description multi-scale dilated encoder network and multiple description decoder networks. Secondly, two entropy estimation networks are learned to estimate the informative amounts of the quantized tensors, which can further supervise the learning of multiple description encoder network to represent the input image delicately. Thirdly, a pair of scalar quantizers accompanied by two importance-indicator maps is automatically learned in an end-to-end self-supervised way. Finally, multiple description structural dissimilarity distance loss is imposed on multiple description decoded images in pixel domain for diversified multiple description generations rather than on feature tensors in feature domain, in addition to multiple description reconstruction loss. Through testing on two commonly used datasets, it is verified that our method is beyond several state-of-the-art multiple description coding approaches in terms of coding efficiency.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Multimedia