Toward Controlled Generation of Text

March 02, 2017 ยท Entered Twilight ยท ๐Ÿ› International Conference on Machine Learning

๐ŸŒ… TWILIGHT: Old Age
Predates the code-sharing era โ€” a pioneer of its time

"Last commit was 5.0 years ago (โ‰ฅ5 year threshold)"

Evidence collected by the PWNC Scanner

Repo contents: .flake8, .gitignore, .pylintrc, .readthedocs.yml, .travis.yml, CHANGELOG.md, LICENSE, README.md, bin, codecov.yml, docs, examples, requirements.txt, scripts, setup.py, tests, texar, travis_key.enc

Authors Zhiting Hu, Zichao Yang, Xiaodan Liang, Ruslan Salakhutdinov, Eric P. Xing arXiv ID 1703.00955 Category cs.LG: Machine Learning Cross-listed cs.AI, cs.CL, stat.ML Citations 1.0K Venue International Conference on Machine Learning Repository https://github.com/asyml/texar/tree/master/examples/text_style_transfer โญ 2391 Last Checked 1 month ago
Abstract
Generic generation and manipulation of text is challenging and has limited success compared to recent deep generative modeling in visual domain. This paper aims at generating plausible natural language sentences, whose attributes are dynamically controlled by learning disentangled latent representations with designated semantics. We propose a new neural generative model which combines variational auto-encoders and holistic attribute discriminators for effective imposition of semantic structures. With differentiable approximation to discrete text samples, explicit constraints on independent attribute controls, and efficient collaborative learning of generator and discriminators, our model learns highly interpretable representations from even only word annotations, and produces realistic sentences with desired attributes. Quantitative evaluation validates the accuracy of sentence and attribute generation.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Machine Learning