- Extracting and composing robust features with denoising autoencoders Pascal Vincent, Hugo Larochelle, Yoshua Bengio, Pierre-Antoine Manzagol
ICML08
[PDF] - Auto-Encoding Variational Bayes Diederik P Kingma, Max Welling
ICLR14
[PDF] - Generating Sentences from a Continuous Space Samuel R. Bowman, Luke Vilnis
CONLL16
[PDF] - Adversarial Autoencoders Alireza Makhzani, Jonathon Shlens, Navdeep Jaitly, Ian Goodfellow, Brendan Frey
ICLR16
[PDF] - Learning Structured Output Representation using Deep Conditional Generative Models Kihyuk Sohn, Xinchen Yan, Honglak Lee
NIPS15
[PDF] - Deep Unsupervised Clustering with Gaussian Mixture Variational Auotoencoders Nat Dilokthanakul, Pedro A. M. Mediano, Marta Garnelo, Matthew C. H. Lee, Hugh Salimbeni, Kai Arulkumaran, Murray Shanahan
ICLR17
[PDF] - Wasserstein Auto-Encoders Ilya Tolstikhin, Olivier Bousquet, Sylvain Gelly, Bernhard Sch¨olkopf
ICLR18
[PDF] [code] - Adversarially Regularized Autoencoders Jake Zhao, Yoon Kim, Kelly Zhang, Alexander M. Rush, Yann LeCun
ICML18
[PDF] [code] - T-CVAE: Transformer-Based Conditioned Variational Autoencoder for Story Completion Tianming Wang, Xiaojun Wan
IJCAI19
[PDF] [code]abstract
Story completion is a very challenging task of generating the missing plot for an incomplete story, which requires not only understanding but also inference of the given contextual clues. In this paper, we present a novel conditional variational autoencoder based on Transformer for missing plot generation. Our model uses shared attention layers for encoder and decoder, which make the most of the contextual clues, and a latent variable for learning the distribution of coherent story plots. Through drawing samples from the learned distribution, diverse reasonable plots can be generated. Both automatic and manual evaluations show that our model generates better story plots than stateof-the-art models in terms of readability, diversity and coherence. - Transformer-based Conditional Variational Autoencoder for Controllable Story Generation Le Fang, Tao Zeng, Chaochun Liu, Liefeng Bo, Wen Dong, Changyou Chen
arXiv
[PDF] [code]abstract
We investigate large-scale latent variable models (LVMs) for neural story generation—an under-explored application for open-domain long text—with objectives in two threads: generation effectiveness and controllability. LVMs, especially the variational autoencoder (VAE), have achieved both effective and controllable generation through exploiting flexible distributional latent representations. Recently, Transformers and its variants have achieved remarkable effectiveness without explicit latent representation learning, thus lack satisfying controllability in generation. In this paper, we advocate to revive latent variable modeling, essentially the power of representation learning, in the era of Transformers to enhance controllability without hurting state-of-the-art generation effectiveness. Specifically, we integrate latent representation vectors with a Transformer-based pre-trained architecture to build conditional variational autoencoder (CVAE). Model components such as encoder, decoder and the variational posterior are all built on top of pre-trained language models—GPT2 specifically in this paper. Experiments demonstrate state-of-the-art conditional generation ability of our model, as well as its excellent representation learning capability and controllability. - APo-VAE: Text Generation in Hyperbolic Space Shuyang Dai, Zhe Gan, Yu Cheng, Chenyang Tao, Lawrence Carin, Jingjing Liu
NAACL21
[PDF] [code]
- NICE: Non-linear Independent Components Estimation Laurent Dinh, David Krueger, Yoshua Bengio
arXiv
[PDF] - Density estimation using Real NVP Laurent Dinh, Jascha Sohl-Dickstein, Samy Bengio
arXiv
[PDF] - Glow: Generative Flow with Invertible 1×1 Convolutions
NeurIPS18
Diederik P. Kingma, Prafulla Dhariwal [PDF] [code] - Masked Autoregressive Flow for Density Estimation
NeurIPS17
George Papamakarios, Theo Pavlakou, Iain Murray [PDF] - Improved Variational Inference with Inverse Autoregressive Flow
NeurIPS16
Diederik P. Kingma, Tim Salimans, Rafal Jozefowicz, Xi Chen - FlowSeq: Non-Autoregressive Conditional Sequence Generation with Generative Flow
EMNLP2019
Xuezhe Ma, Chunting Zhou, Xian Li, Graham Neubig, Eduard Hovy [PDF] [code] - Continuous Language Generative Flow
ACL21
Zineng Tang, Shiyue Zhang, Hyounghun Kim, Mohit Bansal [PDF] [code]
- Deep Unsupervised Learning using Nonequilibrium Thermodynamics
ICML15
Jascha Sohl-Dickstein, Eric Weiss, Niru Maheswaranathan, Surya Ganguli [PDF] - Generative Modeling by Estimating Gradients of the Data Distribution
NeurIPS19
Yang Song, Stefano Ermon [PDF] - Improved Techniques for Training Score-Based Generative Models
NeurIPS20
Yang Song, Stefano Ermon [PDF] - Denoising Diffusion Probabilistic Models
NeurIPS20
Jonathan Ho, Ajay Jain, Pieter Abbeel [PDF] [code] - Diffusion Models Beat GANs on Image Synthesis
NeurIPS21
Prafulla Dhariwal, Alex Nichol [PDF] - GLIDE: Towards Photorealistic Image Generation and Editing with Text-Guided Diffusion Models
arXiv
Alex Nichol, Prafulla Dhariwal, Aditya Ramesh, Pranav Shyam, Pamela Mishkin, Bob McGrew, Ilya Sutskever, Mark Chen [PDF] - Diffusion-LM Improves Controllable Text Generation
NeurIPS22
Xiang Lisa Li, John Thickstun, Ishaan Gulrajani, Percy Liang, Tatsunori B. Hashimoto [PDF] [code]
- BERT, GPT, T5, BART, ...
- Plug and Play Language Models: A Simple Approach to Controlled Text Generation Sumanth Dathathri, Andrea Madotto, Janice Lan, Jane Hung, Eric Frank, Piero Molino, Jason Yosinski, Rosanne Liu
ICLR20
[PDF] [code] - FUDGE: Controlled Text Generation With Future Discriminators Kevin Yang, Dan Klein
NAACL21
[PDF] [code] - GeDi: Generative Discriminator Guided Sequence Generation Ben Krause, Akhilesh Deepak Gotmare, Bryan McCann, Nitish Shirish Keskar, Shafiq Joty, Richard Socher, Nazneen Fatema Rajani
EMNLP21 findings
[PDF] [code] - DExperts: Decoding-Time Controlled Text Generation with Experts and Anti-Experts Alisa Liu, Maarten Sap, Ximing Lu, Swabha Swayamdipta, Chandra Bhagavatula, Noah A. Smith, Yejin Choi
ACL21
[PDF] - Plug-and-Blend: A Framework for Controllable Story Generation with Blended Control Codes Zhiyu Lin, Mark Riedl
arXiv21
[PDF] [code] - Improving Controllable Text Generation with Position-Aware Weighted Decoding
ACL22 findings
Yuxuan Gu, Xiaocheng Feng, Sicheng Ma, Jiaming Wu, Heng Gong, Bing Qin [PDF] - Bridging the Gap Between Training and Inference of Bayesian Controllable Language Models
arXiv
Han Liu, Bingning Wang, Ting Yao, Haijin Liang, Jianjin Xu, Xiaolin Hu [PDF] - Classifiers are Better Experts for Controllable Text Generation
arXiv
Askhat Sitdikov, Nikita Balagansky, Daniil Gavrilov, Alexander Markov [PDF] - Gamma Sampling: Fine-grained Controlling Language Models without Training
arXiv
Shangda Wu, Maosong Sun [PDF] - Controllable Text Generation with Neurally-Decomposed Oracle
NeurIPS22
Tao Meng, Sidi Lu, Nanyun Peng, Kai-Wei Chang [PDF]
- Controlled Text Generation as Continuous Optimization with Multiple Constraints
NeurIPS21
Sachin Kumar, Eric Malmi, Aliaksei Severyn, Yulia Tsvetkov [PDF] - Mix and Match: Learning-free Controllable Text Generation using Energy Language Models
ACL22
Fatemehsadat Mireshghallah, Kartik Goyal, Taylor Berg-Kirkpatrick [PDF] - COLD Decoding: Energy-based Constrained Text Generation with Langevin Dynamics
arXiv
Lianhui Qin, Sean Welleck, Daniel Khashabi, Yejin Choi [PDF] - Constrained Sampling from Language Models via Langevin Dynamics in Embedding Spaces
arXiv
Sachin Kumar, Biswajit Paria, Yulia Tsvetkov [PDF] - Composable Text Controls in Latent Space with ODEs
arXiv
Guangyi Liu, Zeyu Feng, Yuan Gao, Zichao Yang, Xiaodan Liang, Junwei Bao, Xiaodong He, Shuguang Cui, Zhen Li, Zhiting Hu [PDF]
- Attribute Alignment: Controlling Text Generation from Pre-trained Language Models
EMNLP21 findings
Dian Yu, Zhou Yu, Kenji Sagae [PDF] - Controllable Natural Language Generation with Contrastive Prefixes
ACL22 findings
Jing Qian, Li Dong, Yelong Shen, Furu Wei, Weizhu Chen [PDF] - Fine-Grained Controllable Text Generation Using Non-Residual Prompting
ACL22
Fredrik Carlsson, Joey Öhman, Fangyu Liu, Severine Verlinden, Joakim Nivre, Magnus Sahlgren [PDF] - Tailor: A Prompt-Based Approach to Attribute-Based Controlled Text Generation
arXiv
Kexin Yang, Dayiheng Liu, Wenqiang Lei, Baosong Yang, Mingfeng Xue, Boxing Chen, Jun Xie [PDF] - Quark: Controllable Text Generation with Reinforced [Un]learning
NeurIPS22
Ximing Lu, Sean Welleck, Liwei Jiang, Jack Hessel, Lianhui Qin, Peter West, Prithviraj Ammanabrolu, Yejin Choi [PDF]
-
Disentangled Representation Learning for Non-Parallel Text Style Transfer Vineet John, Lili Mou, Hareesh Bahuleyan, Olga Vechtomova
ACL19
[PDF] [code]motivation
To tackle the problem of disentangling the latent representations of style and content in language models. -
Pre-train and Plug-in: Flexible Conditional Text Generation with Variational Auto-Encoders Yu Duan, Canwen Xu, Jiaxin Pei, Jialong Han, Chenliang Li
ACL20
[PDF] [code]motivation
Flexible when new conditions added to a well trained VAE which requires no more retraining.abstract
Conditional Text Generation has drawn much attention as a topic of Natural Language Generation (NLG) which provides the possibility for humans to control the properties of generated contents. Current conditional generation models cannot handle emerging conditions due to their joint end-to-end learning fashion. When a new condition added, these techniques require full retraining. In this paper, we present a new framework named Pre-train and Plug-in Variational Auto-Encoder (PPVAE) towards flexible conditional text generation. PPVAE decouples the text generation module from the condition representation module to allow “one-to-many” conditional generation. When a fresh condition emerges, only a lightweight network needs to be trained and works as a plug-in for PPVAE, which is efficient and desirable for real-world applications. Extensive experiments demonstrate the superiority of PPVAE against the existing alternatives with better conditionality and diversity but less training effort. -
Plug and Play Autoencoders for Conditional Text Generation Florian Mai, Nikolaos Pappas, Ivan Montero, Noah A. Smith, James Henderson
EMNLP20
[PDF] [code] -
Stylized Dialogue Response Generation Using Stylized Unpaired Texts Yinhe Zheng, Zikai Chen, Rongsheng Zhang, Shilei Huang, Xiaoxi Mao, Minlie Huang
AAAI 21
[PDF] [code] -
Plug and Play Language Models: A Simple Approach to Controlled Text Generation Sumanth Dathathri, Andrea Madotto, Janice Lan, Jane Hung, Eric Frank, Piero Molino, Jason Yosinski, Rosanne Liu
ICLR20
[PDF] [code] -
Hooks in the Headline: Learning to Generate Headlines with Controlled Styles Di Jin, Zhijing Jin, Joey Tianyi Zhou, Lisa Orii, Peter Szolovits
ACL20
[PDF] [code]
- CTRL: A Conditional Transformer Language Model for Controllable Generation Nitish Shirish Keskar, Bryan McCann, Lav R. Varshney, Caiming Xiong, Richard Socher
arXiv
[PDF] [code]control code type
Style by domain
More complex control codes
Triggering specific tasks
Zero-shot code-mixing
- A Controllable Model of Grounded Response Generation Zeqiu Wu, Michel Galley, Chris Brockett, Yizhe Zhang, Xiang Gao, Chris Quirk, Rik Koncel-Kedziorski, Jianfeng Gao, Hannaneh Hajishirzi, Mari Ostendorf, Bill Dolan
AAAI21
[PDF] [code] - Generate Your Counterfactuals: Towards Controlled Counterfactual Generation for Text Nishtha Madaan, Inkit Padhi, Naveen Panwar, Diptikalyan Saha
AAAI21
[PDF] [code] - GEDI: GENERATIVE DISCRIMINATOR GUIDED SEQUENCE GENERATION Ben Krause, Akhilesh Deepak Gotmare, Bryan McCann, Nitish Shirish Keskar, Shafiq Joty, Richard Socher, Nazneen Fatema Rajani
arXiv
[PDF] [code] - MEGATRON-CNTRL: Controllable Story Generation with External Knowledge Using Large-Scale Language Models Peng Xu, Mostofa Patwary, Mohammad Shoeybi, Raul Puri, Pascale Fung, Anima Anandkumar, Bryan Catanzaro
EMNLP20
[PDF] [code]
- Syntax-guided Controlled Generation of Paraphrases Ashutosh Kumar, Kabir Ahuja, Raghuram Vadapalli, Partha Talukdar
TACL20
[PDF] [code] - Transformer-Based Neural Text Generation with Syntactic Guidance Yinghao Li, Rui Feng, Isaac Rehg, Chao Zhang
arXiv
[PDF] [code]
- Plug and Play Language Models: A Simple Approach to Controlled Text Generation Sumanth Dathathri, Andrea Madotto, Janice Lan, Jane Hung, Eric Frank, Piero Molino, Jason Yosinski, Rosanne Liu
ICLR20
[PDF] [code] - A Distributional Approach To Controlled Text Generation Muhammad Khalifa, Hady Elsahar, Marc Dymetman
ICLR21
[PDF] [code] - Plug-and-Blend: A Framework for Controllable Story Generation with Blended Control Codes Zhiyu Lin, Mark O. Riedl
arXiv
[PDF] [code]
- Fast Lexically Constrained Decoding with Dynamic Beam Allocation for Neural Machine Translation Matt Post, David Vilar
NAACL18
[PDF] [code] - CGMH: Constrained Sentence Generation by Metropolis-Hastings Sampling Ning Miao, Hao Zhou, Lili Mou, Rui Yan, Lei Li
AAAI19
[PDF] [code] - Mask and Infill: Applying Masked Language Model to Sentiment Transfer Xing Wu, Tao Zhang, Liangjun Zang, Jizhong Han, Songlin Hu
IJCAI19
[PDF] [code] - Unsupervised Text Style Transfer with Padded Masked Language Models Eric Malmi, Aliaksei Severyn, Sascha Rothe
EMNLP20
[PDF] [code] - LEWIS: Levenshtein Editing for Unsupervised Text Style Transfer Machel Reid, Victor Zhong
ACL21 findings
[PDF] [code] - Plug and Play Autoencoders for Conditional Text Generation Florian Mai, Nikolaos Pappas, Ivan Montero, Noah A. Smith, James Henderson
EMNLP20
[PDF] [code] - Pre-train and Plug-in: Flexible Conditional Text Generation with Variational Auto-Encoders Yu Duan, Canwen Xu, Jiaxin Pei, Jialong Han, Chenliang Li
ACL20
[PDF] [code] - Plug and Play Language Models: A Simple Approach to Controlled Text Generation Sumanth Dathathri, Andrea Madotto, Janice Lan, Jane Hung, Eric Frank, Piero Molino, Jason Yosinski, Rosanne Liu
ICLR20
[PDF] [code] - GeDi: Generative Discriminator Guided Sequence Generation Ben Krause, Akhilesh Deepak Gotmare, Bryan McCann, Nitish Shirish Keskar, Shafiq Joty, Richard Socher, Nazneen Fatema Rajani
arXiv20
[PDF] [code] - Directed Beam Search: Plug-and-Play Lexically Constrained Language Generation Damian Pascual, Beni Egressy, Florian Bolli, Roger Wattenhofer
arXiv20
[PDF] [code] - Plug-and-Blend: A Framework for Controllable Story Generation with Blended Control Codes Zhiyu Lin, Mark Riedl
arXiv21
[PDF] [code] - Plug-and-Play Conversational Models Andrea Madotto, Etsuko Ishii, Zhaojiang Lin, Sumanth Dathathri, Pascale Fung
EMNLP20 findings
[PDF] [code] - FUDGE: Controlled Text Generation With Future Discriminators Kevin Yang, Dan Klein
NAACL21
[PDF] [code] - DExperts: Decoding-Time Controlled Text Generation with Experts and Anti-Experts Alisa Liu, Maarten Sap, Ximing Lu, Swabha Swayamdipta, Chandra Bhagavatula, Noah A. Smith, Yejin Choi
ACL21
[PDF] [code] - Zero-Shot Controlled Generation with Encoder-Decoder Transformers Devamanyu Hazarika, Mahdi Namazifar, Dilek Hakkani-Tur
arXiv21
[PDF] [code]
- Lexically Constrained Decoding for Sequence Generation Using Grid Beam Search Chris Hokamp, Qun Liu
ACL17
[PDF] [code] - Gradient-guided Unsupervised Lexically Constrained Text Generation Lei Sha
EMNLP20
[PDF] [code] - CGMH: Constrained Sentence Generation by Metropolis-Hastings Sampling Ning Miao, Hao Zhou, Lili Mou, Rui Yan, Lei Li
AAAI19
[PDF] [code] - Fast Lexically Constrained Decoding with Dynamic Beam Allocation for Neural Machine Translation Matt Post, David Vilar
NAACL18
[PDF] [code] - Improved Lexically Constrained Decoding for Translation and Monolingual Rewriting J Edward Hu, Huda Khayrallah, Ryan Culkin, Patrick Xia, Tongfei Chen, Matt Post, Benjamin Van Durme
NAACL19
[PDF] [code] - Directed Beam Search: Plug-and-Play Lexically Constrained Language Generation Damian Pascual, Beni Egressy, Florian Bolli, Roger Wattenhofer
arXiv20
[PDF] [code] - Language Generation via Combinatorial Constraint Satisfaction: A Tree Search Enhanced Monte-Carlo Approach Maosen Zhang, Nan Jiang, Lei Li, and Yexiang Xue
EMNLP20 findings
[PDF] [code] - POINTER: Constrained Progressive Text Generation via Insertion-based Generative Pre-training Yizhe Zhang, Guoyin Wang, Chunyuan Li, Zhe Gan, Chris Brockett, Bill Dolan
EMNLP20
[PDF] [code]
- Evaluating Style Transfer for Text Remi Mir, Bjarke Felbo, Nick Obradovich, Iyad Rahwan
NAACL19
[PDF] [code] - A Review of Human Evaluation for Style Transfer Eleftheria Briakou, Sweta Agrawal, Ke Zhang, Joel Tetreault, Marine Carpuat
GEM20201
[PDF] [code]
- Expertise Style Transfer: A New Task Towards Better Communication between Experts and Laymen Yixin Cao, Ruihao Shui, Liangming Pan, Min-Yen Kan, Zhiyuan Liu, Tat-Seng Chua
ACL20
[PDF] [code] - StylePTB: A Compositional Benchmark for Fine-grained Controllable Text Style Transfer Yiwei Lyu, Paul Pu Liang, Hai Pham, Eduard Hovy, Barnabás Póczos, Ruslan Salakhutdinov, Louis-Philippe Morency
NAACL21
[PDF] [code]