TY - GEN
T1 - Diversity regularized autoencoders for text generation
AU - Ko, Hyeseon
AU - Lee, Junhyuk
AU - Kim, Jinhong
AU - Lee, Jongwuk
AU - Shim, Hyunjung
N1 - Publisher Copyright:
© 2020 ACM.
PY - 2020/3/30
Y1 - 2020/3/30
N2 - In this paper, we propose a simple yet powerful text generation model, called diversity regularized autoencoders (DRAE). The key novelty of the proposed model lies in its ability to handle various sentence modifications such as insertions, deletions, substitutions, and maskings, and to take them as input. Because the noise-injection strategy enables an encoder to make the latent distribution smooth and continuous, the proposed model can generate more diverse and coherent sentences. Also, we adopt the Wasserstein generative adversarial networks with a gradient penalty to achieve stable adversarial training of the prior distribution. We evaluate the proposed model using quantitative, qualitative, and human evaluations on two public datasets. Experimental results demonstrate that our model using a noise-injection strategy produces more natural and diverse sentences than several baseline models. Furthermore, it is found that our model shows the synergistic effect of grammar correction and paraphrase generation in an unsupervised way.
AB - In this paper, we propose a simple yet powerful text generation model, called diversity regularized autoencoders (DRAE). The key novelty of the proposed model lies in its ability to handle various sentence modifications such as insertions, deletions, substitutions, and maskings, and to take them as input. Because the noise-injection strategy enables an encoder to make the latent distribution smooth and continuous, the proposed model can generate more diverse and coherent sentences. Also, we adopt the Wasserstein generative adversarial networks with a gradient penalty to achieve stable adversarial training of the prior distribution. We evaluate the proposed model using quantitative, qualitative, and human evaluations on two public datasets. Experimental results demonstrate that our model using a noise-injection strategy produces more natural and diverse sentences than several baseline models. Furthermore, it is found that our model shows the synergistic effect of grammar correction and paraphrase generation in an unsupervised way.
KW - Adversarial training
KW - Data augmentation
KW - Variational autoencoder
UR - https://www.scopus.com/pages/publications/85083029250
U2 - 10.1145/3341105.3373998
DO - 10.1145/3341105.3373998
M3 - Conference contribution
AN - SCOPUS:85083029250
T3 - Proceedings of the ACM Symposium on Applied Computing
SP - 883
EP - 891
BT - 35th Annual ACM Symposium on Applied Computing, SAC 2020
PB - Association for Computing Machinery
T2 - 35th Annual ACM Symposium on Applied Computing, SAC 2020
Y2 - 30 March 2020 through 3 April 2020
ER -