TY - GEN
T1 - Sequence to Sequence CycleGAN for Non-Parallel Sentiment Transfer with Identity Loss Pretraining
AU - Crisdayanti, Ida Ayu Putu Ari
AU - Lee, Jee Hyong
N1 - Publisher Copyright:
© 2020 ACM.
PY - 2020/10/13
Y1 - 2020/10/13
N2 - Sentiment transfer has been explored as non-parallel transfer tasks in natural language processing. Previous works depend on a single encoder to disentangle either positive or negative style from its content and rely on a style representation to transfer the style attributes. Utilizing a single encoder to learn disentanglement in both styles might not sufficient due to the different characteristics of each sentiment represented by various vocabularies in the corresponding style. To this end, we propose a sequence to sequence CycleGAN which trains different text generators (encoder-decoder) for each style transfer direction. Learning disentangled latent representations leads previous works to high sentiment accuracy but suffer to preserve the content of the original sentences. In order to manage the content preservation, we pretrained our text generator as autoencoder using the identity loss. The model shows an improvement in sentiment accuracy and BLEU score which indicates better content preservation. It leads our model to a better overall performance compared to baselines.
AB - Sentiment transfer has been explored as non-parallel transfer tasks in natural language processing. Previous works depend on a single encoder to disentangle either positive or negative style from its content and rely on a style representation to transfer the style attributes. Utilizing a single encoder to learn disentanglement in both styles might not sufficient due to the different characteristics of each sentiment represented by various vocabularies in the corresponding style. To this end, we propose a sequence to sequence CycleGAN which trains different text generators (encoder-decoder) for each style transfer direction. Learning disentangled latent representations leads previous works to high sentiment accuracy but suffer to preserve the content of the original sentences. In order to manage the content preservation, we pretrained our text generator as autoencoder using the identity loss. The model shows an improvement in sentiment accuracy and BLEU score which indicates better content preservation. It leads our model to a better overall performance compared to baselines.
KW - content preservation
KW - CycleGAN
KW - identity loss
KW - sentiment transfer
KW - sequence to sequence
UR - https://www.scopus.com/pages/publications/85097405291
U2 - 10.1145/3400286.3418249
DO - 10.1145/3400286.3418249
M3 - Conference contribution
AN - SCOPUS:85097405291
T3 - ACM International Conference Proceeding Series
SP - 26
EP - 30
BT - Proceedings of the 2020 Research in Adaptive and Convergent Systems, RACS 2020
PB - Association for Computing Machinery
T2 - 2020 Research in Adaptive and Convergent Systems, RACS 2020
Y2 - 13 October 2020 through 16 October 2020
ER -