论文标题
如此不同,但是如此相似!受限的无监督文本样式转移
So Different Yet So Alike! Constrained Unsupervised Text Style Transfer
论文作者
论文摘要
在域之间的自动传输文本已在最近变得很流行。它的目的之一是保留文本的语义内容从源到目标域。但是,它不能明确维护源和翻译文本之间的其他属性,例如文本长度和描述性。在转移中维持限制有多个下游应用程序,包括数据增强和偏见。我们通过向生成对抗网络(GAN)模型家族引入两个互补损失,引入了这种约束无监督文本样式转移的方法。与GAN中使用的竞争损失不同,我们引入了合作损失,歧视者和发电机合作并减少了相同的损失。第一个是对比损失,第二个是分类损失,旨在进一步使潜在空间正规化,并在范围内将类似的句子仔细地结合在一起。我们证明,这种训练保留了多个基准数据集的域之间的词汇,句法和域特异性约束,包括一个以上属性更改的数据集。我们表明,根据自动化和人类评估措施,互补的合作损失可改善文本质量。
Automatic transfer of text between domains has become popular in recent times. One of its aims is to preserve the semantic content of text being translated from source to target domain. However, it does not explicitly maintain other attributes between the source and translated text, for e.g., text length and descriptiveness. Maintaining constraints in transfer has several downstream applications, including data augmentation and de-biasing. We introduce a method for such constrained unsupervised text style transfer by introducing two complementary losses to the generative adversarial network (GAN) family of models. Unlike the competing losses used in GANs, we introduce cooperative losses where the discriminator and the generator cooperate and reduce the same loss. The first is a contrastive loss and the second is a classification loss, aiming to regularize the latent space further and bring similar sentences across domains closer together. We demonstrate that such training retains lexical, syntactic, and domain-specific constraints between domains for multiple benchmark datasets, including ones where more than one attribute change. We show that the complementary cooperative losses improve text quality, according to both automated and human evaluation measures.