论文标题
变压器GCRF:带有一般条件随机字段的中文掉落代词
Transformer-GCRF: Recovering Chinese Dropped Pronouns with General Conditional Random Fields
论文作者
论文摘要
代词通常会在中文对话中删除,并且恢复掉落的代词对于NLP应用程序(例如机器翻译)很重要。现有的方法通常将其作为序列标记任务提出,以预测每个令牌及其类型之前是否有一个掉落的代词。每种话语都被认为是一个序列,并独立标记。尽管这些方法已经显示出希望,但标记每种话语都独立地忽略了邻近话语中代词之间的依赖性。建模这些依赖关系对于改善掉落代词恢复的性能至关重要。在本文中,我们提出了一个新颖的框架,该框架将变压器网络的强度与一般条件随机场(GCRF)结合在一起,以建模相邻话语中代词之间的依赖关系。三个中国对话数据集中的结果表明,变压器GCRF模型的表现优于最先进的删除代词恢复模型。探索性分析还表明,GCRF确实有助于捕获邻近话语中代词之间的依赖性,从而有助于改善性能。
Pronouns are often dropped in Chinese conversations and recovering the dropped pronouns is important for NLP applications such as Machine Translation. Existing approaches usually formulate this as a sequence labeling task of predicting whether there is a dropped pronoun before each token and its type. Each utterance is considered to be a sequence and labeled independently. Although these approaches have shown promise, labeling each utterance independently ignores the dependencies between pronouns in neighboring utterances. Modeling these dependencies is critical to improving the performance of dropped pronoun recovery. In this paper, we present a novel framework that combines the strength of Transformer network with General Conditional Random Fields (GCRF) to model the dependencies between pronouns in neighboring utterances. Results on three Chinese conversation datasets show that the Transformer-GCRF model outperforms the state-of-the-art dropped pronoun recovery models. Exploratory analysis also demonstrates that the GCRF did help to capture the dependencies between pronouns in neighboring utterances, thus contributes to the performance improvements.