TransCP: A Transformer Pointer Network for Generic Entity Description Generation with Explicit Content-Planning

Research output: Contribution to journalArticlepeer-review

1 Citation (Scopus)

Abstract

We study neural data-to-text generation to generate a sentence to describe a target entity based on its attributes. Specifically, we address two problems of the encoder-decoder framework for data-to-text generation: i) how to encode a non-linear input (e.g., a set of attributes); and ii) how to order the attributes in the generated description. Existing studies focus on the encoding problem but do not address the ordering problem, i.e., they learn the content-planning implicitly. The other approaches focus on two-stage models but overlook the encoding problem. To address the two problems at once, we propose a model named TransCP to explicitly learn content-planning and integrate them into a description generation model in an end-to-end fashion. We propose a novel Transformer-based Pointer Network with gated residual attention and importance masking to learn a content-plan. To integrate the content-plan with a description generator, we propose a tracking mechanism to trace the extent to which the content-plan is exposed in the previous decoding time-step. This helps the description generator select the attributes to be mentioned in proper order. Experimental results show that our model consistently outperforms state-of-the-art baselines by up to 2% and 3% in terms of BLEU score on two real-world datasets.
Original languageEnglish
Pages (from-to)13070 - 13082
JournalIEEE Transactions on Knowledge and Data Engineering
Volume35
Issue number12
DOIs
Publication statusPublished - May 2023

Keywords

  • Knowledge base
  • natural language generation
  • entity description
  • content planning

Fingerprint

Dive into the research topics of 'TransCP: A Transformer Pointer Network for Generic Entity Description Generation with Explicit Content-Planning'. Together they form a unique fingerprint.

Cite this