Hostname: page-component-77f85d65b8-45ctf Total loading time: 0 Render date: 2026-03-29T17:38:10.603Z Has data issue: false hasContentIssue false

SAN-T2T: An automated table-to-text generator based on selective attention network

Published online by Cambridge University Press:  05 May 2023

Haijie Ding
Affiliation:
Jiangsu Key Laboratory of Big Data Security & Intelligent Processing, Nanjing University of Posts and Telecommunications, Nanjing, China
Xiaolong Xu*
Affiliation:
School of Computer Science, Nanjing University of Posts and Telecommunications, Nanjing, China
*
Corresponding author: Xiaolong Xu; Email: xuxl@njupt.edu.cn
Rights & Permissions [Opens in a new window]

Abstract

Table-to-text generation aims to generate descriptions for structured data (i.e., tables) and has been applied in many fields like question-answering systems and search engines. Current approaches mostly use neural language models to learn alignment between output and input based on the attention mechanisms, which are still flawed by the gradual weakening of attention when processing long texts and the inability to utilize the records’ structural information. To solve these problems, we propose a novel generative model SAN-T2T, which consists of a field-content selective encoder and a descriptive decoder, connected with a selective attention network. In the encoding phase, the table’s structure is integrated into its field representation, and a content selector with self-aligned gates is applied to take advantage of the fact that different records can determine each other’s importance. In the decoding phase, the content selector’s semantic information enhances the alignment between description and records, and a featured copy mechanism is applied to solve the rare word problem. Experiments on WikiBio and WeatherGov datasets show that SAN-T2T outperforms the baselines by a large margin, and the content selector indeed improves the model’s performance.

Information

Type
Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2023. Published by Cambridge University Press
Figure 0

Figure 1. Architecture of SAN-T2T.

Figure 1

Figure 2. The wiki infobox of Penny Ramsey and the preprocessed representation.

Figure 2

Figure 3. The structure of LSTM unit.

Figure 3

Algorithm 1. Gated Content-Selection Algorithm

Figure 4

Figure 4. Architecture of the gated content selector.

Figure 5

Table 1. The Wikipedia infobox of Frederick Parker–Rhodes

Figure 6

Table 2. Results on WikiBio (test set)

Figure 7

Table 3. Effects of text length on WikiBio (test set)

Figure 8

Table 4. Effects of different beam width on WikiBio

Figure 9

Table 5. Results on RotoWire (test set)

Figure 10

Table 6. Results on WeatherGov (test set)

Figure 11

Table 7. Effects of different $\gamma$ on WeatherGov

Figure 12

Table 8. The Wikipedia infobox of Alonzo H. Cushing

Figure 13

Table 9. The generated descriptions for Alonzo H. Cushing

Figure 14

Table 10. Human Evaluation on WikiBio (test set)

Figure 15

Table 11. Weather data for Northfield, Minnesota on 2009-08-1

Figure 16

Table 12. The generated descriptions for Northfield, Minnesota on 2009-08-1

Figure 17

Table 13. The box score and line score for Clippers - Bucks in 2014-21-20

Figure 18

Table 14. The generated descriptions for Clippers - Bucks

Figure 19

Figure 5. An example of visualization for the attention. The image above is the attention from Seq2Seq model, while the below one is from SAN-T2T. The vertical axis represents the text generated by the model, while the horizontal represents the fields’ value. -lrb- and -rrb- indicate brackets (). Deeper colors depict a higher attention score.