Hostname: page-component-77f85d65b8-v2srd Total loading time: 0 Render date: 2026-03-28T09:46:28.314Z Has data issue: false hasContentIssue false

Constrained BERT BiLSTM CRF for understanding multi-sentence entity-seeking questions

Published online by Cambridge University Press:  13 February 2020

Danish Contractor*
Affiliation:
IBM Research AI Indian Institute of Technology Delhi, New Delhi, India
Barun Patra
Affiliation:
Microsoft Corporation, Redmond, WA, USA
Mausam
Affiliation:
Indian Institute of Technology Delhi, New Delhi, India
Parag Singla
Affiliation:
Indian Institute of Technology Delhi, New Delhi, India
*
*Corresponding author. E-mail: dcontrac@in.ibm.com
Rights & Permissions [Opens in a new window]

Abstract

We present the novel task of understanding multi-sentence entity-seeking questions (MSEQs), that is, the questions that may be expressed in multiple sentences, and that expect one or more entities as an answer. We formulate the problem of understanding MSEQs as a semantic labeling task over an open representation that makes minimal assumptions about schema or ontology-specific semantic vocabulary. At the core of our model, we use a BiLSTM (bidirectional LSTM) conditional random field (CRF), and to overcome the challenges of operating with low training data, we supplement it by using BERT embeddings, hand-designed features, as well as hard and soft constraints spanning multiple sentences. We find that this results in a 12–15 points gain over a vanilla BiLSTM CRF. We demonstrate the strengths of our work using the novel task of answering real-world entity-seeking questions from the tourism domain. The use of our labels helps answer 36% more questions with 35% more (relative) accuracy as compared to baselines. We also demonstrate how our framework can rapidly enable the parsing of MSEQs in an entirely new domain with small amounts of training data and little change in the semantic representation.

Information

Type
Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution, and reproduction in any medium, provided the original work is properly cited.
Copyright
© The Author(s), 2020. Published by Cambridge University Press
Figure 0

Fig 1. An MSEQ annotated with our semantic labels.

Figure 1

Fig. 2. Schematic representation of the QA system.

Figure 2

Table 1. Related work: QA

Figure 3

Fig. 3. BERT BiLSTM CCM with features for sequence labeling.

Figure 4

Fig. 4. Snippet of the second questionnaire given to AMT workers.

Figure 5

Table 2. Agreement for entity labels on AMT

Figure 6

Table 3. Sequence tagger F1 scores using CRF with all features (feat), CCM with all features and constraints, and partially supervised CCM over partially labeled crowd data. The second set of results mirrors these settings using a bidirectional LSTM CRF. Results are statistically significant (paired t test, p value $<0.02$ for aggregate F1 for each CRF and corresponding CCM model pair). Models with “PS” as a prefix use partial supervision

Figure 7

Table 4. (i) Precision and recall of entity.type with and without CCM inference

Figure 8

Table 5. Performance of negation detection using gold sequence labels and system generated labels

Figure 9

Table 6. QA task results using the Google places web API as knowledge source

Figure 10

Table 7. Some sample questions from our test set and the answers returned by our system. Answers in green are identified as correct while those in red are incorrect

Figure 11

Table 8. Classification of errors made by our MSEQ-labels-based answering system (using Google Places web API as knowledge source)

Figure 12

Table 9. Labeling performance for book recommendation questions (paired t test, p value $\lt 0.01$ for aggregate F1 in vanilla CRF and CCM model pairs & BiLSTM CRF and CCM model pairs)