Hostname: page-component-77f85d65b8-g4pgd Total loading time: 0 Render date: 2026-03-28T03:46:01.142Z Has data issue: false hasContentIssue false

A transformation-driven approach for recognizing textual entailment

Published online by Cambridge University Press:  16 June 2016

ROBERTO ZANOLI
Affiliation:
Human Language Technology, Fondazione Bruno Kessler, 38123 Trento, Italy e-mail: zanoli@fbk.eu
SILVIA COLOMBO
Affiliation:
Edinburgh University School of Informatics, 11 Crichton St, Edinburgh EH8 9LE, UK e-mail: s1132418@sms.ed.ac.uk
Rights & Permissions [Opens in a new window]

Abstract

Textual Entailment is a directional relation between two text fragments. The relation holds whenever the truth of one text fragment, called Hypothesis (H), follows from another text fragment, called Text (T). Up until now, using machine learning approaches for recognizing textual entailment has been hampered by the limited availability of data. We present an approach based on syntactic transformations and machine learning techniques which is designed to fit well with a new type of available data sets that are larger but less complex than data sets used in the past. The transformations are not predefined, but calculated from the data sets, and then used as features in a supervised learning classifier. The method has been evaluated using two data sets: the SICK data set and the EXCITEMENT English data set. While both data sets are of a larger order of magnitude than data sets such as RTE-3, they are also of lower levels of complexity, each in its own way. SICK consists of pairs created by applying a predefined set of syntactic and lexical rules to its T and H pairs, which can be accurately captured by our transformations. The EXCITEMENT English data contains short pieces of text that do not require a high degree of text understanding to be annotated. The resulting AdArte system is simple to understand and implement, but also effective when compared with other existing systems. AdArte has been made freely available with the EXCITEMENT Open Platform, an open source platform for textual inference.

Information

Type
Articles
Copyright
Copyright © Cambridge University Press 2016 
Figure 0

Table 1. SICK: examples of annotated T–H pairs

Figure 1

Table 2. EXCITEMENT: examples of annotated T–H pairs

Figure 2

Fig. 1. Dependency tree of the Text: The girl is spraying the plants with water (on the left). Dependency tree of the Hypothesis: The boy is spraying the plants with water (on the right).

Figure 3

Table 3. Transformations for some T–H pairs in SICK with similar meaning. On the left, the pairs with the rules applied to produce them. On the right, the transformations produced by our method. An index on T and H highlights the words in the transformation, e.g., 3 (Ins(H3:auxpass)) means that the transformation regards the word at position 3 in H (is)

Figure 4

Table 4. Transformations for some T–H pairs in SICK with contrasting meaning or where the truth of one of its fragments cannot be inferred from the other one. On the left, the pairs with the rules applied to produce them. On the right, the transformations produced by our method

Figure 5

Fig. 2. System architecture.

Figure 6

Table 5. Accuracy measure for the baseline and the basic system configurations. Computing time in seconds for training and for test (in parentheses)

Figure 7

Fig. 3. Learning curve (semi-log scale) calculated with basic-system#1 (SVM).

Figure 8

Table 6. Values in brackets highlight improvements or deterioration in performance with respect to values obtained by basic-system#1 (SVM). Computing time in seconds for training and for test (in parentheses)

Figure 9

Table 7. One resource removed at a time

Figure 10

Table 8. System accuracy for all three entailment relations. In parentheses, the number of examples in the data set

Figure 11

Table 9. Confusion matrix on the SICK test set

Figure 12

Table 10. System accuracy on the T–H pairs generated by applying the preserving, negative and scrambling rules

Figure 13

Table 11. Results of different configurations for each data set of EXCITEMENT. Precision, Recall and F1 measure are calculated on the positive class

Figure 14

Table 12. AdArte as compared with the SemEval-2014 Task#1 systems and the Bowman’s neural network model. Below, AdArte, TIE and P1EDA are evaluated on two-class entailment problem

Figure 15

Table 13. Systems ranked by Accuracy values. A number after the systems shows their position in the ranking when the F1 for the positive class is considered