### Table of Contents

# Natural Logic for Textual Inference

**Bill MacCartney, Christopher D. Manning (2007)**

## Introduction

This paper deals with “**natural logic**” which is a system of logical inference that operates over natural language. Usually the approaches for natural language inference are either robust but shallow or deep but brittle. The system proposed in this paper aims to be in the middle of the existent approaches and avoids, for instance, the error when translating a natural language to first-order logic.

One key concept in the theory of natural logic is “monotonicity” in which, instead of using quantifiers, the concepts or constraints expressed in a sentence are expanded or contracted. This way, linguistics expressions can be represented as *upward-monotone*, *downward monotone*, or *non-monotone* semantic functions.

The developed system is called **NatLog** and has an architecture with three main stages:

- Linguistic preprocessing - parsing input sentences and monotonicity marking;
- Alignment - alignment between the premise and the hypothesis in terms of atomic edits;
- Entailment classification - predicting the final entailment relation based on the entailment relation of atomic edits, where the later are predicted based on the following features: a type of an atomic edit, the effective monotonicity at the effected token span, and various lexical features.

## Comments

- This work represents the first computational model of natural logic;
- In natural logic, entailment is a semantic containment relation over expressions of all types, including words and phrases as well as sentences. Authors define the entailment relation recursively over the semantic types familiar from Montague semantics;
- The training data used to predict the entailment relation of atomic edits was created for this specific experiment.
- The system was tested on the portion (having a single premise) of
*FraCaS test suite*containing 346 inference problems reminiscent of a textbook on formal semantics. Each problem has 3 possible answers: yes, no, unknown; - It was also tested on the
*RTE3 test suite*which contains much longer and “natural” premises;

## Discussion

- How much does this approach contribute to the existent logical inference approaches for natural language?
- Language is fuzzy and this approach captures simple sentences. We are not sure that it can be generalized easily.
- It is good that the examples in test data contain 3 different answers.
- Disadvantage: If you combine a lot of atomic edits, the probability of getting the right answer is lower.
- We liked the evaluation presented in the paper and the results interpretation (that is not very usual in semantics).
- Disadvantage: In composition of entailment relations of atomic edits, the authors while mentioning the obvious compositions do not mention non-obvious logically valid compositions, e.g. exclusivereverseexclusive () or forwardexclusiveexclusive (). This fact puts the question whether they have implemented these two compositions or not in the system.

Written by Ximena Gutiérrez.