To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
In almost every science fiction movie you’ll see people conversing with machines. Of course, the rise of intelligent personal assistants means you probably do this yourself already. This posting asks: what’s the difference? Also, recent news on Facebook acquisitions, spoken language translation, and sentiment analysis.
A spelling error detection and correction application is typically based on three main components: a dictionary (or reference word list), an error model and a language model. While most of the attention in the literature has been directed to the language model, we show how improvements in any of the three components can lead to significant cumulative improvements in the overall performance of the system. We develop our dictionary of 9.2 million fully-inflected Arabic words (types) from a morphological transducer and a large corpus, validated and manually revised. We improve the error model by analyzing error types and creating an edit distance re-ranker. We also improve the language model by analyzing the level of noise in different data sources and selecting an optimal subset to train the system on. Testing and evaluation experiments show that our system significantly outperforms Microsoft Word 2013, OpenOffice Ayaspell 3.4 and Google Docs.
In this article we tackle the problem of multi-party conversation summarization. We investigate the role of discourse segmentation of a conversation on meeting summarization. First, an unsupervised function segmentation algorithm is proposed to segment the transcript into functionally coherent parts, such as Monologuei (which indicates a segment where speaker i is the dominant speaker, e.g., lecturing all the other participants) or Discussionx1x2, . . ., xn (which indicates a segment where speakers x1 to xn involve in a discussion). Then the salience score for a sentence is computed by leveraging the score of the segment containing the sentence. Performance of our proposed segmentation and summarization algorithms is evaluated using the AMI meeting corpus. We show better summarization performance over other state-of-the-art algorithms according to different metrics.
The two main classes of grammars are (a) hand-crafted grammars, which are developed by language experts, and (b) data-driven grammars, which are extracted from annotated corpora. This paper introduces a statistical method for mapping the elementary structures of a data-driven grammar onto the elementary structures of a hand-crafted grammar in order to combine their advantages. The idea is employed in the context of Lexicalized Tree-Adjoining Grammars (LTAG) and tested on two LTAGs of English: the hand-crafted LTAG developed in the XTAG project, and the data-driven LTAG, which is automatically extracted from the Penn Treebank and used by the MICA parser. We propose a statistical model for mapping any elementary tree sequence of the MICA grammar onto a proper elementary tree sequence of the XTAG grammar. The model has been tested on three subsets of the WSJ corpus that have average lengths of 10, 16, and 18 words, respectively. The experimental results show that full-parse trees with average F1-scores of 72.49, 64.80, and 62.30 points could be built from 94.97%, 96.01%, and 90.25% of the XTAG elementary tree sequences assigned to the subsets, respectively. Moreover, by reducing the amount of syntactic lexical ambiguity of sentences, the proposed model significantly improves the efficiency of parsing in the XTAG system.
The Journal of Natural Language Engineering (JNLE) has enjoyed another very successful year. Two years after being accepted into Thomson Reuters Citation Index and being indexed in many of their products (including both the Science and the Social Science editions of the Journals Citation Rankings (JCR)), the journal further established itself as a leading forum for high-quality articles covering all aspects of Natural Language Processing research, including, but not limited to, the engineering of natural language methods and applications. I am delighted to report an increased number of submissions reaching a total of 92 between January–September 2014.
Prof. Geoffrey Leech, an influential scholar who left an indelible mark across many linguistic disciplines and inspired a generation of linguists, died suddenly on 19 August 2014.
This paper describes the Kestrel text normalization system, a component of the Google text-to-speech synthesis (TTS) system. At the core of Kestrel are text-normalization grammars that are compiled into libraries of weighted finite-state transducers (WFSTs). While the use of WFSTs for text normalization is itself not new, Kestrel differs from previous systems in its separation of the initial tokenization and classification phase of analysis from verbalization. Input text is first tokenized and different tokens classified using WFSTs. As part of the classification, detected semiotic classes – expressions such as currency amounts, dates, times, measure phases, are parsed into protocol buffers (https://code.google.com/p/protobuf/). The protocol buffers are then verbalized, with possible reordering of the elements, again using WFSTs. This paper describes the architecture of Kestrel, the protocol buffer representations of semiotic classes, and presents some examples of grammars for various languages. We also discuss applications and deployments of Kestrel as part of the Google TTS system, which runs on both server and client side on multiple devices, and is used daily by millions of people in nineteen languages and counting.
This book contains some fundamental results and insights relating to the automation of negotiation. The focus is on competitive negotiations where the outcome (i.e., whether or not an agreement is reached and if so what agreement is reached) depends on the strategies of all parties. We studied how to model and analyse such strategic negotiations using game theory.
From a strategic point of view, there are four key determinants of the outcome of a negotiation:
the negotiation deadline,
the discount factor,
the information that the players have about the negotiation parameters,
the protocol/procedure used for negotiation, and
the negotiation agenda.
We explored a number of procedures and compared them in terms of the properties (i.e., time of agreement, Pareto optimality, uniqueness, symmetry) of their equilibrium outcomes. But this book is not just about strategic negotiations, it is about their automation. In this vein, we examined the negotiation procedures and their equilibria from a computational perspective. The following are the main sources of complexity that can make the automation of negotiation difficult:
The type of issues. In terms of the computational complexity of calculating offers, divisible issues are easier to deal with than indivisible or discretely divisible issues.
The form of utility functions. In terms of the computational complexity of calculating offers, linear utility functions are easier to deal with than nonlinear ones.
While normal-form games capture the strategic structure of decision-making settings, they abstract away one key aspect of playing games that seems quite central to their character. Specifically, they assume that players make choices and act simultaneously, with no knowledge of the choices of their counterparts. But in many strategic situations, the players do not move simultaneously: they take turns in making their moves. An obvious example of such a situation is the game of chess. Another example is a bargaining scenario where a buyer and a seller take turns in making offers to each other in order to reach an agreement on the price of a commodity. Such situations may be modelled using sequential-moves games. In this chapter, we will study how these games are represented, and what notions of equilibria are used to analyse and solve them.
In a purely sequential-moves game, the players not only take turns in making their moves, but they typically know what the players did in all the previous moves. In contrast, when a player makes a move in a simultaneous-moves game, he does not know the other players' moves.
Purely sequential-moves and purely simultaneous-moves situations rarely arise in practice; many interesting real-world situations involve both simultaneous and sequential moves. Thus, we will learn how to model such situations with games and how to analyse and solve those games.
So far in this book, we have been focusing on bilateral (i.e., one-to-one) negotiations. One can easily imagine situations where more than two agents might need to negotiate with one another. For example, in a typical trading scenario, a seller might want to negotiate with multiple potential buyers. For such scenarios, we need multilateral negotiation protocols, that is protocols allowing one-to-many and many-to-many negotiations. In this chapter, we introduce a number of such protocols.
8.1 Alternating offers protocol with multiple bargainers
Consider the following many-to-many negotiation scenario. A group of two or more agents are together able to jointly produce a surplus, that is some joint gains. However, no subset of the group is able to do this. For example, say we have a group of five musicians that come together and create a piece of music that no subset could produce. By selling their music, they make a joint profit of £500,000. The key question here is “how should the joint gains be split between the individuals?”. We can think of the joint gains as a pie of unit size. The individuals must decide how they will divide the pie between themselves. An agreement requires the approval of all the players; no subset is allowed to reach an agreement.
Since there are more than two agents, we cannot use the alternating offers protocol described in Chapter 5. The protocol must be extended to deal with multiple players.
So far in this book, we have studied negotiation using two approaches: strategic and heuristic. The strategic approach is a non-cooperative game-theoretic approach. There is, however, another possibility: an axiomatic approach. The axiomatic approach will be the main focus of this chapter. Among the three approaches, the strategic and heuristic ones are better suited to the design of negotiating agents. Nevertheless, we include this chapter in order to explain the key concepts that underlie an axiomatic approach and, in the spirit of the Nash program (Nash, 1953; Binmore, 1985; Binmore and Dasgupta, 1987), show that the outcomes of some of the strategic models of bargaining can be very close to the outcomes of some axiomatic models.
We begin by understanding the key differences between cooperative and non-cooperative games. Following this, we introduce some of the prominent axiomatic models for single and multi-issue negotiation and show how some of their outcomes relate to those of strategic models. In the end, we give a comparative account of the axiomatic and strategic approaches in terms of their similarities and differences.
11.1 Background
Game-theoretic analysis of negotiation can be done using one of two possible approaches: axiomatic or strategic. In the former approach, negotiation is modelled as a cooperative game, while in the latter, it is modelled as a non-cooperative game.