To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
The first-order isomorphism problem is to decide whether two non-recursive types using product- and function-type constructors are isomorphic under the axioms of commutative and associative products, and currying and distributivity of functions over products. We show that this problem can be solved in $O(n \log^2 n)$ time and $O(n)$ space, where $n$ is the input size. This result improves upon the $O(n^2 \log n)$ time and $O(n^2)$ space bounds of the best previous algorithm. We also describe an $O(n)$ time algorithm for the linear isomorphism problem, which does not include the distributive axiom, thereby improving upon the $O(n \log n)$ time of the best previous algorithm for this problem.
It is interesting to note that the very first papers related to isomorphism of types were written before the notion itself appeared. Their subjects were the study of equality of terms defined on numbers, the isomorphism of objects in certain categories and the invertibility of $\lambda$-terms, but not the isomorphism of types ‘as such’. One may cite the so-called ‘Tarsky High School Algebra Problem’: whether all identities between terms built from $+, x, \uparrow$, variables and constants are derivable from basic ‘high school equalities’, like $(xy)^z=x^zy^z$. The earliest publications related to this problem date from the 1940s, cf. Birkhoff (1940) – an extensive bibliography may be found in Burris and Yeats (2002).
We consider shifted equality sets of the form EG(a,g1,g2) = {ω | g1(ω) = ag2(ω)}, where g1 and g2 are nonerasingmorphisms and a is a letter. We are interested in the familyconsisting of the languages h(EG(J)), where h is a coding and(EG(J)) is a shifted equality set. We prove several closureproperties for this family. Moreover, we show that everyrecursively enumerable language L ⊆ A* is a projectionof a shifted equality set, that is, L = πA(EG(a,g1,g2)) for some (nonerasing) morphisms g1 and g2 and aletter a, where πA deletes the letters not in A. Thenwe deduce that recursively enumerable star languages coincide withthe projections of equality sets.
We investigate the structure of “worst-case” quasi reduced ordered decision diagrams and Boolean functions whose truth tables are associated to: we suggest different ways to count and enumerate them. We, then, introduce a notion of complexity which leads to the concept of “hard” Boolean functions as functions whose QROBDD are “worst-case” ones. So we exhibit the relation between hard functions and the Storage Access function (also known as Multiplexer).
The paper presents an elementary approach for the calculation of the entropyof a class of languages. This approach is based on the consideration ofroots of a real polynomial and is also suitable for calculating theBernoulli measure. The class of languages we consider here is ageneralisation of the Łukasiewicz language.
Dans cet article, nous exploitons la réductibilité d'unpolynômed'une variable pour calculer efficacement l'idéal des relationsalgébriques entre ses racines.
We study deterministic one-way communication complexity of functions with Hankel communication matrices. Some structural properties of such matrices are establishedand applied to the one-way two-party communication complexity of symmetric Boolean functions.It is shown that the number of required communication bits does not depend on the communication direction, provided thatneither direction needs maximum complexity. Moreover, in order to obtain an optimal protocol, it is in any case sufficient to consider only the communication directionfrom the party with the shorter input to the other party. These facts do not hold for arbitrary Boolean functions in general. Next, gaps between one-way and two-way communication complexity for symmetric Boolean functions are discussed.Finally, we give some generalizations to the case of multiple parties.
In previous work (Gough and Way 2004), we showed that our Example-Based Machine Translation (EBMT) system improved with respect to both coverage and quality when seeded with increasing amounts of training data, so that it significantly outperformed the on-line MT system Logomedia according to a wide variety of automatic evaluation metrics. While it is perhaps unsurprising that system performance is correlated with the amount of training data, we address in this paper the question of whether a large-scale, robust EBMT system such as ours can outperform a Statistical Machine Translation (SMT) system. We obtained a large English-French translation memory from Sun Microsystems from which we randomly extracted a near 4K test set. The remaining data was split into three training sets, of roughly 50K, 100K and 200K sentence-pairs in order to measure the effect of increasing the size of the training data on the performance of the two systems. Our main observation is that contrary to perceived wisdom in the field, there appears to be little substance to the claim that SMT systems are guaranteed to outperform EBMT systems when confronted with ‘enough’ training data. Our tests on a 4.8 million word bitext indicate that while SMT appears to outperform our system for French-English on a number of metrics, for English-French, on all but one automatic evaluation metric, the performance of our EBMT system is superior to the baseline SMT model.
This paper presents a very simple and effective approach to using parallel corpora for automatic bilingual lexicon acquisition. The approach, which uses the Random Indexing vector space methodology, is based on finding correlations between terms based on their distributional characteristics. The approach requires a minimum of preprocessing and linguistic knowledge, and is efficient, fast and scalable. In this paper, we explain how our approach differs from traditional cooccurrence-based word alignment algorithms, and we demonstrate how to extract bilingual lexica using the Random Indexing approach applied to aligned parallel data. The acquired lexica are evaluated by comparing them to manually compiled gold standards, and we report overlap of around 60%. We also discuss methodological problems with evaluating lexical resources of this kind.
Parallel texts have become a vital element for natural language processing. We present a panorama of current research activities related to parallel texts, and offer some thoughts about the future of this rich field of investigation.
Statistical, linguistic, and heuristic clues can be used for the alignment of words and multi-word units in parallel texts. This article describes the clue alignment approach and the optimization of its parameters using a genetic algorithm. Word alignment clues can come from various sources such as statistical alignment models, co-occurrence tests, string similarity scores and static dictionaries. A genetic algorithm implementing an evolutionary procedure can be used to optimize the parameters necessary for combining available clues. Experiments on English/Swedish bitext show a significant improvement of about 6% in F-scores compared to the baseline produced by statistical word alignment.Most of the work described in this paper was carried out at the Department of Linguistics and Philology at Uppsala University. I would like to acknowledge technical and scientific support by people at the department in Uppsala.
Broad coverage, high quality parsers are available for only a handful of languages. A prerequisite for developing broad coverage parsers for more languages is the annotation of text with the desired linguistic representations (also known as “treebanking”). However, syntactic annotation is a labor intensive and time-consuming process, and it is difficult to find linguistically annotated text in sufficient quantities. In this article, we explore using parallel text to help solving the problem of creating syntactic annotation in more languages. The central idea is to annotate the English side of a parallel corpus, project the analysis to the second language, and then train a stochastic analyzer on the resulting noisy annotations. We discuss our background assumptions, describe an initial study on the “projectability” of syntactic relations, and then present two experiments in which stochastic parsers are developed with minimal human intervention via projection from English.
In this article we illustrate and evaluate an approach to create high quality linguistically annotated resources based on the exploitation of aligned parallel corpora. This approach is based on the assumption that if a text in one language has been annotated and its translation has not, annotations can be transferred from the source text to the target using word alignment as a bridge. The transfer approach has been tested and extensively applied for the creation of the MultiSemCor corpus, an English/Italian parallel corpus created on the basis of the English SemCor corpus. In MultiSemCor the texts are aligned at the word level and word sense annotated with a shared inventory of senses. A number of experiments have been carried out to evaluate the different steps involved in the methodology and the results suggest that the transfer approach is one promising solution to the resource bottleneck. First, it leads to the creation of a parallel corpus, which represents a crucial resource per se. Second, it allows for the exploitation of existing (mostly English) annotated resources to bootstrap the creation of annotated corpora in new (resource-poor) languages with greatly reduced human effort.
Standard parameter estimation schemes for statistical translation models can struggle to find reasonable settings on some parallel corpora. We show how auxiliary information can be used to constrain the procedure directly by restricting the set of alignments explored during parameter estimation. This enables the integration of bilingual and monolingual knowledge sources while retaining the flexibility of the underlying models. We demonstrate the effectiveness of this approach for incorporating linguistic and domain-specific constraints on various parallel corpora, and consider the importance of using the context of the parallel text to guide the application of such constraints.
There are large gaps between theory, simulations, laboratory experiments, field experiments, and field deployments, essentially related to the acceptable degree of human intervention permitted to change the system, i.e., getting things to work. A research program typically loops through these stages as experimental results suggest new directions for theory. The basic assumption of this chapter is that since ENS are fundamentally about connecting the physical world to large scale networks, therefore physical experiments must be performed to verify their proper operation. It is necessary to make simplifying assumptions in early design stages, but no system that is designed to interact with the real world can really be trusted without a set of real deployments that test these assumptions. This chapter describes the basic design stages of the development of experimental systems from the formulation of objectives and assembly of the design and application team through deployment. Throughout, a robotic NIMS is used as an example (see Chapter 12 for background).
Deciding on priorities
The most important decision in the whole process is the set of priorities. What particular problems will the system solve? Can slightly different problems be addressed at much less cost and effort? Who will need to be consulted or be on the design team to help determine the problem formulation? The notion of an experiment most of us carry around is based on the formal high school/university laboratory report with objectives, theory, method, results, analysis, and conclusion.
This chapter treats two related topics: detection and estimation of multiple sources, and the sharing of a communication medium among multiple users. In the former, knowledge of the statistical properties of the sources together with control of the layout of sensors can be used to achieve a degree of source separation. These techniques can also be applied in multiple access communications, but in addition the communicators can cooperate in terms of channel assignments and modulation type to make the task much easier. The chapter begins with a general discussion of interference models and mitigation methods before describing some source separation techniques. Then multiplexing methods for communications systems are discussed, before presentation of the effects of the available communication infrastructure and network topology on the choices of multiple access schemes. Adaptive power control is revisited from the point of view of its effect on the capacity of multiple access schemes.
Interference models
Interference is an unwanted signal sharing a channel being monitored, whether for sensing purposes or for communications. In contrast with thermal noise which is added in the receiver, interference is generally time-varying, e.g., another source or communicator. Moreover, it is confined to a particular band of frequencies, unlike Gaussian noise which is modeled as being uniform over the whole range of communication frequencies. If the interference is deliberate, it is termed jamming. In the design of systems that are potentially subject to jamming, the jamming is assumed to be of the worst possible form, i.e., energy is concentrated in such a way as to maximize error probability
Conventional computing platforms include not just processors, memory, storage peripherals, and other peripherals associated with data processing, but also many interfaces (such as sound, imager, and mechanical input sensor systems) that have characteristics similar to systems specifically designed to interact with the physical world. An understanding of the architectures of such platforms is thus a good place to begin, in particular as many of the hardware and software design abstractions may usefully be applied to ENS nodes, just as many networking techniques and abstractions for general purpose computers apply to ENS. However, ENS applied to sensing applications requires architectural solutions that differ dramatically in several respects. For example, ENS platforms may be isolated from human users, may operate at low duty cycles, with deterministic schedules, and may be tolerant to a high communication error rate and long latency. At the same time, ENS may also be required to operate autonomously and unattended for extended periods on minimal battery energy. This is in marked contrast to embedded computing platforms of the past that directly support human users (e.g., cellular telephony platforms or handheld computing systems). For these applications, operations may not conform to deterministic schedules, operating lifetimes may be short, and administration by users is available. Of course, this example of low-duty-cycle operation represents merely one instance and others require widely varying ENS operating requirements; e.g., an image sensor may require a high sensor data rate and a high communication data rate.
Source detection, localization, and identification begin with the realization that the sources are by their nature probabilistic. The source, the coupling of source to medium, the medium itself, and the noise processes are each variable to some degree. Indeed, it is this very randomness that presents the fundamental barrier to rapid and accurate estimation of signals. Consequently, applied probability is essential to the study of communications and other detection and estimation problems. This chapter is concerned with the description of signals as random processes, how signals are transformed from continuous (real-world) signals into discrete representations, and the fundamental limits on the loss of information that result from such transformations. Three broad topics will be touched upon: basic probability theory, representation of stochastic processes, and information theory.
Probability
Discrete random variables
A discrete random variable (RV) X takes on values in a finite set X = {x1, x2, … xm}. The probability of any instance x being xi, written P(x = xi), is pi. Any probability distribution must satisfy the following axioms:
P(xi) ≥ 0.
The probability of an event which is certain is 1.
If xi and xj are mutually exclusive events, then P(xi + xj) = P(xi) + P(xj).
That is, probabilities are always positive, the maximum probability is 1, and probabilities are additive if one event excludes the occurrence of another. Throughout the book, a random variable is denoted by a capital letter, while any given sample of the distribution is denoted by a lower-case letter.