To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
In this chapter, I consider methods of decreasing peak power in MC signals. The simplest method is to clip the MC signal deliberately before amplification. This method is very simple to implement and provides essential PMEPR reduction. However, it suffers some performance degradation, as estimated in Section 8.1. In selective mapping (SLM), discussed in Section 8.2, one favorable signal is selected from a set of different signals that all represent the same information. One possibility for SLM is to choose the best signal from those obtained by inverting any of the coordinates of the coefficient vector. The method of deciding which of the coordinates should be inverted is described in Section 8.3. Further, in Section 8.4 a modification of SLM is analyzed. There the favorable vector is chosen from a coset of a code of given strength. Trellis shaping, where the relevant modification is chosen based on a search on a trellis, is described in Section 8.5. In Section 8.6, the method of tone injection is discussed. Here, instead of using a constellation point its appropriately shifted version can be used. In active constellation extension (ACE), described in Section 8.7, some of the outer constellation points can be extended, yielding PMEPR reduction. In Section 8.8, a method of finding a constellation in the frequency domain is described, such that the resulting region in the time domain has a low PMEPR. In partial transmit sequences (PTS), the transmitted signal is made to have a low PMEPR by partitioning the information-bearing vector to sub-blocks followed by multiplying by a rotating factor the coefficients belonging to the same sub-block.
In this chapter I collect the basic mathematical tools which are used throughout the book. Most of them are given with rigorous proofs. I mainly concentrate here on results and methods that do not appear in the standard engineering textbooks and omit those that happen to be common technical knowledge. On the other hand, I have included some material that is not directly used in further arguments, but I feel that it might prove useful in further research on peak power control problems. It should be advised that the chapter is mainly for reference purposes and may be omitted in the first reading.
The chapter is organized as follows. Section 3.1 deals with harmonic analysis. In Section 3.1.1 I describe the Parseval equality and its generalizations. Section 3.1.2 introduces some useful trigonometric relations. Chebyshev polynomials and interpolation are described in Section 3.1.3. Finally, in Section 3.1.4, I prove Bernstein's inequality relating the maximum of the absolute value of a trigonometric polynomial and its derivative. In Section 3.2 I deal with some notions related to probability. I prove the Chernoff bound on the probability of deviations of values of random variables. In Section 3.3, I introduce tools from algebra. In Section 3.3.1 groups, rings, and fields are defined. Section 3.3.2 describes exponential sums in finite fields and rings. A short account of results from coding theory is presented in Section 3.4. Section 3.4.1 deals with properties of the Hamming space. In Section 3.4.2, definitions related to error-correction codes are introduced. Section 3.4.3 deals with the distance distributions of codes. In Section 3.4.4, I analyze properties of Krawtchouk polynomials playing an important role in the Mac Williams transform of the distance distributions.
In many situations it is beneficial to deal with a discrete-time “sampled” version of multicarrier signals. This reduction allows passing from the continuous setting to an easier-to-handle discrete one. However, we have to estimate the inaccuracies stemming from the approach. In this chapter, I analyze the ratio between the maximum of the absolute value of a continuous MC signal and the maximum over a set of the signal's samples. We start with considering the ratio when the signal is sampled at the Nyquist frequency, i.e. the number of sampling points equals the number of tones. In this case I show that the maximum of the ratio over all MC signals grows with the number of subcarriers (Theorem 4.2). However, if one computes a weighted sum of the maximum of the signal's samples and the maximum of the signal derivative's samples the ratio already is, at most, a constant (Theorem 4.5). I further show that actually the ratio depends on the maximum of the signal; the larger the maximum is the smaller is the ratio (Theorem 4.6). An even better strategy is to use over sampling. Then the ratio becomes constant tending to 1 when the over sampling rate grows (Theorems 4.8, 4.9, 4.10, and 4.11). Furthermore, I tackle the case when we have to use the maximum estimation, projections on specially chosen measuring axes instead of the absolute values of the signal (Theorem 4.14). Finally, I address the problem of relation between the PAPR and the PMEPR and show that the PMEPR estimates the PAPR quite accurately for large values of the carrier frequency (Theorem 4.19).
Chapter 2 kicks off the presentation with an overview of the standard metrics and methodologies followed by a description of specialized tools employed for obtaining subjective voice-quality scores through genuine opinion surveys and via computer modeling emulating human perceptive evaluation of speech quality. It then relates voice-quality scores obtained via surveys or computer evaluations to the perception of worth. It elaborates on the relationships between opinion scores and the potential return on investment in voice-quality technology. It examines results of voice-quality studies with reference to the three popular GSM codecs – full rate (FR), enhanced full rate (EFR) and half rate (HR). The presentation includes a discussion of the effect of noise and transmission errors on the relative performance of these codecs.
Introduction
It is widely agreed that the most vital element affecting the performance of voice quality in wireless networks is the type of codec used in the communication session. It is also well known that spectrum is the most expensive (per channel) building block contributing to the viability of wireless infrastructure. Yet most European GSM wireless operators, throughout the nineties and the first half of the following decade, have opted to restrict their service offerings to full-rate (FR) or enhanced full-rate (EFR), rather than embracing the half-rate (HR) option, which could cut their spectrum consumption in half and save them millions of dollars in operating expenses.
Chapter 6 is dedicated to the subject of level-control optimization. The presentation is divided into three parts and an introduction. The chapter starts the ball rolling in the introduction by defining standard methodologies for measuring and quantifying signal levels. The first part deals with automatic level control (ALC), how it works, and its placement within the network. The second part describes the adaptive level control, a.k.a. noise compensation (NC), how it works under different codecs, and where it is placed in the network. The third part describes the high-level compensation procedure along the same outline.
Basic signal-level measurements and definitions
One of the most critical speech-quality attributes is the perceived-speech level. When it is too loud, it may either overload or hurt the eardrum. When it is too soft, the listener or even the codec may have difficulties picking up words. There are several different measures and metrics used to measure speech levels. The common thread linking them together to a standardized scale is the unit of measurement defined as decibel and abbreviated as dB.
The human auditory system has a dynamic range of 100,000,000,000,000 (1014) intensity units. This dynamic range is better represented by a logarithmic scale, as a ratio of two intensities, P1 and P2. The expression “log(P1/P2)” is labeled a bel. The resulting number is still too large so, instead, the measure employed is one tenth of a bel or a decibel (dB).
Chapter 3 provides an overview of echo in telecommunications networks, its root causes, and its parameters. It follows the presentation with the methods used for controlling electrical echo, including network loss, echo suppression, linear convolution, non-linear processing, and comfort noise injection. The chapter covers the application of echo cancelation in wireless communications. And, in view of the fact that today's wireless networks include long-distance circuit switched, VoIP, and VoATM infrastructures (specifically as part of third-generation architectures), the chapter covers echo cancelation in long-distance and voice-over-packet applications.
Electrical echo
Many people have never experienced echo on a telephone call. They either never made calls outside their vicinity, or never encountered a malfunctioning echo canceler on their long-distance or wireless calls. In reality, echo does exist in the network. It accompanies every call involving an analog PSTN phone, but in most cases it either gets canceled before reaching its listening ear or it arrives too quickly with little delay, so it can sneak in undetected.
To hear an echo, we must first generate a sound. Then the sound must travel over a substantial distance or a slow terrain (a.k.a. complex processing) to accumulate delay. Next, the sound must be reflected and then travel back towards us. Finally, when the reflected sound reaches our ear, it must be loud enough to be heard. In addition, the amount of delay that the original sound signal incurs influences our perception of “reverberated echo” (i.e., increasing delay produces a greater echo effect).
Most people view sleep as a biological necessity designed to rejuvenate and invigorate body and mind so that human beings can function effectively when they are awake. While awake we tend to eat, dispose of waste, spawn the new generation, and take care of business, so that when the day is over we can go back to sleep.
When asked: “What is life's purpose?” we immediately think of life as the time we spend on earth while being awake.
Now let's reverse the paradigm. Let's pretend for a brief moment that life's purpose is a good sleep, and that all of the energetic activities above are designed to sustain our ability to go back and “live” (while asleep) the next day (see Figure 16.1).
I know. This is a weird thought, but let's see if we can apply the concept to voice-quality engineering in wireless networks, so that it makes sense.
Voice-quality systems are designed to operate on voice and to disable themselves on detection of data transmission. In other words, when turning on the VQS, voice-enhancement applications are enabled automatically while the system goes on guard, watching for any data that might require a disabling action.
Now let's reverse the paradigm. Let's disable all or a particular VQ application upon system turn-on, and continue to monitor the channel until voice is detected. When this occurs, the system enables the VQ application, and then puts it back to sleep when the monitored signal no longer looks like voice.
In Chapters 2 and 9, I discussed studies that illustrate a considerable weakness of the GSM half-rate codec in the face of noise. I also brought to light study results showing that an effective NR algorithm may be able to remedy some of the HR weakness and lift the voice-quality performance back to par for signals with a relatively poor SNR (see Figure 9.13).
As new wireless codecs are introduced at the market place, many service providers wonder whether VQ systems are effective in enhancing quality, as higher compression ratios are used in an attempt to reduce spectrum requirements and augment air capacity.
In this chapter, I introduce a procedure for testing the hypothesis that a new codec with a higher compression ratio offers an inferior (or equivalent) voice-quality performance in comparison to an existing codec, specifically under noisy conditions.
If the hypothesis proves to be true, then there is a need to test the hypothesis that VQS can lift the performance level of the new codec to a close proximity (or beyond) of the existing codec (under noisy conditions and without VQS).
Procedure
The procedure outlined here involves two phases.
Phase 1 (see Figure 15.1)
(1) Send pre-recorded speech signals from point B without noise. Capture and record speech signals (pre-recorded) at point A.
(2) Send pre-recorded speech signals from point E with noise. Capture and record speech signals (pre-recorded) at point D.
(3) Send pre-recorded speech signals from point B without noise. Capture and record speech signals (pre-recorded) at point A while sending speech signals from point A (double talk).
In survey after survey potential and actual users of wireless communications indicated that voice quality topped their reasons for selecting a specific service provider. While providers have been well aware of this key component powering their offering, they have not always been certain as to the specific methodology, resolution elements, equipment type, architecture, trade-offs, and rate of return on their particular investment that elevate the perceived voice-quality performance in their network.
It is only natural that voice quality in wireless networks has become a key differentiator among the competing service vendors. Network operators, network infrastructure planners, sales representatives of equipment vendors, their technical and sales support staff, and students of telecommunications seek information and knowledge continually that may help them understand the components of high-fidelity communicated sound.
Throughout the 1990s applications involving voice-quality enhancements, and specifically echo cancelation, have induced fresh inventions, new technology, and startling innovations in the area of enhanced voice performance. The initial echo canceler (EC) product implementations existed for about a decade before a diverse array of voice-quality enhancement realizations emerged to meet the evolving needs of digital wireless communications applications.
Early EC implementations were limited to very long distance (e.g., international) circuit-switched voice and fax applications where echo was perceived (in voice conversations) due to delays associated with signal propagation. The EC application soon expanded beyond strictly very-long-distance applications as further signal processing and dynamic routing along the communications path added delay to end-to-end voice transport.
Part I reviews the major voice codecs, their history, and their relative perceived quality. Voice-coding architectures are the building blocks of transmitted voice. They are the core that shapes the characteristics and quality of transmitted speech. Nevertheless, they are treated in this book only as background to the main subject, which deals with impairments due to transmission architecture and environment, and their corresponding remedies that immunize and repair any potential or actual spoil. Since the effectiveness of the various remedies depends on that underlying coding, it is essential that these designs be understood so that remedies can be fine tuned and customized to suit the particular characteristics of the underlying voice architecture.
Chapter 5 is devoted to the subject of noise reduction. Noise reduction is the most complicated feature among the voice-quality-assurance class of applications. It also requires a higher-level understanding of mathematics. This discussion, however, substitutes numerous mathematical expressions for intuition, ordinary analogies, and logical reasoning, supplemented by graphical and audio illustrations.
The analysis gets underway with the definition of noise, a definition consistent with the principles and characterization employed by a typical noise-reduction algorithm. It then introduces and explains the mathematical concept of time and frequency domains and the transformation process between the two. Once the reader is armed with the understanding of time- and frequency-domain representations, the analysis proceeds to a discussion of the noise-estimation process. The presentation then moves ahead to examine the suppression algorithm, which employs the noise-estimation results in its frequency-band attenuation procedures. The next segment contains a presentation covering the final algorithmic steps, which involve scaling and inverse transformation from frequency to time domains.
The next section in Chapter 5 reflects on key potential side effects associated with noise-reduction algorithms including treatment of non-voice signals. It points to key trade-offs and adverse-feature interactions that may occur in various GSM and CDMA networks – a subject that is covered much more thoroughly in Part V – Managing the network. The final section offers an examination of the network topology and placement of the noise-reduction application within it.
Noise in wireless networks
Background acoustic noise is a major voice-quality irritant that is, unfortunately, abundant in wireless communications.