To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
Weighing in at about three pounds, the human brain is a marvel of evolutionary engineering. The brain transforms signals from millions of sensors located all over the body into appropriate muscle commands to enact a behavior suitable to the task at hand. This closed-loop, real-time control system remains unsurpassed by any artificially created system despite decades of attempts by computer scientists and engineers.
The brain’s unique information processing capabilities arise from its massively parallel and distributed way of computing. The workhorse of the brain is a type of cell known as a neuron, a complex electrochemical device that receives information from hundreds of other neurons, processes this information, and conveys its output to hundreds of other neurons. Furthermore, the connections between neurons are plastic, allowing the brain’s networks to adapt to new inputs and changing circumstances. This adaptive and distributed mode of computation sets the brain apart from traditional computers, which are based on the von Neumann architecture with a separate central processing unit, memory units, fixed connections between components, and a serial mode of computation.
In the previous chapter, we learned about BCIs that required placing electrodes inside the brain. While such an approach provides a high-fidelity window into the spiking activity of neurons, it also comes with significant risks: (1) possible infections due to penetration of the blood-brain barrier, (2) encapsulation of electrodes by immunologically reactive tissue, which can degrade signal quality over time, and (3) the potential for damage to intact brain circuits during implantation.
To counter these risks, researchers have investigated the use of BCIs that do not penetrate the brain surface. Such BCIs can be regarded as semi-invasive BCIs. We will focus on two types of semi-invasive BCIs: electrocorticographic (ECoG) BCIs and BCIs based on recording from nerves outside the brain. As discussed in Chapter 3, ECoG requires surgical placement of electrodes underneath the skull, either under the dura mater (subdural ECoG) or outside the dura mater (epidural ECoG). The procedure is invasive but less so than the methods of the previous chapter. In this chapter, we explore the ability of ECoG BCIs to control cursors and prosthetic devices.
As described in the previous chapter, the brain communicates using spikes, which are all-or-none electrical pulses produced when the neuron receives a sufficient amount of input current from other neurons via synaptic connections. It is therefore not surprising that some of the first technologies for recording brain activity were based on detecting changes in electrical potentials in neurons (invasive techniques based on electrodes) or large populations of neurons (noninvasive techniques such as electroencephalography or EEG). More recent techniques have been based on detecting neural activity indirectly by measuring changes in blood flow that result from increased neural activity in a region (fMRI) or by measuring the minute changes in the magnetic field around the skull caused by neural activity (MEG).
In this chapter, we review some of these technologies that serve as sources of input signals for BCIs. We also briefly describe technologies that can be used to stimulate neurons or brain regions, thereby allowing BCIs to potentially provide direct feedback to the brain based on interactions with the physical world.
The preceding chapters introduced you to the basic concepts in neuroscience, recording and stimulation technologies, signal processing, and machine learning. We are now ready to put it all together to consider the process of building an actual BCI.
Major Types of BCIs
BCIs today can be broadly divided into three major types:
Invasive BCIs: These involve recording from or stimulating neurons inside the brain.
Semi-invasive BCIs: These involve recording from or stimulating the brain surface or nerves.
Noninvasive BCIs: These employ techniques for recording from or stimulating the brain without penetrating the skin or skull.
Within each of these types, we can have BCIs that:
Only record from the brain (and translate the neural data into control signals for output devices).
Only stimulate the brain (and cause certain desired patterns of neural activity in the brain).
Both record and stimulate the brain.
In the next i ve chapters, we will encounter concrete examples of the major typesof BCIs defined above. Before we proceed to these concrete BCI examples, it is usefulto first discuss some of the major types of brain responses that researchers haveexploited for building BCIs.
The field of brain-computer interfacing has witnessed tremendous growth over the past decade. Invasive BCIs based on multielectrode arrays have allowed laboratory animals to precisely control the movement of robotic arms. Implants and semi-invasive BCIs have enabled human subjects to quickly acquire control of computer cursors and simple devices. Noninvasive BCIs, particularly those based on EEG, have allowed humans to control cursors in multiple dimensions and issue commands to semi-autonomous robots. Commercially available BCIs such as cochlear implants and deep brain stimulators have helped improve the quality of life of hundreds of hearing-impaired individuals and patients suffering from debilitating neurological diseases.
The achievements of the field thus far are impressive, but many obstacles remain. As pointed out by Gilja, Shenoy, and colleagues (2011), invasive BCIs have yet to achieve the same levels of performance, multidecade robustness, and naturalistic proprioception and somatosensation as able-bodied people. Furthermore, invasive BCIs remain risky for humans and are used only as a last resort in severely disabled patients. The most popular noninvasive BCIs, based on EEG, suffer from a number of problems:
Electrode placement is cumbersome and setup time is typically long (up to half an hour depending on the number of electrodes).
Results of training and learning may not be transferable from one day to the next due to shifts in electrode locations, noisy contacts with scalp, etc.
Low signal-to-noise ratio and on-line adaptation in subjects necessitate the availability of powerful amplifiers as well as efficient machine-learning and signal processing algorithms.
Signal attenuation and summation between the brain and the scalp, together with sparse sampling of activity, limits the range of useful control signals that can be extracted.
We have thus far studied BCIs that either record from the brain to control an external device (Chapters 7–9) or stimulate the brain to restore sensory or motor function (Chapter 10). The most general type of BCI is one that can simultaneously record from and stimulate different parts of the brain. Such BCIs are called bidirectional (or recurrent) BCIs. Bidirectional BCIs can provide direct feedback to the brain by stimulating sensory neurons to convey the consequences of operating a prosthetic device using motor signals recorded from the same brain. Furthermore, signals recorded from one part of the brain can be used to modulate the neural activity or induce plasticity in a different part of the brain.
In Chapter 1, we discussed the pioneering work of Delgado (1969) on an implantable BCI called the stimoceiver, which can be regarded as the first example of a bidirectional BCI. In this chapter, we briefly review a few more recent examples to illustrate the possibilities opened up by bidirectional BCIs and conclude by noting that the most flexible BCIs of the future will likely be bidirectional, though this flexibility will likely come at the cost and the associated risk of being invasive.
The field of machine learning has played an important role in the development of brain-computer interfaces by providing techniques that can learn to map neural activity to appropriate control commands. Algorithms for machine learning can be broadly divided into two classes: supervised learning and unsupervised learning. In supervised learning, we are given training data that consists of a set of inputs and corresponding outputs. The goal is to learn the underlying function from the training data such that new test inputs are mapped to the correct outputs. If the outputs are discrete classes, the problem is called classification. If the outputs are continuous, the problem is equivalent to regression. Given the emphasis on discovering an underlying function, supervised learning is sometimes also called function approximation. Unsupervised learning, on the other hand, emphasizes discovery of hidden statistical structure in unlabeled data: the training data consists of inputs, which are typically high-dimensional vectors, and the goal is to learn a statistical model that may be compact or useful for subsequent analysis. We have already discussed two prominent unsupervised learning techniques (PCA and ICA) in the previous chapter.
In this chapter, we focus on the two major types of supervised learning techniques: classification and regression. Classification is the problem of assigning one of N labels to a new input signal, given labeled training data consisting of known inputs and their corresponding output labels. Regression is the problem of mapping input signals to a continuous output signal. Many BCIs based on EEG, ECoG, fMRI, and fNIR have relied on classification to generate discrete control outputs (e.g., move a cursor up or down by a small amount). BCIs based on neuronal recordings, on the other hand, have predominantly utilized regression to generate continuous output signals, such as position or velocity signals for a prosthetic device. In general, the choice of whether to use classification or regression when designing a BCI will depend on both the type of brain signal being recorded and the type of application being controlled.
Some of the most important developments in brain-computer interfacing have come from BCIs based on invasive recordings. As reviewed in Chapter 3, invasive recording techniques allow the activities of single neurons or populations of neurons to be recorded. This chapter describes some of the achievements of invasive BCIs in animals and humans.
Two Major Paradigms in Invasive Brain- Computer Interfacing
BCIs Based on Operant Conditioning
A number of BCIs in animals have been based on operant conditioning, a phenomenon discussed in Section 6.2.1. In operant conditioning, an animal receives a reward upon selection of an appropriate action, e.g., pressing a lever. After repetitive pairing, the animal learns to execute the action in anticipation of the reward. In a BCI paradigm, the animal is rewarded if it selectively activates a neuron or population of neurons to move a cursor or prosthetic device in an appropriate manner.
Early BCI Studies
In the late 1960s, in one of the earliest demonstrations of brain- computer interfacing,Eberhard Fetz at the University of Washington in Seattle utilized the idea of operantconditioning to demonstrate that the activity of a single neuron in a primate’s motorcortex can be conditioned to control the needle of an analog meter (Fetz, 1969). Themovement of the needle was directly coupled to the firing rate of the neuron: whenthe needle crossed a threshold, the monkey was rewarded. After several trials, themonkey learned to consistently move the needle past the threshold by increasingthe firing rate of the recorded neuron (Figure 7.1). In this example of operant conditioning, the action (needle movement) that produces reward is coupled to increasedactivity in the recorded neuron (the conditioned response).
Arabic language is strongly structured and considered as one of the most highly inflected and derivational languages. Learning Arabic morphology is a basic step for language learners to develop language skills such as listening, speaking, reading, and writing. Arabic morphology is non-concatenative and provides the ability to attach a large number of affixes to each root or stem that makes combinatorial increment of possible inflected words. As such, Arabic lexical (morphological and phonological) rules may be confusing for second language learners. Our study indicates that research and development endeavors on spelling, and checking of grammatical errors does not provide adequate interpretations to second language learners’ errors. In this paper we address issues related to error diagnosis and feedback for second language learners of Arabic verbs and how they impact the development of a web-based intelligent language tutoring system. The major aim is to develop an Arabic intelligent language tutoring system that solves these issues and helps second language learners to improve their linguistic knowledge. Learners are encouraged to produce input freely in various situations and contexts, and are guided to recognize by themselves the erroneous functions of their misused expressions. Moreover, we proposed a framework that allows for the individualization of the learning process and provides the intelligent feedback that conforms to the learner's expertise for each class of error. Error diagnosis is not possible with current Arabic morphological analyzers. So constraint relaxation and edit distance techniques are successfully employed to provide error-specific diagnosis and adaptive feedback to learners. We demonstrated the capabilities of these techniques in diagnosing errors related to Arabic weak verbs formed using complex morphological rules. As a proof of concept, we have implemented the components that diagnose learner's errors and generate feedback which have been effectively evaluated against test data acquired from real teaching environment. The experimental results were satisfactory, and the performance achieved was 74.34 percent in terms of recall rate.
Paraphrase corpora are an essential but scarce resource in Natural Language Processing. In this paper, we present the Wikipedia-based Relational Paraphrase Acquisition (WRPA) method, which extracts relational paraphrases from Wikipedia, and the derived WRPA paraphrase corpus. The WRPA corpus currently covers person-related and authorship relations in English and Spanish, respectively, suggesting that, given adequate Wikipedia coverage, our method is independent of the language and the relation addressed. WRPA extracts entity pairs from structured information in Wikipedia applying distant learning and, based on the distributional hypothesis, uses them as anchor points for candidate paraphrase extraction from the free text in the body of Wikipedia articles. Focussing on relational paraphrasing and taking advantage of Wikipedia-structured information allows for an automatic and consistent evaluation of the results. The WRPA corpus characteristics distinguish it from other types of corpora that rely on string similarity or transformation operations. WRPA relies on distributional similarity and is the result of the free use of language outside any reformulation framework. Validation results show a high precision for the corpus.
We present a methodology for the extraction of narrative information from a large corpus. The key idea is to transform the corpus into a network, formed by linking the key actors and objects of the narration, and then to analyse this network to extract information about their relations. By representing information into a single network it is possible to infer relations between these entities, including when they have never been mentioned together. We discuss various types of information that can be extracted by our method, various ways to validate the information extracted and two different application scenarios. Our methodology is very scalable, and addresses specific research needs in social sciences.
Comparisons sort objects based on their superiority or inferiority and they may have major effects on a variety of evaluation processes. The Web facilitates qualitative and quantitative comparisons via online debates, discussion forums, product comparison sites, etc., and comparison analysis is becoming increasingly useful in many application areas. This study develops a method for classifying sentences in Korean text documents into several different comparative types to facilitate their analysis. We divide our study into two tasks: (1) extracting comparative sentences from text documents and (2) classifying comparative sentences into seven types. In the first task, we investigate many actual comparative sentences by referring to previous studies and construct a lexicon of comparisons. Sentences that contain elements from the lexicon are regarded as comparative sentence candidates. Next, we use machine learning techniques to eliminate non-comparative sentences from the candidates. In the second task, we roughly classify the comparative sentences using keywords and use a transformation-based learning method to correct initial classification errors. Experimental results show that our method could be suitable for practical use. We obtained an F1-score of 90.23% in the first task, an accuracy of 81.67% in the second task, and an overall accuracy of 88.59% for the integrated system with both tasks.
An open problem in natural language processing is word sense disambiguation (WSD). A word may have several meanings, but WSD is the task of selecting the correct sense of a polysemous word based on its context. Proposed solutions are based on supervised and unsupervised learning methods. The majority of researchers in the area focused on choosing proper size of ‘n’ in n-gram that is used for WSD problem. In this research, the concept has been taken to a new level by using variable ‘n’ and variable size window. The concept is based on the iterative patterns extracted from the text. We show that this type of sequential pattern is more effective than many other solutions for WSD. Using regular data mining algorithms on the extracted features, we significantly outperformed most monolingual WSD solutions. The state-of-the-art results were obtained using external knowledge like various translations of the same sentence. Our method improved the accuracy of the multilingual system more than 4 percent, although we were using monolingual features.