To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
The exploitation of hydrocarbon reservoirs may potentially lead to contamination of soils, shallow water resources, and greenhouse gas emissions. Fluids such as methane or CO2 may in some cases migrate toward the groundwater zone and atmosphere through and along imperfectly sealed hydrocarbon wells. Field tests in hydrocarbon-producing regions are routinely conducted for detecting serious leakage to prevent environmental pollution. The challenge is that testing is costly, time-consuming, and sometimes labor-intensive. In this study, machine learning approaches were applied to predict serious leakage with uncertainty quantification for wells that have not been field tested in Alberta, Canada. An improved imputation technique was developed by Cholesky factorization of the covariance matrix between features, where missing data are imputed via conditioning of available values. The uncertainty in imputed values was quantified and incorporated into the final prediction to improve decision-making. Next, a wide range of predictive algorithms and various performance metrics were considered to achieve the most reliable classifier. However, a highly skewed distribution of field tests toward the negative class (nonserious leakage) forces predictive models to unrealistically underestimate the minority class (serious leakage). To address this issue, a combination of oversampling, undersampling, and ensemble learning was applied. By investigating all the models on never-before-seen data, an optimum classifier with minimal false negative prediction was determined. The developed methodology can be applied to identify the wells with the highest likelihood for serious fluid leakage within producing fields. This information is of key importance for optimizing field test operations to achieve economic and environmental benefits.
Our last emerging trend article introduced Risks 1.0 (fairness and bias) and Risks 2.0 (addictive, dangerous, deadly, and insanely profitable). This article introduces Risks 3.0 (spyware and cyber weapons). Risks 3.0 are less profitable, but more destructive. We will summarize two recent books, Pegasus: How a Spy in Your Pocket Threatens the End of Privacy, Dignity, and Democracy and This is How They Tell Me the World Ends: The Cyberweapons Arms Race. The first book starts with a leak of 50,000 phone numbers, targeted by spyware named Pegasus. Pegasus uses a zero-click exploit to obtain root access to your phone, taking control of the microphone, camera, GPS, text messages, etc. The list of 50,000 numbers includes journalists, politicians, and academics, as well as their friends and family. Some of these people have been murdered. The second book describes the history of cyber weapons such as Stuxnet, which is described as crossing the Rubicon. In the short term, it sets back Iran’s nuclear program for less than the cost of conventional weapons, but it did not take long for Iran to build the fourth-biggest cyber army in the world. As spyware continues to proliferate, we envision a future dystopia where everyone spies on everyone. Nothing will be safe from hacking: not your identity, or your secrets, or your passwords, or your bank accounts. When the endpoints (phones) have been compromised, technologies such as end-to-end encryption and multi-factor authentication offer a false sense of security; encryption and authentication are as pointless as closing the proverbial barn door after the fact. To address Risks 3.0, journalists are using the tools of their trade to raise awareness in the court of public opinion. We should do what we can to support them. This paper is a small step in that direction.
This article analyzes raw driving data of passenger cars in the city of Semnan in Iran, with the objective of understanding the impact of traffic conditions at different times of day (morning, noon, evening, and night). For this study, two cars, the Toyota Prius and the Peugeot Pars (or the IKCO Persia), were used, and the data of speed, longitude, latitude, and altitude of the vehicles were acquired. This data was collected over a week (July 21–28, 2022) for a distance of 670 km (13 hr), with the help of the Global Positioning System application, and were presented for both cars. In addition to this, the data on fuel consumption and average speed, based on the Electronic Control Unit in the Prius, was also collected. Finally, a sensitivity analysis was done on the features of the raw data, based on the Principal Component Analysis method.
The Handbook of Augmented Reality Training Design Principles is for anyone interested in using augmented reality and other forms of simulation to design better training. It includes eleven design principles aimed at training recognition skills for combat medics, emergency department physicians, military helicopter pilots, and others who must rapidly assess a situation to determine actions. Chapters on engagement, creating scenario-based training, fidelity and realism, building mental models, and scaffolding and reflection use real-world examples and theoretical links to present approaches for incorporating augmented reality training in effective ways. The Learn, Experience, Reflect framework is offered as a guide to applying these principles to training design. This handbook is a useful resource for innovative design training that leverages the strengths of augmented reality to create an engaging and productive learning experience.
Latent semantic analysis (LSA) and correspondence analysis (CA) are two techniques that use a singular value decomposition for dimensionality reduction. LSA has been extensively used to obtain low-dimensional representations that capture relationships among documents and terms. In this article, we present a theoretical analysis and comparison of the two techniques in the context of document-term matrices. We show that CA has some attractive properties as compared to LSA, for instance that effects of margins, that is, sums of row elements and column elements, arising from differing document lengths and term frequencies are effectively eliminated so that the CA solution is optimally suited to focus on relationships among documents and terms. A unifying framework is proposed that includes both CA and LSA as special cases. We empirically compare CA to various LSA-based methods on text categorization in English and authorship attribution on historical Dutch texts and find that CA performs significantly better. We also apply CA to a long-standing question regarding the authorship of the Dutch national anthem Wilhelmus and provide further support that it can be attributed to the author Datheen, among several contenders.
We study the problem of determining the minimum number $f(n,k,d)$ of affine subspaces of codimension $d$ that are required to cover all points of $\mathbb{F}_2^n\setminus \{\vec{0}\}$ at least $k$ times while covering the origin at most $k - 1$ times. The case $k=1$ is a classic result of Jamison, which was independently obtained by Brouwer and Schrijver for $d = 1$. The value of $f(n,1,1)$ also follows from a well-known theorem of Alon and Füredi about coverings of finite grids in affine spaces over arbitrary fields. Here we determine the value of this function exactly in various ranges of the parameters. In particular, we prove that for $k\geq 2^{n-d-1}$ we have $f(n,k,d)=2^d k-\left\lfloor{\frac{k}{2^{n-d}}}\right\rfloor$, while for $n \gt 2^{2^d k-k-d+1}$ we have $f(n,k,d)=n + 2^d k -d-2$, and obtain asymptotic results between these two ranges. While previous work in this direction has primarily employed the polynomial method, we prove our results through more direct combinatorial and probabilistic arguments, and also exploit a connection to coding theory.
We prove a new sufficient pair degree condition for tight Hamiltonian cycles in $3$-uniform hypergraphs that (asymptotically) improves the best known pair degree condition due to Rödl, Ruciński, and Szemerédi. For graphs, Chvátal characterised all those sequences of integers for which every pointwise larger (or equal) degree sequence guarantees the existence of a Hamiltonian cycle. A step towards Chvátal’s theorem was taken by Pósa, who improved on Dirac’s tight minimum degree condition for Hamiltonian cycles by showing that a certain weaker condition on the degree sequence of a graph already yields a Hamiltonian cycle.
In this work, we take a similar step towards a full characterisation of all pair degree matrices that ensure the existence of tight Hamiltonian cycles in $3$-uniform hypergraphs by proving a $3$-uniform analogue of Pósa’s result. In particular, our result strengthens the asymptotic version of the result by Rödl, Ruciński, and Szemerédi.
Web search is an experience that naturally lends itself to recommendations, including query suggestions and related entities. In this article, we propose to recommend specific tasks to users, based on their search queries, such as planning a holiday trip or organizing a party. Specifically, we introduce the problem of query-based task recommendation and develop methods that combine well-established term-based ranking techniques with continuous semantic representations, including sentence representations from several transformer-based models. Using a purpose-built test collection, we find that our method is able to significantly outperform a strong text-based baseline. Further, we extend our approach to using a set of queries that all share the same underlying task, referred to as search mission, as input. The study is rounded off with a detailed feature and query analysis.
Owing to inherent flexibility, compliant actuator with physical elastic elements can implement compliant interactions between robot and environment. Nonlinear stiffness elastic actuators (NSEAs) use one motor to adjust position and stiffness, which improve compactness of variable stiffness actuators, and consider high torque/force resolution and high bandwidth. The primary challenge of designing a NSEA is designing a reliable compliant mechanism with a given torque-deflection profile. In this study, a general design method of torsional compliant mechanism through given discrete torque-deflection points was proposed, where a chain algorithm based on 2R pseudo-rigid-body model was used to describe the large deformation in elastic elements. Moreover, the theoretical model of stiffness of a compliant mechanism based on series flexible beams was developed. Based on the proposed model, the dimensional parameters of proposed compliant mechanism were designed satisfying the condition of the given torque-deflection points. Finally, simulation and physical experiment of the designed compliant mechanism through three given torque-deflection points are conducted to show the effectiveness of the proposed method. The designed compliant mechanism was tested and evaluated in the self-developed NSEA.
A robotic system constructed as a wheeled inverted pendulum (WIP) serves as an impressive demonstrator, since this kind of system is inherently nonlinear, unstable, and nonminimum phase. These properties may pose several difficulties, when it comes to control and trajectory planning. This paper shows a method for deriving a highly dynamic trajectory compliant with the system dynamics by means of solving an optimal control problem (OCP) using multiple shooting. The assumed task includes that the WIP should pass a height-restricting barrier. This can be achieved by leaning back or forth, in order to reduce the overall height of the WIP. The constraints inherent to the definition of this trajectory are nonconvex due to the shape of the robot. The constraint functions have a local minimum in an infeasible region. This can cause problems when the initial guess is within this infeasible region. To overcome this, a multistage approach is proposed for this special OCP to evade the infeasible local minimum. After solving four stages of subsequent optimization problems, the optimal trajectory is obtained and can be used as feedforward for the real system.
Devices known as kinematic couplings offer accurate, repeatable, and stiff connections between two parts. They are characterized by point contacts and enable great repeatability with errors less than 1 micron, in contrast to conventional coupling systems like alignment pins or those based on elastic deformation. In this study, a robotized laser-cutting machine is equipped with a three-groove kinematic coupling design to increase the precision of workpiece placement. Given the application’s requirements, a preliminary design of the coupling is defined. An analytical approach is provided for calculating stresses and deflections at the locations where balls and grooves make contact, and it is then utilized to calculate positioning errors caused by the mechanical structure’s elastic deformation under various loading conditions. The outcomes of the simulations are finally discussed and highlight the efficacy of the solution tested.
We extend the notion of universal graphs to a geometric setting. A geometric graph is universal for a class $\mathcal H$ of planar graphs if it contains an embedding, that is, a crossing-free drawing, of every graph in $\mathcal H$. Our main result is that there exists a geometric graph with $n$ vertices and $O\!\left(n \log n\right)$ edges that is universal for $n$-vertex forests; this generalises a well-known result by Chung and Graham, which states that there exists an (abstract) graph with $n$ vertices and $O\!\left(n \log n\right)$ edges that contains every $n$-vertex forest as a subgraph. The upper bound of $O\!\left(n \log n\right)$ edges cannot be improved, even if more than $n$ vertices are allowed. We also prove that every $n$-vertex convex geometric graph that is universal for $n$-vertex outerplanar graphs has a near-quadratic number of edges, namely $\Omega _h(n^{2-1/h})$, for every positive integer $h$; this almost matches the trivial $O(n^2)$ upper bound given by the $n$-vertex complete convex geometric graph. Finally, we prove that there exists an $n$-vertex convex geometric graph with $n$ vertices and $O\!\left(n \log n\right)$ edges that is universal for $n$-vertex caterpillars.
We examine the binary classification of sentiment views for verbal multiword expressions (MWEs). Sentiment views denote the perspective of the holder of some opinion. We distinguish between MWEs conveying the view of the speaker of the utterance (e.g., in “The company reinvented the wheel” the holder is the implicit speaker who criticizes the company for creating something already existing) and MWEs conveying the view of explicit entities participating in an opinion event (e.g., in “Peter threw in the towel” the holder is Peter having given up something). The task has so far been examined on unigram opinion words. Since many features found effective for unigrams are not usable for MWEs, we propose novel ones taking into account the internal structure of MWEs, a unigram sentiment-view lexicon and various information from Wiktionary. We also examine distributional methods and show that the corpus on which a representation is induced has a notable impact on the classification. We perform an extrinsic evaluation in the task of opinion holder extraction and show that the learnt knowledge also improves a state-of-the-art classifier trained on BERT. Sentiment-view classification is typically framed as a task in which only little labeled training data are available. As in the case of unigrams, we show that for MWEs a feature-based approach beats state-of-the-art generic methods.
Recently, neural abstractive text summarization (NATS) models based on sequence-to-sequence architecture have drawn a lot of attention. Real-world texts that need to be summarized range from short news with dozens of words to long reports with thousands of words. However, most existing NATS models are not good at summarizing long documents, due to the inherent limitations of their underlying neural architectures. In this paper, we focus on the task of long document summarization (LDS). Based on the inherent section structures of source documents, we divide an abstractive LDS problem into several smaller-sized problems. In this circumstance, how to provide a less-biased target summary as the supervision for each section is vital for the model’s performance. As a preliminary, we formally describe the section-to-summary-sentence (S2SS) alignment for LDS. Based on this, we propose a novel NATS framework for the LDS task. Our framework is built based on the theory of unbalanced optimal transport (UOT), and it is named as UOTSumm. It jointly learns three targets in a unified training objective, including the optimal S2SS alignment, a section-level NATS summarizer, and the number of aligned summary sentences for each section. In this way, UOTSumm directly learns the text alignment from summarization data, without resorting to any biased tool such as ROUGE. UOTSumm can be easily adapted to most existing NATS models. And we implement two versions of UOTSumm, with and without the pretrain-finetune technique. We evaluate UOTSumm on three publicly available LDS benchmarks: PubMed, arXiv, and GovReport. UOTSumm obviously outperforms its counterparts that use ROUGE for the text alignment. When combined with UOTSumm, the performance of two vanilla NATS models improves by a large margin. Besides, UOTSumm achieves better or comparable performance when compared with some recent strong baselines.
The rich, multi-faceted and multi-disciplinary field of matching-based market design is an active and important one due to its highly successful applications with economic and sociological impact. Its home is economics, but with intimate connections to algorithm design and operations research. With chapters contributed by over fifty top researchers from all three disciplines, this volume is unique in its breadth and depth, while still being a cohesive and unified picture of the field, suitable for the uninitiated as well as the expert. It explains the dominant ideas from computer science and economics underlying the most important results on market design and introduces the main algorithmic questions and combinatorial structures. Methodologies and applications from both the pre-Internet and post-Internet eras are covered in detail. Key chapters discuss the basic notions of efficiency, fairness and incentives, and the way market design seeks solutions guided by normative criteria borrowed from social choice theory.
Despite the growing interest in addiction research, which demonstrates the potential predictive role of adverse childhood experiences (ACEs), little is known about their impact on the psychological symptoms of craving.
Methods
After reviewing the relevant diagnostic criteria for addiction and comorbid mental disorders along with routinely collected clinical and service-use data, 208 outpatients were assessed on the study protocol. Following the recruitment phase, nominal and ordinal data were analyzed using nonparametric methods.
Results
Most of the outpatients reported ACEs (89.1%) and experienced cravings (73.4–95.7%). A positive association between ACEs and either intention and preplanning (r = .14, p < .05) or lack of control (r = .15; p < .05) of the craving behavior was found.
Conclusion
Craving behavior in addiction remains a subject of debate. Although correlation analyses showed significant associations between reported ACEs and measures of craving, they were relatively small.
In concurrent and distributed systems, processes can complete tasks together by playing their parts in a joint plan. The plan, or protocol, can be written as a choreography: a formal description of overall behaviour that processes should collaborate to implement, like authenticating a user or purchasing an item online. Formality brings clarity, but not only that. Choreographies can contribute to important safety and liveness properties. This book is an ideal introduction to theory of choreographies for students, researchers, and professionals in computer science and applied mathematics. It covers languages for writing choreographies and their semantics, and principles for implementing choreographies correctly. The text treats the study of choreographies as a discipline in its own right, following a systematic approach that starts from simple foundations and proceeds to more advanced features in incremental steps. Each chapter includes examples and exercises aimed at helping with understanding the theory and its relation to practice.