To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
Scientific progress relies on reproducibility, replicability, and robustness of research outcomes. After briefly discussing these terms and their significance for reliable scientific discovery, we argue for the importance of investigating robustness of outcomes to experimental protocol variations. We highlight challenges in achieving robust, replicable results in multi-step plant science experiments, using split-root assays in Arabidopsis thaliana as a case study. These experiments are important for unraveling the contributions of local, systemic and long-distance signalling in plant responses and play a central role in nutrient foraging research. The complexity of these experiments allows for extensive variation in protocols. We investigate what variations do or do not result in similar outcomes and provide concrete recommendations for enhancing the replicability and robustness of these and other complex experiments by extending the level of detail in research protocols.
This chapter explores the potential of digital history, geographic information systems (GIS), and spatial humanities in Ottoman studies, with a focus on a historical geographic information system (HGIS) application. It highlights the transformative impact of digital humanities (DH) on historical knowledge production, enabling replication and deeper research. Incorporating GIS into DH has led to geospatial humanities and spatial history, opening new research avenues. Ottoman studies are relatively new to these approaches, with limited data-driven research. The chapter addresses challenges arising from the historical disconnect between history and geography in Ottoman studies, emphasizing the significance of gazetteers and historical population data for large-scale HGIS applications. Presenting a case study analyzing historical census data for two Bulgarian regions, it assesses HGIS benefits and limitations. The chapter advocates a transparent, replicable, and cautious interpretation of digital and spatial historical analyses, calling for the continued development of geospatial methods in south-east Europe for long-term historical population geography insights.
Julianne House, Universität Hamburg/Hun-Ren Hungarian Research Centre for Linguistics /Hellenic American University,Dániel Z. Kádár, Dalian University of Foreign Languages/Hun-Ren Hungarian Research Centre for Linguistics/University of Maribor
In Chapter 1, we provide an introduction for the present book. First we present our take on language and politics, by outlining how and why our bottom-up, strictly language-anchored and replicable analysis contributes to previous inquiries in the field. Here we also outline the elements of our framework and the ways they are interconnected, as well as the analytic pitfalls that our approach helps to avoid. Following this, we present the contents of the book.
We comment on Sijtsma’s (2014) thought-provoking essay on how to minimize questionable research practices (QRPs) in psychology. We agree with Sijtsma that proactive measures to decrease the risk of QRPs will ultimately be more productive than efforts to target individual researchers and their work. In particular, we concur that encouraging researchers to make their data and research materials public is the best institutional antidote against QRPs, although we are concerned that Sijtsma’s proposal to delegate more responsibility to statistical and methodological consultants could inadvertently reinforce the dichotomy between the substantive and statistical aspects of research. We also discuss sources of false-positive findings and replication failures in psychological research, and outline potential remedies for these problems. We conclude that replicability is the best metric of the minimization of QRPs and their adverse effects on psychological research.
When analyzing data, researchers make some choices that are either arbitrary, based on subjective beliefs about the data-generating process, or for which equally justifiable alternative choices could have been made. This wide range of data-analytic choices can be abused and has been one of the underlying causes of the replication crisis in several fields. Recently, the introduction of multiverse analysis provides researchers with a method to evaluate the stability of the results across reasonable choices that could be made when analyzing data. Multiverse analysis is confined to a descriptive role, lacking a proper and comprehensive inferential procedure. Recently, specification curve analysis adds an inferential procedure to multiverse analysis, but this approach is limited to simple cases related to the linear model, and only allows researchers to infer whether at least one specification rejects the null hypothesis, but not which specifications should be selected. In this paper, we present a Post-selection Inference approach to Multiverse Analysis (PIMA) which is a flexible and general inferential approach that considers for all possible models, i.e., the multiverse of reasonable analyses. The approach allows for a wide range of data specifications (i.e., preprocessing) and any generalized linear model; it allows testing the null hypothesis that a given predictor is not associated with the outcome, by combining information from all reasonable models of multiverse analysis, and provides strong control of the family-wise error rate allowing researchers to claim that the null hypothesis can be rejected for any specification that shows a significant effect. The inferential proposal is based on a conditional resampling procedure. We formally prove that the Type I error rate is controlled, and compute the statistical power of the test through a simulation study. Finally, we apply the PIMA procedure to the analysis of a real dataset on the self-reported hesitancy for the COronaVIrus Disease 2019 (COVID-19) vaccine before and after the 2020 lockdown in Italy. We conclude with practical recommendations to be considered when implementing the proposed procedure.
By making building instructions freely accessible to everyone, open-source machine tools (OSMTs) promise to democratize manufacturing by enabling users in marginalized settings to build machines tools by themselves. There is, however, a lack of empirical evidence of the replicability of OSMT designs in low-resource contexts. This article explores OSMT replicability through qualitative and empirical methods to answer the central research question: Are designs that are fully open source also globally replicable? A comparative experiment was carried out by replicating an open-source 3D printer in two different locations: in Germany (resource-rich) and in Oman (resource-poor). The experiment aimed to determine the barriers faced with the replication in each location. It was significantly more challenging to replicate the 3D printer in Oman, primarily due to difficulties in sourcing and manufacturing, necessitating extensive modifications, which demanded greater skills and dexterity from users compared to those in Germany. Qualitative interviews found that limited digital literacy posed a significant barrier for microenterprise owners in replicating OSMT. Finally, design guidelines were proposed to enhance the global replicability of contextualized OSMT designs.
With its promise of nondestructive processing, rapid low-cost sampling, and portability to any field site or museum in the world, portable X-ray fluorescence (pXRF) spectrometry is rapidly becoming a standard piece of equipment for archaeologists. Even though the use of pXRF is becoming standard, the publication of pXRF analytical methods and the resulting data remains widely variable. Despite validation studies that demonstrate the importance of sample preparation, data collection settings, and data processing, there remains no standard for how to report pXRF results. In this article, we address the need for best practices in publishing pXRF analyses. We outline information that should be published alongside interpretive results in any archaeological application of pXRF. By publishing this basic information, archaeologists will increase the transparency and replicability of their analyses on an inter-analyst/inter-analyzer basis and provide clarity for journal editors and peer reviewers on publications and grant proposals for studies that use pXRF. The use of these best practices will result in better science in the burgeoning use of pXRF in archaeology.
This chapter details the practical, theoretical, and philosophical aspects of experimental science. It discusses how one chooses a project, performs experiments, interprets the resulting data, makes inferences, and develops and tests theories. It then asks the question, "are our theories accurate representations of the natural world, that is, do they reflect reality?" Surprisingly, this is not an easy question to answer. Scientists assume so, but are they warranted in this assumption? Realists say "yes," but anti-realists argue that realism is simply a mental representation of the world as we perceive it, that is, metaphysical in nature. Regardless of one's sense of reality, the fact remains that science has been and continues to be of tremendous practical value. It would have to be a miracle if our knowledge and manipulation of the nature were not real. Even if they were, how do we know they are true in an absolute sense, not just relative to our own experience? This is a thorny philosophical question, the answer to which depends on the context in which it is asked. The take-home message for the practicing scientist is "never assume your results are true."
Willis J. Edmondson,Juliane House, Universität Hamburg and the Hungarian Research Centre for Linguistics,Daniel Z. Kadar, Dalian University of Foreign Languages, China and Hungarian Research Centre for Linguistics
Chapter 6 presents a key component of this interactional grammar: illocutionary acts. In this grammar, we use the expressions ‘illocutionary act’ and ‘speech act’ interchangeably. The chapter provides a systematic and replicable interactional typology of illocutionary acts. This typology is particularly suitable for analysing discourse and understanding the role of illocutionary acts in any types of data and any language.
Religious belief is a topic of longstanding interest to psychological science, but the psychology of religious disbelief is a relative newcomer. One prominently discussed model is analytic atheism, wherein cognitive reflection, as measured with the Cognitive Reflection Test, overrides religious intuitions and instruction. Consistent with this model, performance-based measures of cognitive reflection predict religious disbelief in WEIRD (Western, Educated, Industrialized, Rich, & Democratic) samples. However, the generality of analytic atheism remains unknown. Drawing on a large global sample (N = 3461) from 13 religiously, demographically, and culturally diverse societies, we find that analytic atheism as usually assessed is in fact quite fickle cross-culturally, appearing robustly only in aggregate analyses and in three individual countries. The results provide additional evidence for culture’s effects on core beliefs.
Social science research on the aims and impacts of Chinese development finance remains in its infancy because Beijing shrouds its overseas portfolio of grants and loans in secrecy. This chapter introduces the Tracking Underreported Financial Flows (TUFF) methodology that the authors have developed to assemble a comprehensive dataset of Chinese aid and debt-financed development projects around the globe. It also provides an overview of previous attempts to quantify Chinese development finance, and explains how the authors’ methods and data are different from those of others. This chapter also tests whether an alternative approach—field-based data collection—might yield more useful and reliable re- sults. Drawing upon evidence from a “ground-truthing” exercise in Uganda and South Africa, the authors demonstrate that field-based and TUFF-based data collection methods produce similar results. However, the TUFF methodology is less vulnerable to detection bias and more readily scalable than field-based data collection.
Bilingualism is hard to define, measure, and study. Sparked by the “replication crisis” in the social sciences, a recent discussion on the advantages of open science is gaining momentum. Here, we join this debate to argue that bilingualism research would greatly benefit from embracing open science. We do so in a unique way, by presenting six fictional stories that illustrate how open science practices – sharing preprints, materials, code, and data; pre-registering studies; and joining large-scale collaborations – can strengthen bilingualism research and further improve its quality.
The diversity of design research studies and their associated methods and reporting style make it difficult for the design research community of practice to leverage its work into further advancing the field. We illustrate how a structured multilevel analysis of diverse studies creates a canonical model that allows for the transfer of insight between studies, enhances their comprehension, and supports improved study designs. The benefits of such an approach will increase if different stakeholders adopt such structured approaches to enrich the design research community of practice.
Climate is an emergent system with many interacting processes and components. Complexity is essential to accurately model the system and make quantitative predictions. But this complexity obscures the different compensating errors inherent in climate models. The Anna Karenina principle, which assumes that these compensating errors are random, is introduced. By using models with different formulations for small-scale processes to make predictions and then averaging them, we can expect to cancel out the random errors. This multimodel averaging can increase the skill of climate predictions, provided the models are sufficiently diverse. Climate models tend to borrow formulations from each other, which can lead to “herd mentality” and reduce model diversity. The need to preserve the diversity of models works against the need for replicability of results from those models. A compromise between these two conflicting goals becomes essential.
Rigor and reproducibility are two important cornerstones of medical and scientific advancement. Clinical and translational research (CTR) contains four phases (T1–T4), involving the translation of basic research to humans, then to clinical settings, practice, and the population, with the ultimate goal of improving public health. Here we provide a framework for rigorous and reproducible CTR.
Methods:
In this paper we define CTR, provide general and phase-specific recommendations for improving quality and reproducibility of CTR with emphases on study design, data collection and management, analyses and reporting. We present and discuss aspects of rigor and reproducibility following published examples of CTR from the literature, including one example that shows the development path of different treatments that address anaplastic lymphoma kinase-positive (ALK+) non-small cell lung cancer (NSCLC).
Results:
It is particularly important to consider robust and unbiased experimental design and methodology for analysis and interpretation for clinical translation studies to ensure reproducibility before taking the next translational step. There are both commonality and differences along the clinical translation research phases in terms of research focuses and considerations regarding study design, implementation, and data analysis approaches.
Conclusions:
Sound scientific practices, starting with rigorous study design, transparency, and team efforts can greatly enhance CTR. Investigators from multidisciplinary teams should work along the spectrum of CTR phases, and identify optimal practices for study design, data collection, data analysis, and results reporting to allow timely advances in the relevant field of research.
In recent decades, empirical research has developed across many areas of intellectual property law. This chapter examines challenges that can arise in conducting, or drawing upon, empirical research in IP law. These include assessing a study’s value in terms of methodology, sample choice and size, execution and reporting, as well as the conclusions that might reasonably be drawn from the research. As IP scholars generate more empirically informed research, there can be value in asking whether the studies are robust and well executed, whether they reveal information or viewpoints previously not recognised, and whether they produce work from which legal or scholarly lessons can be drawn. Robust, well-conducted and analysed empirical studies may provide insights that develop IP scholarship in new ways and potentially improve policy and decision making. This underlines the importance of not being complacent about empirical analysis but being open to rigorous questioning, both individually and collectively, about our practices.
This article provides an accessible tutorial with concrete guidance for how to start improving research methods and practices in your lab. Following recent calls to improve research methods and practices within and beyond the borders of psychological science, resources have proliferated across book chapters, journal articles, and online media. Many researchers are interested in learning more about cutting-edge methods and practices but are unsure where to begin. In this tutorial, we describe specific tools that help researchers calibrate their confidence in a given set of findings. In Part I, we describe strategies for assessing the likely statistical power of a study, including when and how to conduct different types of power calculations, how to estimate effect sizes, and how to think about power for detecting interactions. In Part II, we provide strategies for assessing the likely type I error rate of a study, including distinguishing clearly between data-independent (“confirmatory”) and data-dependent (“exploratory”) analyses and thinking carefully about different forms and functions of preregistration.
Psychology and neighboring disciplines are currently consumed with a replication crisis. Recent work has shown that replication can have the unintended consequence of perpetuating unwarranted conclusions when repeating an incorrect line of scientific reasoning from one study to another. This tutorial shows how decision researchers can derive logically coherent predictions from their theory by keeping track of the heterogeneity of preference the theory permits, rather than dismissing such heterogeneity as a nuisance. As an illustration, we reanalyze data of Barron and Ursino (2013). By keeping track of the heterogeneity of preferences permitted by Cumulative Prospect Theory, we show how the analysis and conclusions of Barron and Ursino (2013) change. This tutorial is intended as a blue-print for graduate student projects that dig deeply into the merits of prior studies and/or that supplement replication studies with a quality check.
The ICD-11 includes a new disorder, complex post-traumatic stress disorder (CPTSD). A network approach to CPTSD will enable investigation of the structure of the disorder at the symptom level, which may inform the development of treatments that target specific symptoms to accelerate clinical outcomes.
Aims
We aimed to test whether similar networks of ICD-11 CPTSD replicate across culturally different samples and to investigate possible differences, using a network analysis.
Method
We investigated the network models of four nationally representative, community-based cross-sectional samples drawn from Germany, Israel, the UK, and the USA (total N = 6417). CPTSD symptoms were assessed with the International Trauma Questionnaire in all samples. Only those participants who reported significant functional impairment by CPTSD symptoms were included (N = 1591 included in analysis; mean age 43.55 years, s.d. 15.10, range 14–99, 67.7% women). Regularised partial correlation networks were estimated for each sample and the resulting networks were compared.
Results
Despite differences in traumatic experiences, symptom severity and symptom profiles, the networks were very similar across the four countries. The symptoms within dimensions were strongly associated with each other in all networks, except for the two symptom indicators assessing aspects of affective dysregulation. The most central symptoms were ‘feelings of worthlessness’ and ‘exaggerated startle response’.
Conclusions
The structure of CPTSD symptoms appears very similar across countries. Addressing symptoms with the strongest associations in the network, such as negative self-worth and startle reactivity, will likely result in rapid treatment response.
Declaration of interest
A.M. and M.C. were members of the World Health Organization (WHO) ICD-11 Working Group on the Classification of Disorders Specifically Associated with Stress, reporting to the WHO International Advisory Group for the Revision of ICD-10 Mental and Behavioural Disorders. The views expressed in this article are those of the authors and do not represent the official policies or positions of the International Advisory Group or the WHO.
This study aimed to assess the consistency and replicability of these process measures during provision of the Italian Medicines Use Review (I-MUR).
Background
Medication review is a common intervention provided by community pharmacists in many countries, but with little evidence of consistency and replicability. The I-MUR utilised a standardised question template in two separate large-scale studies. The template facilitated pharmacists in recording medicines and problems reported by patients, the pharmaceutical care issues (PCIs) they found and actions they took to improve medicines use.
Methods
Community pharmacists from four cities and across 15 regions were involved in the two studies. Patients included were adults with asthma. Medicines use, adherence, asthma problems, PCIs and actions taken by pharmacists were compared across studies to assess consistency and replicability of I-MUR.
Findings
The total number of pharmacists and patients completing the studies was 275 and 1711, respectively. No statistically significant differences were found between the studies in the following domains: patients’ demographic, patients’ perceived problems, adherence, asthma medicines used and healthy living advice provided by pharmacists. The proportion of patients in which pharmacists identified PCIs was similar across both studies. There were differences only in the incidence of non-steroidal anti-inflammatory drug use, the frequency of potential drug-disease interactions and in the types of advice given to patients and GPs.
Conclusions
The use of a standardised template for the I-MUR may have contributed to a degree of consistency in the issues found, which suggests this intervention could have good replicability.