To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
The last decade has witnessed a significant increase in academic and journalistic interest in the United States Supreme Court’s “shadow docket.” Yet despite this increased interest, there exists no systematic database of the shadow docket. This paper summarizes the Supreme Court Shadow Docket Database, which we created by parsing the Journal of the Supreme Court from the 1993 through 2024 terms into individual orders. We categorize these orders into a set of usable categories, including cert denials, injunctions, summary reversals, mandamus petitions, and grant, vacate, and remands. We illustrate some potential uses of the data by examining several interesting trends in the Court’s use of the shadow docket over time.
International relations (IR) as a discipline have had a troubled history in Italy. Indeed, the previous academic literature on the topic has highlighted how the lack of critical mass and influence of Italian IR scholarship have negatively impacted its visibility at the international level (Lucarelli and Menotti in Rivista Italiana di Scienza Politica 32(1):32–82, 2002; Friedrichs in European approaches to international relations theory: a house with many mansions, Routledge, London, 2004). However, there are few systematic studies that focus on the scientific publications of Italian scholars in IR. In order to fill this gap, the article presents the results of a study assessing Italian scholars’ impact in peer-reviewed international journals. Specifically, this work aims to make a broader assessment of Italian scholars’ publications from 2011 to 2017 using a database that includes 25 high impact-factor international peer-reviewed journals and five Italian journals. It also aims to identify the gender, geographic origin, affiliation and academic role of Italian scholars, as well as the topic, area, theoretical approach and methodological underpinnings of each article, so as to generate previously unexplored findings on the solidity and impact of Italian IR research both in Italy and abroad.
The components of a continuous quality improvement (CQI) program, a database, documentation, metrics, benchmarks, protocols, pathways, and standardized order sets are discussed.
Electronic Health Records (EHR) analysis is pivotal in advancing medical research. Numerous real-world EHR data providers offer data access through exported datasets. While enabling profound research possibilities, exported EHR data requires quality control and restructuring for meaningful analysis. Challenges arise in medical events (e.g., diagnoses or procedures) sequence analysis, which provides critical insights into conditions, treatments, and outcomes progression. Identifying causal relationships, patterns, and trends requires a more complex approach to data mining and preparation.
Methods:
This paper introduces EHRchitect – an application written in Python that addresses the quality control challenges by automating dataset transformation, facilitating the creation of a clean, formatted, and optimized MySQL database (DB), and sequential data extraction according to the user’s configuration.
Results:
The tool creates a clean, formatted, and optimized DB, enabling medical event sequence data extraction according to users’ study configuration. Event sequences encompass patients’ medical events in specified orders and time intervals. The extracted data are presented as distributed Parquet files, incorporating events, event transitions, patient metadata, and events metadata. The concurrent approach allows effortless scaling for multi-processor systems.
Conclusion:
EHRchitect streamlines the processing of large EHR datasets for research purposes. It facilitates extracting sequential event-based data, offering a highly flexible framework for configuring event and timeline parameters. The tool delivers temporal characteristics, patient demographics, and event metadata to support comprehensive analysis. The developed tool significantly reduces the time required for dataset acquisition and preparation by automating data quality control and simplifying event extraction.
This article presents an original database on international standards, constructed using modern data gathering methods. StanDat facilitates studies into the role of standards in the global political economy by (1) being a source for descriptive statistics, (2) enabling researchers to assess scope conditions of previous findings, and (3) providing data for new analyses, for example the exploration of the relationship between standardization and trade, as demonstrated in this article. The creation of StanDat aims to stimulate further research into the domain of standards. Moreover, by exemplifying data collection and dissemination techniques applicable to investigating less-explored subjects in the social sciences, it serves as a model for gathering, systematizing, and sharing data in areas where information is plentiful yet not readily accessible for research.
Flavonoids are a key class of polyphenols, i.e., phytochemical compounds present in foods and beverages, which have been described as having health benefits in preventing several chronic diseases. Estimating flavonoid intake has already been conducted in several countries but has yet to be performed in Portugal. This study included 5005 participants aged 3–84 years and aimed to estimate dietary flavonoid intake in the Portuguese population, using data from the National Food and Physical Activity Survey 2015–2016, providing information on intake, main food contributors and the socio-demographic factors associated with the intake. Food intake data from the survey was converted to flavonoid intake using a database built to include the most updated USDA databases on flavonoids, isoflavones and proanthocyanidins and the Phenol-Explorer database. The rationale for combining food consumption data and different flavonoid databases using the FoodEx2 classification system was established. Linear regressions assessed the associations between socio-demographic factors and dietary flavonoid intake. The total flavonoid intake of the Portuguese population was estimated to be 107·3 mg/d. Flavanols were the most representative subclass, followed by flavonols, anthocyanidins, flavanones, flavones and isoflavones. Fruits and vegetables were the primary food contributors, providing 31·5 % and 12·4 % of the total flavonoid intake. Adolescents had the lowest total flavonoid intake, and older adults had the highest. This study provides information on the Portuguese population’s dietary flavonoids, allowing for international comparisons. It can also streamline forthcoming investigations into the link between flavonoid consumption and its impact on health, contributing to the future establishment of dietary reference values.
In the past decades, a backlash against globalization has been brewing, especially in advanced economies. Despite this backlash being only partly determined by trade, we observe an increasing demand for transparency on procedures, methodologies, and results. Impact assessments (IAs) aim at identifying expected effects of trade agreements and at highlighting policymakers' concerns, thus representing an important tool to foster public acceptance. To help us identify spillovers of trade liberalization, we construct a country and sector-specific database of impact assessments. This database provides an overview of the evolution of the coverage and methodological approaches taken by the EU and US for their IAs. We rely on official EU and US sources over the period 1990–2023. We first observe differences in terms of methodology and institutional framework within and between the two regions. Secondly, the coverage of non-trade outcomes has evolved over time both for the EU and the US, with the inclusion of more labour, environmental, and human rights indicators as well as cross-cutting issues. We observe that the depth of the evaluation is correlated with the partner country's social protection and environmental performance. Lastly, we find that the inclusion of a sector in the analysis is driven by economic reasons in the EU but by political reasons in the US.
There has been a growing recognition of the significant role played by the human gut microbiota in altering the bioavailability as well as the pharmacokinetic and pharmacodynamic aspects of orally ingested xenobiotic and biotic molecules. The determination of species-specific contributions to the metabolism of biotic and xenobiotic molecules has the potential to aid in the development of new therapeutic and nutraceutical molecules that can modulate human gut microbiota. Here we present “GutBugDB,” an open-access digital repository that provides information on potential gut microbiome-mediated biotransformation of biotic and xenobiotic molecules using the predictions from the GutBug tool. This database is constructed using metabolic proteins from 690 gut bacterial genomes and 363,872 protein enzymes assigned with their EC numbers (with representative Expasy ID and domains present). It provides information on gut microbiome enzyme-mediated metabolic biotransformation for 1439 FDA-approved drugs and nutraceuticals. GutBugDB is publicly available at https://metabiosys.iiserb.ac.in/gutbugdb/.
It is vital that horizon scanning organizations can capture and disseminate intelligence on new and repurposed medicines in clinical development. To our knowledge, there are no standardized classification systems to capture this intelligence. This study aims to create a novel classification system to allow new and repurposed medicines horizon scanning intelligence to be disseminated to healthcare organizations.
Methods
A multidisciplinary working group undertook literature searching and an iterative, three-stage piloting process to build consensus on a classification system. Supplementary data collection was carried out to facilitate the implementation and validation of the system on the National Institute of Health and Care Research (NIHR) Innovation Observatory (IO)‘s horizon scanning database, the Medicines Innovation Database (MInD).
Results
Our piloting process highlighted important issues such as the patency and regulatory approval status of individual medicines and how combination therapies interact with these characteristics. We created a classification system with six values (New Technology, Repurposed Technology (Off-patent/Generic), Repurposed Technology (On-patent/Branded), Repurposed Technology (Never commercialised), New + Repurposed Technology (Combinations-only), Repurposed Technology (Combinations-only)) that account for these characteristics to provide novel horizon scanning insights. We validated our system through application to over 20,000 technology records on the MInD.
Conclusions
Our system provides the opportunity to deliver concise yet informative intelligence to healthcare organizations and those studying the clinical development landscape of medicines. Inbuilt flexibility and the use of publicly available data sources ensure that it can be utilized by all, regardless of location or resource availability.
Conversation Analysis usually involves collecting, organizing, and analyzing audiovisual data clips and transcripts. In this chapter, we provide guidance based on common CA research practices for making, naming, and organizing clips. We provide examples of both digital and analog tools and methods for preparing, manipulating, and reviewing transcripts and data throughout the analytic research cycle. Finally, we discuss common data management techniques for protecting participant privacy by masking voices, faces, and other identifiable features before sharing clips and transcripts e.g., during CA data sessions. This chapter aims to support CA researchers who have already collected and organized their field recordings, and are ready to start making, sharing, and analyzing collections of clips.
Historically, information about radio astronomy observatories to be used in sharing and compatibility studies was provided by the relevant national administrations responsible for their radio astronomy observatories. The IAU CPS provides a streamlined platform for radio astronomy observatories and satellite operators to collaborate on sharing and compatibility studies prior to enrolling on the national or international regulatory framework, reducing the burden on national administrations. This leads to the requirement to have an alternative solution to have information about radio astronomy observatories. We discussed currently used alternative data collection as well as shared the information on the creation of a CPS-based database to serve as a source of radio astronomy observatories characteristics for sharing and compatibility studies to be carried out within IAU CPS.
The purpose of this short research note is to draw attention to two major pitfalls of working with databases of decisions of the Court of Justice of the European Union. The first one is technical in nature and relates to the discrepant coverage of the Curia and Eur-Lex databases. The second one is linguistic in nature and relates to the fact that most scholars using these databases work in English. New work on this front is capable of addressing the first issue but a change to research practices would be required to address the second.
Advances in comparative ageing research strongly depend on data quality and quantity. Across the world, zoos and aquariums gather data on the physiology, morphology, health and demography of the animals under their care to facilitate their management. Many of these data are hosted in a centralized database, the Species360 Zoological Information Management System (ZIMS). As of 2022, ZIMS held records on ~10 million individuals across 22,000 species and over 1200 member institutions, with historical animal records dating back to the mid-1800s. These millions of age-specific data could enable analyses testing hypotheses at individual and species levels and between species with vastly different life history strategies. This chapter summarizes the diversity of questions (ranging from evolutionary theories to mechanistic hypotheses) for ageing research that could be addressed using data from zoo and aquarium populations. In addition, many of these studies could inform the management and conservation of animals, not only in zoos and aquariums, but also in the wild.
This chapter introduces a database on the international spread of uranium enrichment and plutonium reprocessing facilities. This database identifies countries with nuclear latency and serves as the basis for the empirical analyses carried out in the book.
This chapter provides an overview of the process of conceiving, researching, editing, and publishing dictionaries, both synchronic (or commercial) and historical. Discussed methods and tools for making dictionaries range from traditional hand-copying of citations from print books and paper-and-pencil editing to sophisticated electronic technologies like databases, corpora, concordances, and networked editing software. The chapter shows how editorial conception of the needs and sophistication of the end user largely determines the dictionary’s length and headword list as well as the format, defining style, and level of detail in entries. The chapter goes on to examine how the pressures of commercial publishing, with its looming deadlines and pressing need to recoup investment by profits from sales, affect the scope of dictionaries and the amount of time editors can devote to a project, and how these pressures differ from those affecting longer-trajectory, typically grant-funded historical dictionaries. Assessing the consequent challenges for managing and motivating people working in these two very different situations, what may be the most important factor in a project’s success, concludes the survey of dictionary editing.
Society of Thoracic Surgeons Congenital Heart Surgery Database is the largest congenital heart surgery database worldwide but does not provide information beyond primary episode of care. Linkage to hospital electronic health records would capture complications and comorbidities along with long-term outcomes for patients with CHD surgeries. The current study explores linkage success between Society of Thoracic Surgeons Congenital Heart Surgery Database and electronic health record data in North Carolina and Georgia.
Methods:
The Society of Thoracic Surgeons Congenital Heart Surgery Database was linked to hospital electronic health records from four North Carolina congenital heart surgery using indirect identifiers like date of birth, sex, admission, and discharge dates, from 2008 to 2013. Indirect linkage was performed at the admissions level and compared to two other linkages using a “direct identifier,” medical record number: (1) linkage between Society of Thoracic Surgeons Congenital Heart Surgery Database and electronic health records from a subset of patients from one North Carolina institution and (2) linkage between Society of Thoracic Surgeons data from two Georgia facilities and Georgia’s CHD repository, which also uses direct identifiers for linkage.
Results:
Indirect identifiers successfully linked 79% (3692/4685) of Society of Thoracic Surgeons Congenital Heart Surgery Database admissions across four North Carolina hospitals. Direct linkage techniques successfully matched Society of Thoracic Surgeons Congenital Heart Surgery Database to 90.2% of electronic health records from the North Carolina subsample. Linkage between Society of Thoracic Surgeons and Georgia’s CHD repository was 99.5% (7,544/7,585).
Conclusions:
Linkage methodology was successfully demonstrated between surgical data and hospital-based electronic health records in North Carolina and Georgia, uniting granular procedural details with clinical, developmental, and economic data. Indirect identifiers linked most patients, consistent with similar linkages in adult populations. Future directions include applying these linkage techniques with other data sources and exploring long-term outcomes in linked populations.
Organizations supporting translational research and translational science, including Clinical and Translational Science Award (CTSA) hubs, provide a diverse and often changing array of resources, support, and services to a myriad of researchers and research efforts. While a wide-ranging scope of programs is essential to the advancement of translational research and science, it also complicates a systematic and unified process for tracking activities, studying research processes, and examining impact. To overcome these challenges, the Duke University School of Medicine’s CTSA hub created a data platform, Translational Research Accomplishment Cataloguer (TRACER), that provides capacity to enhance strategic decision-making, impact assessment, and equitable resource distribution. This article reviews TRACER development processes, provides an overview of the TRACER platform, addresses challenges in the development process, and describes avenues for addressing or overcoming these challenges. TRACER development allowed our hub to conceptually identify key processes and goals within programs and linkages between programs, and it sets the stage for advancing evidence-based improvement across our hub. This platform development provides key insight into facilitators that can inform other initiatives seeking to collect and align organizational data for strategic decision-making and impact assessment. TRACER or similar platforms are additionally well positioned to advance the study of translational science.
Archaeologists seek to improve our understanding of the past by studying, preserving, protecting, and sharing nonreplaceable archaeological resources. Archaeological collections hold information that can assist these aims as long as they are properly cared for, identified, and accessible. One of the most serious barriers is the lack of large-scale coordinated efforts to make archaeological collections findable and accessible. This article suggests that developing and implementing the use of a standardized set of attributes regarding collections provides solutions and strategies to find collections. These attributes can connect and standardize existing archaeological collections from a variety of sources (federal and state agencies, CRM firms, Indigenous and descendant communities, and academic departments), serving the profession in multiple ways. Most critically, the baseline data can be synthesized to inform and direct priorities for future fieldwork, thereby decreasing redundancy in archaeological collections and improving curation efforts nationwide. Such efforts would also provide a resource to students and researchers looking to understand and interpret the past at multiple scales by encouraging more collections-based research and less archaeological site destruction. Access for descendant communities will also be improved with information about their cultural heritage. This, in turn, encourages transparency and collaboration between those communities and archaeologists.
Informal borrowings constitute an important linguistic phenomenon, yet they remain underrepresented in scholarly literature. This book is to remedy the situation. Drawing from the methodological framework of documentary linguistics and sociolinguistics, it relies on lexical material from a large database of citations from diverse sources – including spoken utterances, films and TV shows, print, and social media – to ensure authenticity and representativeness. Much space is devoted to the presentation, explanation, interpretation, and illustration of language data; the format of description is designed to be extensive, covering a wide range of themes which allow an examination from various perspectives. The description is amply supported throughout the text with usage examples that illustrate linguistic patterns, show the sociocultural context in which they are used, and attest to the very existence of these expressions.
We present the ‘SISAL webApp’—a web-based tool to query the Speleothem Isotope Synthesis and AnaLysis (SISAL) database. The software provides an easy-to-use front-end interface to mine data from the SISAL database while providing the SQL code alongside as a learning tool. It allows for simple and increasingly complex querying of the SISAL database based on various data and metadata fields. The SISAL webApp version currently hosts SISALv2 of the database with 691 records from 294 sites, 512 of which have standardized chronologies. The SISAL webApp has sufficient flexibility to host future versions of the SISAL database, which may include allied speleothem information such as trace elements and cave-monitoring records. The SISAL webApp will increase accessibility to the SISAL database while also functioning as a learning tool for more advanced ways of querying paleoclimate databases. The SISAL webApp is available at http://geochem.hu/SISAL_webApp.