This journal utilises an Online Peer Review Service (OPRS) for submissions. By clicking "Continue" you will be taken to our partner site
https://mc.manuscriptcentral.com/dataandpolicy.
Please be aware that your Cambridge account is not valid for this OPRS and registration is required. We strongly advise you to read all "Author instructions" in the "Journal information" area prior to submitting.
To save this undefined to your undefined account, please select one or more formats and confirm that you agree to abide by our usage policies. If this is the first time you used this feature, you will be asked to authorise Cambridge Core to connect with your undefined account.
Find out more about saving content to .
To send this article to your Kindle, first ensure no-reply@cambridge.org is added to your Approved Personal Document E-mail List under your Personal Document Settings on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part of your Kindle email address below. Find out more about sending to your Kindle.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations. ‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi. ‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
The REDATAM (retrieval of data for small areas by microcomputer) statistical package and format, developed by ECLAC, has been a critical tool for disseminating census data across Latin America since the 1990s. However, significant limitations persist, including its proprietary nature, lack of documentation, and restricted flexibility for advanced data analysis. These challenges hinder the transformation of raw census data into actionable information for policymakers, researchers, and advocacy groups. To address these issues, we developed Open REDATAM, an open-source and multiplatform tool that converts REDATAM data into widely supported CSV files and native R and Python data structures. By providing integration with R and Python, Open REDATAM empowers users to work with the tools they already know and perform data analyses without leaving their R or Python window. Our work emphasizes the need for a REDATAM official format specification to further enable informed policy debates that can improve policy processes’ implementation and feedback.
As data becomes a key component of urban governance, the night-time economy is still barely visible in datasets or in policies to improve urban life. In the last 20 years, over 50 cities worldwide appointed night mayors and governance mechanisms to tackle conflicts, foster innovation, and help the night-time economy sector grow. However, the intersection of data, digital rights, and 24-hour cities still needs more studies, examples, and policies. Here, the key argument is that the increasing importance of the urban night in academia and local governments claims for much-needed responsible data practices to support and protect nightlife ecosystems. By understanding these ecosystems and addressing data invisibilities, it is possible to develop a robust framework anchored in safeguarding human rights in the digital space and create comprehensive policies to help such ecosystems thrive. Night-time governance matters for the data policy community for three reasons. First, it brings together issues covered in different disciplines by various stakeholders. We need to build bridges between sectors to avoid siloed views of urban data governance. Second, thinking about data in cities also means considering the social, economic, and cultural impact of datafication and artificial intelligence on a 24-hour cycle. Creating a digital rights framework for the night means putting into practice principles of justice, ethics, and responsibility. Third, as Night Studies is an emerging field of research, policy and advocacy, there is an opportunity to help shape how, why, and when data about the night is collected and made available to society.
The complex socioeconomic landscape of conflict zones demands innovative approaches to assess and predict vulnerabilities for crafting and implementing effective policies by the United Nations (UN) institutions. This article presents a groundbreaking Augmented Intelligence-driven Prediction Model developed to forecast multidimensional vulnerability levels (MVLs) across Afghanistan. Leveraging a symbiotic fusion of human expertise and machine capabilities (e.g., artificial intelligence), the model demonstrates a predictive accuracy ranging between 70% and 80%. This research not only contributes to enhancing the UN Early Warning (EW) Mechanisms but also underscores the potential of augmented intelligence in addressing intricate challenges in conflict-ridden regions. This article outlines the use of augmented intelligence methodology applied to a use case to predict MVLs in Afghanistan. It discusses the key findings of the pilot project, and further proposes a holistic platform to enhance policy decisions through augmented intelligence, including an EW mechanism to significantly improve EW processes, thereby supporting decision-makers in formulating effective policies and fostering sustainable development within the UN.
Artificial intelligence (AI) requires new ways of evaluating national technology use and strategy for African nations. We conduct a survey of existing “readiness” assessments both for general digital adoption and AI policy in particular. We conclude that existing global readiness assessments do not fully capture African states’ progress in AI readiness and lay the groundwork for how assessments can be better used for the African context. We consider the extent to which these indicators map to the African context and what these indicators miss in capturing African states’ on-the-ground work in meeting AI capability. Through case studies of four African nations of diverse geographic and economic dimensions, we identify nuances missed by global assessments and offer high-level policy considerations for how states can best improve their AI readiness standards and prepare their societies to capture the benefits of AI.
Sexual and gender–based violence (SGBV) is a multifaceted, endemic, and nefarious phenomenon that remains poorly measured and understood, despite greater global awareness of the issue. While efforts to improve data collection methods have increased–including the implementation of the Demographic and Health Survey (DHS) in some countries–the lack of reliable SGBV data remains a significant challenge to developing targeted policy interventions and advocacy initiatives. Using a recent mixed–methods research project conducted by the authors in Sierra Leone as a case study, this paper discusses the current status of SGBV data, challenges faced, and potential research a pproaches.
Climate change exacerbates existing risks and vulnerabilities for people globally, and migration is a longstanding adaptation response to climate risk. The mechanisms through which climate change shapes human mobility are complex, however, and gaps in data and knowledge persist. In response to these gaps, the United Nations Development Programme’s (UNDP) Predictive Analytics, Human Mobility, and Urbanization Project employed a hybrid approach that combined predictive analytics with participatory foresight to explore climate change-related mobility in Pakistan and Viet Nam from 2020 to 2050. Focusing on Karachi and Ho Chi Minh City, the project estimated temporal and spatial mobility patterns under different climate change scenarios and evaluated the impact of such in-migration across key social, political, economic, and environmental domains. Findings indicate that net migration into these cities could significantly increase under extreme climate scenarios, highlighting both the complex spatial patterns of population change and the potential for anticipatory policies to mitigate these impacts. While extensive research exists on foresight methods and theory, process reflections are underrepresented. The innovative approach employed within this project offers valuable insights on foresight exercise design choices and their implications for effective stakeholder engagement, as well as the applicability and transferability of insights in support of policymaking. Beyond substantive findings, this paper offers a critical reflection on the methodological alignment of data-driven and participatory foresight with the aim of anticipatory policy ideation, seeking to contribute to the enhanced effectiveness of foresight practices.
This study examines Nigeria’s National Information Technology Development Agency Code of Practice for Interactive Computer Service Platforms as one of Africa’s first push towards digital and social media co-regulation. Already established as a regulatory practice in Europe, co-regulation emphasises the need to impose duties of care on platforms and hold them, instead of users, accountable for safe online experiences. It is markedly different from the prior (and existing) regulatory paradigm in Nigeria, which is based on direct user regulation. By analysing the Code of Practice, therefore, this study considers what Nigeria’s radical turn towards co-regulation means for digital policy and social media regulation in relation to standards, information-gathering, and enforcement. It further sheds light on what co-regulation entails for digital regulatory practice in the wider African context, particularly in terms of the balance of power realities between Global North platforms and Global South countries.
In this paper, we provide a systematic review of existing artificial intelligence (AI) regulations in Europe, the United States, and Canada. We build on the qualitative analysis of 129 AI regulations (enacted and not enacted) to identify patterns in regulatory strategies and in AI transparency requirements. Based on the analysis of this sample, we suggest that there are three main regulatory strategies for AI: AI-focused overhauls of existing regulation, the introduction of novel AI regulation, and the omnibus approach. We argue that although these types emerge as distinct strategies, their boundaries are porous as the AI regulation landscape is rapidly evolving. We find that across our sample, AI transparency is effectively treated as a central mechanism for meaningful mitigation of potential AI harms. We therefore focus on AI transparency mandates in our analysis and identify six AI transparency patterns: human in the loop, assessments, audits, disclosures, inventories, and red teaming. We contend that this qualitative analysis of AI regulations and AI transparency patterns provides a much needed bridge between the policy discourse on AI, which is all too often bound up in very detailed legal discussions and applied sociotechnical research on AI fairness, accountability, and transparency.
Research in decentralized computing, specifically in consensus algorithms, has focused on providing resistance to an adversary with a minority stake. This has resulted in systems that are majoritarian in the extreme, ignoring valuable lessons learned in law and politics over centuries. In this article, we first detail this phenomenon of majoritarianism and point out how minority protections in the nondigital world have been implemented. We motivate adding minority protections to collaborative systems with examples. We also show how current software deployment models exacerbate majoritarianism, highlighting the problem of monoculture in client software in particular. We conclude by giving some suggestions on how to make decentralized computing less hostile to those in the minority.
We discuss the emerging technology of digital twins (DTs) and the expected demands as they scale to represent increasingly complex, interconnected systems. Several examples are presented to illustrate core use cases, highlighting a progression to represent both natural and engineered systems. The forthcoming challenges are discussed around a hierarchy of scales, which recognises systems of increasing aggregation. Broad implications are discussed, encompassing sensing, modelling, and deployment, alongside ethical and privacy concerns. Importantly, we endorse a modular and peer-to-peer view for aggregate (interconnected) DTs. This mindset emphasises that DT complexity emerges from the framework of connections (Wagg et al. [2024, The philosophical foundations of digital twinning, Preprint]) as well as the (interpretable) units that constitute the whole.
Displacement continues to increase at a global scale and is increasingly happening in complex, multicrisis settings, leading to more complex and deeper humanitarian needs. Humanitarian needs are therefore increasingly outgrowing the available humanitarian funding. Thus, responding to vulnerabilities before disaster strikes is crucial but anticipatory action is contingent on the ability to accurately forecast what will happen in the future. Forecasting and contingency planning are not new in the humanitarian sector, where scenario-building continues to be an exercise conducted in most humanitarian operations to strategically plan for coming events. However, the accuracy of these exercises remains limited. To address this challenge and work with the objective of providing the humanitarian sector with more accurate forecasts to enhance the protection of vulnerable groups, the Danish Refugee Council has already developed several machine learning models. The Anticipatory Humanitarian Action for Displacement uses machine learning to forecast displacement in subdistricts in the Liptako-Gourma region in Sahel, covering Burkina Faso, Mali, and Niger. The model is mainly built on data related to conflict, food insecurity, vegetation health, and the prevalence of underweight to forecast displacement. In this article, we will detail how the model works, the accuracy and limitations of the model, and how we are translating the forecasts into action by using them for anticipatory action in South Sudan and Burkina Faso, including concrete examples of activities that can be implemented ahead of displacement in the place of origin, along routes and in place of destination.
This study analyzes National Cyber Security Strategies (NCSSs) of G20 countries through a novel combination of qualitative and quantitative methodologies. It focuses on delineating the shared objectives, distinct priorities, latent themes, and key priorities within the NCSSs. Latent dirichlet allocation topic modeling technique was used to identify implicit themes in the NCSSs to augment the explicitly articulated strategies. By exploring the latest versions of NCSS documents, the research uncovers a detailed panorama of multinational cybersecurity dynamics, offering insights into the complexities of shared and unique national cybersecurity challenges. Although challenged by the translation of non-English documents and the intrinsic limitations of topic modeling, the study significantly contributes to the cybersecurity policy domain, suggesting directions for future research to broaden the analytical scope and incorporate more diverse national contexts. In essence, this research underscores the indispensability of a multifaceted, analytical approach in understanding and devising NCSSs, vital for navigating the complex, and ever-changing digital threat environment.
Effective enforcement of laws and regulations hinges heavily on robust inspection policies. While data-driven approaches to testing the effectiveness of these policies are gaining popularity, they suffer significant drawbacks, particularly a lack of explainability and generalizability. This paper proposes an approach to crafting inspection policies that combines data-driven insights with behavioral theories to create an agent-based simulation model that we call a theory-infused phenomenological agent-based model (TIP-ABM). Moreover, this approach outlines a systematic process for combining theories and data to construct a phenomenological ABM, beginning with defining macro-level empirical phenomena. Illustrated through a case study of the Dutch inland shipping sector, the proposed methodology enhances explainability by illuminating inspectors’ tacit knowledge while iterating between statistical data and underlying theories. The broader generalizability of the proposed approach beyond the inland shipping context requires further research.
The global number of individuals experiencing forced displacement has reached its highest level in the past decade. In this context, the provision of services for those in need requires timely and evidence-based approaches. How can mobile phone data (MPD) based analyses address the knowledge gap on mobility patterns and needs assessments in forced displacement settings? To answer this question, in this paper, we examine the capacity of MPD to function as a tool for anticipatory analysis, particularly in response to natural disasters and conflicts that lead to internal or cross-border displacement. The paper begins with a detailed review of the processes involved in acquiring, processing, and analyzing MPD in forced displacement settings. Following this, we critically assess the challenges associated with employing MPD in policy-making, with a specific focus on issues of user privacy and data ethics. The paper concludes by evaluating the potential benefits of MPD analysis for targeted and effective policy interventions and discusses future research avenues, drawing on recent studies and ongoing collaborations with mobile network operators.
In 2020, the COVID-19 pandemic resulted in a rapid response from governments and researchers worldwide, but information-sharing mechanisms were variable, and many early efforts were insufficient for the purpose. We interviewed fifteen data professionals located around the world, working with COVID-19-relevant data types in semi-structured interviews. Interviews covered both challenges and positive experiences with data in multiple domains and formats, including medical records, social deprivation, hospital bed capacity, and mobility data. We analyze this qualitative corpus of experiences for content and themes and identify four sequential barriers a researcher may encounter. These are: (1) Knowing data exists, (2) being able to access that data, (3) data quality, and (4) ability to share data onwards. A fifth barrier, (5) human throughput capacity, is present throughout all four stages. Examples of these barriers range from challenges faced by single individuals to non-existent records of historic mingling/social distance laws, and up to systemic geopolitical data suppression. Finally, we recommend that governments and local authorities explicitly create machine-readable temporal “law as code” for changes in laws such as mobility/mingling laws and changes in geographical regions.
Forecasting international migration is a challenge that, despite its political and policy salience, has seen a limited success so far. In this proof-of-concept paper, we employ a range of macroeconomic data to represent different drivers of migration. We also take into account the relatively consistent set of migration policies within the European Common Market, with its constituent freedom of movement of labour. Using panel vector autoregressive (VAR) models for mixed-frequency data, we forecast migration in the short- and long-term horizons for 26 of the 32 countries within the Common Market. We demonstrate how the methodology can be used to assess the possible responses of other macroeconomic variables to unforeseen migration events—and vice versa. Our results indicate reasonable in-sample performance of migration forecasts, especially in the short term, although with varying levels of accuracy. They also underline the need for taking country-specific factors into account when constructing forecasting models, with different variables being important across the regions of Europe. For the longer term, the proposed methods, despite high prediction errors, can still be useful as tools for setting coherent migration scenarios and analysing responses to exogenous shocks.
Rapid urbanization poses several challenges, especially when faced with an uncontrolled urban development plan. Therefore, it often leads to anarchic occupation and expansion of cities, resulting in the phenomenon of urban sprawl (US). To support sustainable decision–making in urban planning and policy development, a more effective approach to addressing this issue through US simulation and prediction is essential. Despite the work published in the literature on the use of deep learning (DL) methods to simulate US indicators, almost no work has been published to assess what has already been done, the potential, the issues, and the challenges ahead. By synthesising existing research, we aim to assess the current landscape of the use of DL in modelling US. This article elucidates the complexities of US, focusing on its multifaceted challenges and implications. Through an examination of DL methodologies, we aim to highlight their effectiveness in capturing the complex spatial patterns and relationships associated with US. This work begins by demystifying US, highlighting its multifaceted challenges. In addition, the article examines the synergy between DL and conventional methods, highlighting the advantages and disadvantages. It emerges that the use of DL in the simulation and forecasting of US indicators is increasing, and its potential is very promising for guiding strategic decisions to control and mitigate this phenomenon. Of course, this is not without major challenges, both in terms of data and models and in terms of strategic city planning policies.
Currently, artificial intelligence (AI) is integrated across various segments of the public sector, in a scattered and fragmented manner, aiming to enhance the quality of people’s lives. While AI adoption has proven to have a great impact, there are several aspects that hamper its utilization in public administration. Therefore, a large set of initiatives is designed to play a pivotal role in promoting the adoption of reliable AI, including documentation as a key driver. The AI community has been proactively recommending a variety of initiatives aimed at promoting the adoption of documentation practices. While currently proposed AI documentation artifacts play a crucial role in increasing the transparency and accountability of various facts about AI systems, we propose a code-bound declarative documentation framework that aims to support the responsible deployment of AI-based solutions. Our proposed framework aims to address the need to shift the focus from data and models being considered in isolation to the reuse of AI solutions as a whole. By introducing a formalized approach to describing adaptation and optimization techniques, we aim to enhance existing documentation alternatives. Furthermore, its utilization in the public administration aims to foster the rapid adoption of AI-based applications due to the open access to common use cases in the public sector. We further showcase our proposal with a public sector-specific use case, such as a legal text classification task, and demonstrate how the AI Product Card enables its reuse through the interactions of the formal documentation specifications with the modular code references.