Hostname: page-component-89b8bd64d-z2ts4 Total loading time: 0 Render date: 2026-05-08T05:29:27.472Z Has data issue: false hasContentIssue false

Muddling-Through and Deep Learning for Managing Large-Scale Uncertain Risks

Published online by Cambridge University Press:  02 September 2019

Tony Cox*
Affiliation:
Cox Associates and University of Colorado, 503 N. Franklin Street, DENVER, CO 80218, USA, e-mail: tcoxdenver@aol.com
*
Rights & Permissions [Opens in a new window]

Abstract

Managing large-scale, geographically distributed, and long-term risks arising from diverse underlying causes – ranging from poverty to underinvestment in protecting against natural hazards or failures of sociotechnical, economic, and financial systems – poses formidable challenges for any theory of effective social decision-making. Participants may have different and rapidly evolving local information and goals, perceive different opportunities and urgencies for actions, and be differently aware of how their actions affect each other through side effects and externalities. Six decades ago, political economist Charles Lindblom viewed “rational-comprehensive decision-making” as utterly impracticable for such realistically complex situations. Instead, he advocated incremental learning and improvement, or “muddling through,” as both a positive and a normative theory of bureaucratic decision-making when costs and benefits are highly uncertain. But sparse, delayed, uncertain, and incomplete feedback undermines the effectiveness of collective learning while muddling through, even if all participant incentives are aligned; it is no panacea. We consider how recent insights from machine learning – especially, deep multiagent reinforcement learning – formalize aspects of muddling through and suggest principles for improving human organizational decision-making. Deep learning principles adapted for human use can not only help participants in different levels of government or control hierarchies manage some large-scale distributed risks, but also show how rational-comprehensive decision analysis and incremental learning and improvement can be reconciled and synthesized.

Information

Type
Symposium on Analysis for Uncertain Futures
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution, and reproduction in any medium, provided the original work is properly cited.
Copyright
© Society for Benefit-Cost Analysis, 2019
Figure 0

Table 1 Some enhancements to reinforcement learning (RL) algorithms.

Figure 1

Table 2 Some principles for safe learning, i.e., learning without risking catastrophic failures.

Figure 2

Table 3 Some MARL variations and extensions.