Hostname: page-component-89b8bd64d-dvtzq Total loading time: 0 Render date: 2026-05-12T16:17:55.091Z Has data issue: false hasContentIssue false

Reversing the logic of generative AI alignment: a pragmatic approach for public interest

Published online by Cambridge University Press:  10 March 2025

Gleb Papyshev*
Affiliation:
Division of Social Science, The Hong Kong University of Science and Technology, Kowloon, Hong Kong, China
*

Abstract

The alignment of artificial intelligence (AI) systems with societal values and the public interest is a critical challenge in the field of AI ethics and governance. Traditional approaches, such as Reinforcement Learning with Human Feedback (RLHF) and Constitutional AI, often rely on pre-defined high-level ethical principles. This article critiques these conventional alignment frameworks through the philosophical perspectives of pragmatism and public interest theory, arguing against their rigidity and disconnect with practical impacts. It proposes an alternative alignment strategy that reverses the traditional logic, focusing on empirical evidence and the real-world effects of AI systems. By emphasizing practical outcomes and continuous adaptation, this pragmatic approach aims to ensure that AI technologies are developed according to the principles that are derived from the observable impacts produced by technology applications.

Information

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2025. Published by Cambridge University Press
Submit a response

Comments

No Comments have been published for this article.