Cultural Differences in People's Reactions and Applications of Robots, Algorithms, and Artificial Intelligence

Abstract Although research in cultural psychology has established that virtually all human behaviors and cognitions are in some ways shaped by culture, culture has been surprisingly absent from the emerging literature on the psychology of technology. In this perspective article, we first review recent findings on machine aversion versus appreciation. We then offer a cross-cultural perspective in understanding how people might react differently to machines. We propose three frameworks – historical, religious, and exposure – to explain how Asians might be more accepting of machines than their Western counterparts. We end the article by discussing three exciting human–machine applications found primarily in Asia and provide future research directions.


Introduction
Robots, algorithms, and artificial intelligence (AI) are becoming rapidly commonplace in today's technology-driven world.In 2020, a record three million industrial robots were operating across the globe (IFR, 2021a).Asia, in particular, took the lead being the largest market for industrial robots, with countries like China, Japan, and South Korea continuing to be the first, second, and fourth biggest market players, respectively.Apart from industrial robots, Asia also has a stronghold on the development of social robots, which is projected to grow 36% in market size in the APAC region alone by 2025, dominating other markets like Europe and the United States (Technavio, 2022).Unlike industrial robots, which are predominantly found in factories, social robots are designed to behave and interact with humans.Japan, for example, is home to some robot-manned hotels (Yam, Bigman, Tang, et al., 2021), robot pets (Craft, 2022), as well as robot caregivers at nursing homes (Lufkin, 2020).In China, the COVID-19 lockdown saw robots deployed to deliver food and medicine, as well as disinfect hospitals (Fannin, 2020).
Outside of robots, Asia has grown substantially in the space of AI technologies as well (International Institute of Communications, 2020).In Singapore and Japan, machine learning algorithms have been adopted by big insurance companies to automate claims processes.In China, AI helps farmers monitor and predict environmental conditions and yields.Despite these promising and exciting developments in technology, however, social scientists have been met with a perplexing finding: people are averse to machines 1 (for a review, see Gray, Yam, Eng, Wilbanks, & Waytz, 2023).Robots may be helpful, but people also find them creepy (MacDorman, 2006).Algorithms may make objective, calculated decisions, but people do not always trust them (von Eschenbach, 2021).Importantly, these effects appear to be culturally dependent (Lim, Rooksby, & Cross, 2021).In this article, we first briefly review recent findings on machine aversion versus appreciation, focusing on the whowho likes machines, and who does not?and whatwhat do machines do that people like or dislike?For the former, past work has looked at how people of varying expertise (Logg, Minson, & Moore, 2019), ages (Chien et al., 2019), gender (Nomura, 2017), and occupations (Reich & Eyssel, 2013) respond differently toward machines.For the latter, research has established that people also appreciate machines differently depending on the nature of the task it is doing.For example, people do not like algorithms making moral decisions (Bigman & Gray, 2018), but outside of the moral domain people perceive algorithms as better at objective than subjective tasks (Castelo, Bos, & Lehmann, 2019).Nevertheless, there is another important variable that may also affect attitudes and behaviors toward machinesculture, defined as an 'untidy and expansive set of material and symbolic concepts, such as world, environment, contexts, cultural systems, social systems, social structures, institutions, practices, policies, meanings, norms, and values, that give form and direction to behavior' (Markus & Kitayama, 2010: 422).
Broadly, culture 2 is who we areit encompasses our common values, beliefs, and norms, distinct from other groups (Lehman, Chiu, & Schaller, 2004), and is shaped by our unique histories, languages, geographies, and so on.Culture is often treated as all-encompassing; it can reach into and influence people's behavior from the way we greet each other (Li, 2009) to the way we negotiate job salaries (Lu, 2023).In this perspective article, we ask: how do cultures shape attitudes and behavior toward machines?To explore these differences, we then delineate a cross-cultural perspective with an attempt to explain culturally divergent responses to machines.We propose three frameworkshistorical, religious 3 , and exposureto help scholars understand the nuanced cultural differences in how people view and interact with machines.Finally, we look ahead and discuss three interesting areas where new technological applications are emerging in Asiasex, religion, and therapyand urge future research to attend to these new and exciting contexts, as well as their implications for human-human interactions.

Machine Aversion or Appreciation?
The phenomenon of machine aversion was first documented by Dietvorst, Simmons, and Massey (2015).In a series of studies, participants either witnessed an algorithm making a forecast, a human making a forecast, or both.Strangely, even after seeing algorithms outperform human forecasters (i.e., made better, more accurate predictions), participants still largely chose human forecasters over algorithmic ones to make future predictions.Furthermore, people seem most averse to machines making moral (Bigman & Gray, 2018) and social (Yeomans, Shah, Mullainathan, & Kleinberg, 2019) decisions, even when machines are better than humans at making these decisions.
Algorithms aside, robots also inspire feelings of threat and anxiety, especially in the context of the workplace.For instance, Yam and colleagues found that adults exposed to robots either physically (e.g., interacting with them at work in real life) or psychologically (e.g., reading an article about the presence of robots in businesses) report higher job insecurity and anxiety (Yam, Tang, Jackson, Su, & Gray, 2023), especially when interacting with humanoid robots (Yam, Bigman, & Gray, 2021).Such job insecurity further predicted employee burnout and workplace incivility.Separately, Gamez-Djokic and Waytz (2020) also found that such automation-related job anxiety can translate into negative sentiments toward out-groups, such as immigrants.Taken together, not only do these findings suggest that people generally fear novel technologies they do not yet understand, but they also hint that such fears toward machines can spill over and lead to negative interpersonal outcomes.
But there is another side of the storyone in which people do not fear, but prefer, machines.Admittedly, this literature is less developed compared to machine aversion research.In a series of studies, Logg et al. (2019) examined participants' preferences for human versus algorithmic predictions regarding people's weight, song popularity, and geopolitical events, among others, without providing any performance accuracy feedback.Interestingly, they found that people relied more heavily on the advice they thought was generated by algorithms than by people (though the forecasts were otherwise identical).Another line of research by Dietvorst, Simmons, and Massey (2018) has found that people are less averse to algorithms if they have the power to modify how algorithms work.
These findings challenge the existence of general machine aversion, instead, suggesting that certain individuals, like experts, may be more machine-averse than others, preferring to rely on their own predictions.Beyond individual differences, the task domain that the machines are operating in influences people's aversion (vs.appreciation) toward them as well.For example, people are highly averse to machines making decisions in the moral domain such as deciding if a criminal gets parole, or if a patient should undergo a risky surgery that may save or kill them (Bigman & Gray, 2018).This is because people do not perceive machines to be able to fully think or feel.People also have less trust in algorithmic advice for tasks that they perceive to be subjective (e.g., dating advice) than tasks that they perceive to be objective (e.g., financial advice) (Castelo et al., 2019).Evidently, people's attitudes toward machines are influenced by a range of factors, including individual user differences and the type of task the machine is doing.
Are Attitudes Toward Machines Culture-Specific?Culture may be a hidden influence on how people respond to machines.Cross-cultural research in the social sciences has paid much attention to the East-West dichotomy in exploring how people in the East diverge from the West in terms of values, beliefs, and behavioral norms (Kagitcibasi & Berry, 1989).For example, scholars have long discussed fundamental cultural differences in how the East and the West organize knowledge (Nisbett, Peng, Choi, & Norenzayan, 2001), use logical rules (Nisbett, 2004), and cooperate in teams (Qin et al., 2023).In terms of interacting with machines, a popular perception is that Eastern cultures are fonder of robots than Western ones (see headlines like 'Why Japanese Love Robots (And Americans Fear Them)'; Mims, 2010; or 'Asia Has Learnt to Love Robots -The West Should, Too'; Thornhill, 2018).
Take Japan for example, known for being the home to Paro, a therapeutic baby seal robot (IEEE, 2022), Geminoid HI-2, a robot doppelganger of human roboticist Hiroshi Ishiguro (Guizzo, 2010), and perhaps, most famously, Astro Boy, a friendly superpowered manga and anime android character who fights evil and saves lives (Robot Hall of Fame, 2004).Japan is a prime example of a globally lauded robot-loving culture.Research on attitudes toward machines in Asianot just Japanhas also reported positive reactions.For instance, Bigman, Yam, Marciano, Reynolds, and Gray (2021) found that participants in Singapore reported an increased preference for algorithm decision-making in healthcare settings when the topic of inequality in healthcare access was highlighted to them.Likewise, Oh et al. (2019) found that Korean doctors and medical students had generally positive attitudes regarding the use of AI in the healthcare sector, largely appreciating its capacity to analyze big amounts of clinical data in a short amount of time.Similarly, in China, a content analysis of social media posts found that most of the general public regard medical AI positively, believing that such technology has the potential to partially or even fully replace human doctors (Gao, He, Chen, Li, & Lai, 2020).A recent field experiment shows that workers in China perceived algorithmic task assignment as fairer than human-based task assignment (Bai, Dai, Zhang, Zhang, & Hu, 2022).
But beyond commonplace assumptions, when looking at cross-cultural research specifically, the findings are more nuanced.Though some studies have reported more positive reactions to machines among people from the East than their counterparts from the West (e.g., Li, Rau, & Li, 2010), others have found the opposite (e.g., Bartneck, Suzuki, Kanda, & Nomura, 2007).Lim et al. (2021Lim et al. ( : 1307) ) recently reviewed the past two decades of research on human-robot interaction (HRI) focusing on cultural influences and noted 'complex and intricate relationships between culture and human cognition', while discussing how national culture and prior experiences with robots affect human-robot interactions.
There may be no easy answer to the question, 'Do Eastern cultures like and accept machines more than Western cultures'?Cultures are multifaceted and dynamic and may affect a wide range of attitudes and behaviors toward machines in diverse yet interactive ways.Rather than determining a simplistic 'yes' or 'no' answer, in this article we aim to further dissect and understand the differences in how the East and West think about and interact with machines at large.There is a need to go beyond the basic assumption that the East just has more affinity for technology and explore the specific differences and underlying mechanisms.Furthermore, most social science research on novel technologies has been conducted using samples from the Westspecifically, samples from Western, educated, industrialized, rich, and democratic (WEIRD) societies (Henrich, Heine, & Norenzayan, 2010).Ironically, as previously discussed, many of these technologies are primarily being deployed and used outside of the West.Thus, it is no longer sufficient to rely on the findings situated in WEIRD societies to fully understand people's attitudes and behaviors toward machines, without considering the significance of culture in shaping our everyday thoughts and behaviors.The time is ripe for us to review the present state of knowledge about cultural attitudes and behaviors toward machines and identify how and why the East and West may differ in this regard (see Table 1 for a summary).

Historical and Religious Perspectives
The different historical legacies of Eastern and Western cultures may shed light on how these cultural traditions perceive machines today.Eastern and Western cultural traditions ultimately stem from the same deep historical roots (Bouckaert et al., 2022;Lee, Han, Byron, & Fan, 2008).Historical and ethnographic studies have found that early societies in both regions may have held widespread belief in animism 4 , in which many non-human agents or spiritual forces animated the natural world (Jackson, Dillion, et al., 2023;Peoples, Duda, & Marlowe, 2016).However, the religious and philosophical traditions in the East and West have diverged in key ways over the last 3000 years.Many Eastern religious traditions such as Buddhism and Shintoism continue to emphasize the co-existence of humans and animistic agents or forces.In contrast, Western religious and philosophical traditions have diverged to emphasize 'human exceptionalism', defined as the perception of human beings having distinct

Religion and History
Animism (Jackson, Dillion, et al., 2023;Peoples et al., 2016) linked to Shintoism and Buddhism Human exceptionalism (Srinivasan & Kasturirangan, 2016), anthropocentrism (Daliot-Bul, 2019)linked to Christianity Government Japan's government encourages robot and technology adoption to help manage decreasing labor force (due to an ageing and declining population) (Wagner, 2009) Japanese people are more exposed to robots in the media (Haring, Mougenot, et al., 2014;MacDorman et al., 2009) People in the US and Europe have less exposure to robots in the media (Haring, Mougenot, et al., 2014;MacDorman et al., 2009) Japan has many positive and helpful depictions of robots in the fictional media (e.g., Astro Boy) Hollywood presents many catastrophic depictions of robots in the fictional media (e.g., Terminator) News reports in China and Japan use fewer negative frames reporting social robots (De Boer et al., 2021) News reports in the US use more negative frames reporting social robots (De Boer et al., 2021) Real-life Exposure Japanese people have more exposure to robots in real life than US participants (MacDorman et al., 2009) Less exposure in the US compared to Japan (MacDorman et al., 2009) However, Haring, Mougenot, et al. (2014) found that Japanese people had less personal contact with robots than European people Haring, Mougenot, et al. (2014) found that European people had more contact with robots than Japanese people minds, rights, and capabilities which are not shared by non-human animals (Srinivasan & Kasturirangan, 2016).Belief in other non-human agents such as spirits or disembodied forces has also declined in Western cultures, contributing further to the view that humans have unique abilities, rights, and privileges (Jackson, Caluori, Gray, & Gelfand, 2021).We argue that this difference in how both cultures perceive and understand non-human entities shapes fundamentally divergent relationships with machinesfor the East, machines can be part of the natural world as much as other life is; they complement it.For the West, machines are Others, too inherently different from humans.As such, machines are more accepted in Eastern cultures as people view them in harmony with themselves, whereas in Western cultures, they are seen more as dissimilar, unfamiliar aliens, possibly posing a threat to people's identities and societies.
The legacy of animistic beliefs in current-day human-machine interaction is particularly visible in Japan.Broadly, in Japanese culture, non-human entities are frequently thought of as having souls or spirits, not unlike humans.Even entities that are not human or even alive are perceived to possess life in their own right.These beliefs can be traced to the Shinto religion which assumes that kami, or divinity, lives in all parts of nature, such as the seas, mountains, and flowers (Asquith & Kalland, 1996).Notably, Shinto is the largest religion in Japan in practice, with over 80% of the Japanese population participating in some Shinto practices (Breen & Teeuwen, 2010).As such, it is not uncommon there to practice rites honoring or respecting non-human entities, like funeral ceremonies and offerings made for pets after they die (Kenney, 2004), or the famous KonMari method in which homeowners are asked to greet their houses and belongings upon entering and converse with them as if they were conscious, living beings (Kahn, 2021).The idea that spirits exist within non-humans is not just accepted but embraced in Shintoism.As Geraci (2006: 236) summarizes, 'robots fit into the natural world [of Japan] as easily as any other object'.Despite not being human, machines are still easily accepted as a part of the natural world, complementing what is already there.
In Western cultures, on the contrary, the Judeo-Christian worldview emphasizes human uniqueness and importance (MacDorman, Vasudevan, & Ho, 2009).Medieval Christianity, for instance, believed in a natural order or hierarchy in the world, known as the 'great chain of being', which has historically dominated Western thought (Nee, 2005).Accordingly, the Universe is ranked linearly, starting with rocks, then plants, animals, humans, angels, and finally, God.Notably, humans are valued over inanimate entities by the order.In a similar vein, Westerners do not typically view non-human entities as possessing agency or moral concerns the same way humans do (Laham, 2009).Though Medieval Christianity may no longer hold the same influence today that it did in the past, scholars argue that the human exceptionalism worldview in the West has only evolved into anthropocentrism (Daliot-Bul, 2019).
Anthropocentrism is the broad belief that humans are the only entities with intrinsic value, whereas the value of all other entities lies only in their instrumental values to serve humans (Goralnik & Nelson, 2012).Unlike human exceptionalism, anthropocentrism orders and arranges the world around humans rather than God.Though distinct in philosophies, anthropocentrism is similar to human exceptionalism in that it continues to hold up the assumption that humans are distinct and important in fact, now, they are the center.Be it either, the dominant Western thought distinguishes neatly between humans and non-humans.But why would this create machine aversion?First, people categorize others into in-groups versus out-groups, in which in-group members share a social identity and out-group members do not.People behave more favorably toward their in-groups while discriminating against out-groups (Tajfel, Billig, Bundy, & Flament, 1971).By human exceptionalism and/or anthropocentrism beliefs, the strict distinction between humans and non-humans likely exacerbates the social categorization of machines as an out-group, evoking unfavorable evaluations and reactions.Second, given the basic assumptions of these philosophies that humans are superior to all other entities (other than the divine), the rise of machines may be perceived as a threat (Floridi, 2014).Humans have always been positioned near the top or at the center of the orderusually assumed to be the most intelligent and exceptional (Finlay & Workman, 2013;Roth & Dicke, 2005).As such, the expanding capabilities and presence of machines in recent times may be seen as a threat of replacement to humans, both in terms of realistic threat (i.e., people's jobs and livelihood are on the line) and symbolic threat (i.e., people's identity as unique and superior entities are endangered) (Gray et al., 2023), contributing to the unease people feel about co-existing among machines.Hence, while animism in the East may promote a complementary and harmonic relationship between humans and machines by highlighting a commonality (that we all have a soul and a place in the world), Western human exceptionalism and anthropocentrism seem to emphasize differences, presenting a more competitive relationship.
But how relevant are these historical and religious perspectives to the human-machine relationship today?Societies have become more secular over recent decades (Jackson et al., 2021), but the history of places and cultures can continue to reflect in people's behaviors today.To illustrate, Talhelm et al. (2014) demonstrated that the agricultural histories of different regions in China (e.g., farming rice vs. farming wheat) which shapes community networks (e.g., interdependent vs. independent) continue to affect people's thinking styles (e.g., holistic vs. analytical) in the modern day.As such, even if these traditional beliefs are no longer as subscribed to today as before, it is nevertheless likely that such opposing historical backdrops of the East and the West have played a significant role in cultivating the different cultural attitudes and behaviors toward machines observed today.But of course, the past alone cannot paint the full story, and in the next section, we discuss a more contemporary perspective that shapes cross-cultural perceptions of machines.
The Role of Exposure Exposure to machines in one's everyday life can influence people's attitudes and behavior toward machines in important ways.Here, we discuss two modes of exposure to machines -(a) through the media and (b) through real-life experiencezooming in on how they differ across the East and the West, and how such different qualities and quantities of exposure to machines can shape culturally divergent attitudes of them.

Cultural representations of technology
First, how a culture represents machines through its artifacts (e.g., toys, the media) can shape people's judgments and preferences for them.This draws from the information processing approach (Entman, 1989), which proposes that salient information about a target object from the media has the power to influence people's pre-existing schemas (i.e., mental representations) of it.Indeed, the media can serve as a means by which cultural values are transmitted.How machines are represented in the mass media (e.g., kind vs. evil) in each culture thus tells people what and how to think of them.
In Japan, the government 5 has long promoted the advancement of social robotics on a national scale to address the manpower needs of a rapidly ageing and declining population (Wagner, 2009).Supporting this push, the Japanese mass media has also heavily promoted robot and automation acceptance, through positive narratives of robots or androids working alongside humans for the benefit of humanity (e.g., Astro Boy, Power Rangers, Doraemon).Such positive depictions of machines tie in withand are likely fostered bythe animistic beliefs of the East that support the idea that machines can co-exist in peace and harmony with humans.Doraemon, for instance, features the stories and adventures of an earless robotic cat that lives happily with the Nobi family, often coming to the rescue of their young son in times of trouble.As such, the concept of robotic help is neither unfamiliar nor jarring; it is actively encouraged.The positive depictions of robots by the government and mass media as helpers and a potential solution to managing a societal problem communicate and reinforce the idea that the appropriate or 'correct' attitude toward machines is acceptance because they are favorable for them and society.Therefore, it is unsurprising that Japanese people may exhibit more preference and intention to use machines.
In contrast, media depictions of technology in the West tend to be a lot more mixedwith some downright negative and threatening.Hollywood blockbuster Terminator (1984), for instance, stars a cyborg assassin antagonist, as well as a threat of an AI-activated nuclear holocaust (Eoin, 2020).Such narratives that machines may well outsmart and overtake humans one day are common in the Western media and contribute to a larger, overarching cultural view of technology as a threatening entity that endangers us.Such narratives are also likely underpinned by the Western philosophy of human exceptionalismthey send a message that machines compete with humans for power and dominance and count on audiences hoping for the humans to ultimately win.Considering the large amount of media discussion about the threats of machines replacing (or displacing) human employees, this leads Westerners to learn to perceive machines as a threat and fear or dislike them.This is in line with Gerbner and Gross (1976) cultivation theory which posits that messages and images depicted in the media 'act like the pull of gravity toward an imagined center' over time (Judge & Cable, 2011: 96).Furthermore, negative events or information have asymmetrically strong effects on the human psyche (compared to positive ones) (Baumeister, Bratslavsky, Finkenauer, & Vohs, 2001), meaning the impact of such negative machine portrayal on people's attitudes likely supersedes the influence of other positive depictions.Given the contrasting dominant narratives both cultures have of machines, it is no surprise that the West might be more ambivalent toward machines than their Eastern counterparts (De Boer, Jansen, Bustos, Prinse, Horwitz, & Hoorn, 2021).

Technology exposure
Second, having real-life experiences of interacting with machines can also affect people's attitudes toward machines.Researchers usually hypothesize that more prior machine exposure translates to more positive attitudes, presumably based on ideas of the mere-exposure hypothesis (that familiarity leads to liking) (Harrison, 1977), or that exposure reduces uncertainty and anxiety toward machines (Bartneck et al., 2007).Indeed, some researchers have demonstrated that having previous interactions or real-life experiences seeing robots in action is correlated with lesser negative attitudes toward them.For example, Nomura, Suzuki, Kanda, and Kato (2006) found that Japanese students who have seen real robots live before had less negative attitudes toward them than those who had not.In another study, Bartneck et al. (2007) found that participants who had directly interacted with Aibo, a robot dog, before rated lower on negative attitudes toward robots than the participants who had not.
Given that the robot density in Asia trumps that of Europe and the Americas (IFR, 2021b), it could be expected that the East affords more opportunities for people to experience and interact with robots than the West, contributing to differential attitudes toward them.This line of reasoning has been explored by Li et al. (2010) who found that German participants liked, trusted, and engaged with social robots less than Chinese and Korean people, proposedly due to Germany's general lack of exposure to social robots (though they are more exposed to industrial robots such as welding robots and robotic arms).In contrast, social robots are presumably more commonplace and utilized in Korea (a culture which prefers 'small and slow', in line with the appearance and mobility of social robots) than in Germany (a culture which prefers 'big and fast', more in line with the features of industrial robots).Separately, Han, Hyun, Kim, Cho, Kanda, and Nomura (2009) found that parents in Spain expressed the most negativity about the use of educator robots for their children compared to parents from Korea and Japan.Korean parents were the least resistant to using educator robots, correlating with the existing prevalence of e-learning in Korea already.These studies suggest that different cultures afford different levels of familiarity with machinesand such differences relate to less negativity and aversion toward machines in the East than in the West.

Further unpacking technology exposure
Although extant research has primarily supported our proposed cultural perspective of humanmachine interactions, there are also exceptions.For example, while Li et al. (2010) and Han et al. (2009) reported fewer negative attitudes and machine aversion in Eastern cultures than in Western ones, a handful of studies have found that attitudes across cultures are more similar than different.For example, MacDorman et al. (2009) found that both Japanese and US participants self-reported preferring humans over robots and implicitly associated robots with weapons more than humans (albeit slightly more for the US than Japan), despite Japanese participants having had more exposure to robot-related content than participants from the US through both the media and real-life interactions.Similarly, Haring, Mougenot, Ono, and Watanabe (2014) found that both Japanese and European participants reported similar assumptions and attitudes toward robots, neither being more positive than the other, as well as a similar level of fear.They also found that Japanese participants only had higher robot exposure compared to European participants through the media (e.g., TV, manga), but less in terms of personal contact.Surprisingly, despite both these studies finding some sort of increased exposure Japan has to robots than the US and Europe, neither found strong support that Japanese people had a stronger preference or more positive attitudes to robots than the participants from the West.A plausible explanation is that although interacting with machines in one's day-to-day life may provide much room for people to observe and learn about their value, the opposite is also true: people may also learn about machine failures, have bad experiences with them, or feel disappointed and disillusioned if the machines do not meet their expectations (Yam, Bigman, Tang, et al., 2021;Yam, Goh, Fehr, Lee, Soh, & Gray, 2022).As Bartneck et al. (2007) suggest, prior exposure to robots may not only have made Japanese participants more aware of their capabilities but also their limitations and weaknesses.
Given these somewhat mixed findings, it is difficult to conclude the true extent of differences in terms of the amount of technology exposure the East versus the West affords or how divergent (or convergent) attitudes toward machines are across both cultures.But what can be said is that the machine exposure discussed so far is only but one piece of the culture story.Yes, there are more robots in the East than in the West (IFR, 2021b)but does everyone in a culture get the same access to experience them?Other factors like age, gender, and occupation may also affect a person's likelihood of having had prior exposure to a machine, and depending on the demographic of the sample a study uses, the findings may or may not represent a cultural group's general level of exposure to machines.And if the East does afford more opportunities for interaction with machines than the West, are these interactions always positive?Do these interactions promote machine acceptance, or do they backfire and create machine aversion?More than how much exposure someone has had to machines in the past, the quality of the exposure needs to be examined, too.As such, future research can clarify the role of technology exposure on cultural attitudes toward machines by studying machine exposure not just as a measure of quantity, but quality, as well.

New Applications of Machines in Asia
Thus far, we have reviewed and proposed some plausible reasons as to how and why different cultures may diverge in their views of machines.Importantly, culture forms another piece of the puzzle to understanding the who of 'who likes machines, and who does not'?But to also return to the 'what' what do machines do that people like or dislike?here we review the benefits and drawbacks of three new and exciting types of machine applications primarily found in Asia.Our review suggests that Asians appear to be more receptive to interacting with machines in the social, moral, and spiritual domains than their Western counterparts.To begin, we suggest that this review is by no means exhaustive but aims to facilitate generative research beyond the traditional paradigm of machine aversion versus appreciation to explore how machines might fundamentally change human-to-human relationships in these unique contexts.

Sex Robots Change Romantic Relationships
The sex robot industry is booming, estimated to be worth $30 billion in 2019 and expected to double by 2026 (Williams, 2021; Figure 1A).Nagoya, Japan, for instance, is home to a 'hyper-realistic' robot brothel, where customers may pay ¥13000 (approximately US$ 100) for an hour's session with one of the brothel's four sex robots (Hicks, 2019).Meanwhile, in China, tech companies have launched sales of customizable, AI-powered sex dolls that can hold simple conversations with users and move their eyes and arms (Song, 2018).On the one hand, such developments might help alleviate loneliness and provide a source of intimacy for people with social anxiety.On the other hand, the surge of sex robots raises ethical challenges.Learning to compromise and think about other people is important for healthy psychological growth and maturity, and partners can provide a valuable source of social support.Seeking companionship with robots rather than humans might increase egoism or reduce human reliance on one another for social connection (Kiron & Unruh, 2018).It might also change the way people think about their romantic interests.If people can get sexual gratification from robots, it might cause them to objectify potential future partners, be they mechanical or flesh and blood.
Clergy Robots Change People's Relationship with the Divine and Religious Leaders For millennia, religious groups around the world have elevated people to elite roles such as shamans, priests, or medicine men.These religious elites have in turn been crucial in maintaining the credibility of the religious beliefs that they espouse (Henrich, 2009;Lanman & Buhrmester, 2017).However, as religious decline has accelerated in multiple world regions (Norris & Inglehart, 2011), some religious groups are using 'robot priests' to try to attract younger and more technologically savvy adherents.
The rise of robot priests has been especially pronounced in Japan, which is consistent with Japan's cultural legacy of elevating non-human agents to anthropocentric roles.For example, Japan's SoftBank Robotics is producing a new line of Pepper robot Buddhist monks which will lead funeral rites (Gibbs, 2017).A 400-year-old Japanese temple has taken this trend one step further by introducing a robot named Mindar to deliver Buddhist sermons (Figure 1B; see also Jackson, Yam, Tang, Liu, & Shariff, 2023).The rise of robots in East Asian religious settings contrasts with a deep aversion to religious robots in many Western cultural contexts.For example, the advent of large language models such as ChatGPT led to a wave of opposition to the possibility of developing automated sermons (Crary, 2023;Gerber, 2023).The 2017 introduction of a robot named 'BlessU-2' in a German Protestant church provoked a wave of media interest, but it had a fairly minor responsibility (reading blessings in different languages) and has not spread in the intervening years.
Time will tell whether robot preachers continue to be accepted in East Asian contexts.Given people's attraction to credible and charismatic religious figures (Lanman & Buhrmester, 2017;Sperber, 2010), robot preachers may fail to inspire the same kind of commitment as human preachers and may eventually drift out of style.But the fact that temples in Japan are adopting these automated agents in religious contexts suggests that they may have less aversion to machines making moral decisions than Westerners (Bigman & Gray, 2018).This may foreshadow an era where machines seep into moral and religious spheres of society in East Asian cultures.This trend could have implications for well-being and ethics.For example, people also see religious leaders as sources of community support during difficult circumstances, and this trust and support may be undermined when robots serve in these roles.Robot Therapists Shape Information Disclosure to Others A final area where robots are transforming jobs is psychotherapy.People are often reluctant to seek help because confiding in a mental health professional makes them vulnerable and is stigmatized in many cultures.In East Asian cultures, for instance, unrestrained expression of emotion is generally frowned upon and sometimes viewed as a threat to social harmony (Ng, 1997).Unsurprisingly, the fear of stigmatization relating to one's mental health can significantly delay people from seeking treatment (Subramaniam et al., 2020).But robot therapists offer a solution to this problem (Figure 1C).
For example, effective psychological therapies often require full disclosure of the patient's darkest fears and secrets, which can be difficult to achieve with human therapists as clients struggle with feelings of embarrassment or shame.Significantly, self-disclosure to a robot does not seem to evoke the same kind or extent of resistanceresearch has found that people engage in more self-disclosure, particularly on negative topics, interacting with robot therapists relative to human therapists (Takahashi, Takahashi, Ban, Shimaya, Yoshikawa, & Ishiguro, 2017).This may be why people have been confiding in and seeking mental health advice from ChatGPT, a popular AI language model chatbot released in late 2022 (Broderick, 2023).Interestingly, ChatGPT was not designed or intended as a therapy chatbot.Yet, chatbots like these can promise anonymity and non-judgment to people, possibly more so than professional human therapists, which may make them more appealing 'listeners' to some audiences.Such trends importantly suggest there is already some openness toor even a demand forrobot therapists among us.In Asia, chatbots designed specifically for counseling have been sprouting up over the past few years, such as Singapore's mindline.sg,launched by the government to support the community during the COVID-19 pandemic (Goh, 2020).It is equipped to guide users through meditative and breathing exercises, among other care techniques.In China, a similar AI chatbot designed to provide counseling services free-of-charge and around-the-clock, Xiaotian, is also in development (Xinhua, 2021).According to its creators, Xiaotian will be able to guide users through 50-minute-long conversations about their feelings and experiences and direct them to professional help resources if needed.
But despite the potential benefits robot therapists may bring, these developments also raise ethical concerns.In addition to the obvious issues of privacy and data security, there is some evidence suggesting that a therapeutic alliance between the patient and the therapist is necessary for effective psychotherapy (Horvath & Luborsky, 1993).With non-human therapists, this might be lost.Moreover, deep and personal disclosures with a robot therapist might ironically further reduce humans' desire to socially connect with other humans.Such robots might create a paradox for social connectedness.On the one hand, they allow people to receive counseling and access mental health resources; but on the other, they can lead to further social alienation and ostracism because people who seek help from robots would no longer feel the need or want to share their deepest disclosures with other people.

Future Research
Having proposed a cross-cultural perspective and discussed unique machine applications in Asia, we now turn to future research that we believe will be fruitful.We organize this section into three board categoriesunexamined populations and underlying mechanisms, boundary conditions within and across cultures, and the paradox of technology development versus adoption and deployment.

Unexamined Populations and Underlying Mechanisms
The examples and research cited pertaining to the East throughout this article have primarily been based on developed economies in Asia such as China, Singapore, Japan, and South Korea.This largely corresponds to the top three countries in terms of robot density -South Korea, Singapore, and Japan (IFR, 2021b).China has also seen tremendous growth in robot applicationgrowth data suggest that robot density per capita grew by over 400% in China between 2015 and 2020 (compared to a less than 45% growth in the US).Still, we call for more research to examine human-machine interactions in other parts of Asia, particularly developing economies in Southeast and South Asia.Approximately one-third of the world's population resides in this region; yet there is little research on how South and Southeast Asians perceive new technologies and machines, a problem noted in many areas of behavioral science research (e.g., Bernardo, Mateo, & Dela Cruz, 2022).That said, we have reasons to believe that some of the aforementioned cross-cultural differences might replicate in these regions.For example, in one rare study, Yam et al. (2023) examined how engineers from India reacted to robots and the results were largely consistent with samples from Singapore.
Although these cross-cultural differences in how people perceive new technologies may replicate, their underlying mechanisms might differ.Throughout this article, we have proposed a framework to explain why there lie cross-cultural differences in machine aversion versus appreciation, but most of these mechanisms remained untested.We also urge future research to unpack further cultural mechanisms undiscussed in this article.For example, cognitive differences between cultures may also explain why Asians are more receptive to machines than Westerners.Possibly, as Asians are more likely to be holistic thinkers (Markus & Kitayama, 1991) and relationship-focused (Qin et al., 2023), they are more likely to perceive machines in terms of their relationships with others.In other words, machines can be thought of as helpful and/or harmful depending on the targeted relationships (e.g., beneficial to consumers but harmful to some employees).On the other hand, Westerners are more likely to be analytical thinkers, perceiving machines in isolation, and hence, as potential threats.As another example, much of the extant research has examined cultural differences in trust toward machines (e.g., Haring, Silvera-Tawil, Matsumoto, Velonaki, & Watanabe, 2014;Wang, Rau, Evers, Robinson, & Hinds, 2010), but this line of work has not examined more nuanced mechanisms such as cognitive versus affective trust (McAllister, 1995).Notably, the latter has been found to be especially important in Asian contexts (Chen, Eberly, Chiang, Farh, & Cheng, 2014), and likely more difficult for machines to cultivate.

Boundary Conditions within and Across Cultures
Although we have argued that Asians generally react more positively to machines than their counterparts in the West, there are important moderators to be considered within cultures.For example, we suggest an interesting industry moderating effect within culturesalthough Asians are generally more accepting of robots than Westerners, they may reject the deployment of robotic caregivers in nursing homes.This is because using robots to take up the responsibilities and tasks of caring for one's aged parents may be at odds with the Confucianist concept of filial piety (Ng, Lee, & Wu, 2022).Conversely, the use of robots for surveillance purposes may be better accepted in the East than in the West, as Western cultures tend to value individual rights to privacy and freedom more strongly than Eastern cultures do.A closely related illustration of this is the pervasive uptake and embrace of digital tracking tools in East Asian nations during the COVID-19 pandemic to support contract tracing compared to Western countries (Cha, 2020).
In addition, there are broader moderators that would be applicable across cultures.One such moderator is age or generational difference.Mahmud, Islam, Ahmed, and Smolander (2022)'s review of algorithm aversion identified that older people find algorithms less useful and trustworthy.However, it also reported some conflicting findings with regard to older people's preference for algorithms over humans - Thurman, Moeller, Helberger, and Trilling (2019) found that older people preferred human news recommenders over algorithmic ones ;Ho, Wheatley, and Scialfa (2005) found that older people depended more on an algorithm to perform a medication management task than younger people did.Looking at robots, some studies have also found that older adults feel more fearful and anxious about having robots at home (Scopelliti, Giuliani, & Fornara, 2005), and tend to hold more negative implicit attitudes toward robots than younger people (Chien et al., 2019).All in all, age is likely a relevant but incomplete piece to understanding algorithm aversion.Older people are indeed less likely to be familiar with new technologies and machines which may account for more anxiety and distrust of them.But, on the other hand, older people may also find that machines are more valuable and useful to them (than younger people would) for certain tasks that are too demanding or difficult for them (e.g., in Ho et al.'s (2005) study, older people relied more on the algorithm because they had less confidence in themselves to perform the task).In sum, we encourage future work to seriously examine age or generational difference, and technology familiarity/affinity as moderators because these factors might trump any cross-cultural differences identified in this article.
The Paradox of Technology Development Versus Adoption and Deployment Finally, our review reveals one interesting paradoxmost new AI innovations are happening in the West and yet, their widespread application and deployment are often found in the East.One possible reason for the slow uptake of AI and machines in real-world settings in the West is the lengthy review and approval processes by regulatory authorities.Interestingly, global surveys have found that concerns about the harms of AI decision-making run highest in regions like Latin America and North America, whereas Southeast Asia and East Asia reported much lower levels of such worries (Neudert, Knuutila, & Howard, 2020).These attitudes of concern and apprehension may reflect in the different societies' relative willingness or resistance to deploying newly innovated technologies, explaining some of the West's relatively slow deployment processes for new technologies, despite the thriving state of innovation there.

Conclusion
In this perspective article, we have provided a brief review of the recent machine aversion versus appreciation debate through a cultural lens, from the perspectives of the historical, religious, and the role of machine exposure.Additionally, we discussed three unique applications of machines found primarily in Asiain the domains of sex, religion, and mental healththat could generate exciting future research beyond the traditional paradigm of just machine aversion versus appreciation.Given the extensive and exciting developments of technology in both the East and the West, the time is ripe for scholarly, cross-cultural collaboration to realize the full potential of human-machine interaction in both the East and the West.

Figure 1 .
Figure 1.Benefits and ethical threats of robots in sex, religion, and mental health.Note: The left panel (A) shows Henry, a sexbot, sold online.From Meet Henry, the Male Sex Robot With Artificial Intelligence and a British Accent [Photograph], by Realbotix, 2018, Allure (https://www.allure.com/story/realbotix-henry-male-sex-robot-with-artificial-intelligence).The middle panel (B) shows Mindar, a robot priest introduced in a 400-year-old temple in Kyoto, Japan.From Kyoto Temple Enlists Android Buddhist Deity To Help People [Photograph], by The Asahi Shimbun, 2019, Getty Images (https://www.gettyimages.com/detail/news-photo/android-kannon-named-minder-is-displayed-at-kodaiji-temple-news-photo/1131988643).The right panel (C) shows NAO, a 60-cm robot piloted in Singapore to engage children with autism in social interactions.From NAO Robot Aims to Help Kids with Autism Become More Social [Photograph], by Nuria Ling, 2013, The Straits Times (https://www.straitstimes.com/singapore/nao-robot-aims-to-help-kids-with-autism-become-more-social).

Table 1 .
Summary of factors proposed to affect East-West differences in people's attitudes and behaviors toward new technologies