Skip to main content

2024 | Buch

Chatbot Research and Design

7th International Workshop, CONVERSATIONS 2023, Oslo, Norway, November 22–23, 2023, Revised Selected Papers

herausgegeben von: Asbjørn Følstad, Theo Araujo, Symeon Papadopoulos, Effie L.-C. Law, Ewa Luger, Morten Goodwin, Sebastian Hobert, Petter Bae Brandtzaeg

Verlag: Springer Nature Switzerland

Buchreihe : Lecture Notes in Computer Science

insite
SUCHEN

Über dieses Buch

This book constitutes the proceedings of the 7th International Workshop on Chatbot Research and Design, CONVERSATIONS 2023, which was held during November 2023.

The 12 regular papers were carefully reviewed and selected for inclusion in the book. They were organized in following topical sections:

Understanding and Enhancing Conversational Interactions, LLM-driven Conversational Design and Analysis, Ethical Perspectives and Bias, Complementing Perspectives.

Inhaltsverzeichnis

Frontmatter

Understanding and Enhancing Conversational Interactions

Frontmatter
Voting Assistant Chatbot for Increasing Voter Turnout at Local Elections: An Exploratory Study
Abstract
Local Dutch elections suffer from a voter turnout decrease with 23% between 1986 and 2022. This raises general quality concerns about the local democracy. A Conversational Agent Voting Advice Application (CAVAA) chatbot was developed to aid in solving this problem. The objective of the study was to analyze the effect of the CAVAA chatbot on the perceived political knowledge and intention to vote in municipal elections.
An experimental within-subject study (N = 95) in the run-up to the Dutch municipal elections of March 16, 2022 among a typical Dutch municipality of 27,000 inhabitants was conducted. The CAVAA chatbot asked the participant’s opinion on 20 local issues, provided further background information if desired in a dialogue, and concluded the interaction with a voting advice.
Participants reported a greater understanding of local political issues after CAVAA use. However, no effect was found on the intention to vote.
The study therefore suggests that a voting assistant chatbot may contribute to a higher voter turnout, but the results should be interpreted with caution and further research is desirable.
Nina van Zanten, Roel Boumans
Conversational Repair Strategies to Cope with Errors and Breakdowns in Customer Service Chatbot Conversations
Abstract
This study aimed to investigate (1) what errors and conversational repair strategies appear during conversations with a real-life customer service chatbot and (2) how people perceive these errors and repair strategies in terms of user satisfaction, brand attitude, and trust. This study involved a corpus study of real-life conversations (N=100) with a customer service chatbot to investigate which errors and repairs occurred to inform a follow-up online experiment (N=150) on the perception of these errors and repairs. The experiment employed a 3 (error; excess of information, unsolvable question, lack of information) by 3 (repair strategy; repeat, options, defer) mixed subject design with the type of error as between-subjects factor and repair strategy as within-subjects factor. The results revealed that the repair strategy defer most positively impacted perceptions of trust and brand attitude, followed by the strategy options, and lastly repeat. In contrast, no significant main effects of error type nor interaction effects were found on user satisfaction, trust, and brand attitude. However, the open-ended questions revealed that there might be a connection between the nature of the customer request and the repair strategy.
Anouck Braggaar, Jasmin Verhagen, Gabriëlla Martijn, Christine Liebrecht
Aragón Open Data Assistant, Lesson Learned of an Intelligent Assistant for Open Data Access
Abstract
Chatbots are becoming more popular on websites. To ensure their widespread adoption and effectiveness, it is crucial that the development of these assistant technologies prioritizes user experience, integrating advanced computational methods without losing the human-centric perspective. This paper provides a comprehensive analysis of the insights obtained from the Aragón Intelligent Assistant Project, highlighting the main key lessons from deploying a chatbot dedicated to facilitating accessibility to open data for the regional government of Aragón. This article presents the difficulties and obstacles faced to meet the needs of real users while modern natural language processing technologies are being incorporated. The discussion underscores that, notwithstanding the sophistication of artificial intelligence, the user experience should be prioritized through ongoing evaluation and improvement. Chatbots must be continually tunned to align with human interaction paradigms if they are used to be as valuable tools for citizens.
Rafael del Hoyo-Alonso, Vega Rodrigalvarez-Chamarro, Jorge Vea-Murgía, Iñigo Zubizarreta, Julián Moyano-Collado

LLM-driven Conversational Design and Analysis

Frontmatter
Saleshat: A LLM-Based Social Robot for Human-Like Sales Conversations
Abstract
Large language models (LLMs) have generated excitement in many areas and may also make human-like conversations with social robots possible. Drawing from human-robot interaction literature and interviews, we developed Saleshat based on the commercial social robot Furhat and the large language model GPT-4. Saleshat emphasizes refined natural language processing and dynamic control of the robot’s physical appearance through the LLM. Responses from the LLM are processed sequentially, enabling the robot to react quickly. The results of our first formative evaluation with six users engaging in a sales conversation about Bluetooth speakers show that Saleshat can provide accurate and detailed responses, maintain a good conversation flow, and show dynamically controlled non-verbal cues. With our findings, we contribute to research on social robots and LLMs by providing design knowledge for LLM-based social robots and by uncovering the benefits and challenges of integrating LLMs into a social robot.
Leon Hanschmann, Ulrich Gnewuch, Alexander Maedche
Leveraging Large Language Models as Simulated Users for Initial, Low-Cost Evaluations of Designed Conversations
Abstract
In this paper, we explore the use of large language models, in this case the ChatGPT API, as simulated users to evaluate designed, rule-based conversations. This type of evaluation can be introduced as a low-cost method to identify common usability issues prior to testing conversational agents with actual users. Preliminary findings show that ChatGPT is good at playing the part of a user, providing realistic testing scenarios for designed conversations even if these involve certain background knowledge or context. GPT-4 shows vast improvements over ChatGPT (3.5). In future work, it is important to evaluate the performance of simulated users in a more structured, generalizable manner, for example by comparing their behavior to that of actual users. In addition, ways to fine-tune the LLM could be explored to improve its performance, and the output of simulated conversations could be analyzed to automatically derive usability metrics such as the number of turns needed to reach the goal. Finally, the use of simulated users with open-ended conversational agents could be explored, where the LLM may also be able to reflect on the user experience of the conversation.
Jan de Wit
Examining Lexical Alignment in Human-Agent Conversations with GPT-3.5 and GPT-4 Models
Abstract
This study employs a quantitative approach to investigate lexical alignment in human-agent interactions involving GPT-3.5 and GPT-4 language models. The research examines alignment performances across different conversational contexts and compares the performance of the two models. The findings highlight the significant improvements in GPT-4’s ability to foster lexical alignment, and the influence of conversation topics on alignment patterns.
Boxuan Wang, Mariët Theune, Sumit Srivastava

Ethical Perspectives and Bias

Frontmatter
In Search of Dark Patterns in Chatbots
Abstract
While Dark Patterns are widely present in graphical user interfaces, in this research we set out to find out whether they are also starting to appear in Chatbots. Dark Patterns are intentionally deceptive designs that trick users into acting contrary to their intention - and in favor of the organization that implements them. Chatbots, as a kind of conversational user interface, can potentially also suffer from Dark Patterns or other poor interaction design, sometimes referred to as Usability Smells. This keeps users from easily achieving their goals and can lead to frustration or limitations for users. To find Dark Patterns and Usability Smells, we analyzed user reports of negative experiences. Since we found no well known dataset of reports, we created the ChIPS dataset with 69 complaints from different web sources, and then classified them as one of 16 established Dark Patterns, potential new Dark Patterns, Usability Smells, or neither. Results show that, even though there are instances of established Dark Patterns, negative experiences usually are caused by chatbot defects, high expectations from users, or non-intuitive interactions.
Verena Traubinger, Sebastian Heil, Julián Grigera, Alejandra Garrido, Martin Gaedke
Language Ideology Bias in Conversational Technology
Abstract
The beliefs that we have about language are called language ideologies and influence how we create and use language technologies. In this paper, we explore language ideologies and their role in the process of language technology design using conversational technology as an illustrative example. We draw on two qualitative studies, both of which aim at discovering common language conceptualisations in the context of language technology design through collaborative work with study participants. In study 1, we use a survey, group discussions and co-design methods with technology developers. In study 2, we use a survey and group work with technology users. We found that standard language ideology is intertwined with a referential (language in its function to convey information) view on language data in the development process, and that a conceptualization of language as referential tool dominates the language technology landscape. However, participants in both qualitative studies are aware of other functions of language. Further we found that language ideologies are intertwined with public discourse about language technology, and upcoming policies on AI regulation will reinforce these ideologies. We argue that non-referential functions of language must be integrated into language models, and that the actual practices of both language and language technologies must be carefully considered for improved conversational AI and effective policies.
Sviatlana Höhn, Bettina Migge, Doris Dippold, Britta Schneider, Sjouke Mauw
Should Conversational Agents Care About Our Gender Identity?
Abstract
Chatbots are increasing their relevance in the global market. Nonetheless, companies are still struggling to develop chatbots that provide their clients with an optimal experience and, so far, few insights have been obtained to improve their related User Experience (UX). This study investigates whether chatbots that consider users’ gender identity result in an improved UX, and whether sensitivity towards this social matter moderates this relationship. Therefore, a one-factor within-subjects experiment was conducted, involving participants interacting with two versions of a buying-assistant chatbot. In one condition, the chatbot provided a more personalised interaction by presuming the user’s gender identity by their sex assigned at birth and conversing with them using a gender-specific language (e.g., ‘women’s clothing’, ‘men’s clothing’). The second condition, instead, used a gender-neutral approach, using gender-neutral language (e.g., ‘clothing’). We hypothesised that the chatbot presuming a cisgender identity of the user would result in a higher UX than the gender-neutral chatbot, and that this effect would be more substantial for users who score low on gender sensitivity. UX was measured by evaluating the chatbot’s Usability, Empathy and Supportiveness. Results indicate that the chatbot presuming a cisgender identity was considered significantly more usable and supportive, but less empathetic. A moderation effect of gender sensitivity on the evaluation of the chatbots was not found.
Arturo Cocchi, Tibor Bosse, Michelle van Pinxteren

Complementing Perspectives

Frontmatter
Conversational Interactions with NPCs in LLM-Driven Gaming: Guidelines from a Content Analysis of Player Feedback
Abstract
The growing capability and availability of large language models (LLMs) have led to their adoption in a number of domains. One application domain that could prove fruitful is to video games, where LLMs could be used to provide conversational responses from non-playable characters (NPCs) that are more dynamic and diverse. Additionally, LLMs could allow players the autonomy to converse in open-ended conversations potentially improving player immersion and agency. However, due to their recent commercial popularity, the consequences (both negative and positive) of using LLMs in video games from a player’s perspective is currently unclear. On from this, we analyse player feedback to the use of LLM-driven NPC responses in a commercially available video game. We discuss findings and implications, and generate guidelines for designers incorporating LLMs into NPC dialogue.
Samuel Rhys Cox, Wei Tsang Ooi
Exploring the Dark Corners of Human-Chatbot Interactions: A Literature Review on Conversational Agent Abuse
Abstract
Agent abuse is emerging as a significant concern in the realm of human-chatbot interactions. Despite the relevance of this phenomenon, from a social and psychological perspective, there have been relatively few published studies on the topic over the years. This calls for special attention and a need for a comprehensive understanding of the challenges posed by abusive behaviors towards conversational agents. Following the PRISMA protocol, this review intends to systematize the knowledge currently available in the scientific domain of chatbot abuse, identifying and evaluating research published between January 1989 and July 2023 across two databases. The review sheds light on the diverse range of studies that have contributed to defining and operationalizing chatbot abuse while exploring avenues for developing evidence-based interventions to discourage verbal mistreatment of conversational agents. By building on empirical, theoretical, and conceptual works, this research promotes awareness and consciousness-raising against chatbot mistreatment, advancing the scientific community's understanding of the complexities surrounding chatbot abuse and its possible implications. In doing so, the study fosters a more ethical, respectful, and empathetic approach toward conversational agents in the digital landscape, and considering the cross-cutting and cross-cultural nature of the issue, the author prompts the need for further empirical research on the topic.
Roberta De Cicco
Anticipating User Needs: Insights from Design Fiction on Conversational Agents for Computational Thinking
Abstract
Computational thinking, and by extension, computer programming, is notoriously challenging to learn. Conversational agents and generative artificial intelligence (genAI) have the potential to facilitate this learning process by offering personalized guidance, interactive learning experiences, and code generation. However, current genAI-based chatbots focus on professional developers and may not adequately consider educational needs. Involving educators in conceiving educational tools is critical for ensuring usefulness and usability. We enlisted nine instructors to engage in design fiction sessions in which we elicited abilities such a conversational agent supported by genAI should display. Participants envisioned a conversational agent that guides students stepwise through exercises, tuning its method of guidance with an awareness of the educational background, skills and deficits, and learning preferences. The insights obtained in this paper can guide future implementations of tutoring conversational agents oriented toward teaching computational thinking and computer programming.
Jacob Penney, João Felipe Pimentel, Igor Steinmacher, Marco A. Gerosa
Backmatter
Metadaten
Titel
Chatbot Research and Design
herausgegeben von
Asbjørn Følstad
Theo Araujo
Symeon Papadopoulos
Effie L.-C. Law
Ewa Luger
Morten Goodwin
Sebastian Hobert
Petter Bae Brandtzaeg
Copyright-Jahr
2024
Electronic ISBN
978-3-031-54975-5
Print ISBN
978-3-031-54974-8
DOI
https://doi.org/10.1007/978-3-031-54975-5

Premium Partner