أنتجت شركات التكنولوجيا ردود متنوعة على المخاوف بشأن آثار تصميم أنظمة AI المحادثة.ادعى البعض أن مساعديهم الصوتي هم في الواقع ليسوا جنسيين أو يشبه الجنسين --- على الرغم من ميزات التصميم التي تشير إلى العكس.قارنا هذه المطالبات بتصورات المستخدمين عن طريق تحليل الضمائر التي يستخدمونها عند الإشارة إلى مساعدين AI.ونحن ندرس أيضا ردود الأنظمة والمدى الذي يولدون الناتج الذي هو الجنس والجنس البشري.نجد أنه، في حين أن بعض الشركات تبدو مخاطبة الشواغل الأخلاقية التي أثيرت، في بعض الحالات، لا يبدو أن مطالباتهم تحمل حقيقة.على وجه الخصوص، تظهر نتائجنا أن مخرجات النظام غامضة فيما يتعلق بانبان النظم، وأن المستخدمين يميلون إلى تخصيصهم ونوع الجنس نتيجة لذلك.
Technology companies have produced varied responses to concerns about the effects of the design of their conversational AI systems. Some have claimed that their voice assistants are in fact not gendered or human-like---despite design features suggesting the contrary. We compare these claims to user perceptions by analysing the pronouns they use when referring to AI assistants. We also examine systems' responses and the extent to which they generate output which is gendered and anthropomorphic. We find that, while some companies appear to be addressing the ethical concerns raised, in some cases, their claims do not seem to hold true. In particular, our results show that system outputs are ambiguous as to the humanness of the systems, and that users tend to personify and gender them as a result.
References used
https://aclanthology.org/
In this paper, we investigate what types of stereotypical information are captured by pretrained language models. We present the first dataset comprising stereotypical attributes of a range of social groups and propose a method to elicit stereotypes
Style transfer aims to rewrite a source text in a different target style while preserving its content. We propose a novel approach to this task that leverages generic resources, and without using any task-specific parallel (source--target) data outpe
We present an error analysis of neural UPOS taggers to evaluate why using gold tags has such a large positive contribution to parsing performance while using predicted UPOS either harms performance or offers a negligible improvement. We also evaluate
Story generation is an open-ended and subjective task, which poses a challenge for evaluating story generation models. We present Choose Your Own Adventure, a collaborative writing setup for pairwise model evaluation. Two models generate suggestions
Natural language processing (NLP) tasks, ranging from text classification to text generation, have been revolutionised by the pretrained language models, such as BERT. This allows corporations to easily build powerful APIs by encapsulating fine-tuned