Diversity and inclusion are critical aspects of the responsible development of artificial intelligence (AI) technologies, including affective computing. Affective computing, which focuses on recognizing, interpreting, and responding to human emotions, can revolutionize various domains, such as health care, education, and human-machine interaction. Capturing subjective states through technical means is challenging, though, and errors can occur, as seen with lie detectors not working adequately or gender classifier systems misgendering users.
If used for ulterior decision-making processes, such inferences could have disastrous consequences for people, the impacts of which may vary depending on the context of an application, i.e., flagging innocent people as potential criminals in border control or detrimentally affecting vulnerable groups in mental health care.
Following this line of thought, Tessa Verhoef from the Creative Intelligence Lab at Leiden University and Eduard Fosch-Villaronga from eLaw—Center for Law and Digital Technologies have written an article posted on the arXiv preprint server highlighting that systems trained on the datasets currently available and used most widely may not work equally well for everyone and will likely have racial biases, biases against users with (mental) health problems, and age biases because they derive from limited samples that do not fully represent societal diversity.
Verhoef and Fosch-Villaronga presented the paper entitled “Towards affective computing that works for everyone” at the conference Affective Computing + Intelligent Interaction (ACII ’23) that was held at the Massachusetts Institute of Technology (MIT) Media Lab Sept. 10–13. The annual Conference of the Association for the Advancement of Affective Computing (AAAC) is the premier international forum for research on affective and multimodal human-machine interaction and systems.
In their paper, they argue that missing diversity, equity, and inclusion elements in affective computing datasets directly affect the accuracy and fairness of emotion recognition algorithms across different groups.
The researchers conducted a literature review revealing how affective computing systems may work differently for different groups due to, for instance, mental health conditions impacting facial expressions and speech or age-related changes in facial appearance and health. To do so, they analyzed existing affective computing datasets and highlighted a disconcerting lack of diversity in current affective computing datasets regarding race, sex/gender, age, and (mental) health representation.
By emphasizing the need for more inclusive sampling strategies and standardized documentation of demographic factors in datasets, the researchers provide recommendations and call for greater attention to inclusivity and consideration of societal consequences in affective computing research to promote ethical and accurate outcomes in this emerging field.
Tessa Verhoef et al, Towards affective computing that works for everyone, arXiv (2023). DOI: 10.48550/arxiv.2309.10780
Toward affective computing that works for everyone (2023, September 25)
retrieved 25 September 2023
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.