Publication:
Implicit Stereotypes in Pre-Trained Classifiers

dc.contributor.authorNassim Dehoucheen_US
dc.contributor.otherMahidol Universityen_US
dc.date.accessioned2022-08-04T08:28:31Z
dc.date.available2022-08-04T08:28:31Z
dc.date.issued2021-01-01en_US
dc.description.abstractPre-trained deep learning models underpin many public-facing applications, and their propensity to reproduce implicit racial and gender stereotypes is an increasing source of concern. The risk of large-scale, unfair outcomes resulting from their use thus raises the need for technical tools to test and audit these systems. In this work, a dataset of 10,000 portrait photographs was generated and classified, using CLIP (Contrastive Language-Image Pretraining), according to six pairs of opposing labels describing a subject's gender, ethnicity, attractiveness, friendliness, wealth, and intelligence. Label correlation was analyzed and significant associations, corresponding to common implicit stereotypes in culture and society, were found at the 99% significance level. A strong positive correlation was notably found between labels Female and Attractive, Male and Rich, as well as White Person and Attractive. These results are used to highlight the risk of more innocuous labels being used as partial euphemisms for protected attributes. Moreover, some limitations of common definitions of algorithmic fairness as they apply to general-purpose, pre-trained systems are analyzed, and the idea of controlling for bias at the point of deployment of these systems rather than during data collection and training is put forward as a possible circumvention.en_US
dc.identifier.citationIEEE Access. Vol.9, (2021), 167936-167947en_US
dc.identifier.doi10.1109/ACCESS.2021.3136898en_US
dc.identifier.issn21693536en_US
dc.identifier.other2-s2.0-85122065960en_US
dc.identifier.urihttps://repository.li.mahidol.ac.th/handle/20.500.14594/76724
dc.rightsMahidol Universityen_US
dc.rights.holderSCOPUSen_US
dc.source.urihttps://www.scopus.com/inward/record.uri?partnerID=HzOxMe3b&scp=85122065960&origin=inwarden_US
dc.subjectComputer Scienceen_US
dc.subjectEngineeringen_US
dc.subjectMaterials Scienceen_US
dc.titleImplicit Stereotypes in Pre-Trained Classifiersen_US
dc.typeArticleen_US
dspace.entity.typePublication
mu.datasource.scopushttps://www.scopus.com/inward/record.uri?partnerID=HzOxMe3b&scp=85122065960&origin=inwarden_US

Files

Collections