Hide metadata

dc.date.accessioned2024-02-18T17:51:22Z
dc.date.available2024-02-18T17:51:22Z
dc.date.created2023-06-07T15:16:54Z
dc.date.issued2023
dc.identifier.citationKenfack, Patrik Ramírez Rivera, Adín Khan, Adil Mazzara, Manuel . Learning Fair Representations through Uniformly Distributed Sensitive Attributes. 2023 IEEE Conference on Secure and Trustworthy Machine Learning (SaTML). 2023 IEEE (Institute of Electrical and Electronics Engineers)
dc.identifier.urihttp://hdl.handle.net/10852/108244
dc.description.abstractMachine Learning (ML) models trained on biased data can reproduce and even amplify these biases. Since such models are deployed to make decisions that can affect people's lives, ensuring their fairness is critical. One approach to mitigate possible unfairness of ML models is to map the input data into a less-biased new space by means of training the model on fair representations. Several methods based on adversarial learning have been proposed to learn fair representation by fooling an adversary in predicting the sensitive attribute (e.g., gender or race). However, adversarial-based learning can be too difficult to optimize in practice; besides, it penalizes the utility of the representation. Hence, in this research effort we train bias-free representations from the input data by inducing a uniform distribution over the sensitive attributes in the latent space. In particular, we propose a probabilistic framework that learns these representations by enforcing the correct reconstruction of the original data, plus the prediction of the attributes of interest while eliminating the possibility of predicting the sensitive ones. Our method leverages the inability of Deep Neural Networks (DNNs) to generalize when trained on a noisy label space to regularize the latent space. We use a network head that predicts a noisy version of the sensitive attributes in order to increase the uncertainty of their predictions at test time. Our experiments in two datasets demonstrated that the proposed model significantly improves fairness while maintaining the prediction accuracy of downstream tasks.
dc.languageEN
dc.publisherIEEE (Institute of Electrical and Electronics Engineers)
dc.titleLearning Fair Representations through Uniformly Distributed Sensitive Attributes
dc.title.alternativeENEngelskEnglishLearning Fair Representations through Uniformly Distributed Sensitive Attributes
dc.typeChapter
dc.creator.authorKenfack, Patrik
dc.creator.authorRamírez Rivera, Adín
dc.creator.authorKhan, Adil
dc.creator.authorMazzara, Manuel
cristin.unitcode185,15,5,47
cristin.unitnameDigital signalbehandling og bildeanalyse
cristin.ispublishedtrue
cristin.fulltextpostprint
dc.identifier.cristin2152772
dc.identifier.bibliographiccitationinfo:ofi/fmt:kev:mtx:ctx&ctx_ver=Z39.88-2004&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.btitle=2023 IEEE Conference on Secure and Trustworthy Machine Learning (SaTML)&rft.spage=&rft.date=2023
dc.identifier.pagecount707
dc.identifier.doihttps://doi.org/10.1109/SaTML54575.2023.00014
dc.type.documentBokkapittel
dc.type.peerreviewedPeer reviewed
dc.source.isbn978-1-6654-6299-0
dc.type.versionAcceptedVersion
cristin.btitle2023 IEEE Conference on Secure and Trustworthy Machine Learning (SaTML)
dc.relation.projectNFR/309439


Files in this item

Appears in the following Collection

Hide metadata