The increasing availability of online data has meant that data-driven models have been applied to more and more tasks in recent years. In some domains and/or applications, such data must be protected before they are used. Hence, one of the problems only partially addressed in the literature is to determine how the performance of Machine Learning models is affected by data protection. More important, the explainability of the results of such models as a consequence of data protection has been even less investigated to date. In this paper, we refer to this very problem by considering non-perturbative data protection, and by studying the explainability of supervised models applied to the data classification task.
Locci, S., Di Caro, L., Livraga, G., Viviani, M. (2023). Explainability of the Effects of Non-Perturbative Data Protection in Supervised Classification. In Proceedings - 2023 22nd IEEE/WIC International Conference on Web Intelligence and Intelligent Agent Technology, WI-IAT 2023 (pp.402-408). IEEE [10.1109/wi-iat59888.2023.00066].
Explainability of the Effects of Non-Perturbative Data Protection in Supervised Classification
Viviani, M
2023
Abstract
The increasing availability of online data has meant that data-driven models have been applied to more and more tasks in recent years. In some domains and/or applications, such data must be protected before they are used. Hence, one of the problems only partially addressed in the literature is to determine how the performance of Machine Learning models is affected by data protection. More important, the explainability of the results of such models as a consequence of data protection has been even less investigated to date. In this paper, we refer to this very problem by considering non-perturbative data protection, and by studying the explainability of supervised models applied to the data classification task.I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.