Show simple item record

dc.contributor.authorVadillo Jueguen, Jon
dc.contributor.authorSantana Hermida, Roberto ORCID
dc.date.accessioned2022-01-25T08:52:34Z
dc.date.available2022-01-25T08:52:34Z
dc.date.issued2022-01
dc.identifier.citationComputers & Security 112 : (2022) // Article ID 102495es_ES
dc.identifier.issn0167-4048
dc.identifier.issn1872-6208
dc.identifier.urihttp://hdl.handle.net/10810/55137
dc.description.abstract[EN] Human-machine interaction is increasingly dependent on speech communication, mainly due to the remarkable performance of Machine Learning models in speech recognition tasks. However, these models can be fooled by adversarial examples, which are inputs in-tentionally perturbed to produce a wrong prediction without the changes being noticeable to humans. While much research has focused on developing new techniques to generate adversarial perturbations, less attention has been given to aspects that determine whether and how the perturbations are noticed by humans. This question is relevant since high fool-ing rates of proposed adversarial perturbation strategies are only valuable if the perturba-tions are not detectable. In this paper we investigate to which extent the distortion metrics proposed in the literature for audio adversarial examples, and which are commonly applied to evaluate the effectiveness of methods for generating these attacks, are a reliable mea-sure of the human perception of the perturbations. Using an analytical framework, and an experiment in which 36 subjects evaluate audio adversarial examples according to different factors, we demonstrate that the metrics employed by convention are not a reliable measure of the perceptual similarity of adversarial examples in the audio domain.es_ES
dc.description.sponsorshipThis work was supported by the Basque Government (PRE_2019_1_0128 predoctoral grant, IT1244-19 and project KK-2020/00049 through the ELKARTEK program); the Spanish Ministry of Economy and Competitiveness MINECO (projects TIN2016-78365-R and PID2019-104966GB-I00); and the Spanish Ministry of Science, Innovation and Universities (FPU19/03231 predoctoral grant). The authors would also like to thank to the Intelligent Systems Group (University of the Basque Country UPV/EHU, Spain) for providing the computational resources needed to develop the project, as well as to all the participants that took part in the experiments.es_ES
dc.language.isoenges_ES
dc.publisherElsevieres_ES
dc.relationinfo:eu-repo/grantAgreement/MICINN/TIN2016-78365-Res_ES
dc.relationinfo:eu-repo/grantAgreement/MICINN/PID2019-104966GB-I00es_ES
dc.rightsinfo:eu-repo/semantics/openAccesses_ES
dc.rights.urihttp://creativecommons.org/licenses/by/3.0/es/*
dc.subjectadversarial exampleses_ES
dc.subjectdeep neural networkses_ES
dc.subjectspeech command classificationes_ES
dc.subjectrobust speech recognitiones_ES
dc.subjecthuman perceptiones_ES
dc.titleOn the human evaluation of universal audio adversarial perturbationses_ES
dc.typeinfo:eu-repo/semantics/articlees_ES
dc.rights.holder(c) 2021 The Authors. Published by Elsevier Ltd. This is an open access article under the CC BY license ( http://creativecommons.org/licenses/by/4.0/ )es_ES
dc.rights.holderAtribución 3.0 España*
dc.relation.publisherversionhttps://www.sciencedirect.com/science/article/pii/S0167404821003199?via%3Dihubes_ES
dc.identifier.doi10.1016/j.cose.2021.102495
dc.departamentoesCiencia de la computación e inteligencia artificiales_ES
dc.departamentoeuKonputazio zientziak eta adimen artifizialaes_ES


Files in this item

Thumbnail
Thumbnail

This item appears in the following Collection(s)

Show simple item record

(c) 2021 The Authors. Published by Elsevier Ltd. This is an open access article under the CC BY license ( http://creativecommons.org/licenses/by/4.0/ )
Except where otherwise noted, this item's license is described as (c) 2021 The Authors. Published by Elsevier Ltd. This is an open access article under the CC BY license ( http://creativecommons.org/licenses/by/4.0/ )