RT Journal Article T1 Robust image classification against adversarial attacks using elastic similarity measures between edge count sequences A1 Oregi, Izaskun A1 Del Ser, Javier A1 Pérez, Aritz A1 Lozano, José A. AB Due to their unprecedented capacity to learn patterns from raw data, deep neural networks have become the de facto modeling choice to address complex machine learning tasks. However, recent works have emphasized the vulnerability of deep neural networks when being fed with intelligently manipulated adversarial data instances tailored to confuse the model. In order to overcome this issue, a major effort has been made to find methods capable of making deep learning models robust against adversarial inputs. This work presents a new perspective for improving the robustness of deep neural networks in image classification. In computer vision scenarios, adversarial images are crafted by manipulating legitimate inputs so that the target classifier is eventually fooled, but the manipulation is not visually distinguishable by an external observer. The reason for the imperceptibility of the attack is that the human visual system fails to detect minor variations in color space, but excels at detecting anomalies in geometric shapes. We capitalize on this fact by extracting color gradient features from input images at multiple sensitivity levels to detect possible manipulations. We resort to a deep neural classifier to predict the category of unseen images, whereas a discrimination model analyzes the extracted color gradient features with time series techniques to determine the legitimacy of input images. The performance of our method is assessed over experiments comprising state-of-the-art techniques for crafting adversarial attacks. Results corroborate the increased robustness of the classifier when using our discrimination module, yielding drastically reduced success rates of adversarial attacks that operate on the whole image rather than on localized regions or around the existing shapes of the image. Future research is outlined towards improving the detection accuracy of the proposed method for more general attack strategies. SN 0893-6080 YR 2020 FD 2020-08 LK https://hdl.handle.net/11556/3453 UL https://hdl.handle.net/11556/3453 LA eng NO Oregi , I , Del Ser , J , Pérez , A & Lozano , J A 2020 , ' Robust image classification against adversarial attacks using elastic similarity measures between edge count sequences ' , Neural Networks , vol. 128 , pp. 61-72 . https://doi.org/10.1016/j.neunet.2020.04.030 NO Publisher Copyright: © 2020 Elsevier Ltd NO This research work has been supported by the Basque Government through the EMAITEK and ELKARTEK funding programs. J. Del Ser receives funding support from the Consolidated Research Group MATHMODE ( IT1294-19 ) granted by the Department of Education of the Basque Government . A. Pérez and J. A. Lozano are supported by the Basque Government through the BERC 2018–2021 program and by the Spanish Ministry of Economy and Competitiveness MINECO through BCAM Severo Ochoa excellence accreditation SEV-2017-0718. A. Pérez also acknowledges funding support from AEI/FEDER (UE) through project TIN2017-82626-R. J. A. Lozano is also supported by Spanish Ministry of Economy and Competitiveness MINECO through TIN2016-78365-R . This research work has been supported by the Basque Government through the EMAITEK and ELKARTEK funding programs. J. Del Ser receives funding support from the Consolidated Research Group MATHMODE (IT1294-19) granted by the Department of Education of the Basque Government. A. P?rez and J. A. Lozano are supported by the Basque Government through the BERC 2018?2021 program and by the Spanish Ministry of Economy and Competitiveness MINECO through BCAM Severo Ochoa excellence accreditation SEV-2017-0718. A. P?rez also acknowledges funding support from AEI/FEDER (UE) through project TIN2017-82626-R. J. A. Lozano is also supported by Spanish Ministry of Economy and Competitiveness MINECO through TIN2016-78365-R. DS TECNALIA Publications RD 31 jul 2024