Adversarial attack vulnerability of medical image analysis systems: Unexplored factors

Research output: Contribution to journalJournal articleResearchpeer-review

Standard

Adversarial attack vulnerability of medical image analysis systems : Unexplored factors. / Bortsova, Gerda; González-Gonzalo, Cristina; Wetstein, Suzanne C; Dubost, Florian; Katramados, Ioannis; Hogeweg, Laurens; Liefers, Bart; van Ginneken, Bram; Pluim, Josien P W; Veta, Mitko; Sánchez, Clara I; de Bruijne, Marleen.

In: Medical Image Analysis, Vol. 73, 102141, 2021.

Research output: Contribution to journalJournal articleResearchpeer-review

Harvard

Bortsova, G, González-Gonzalo, C, Wetstein, SC, Dubost, F, Katramados, I, Hogeweg, L, Liefers, B, van Ginneken, B, Pluim, JPW, Veta, M, Sánchez, CI & de Bruijne, M 2021, 'Adversarial attack vulnerability of medical image analysis systems: Unexplored factors', Medical Image Analysis, vol. 73, 102141. https://doi.org/10.1016/j.media.2021.102141

APA

Bortsova, G., González-Gonzalo, C., Wetstein, S. C., Dubost, F., Katramados, I., Hogeweg, L., Liefers, B., van Ginneken, B., Pluim, J. P. W., Veta, M., Sánchez, C. I., & de Bruijne, M. (2021). Adversarial attack vulnerability of medical image analysis systems: Unexplored factors. Medical Image Analysis, 73, [102141]. https://doi.org/10.1016/j.media.2021.102141

Vancouver

Bortsova G, González-Gonzalo C, Wetstein SC, Dubost F, Katramados I, Hogeweg L et al. Adversarial attack vulnerability of medical image analysis systems: Unexplored factors. Medical Image Analysis. 2021;73. 102141. https://doi.org/10.1016/j.media.2021.102141

Author

Bortsova, Gerda ; González-Gonzalo, Cristina ; Wetstein, Suzanne C ; Dubost, Florian ; Katramados, Ioannis ; Hogeweg, Laurens ; Liefers, Bart ; van Ginneken, Bram ; Pluim, Josien P W ; Veta, Mitko ; Sánchez, Clara I ; de Bruijne, Marleen. / Adversarial attack vulnerability of medical image analysis systems : Unexplored factors. In: Medical Image Analysis. 2021 ; Vol. 73.

Bibtex

@article{8236e9d80cc9404f8a12cb2edadf213a,
title = "Adversarial attack vulnerability of medical image analysis systems: Unexplored factors",
abstract = "Adversarial attacks are considered a potentially serious security threat for machine learning systems. Medical image analysis (MedIA) systems have recently been argued to be vulnerable to adversarial attacks due to strong financial incentives and the associated technological infrastructure. In this paper, we study previously unexplored factors affecting adversarial attack vulnerability of deep learning MedIA systems in three medical domains: ophthalmology, radiology, and pathology. We focus on adversarial black-box settings, in which the attacker does not have full access to the target model and usually uses another model, commonly referred to as surrogate model, to craft adversarial examples that are then transferred to the target model. We consider this to be the most realistic scenario for MedIA systems. Firstly, we study the effect of weight initialization (pre-training on ImageNet or random initialization) on the transferability of adversarial attacks from the surrogate model to the target model, i.e., how effective attacks crafted using the surrogate model are on the target model. Secondly, we study the influence of differences in development (training and validation) data between target and surrogate models. We further study the interaction of weight initialization and data differences with differences in model architecture. All experiments were done with a perturbation degree tuned to ensure maximal transferability at minimal visual perceptibility of the attacks. Our experiments show that pre-training may dramatically increase the transferability of adversarial examples, even when the target and surrogate's architectures are different: the larger the performance gain using pre-training, the larger the transferability. Differences in the development data between target and surrogate models considerably decrease the performance of the attack; this decrease is further amplified by difference in the model architecture. We believe these factors should be considered when developing security-critical MedIA systems planned to be deployed in clinical practice. We recommend avoiding using only standard components, such as pre-trained architectures and publicly available datasets, as well as disclosure of design specifications, in addition to using adversarial defense methods. When evaluating the vulnerability of MedIA systems to adversarial attacks, various attack scenarios and target-surrogate differences should be simulated to achieve realistic robustness estimates. The code and all trained models used in our experiments are publicly available.3.",
author = "Gerda Bortsova and Cristina Gonz{\'a}lez-Gonzalo and Wetstein, {Suzanne C} and Florian Dubost and Ioannis Katramados and Laurens Hogeweg and Bart Liefers and {van Ginneken}, Bram and Pluim, {Josien P W} and Mitko Veta and S{\'a}nchez, {Clara I} and {de Bruijne}, Marleen",
note = "Copyright {\textcopyright} 2021 The Author(s). Published by Elsevier B.V. All rights reserved.",
year = "2021",
doi = "10.1016/j.media.2021.102141",
language = "English",
volume = "73",
journal = "Medical Image Analysis",
issn = "1361-8415",
publisher = "Elsevier",

}

RIS

TY - JOUR

T1 - Adversarial attack vulnerability of medical image analysis systems

T2 - Unexplored factors

AU - Bortsova, Gerda

AU - González-Gonzalo, Cristina

AU - Wetstein, Suzanne C

AU - Dubost, Florian

AU - Katramados, Ioannis

AU - Hogeweg, Laurens

AU - Liefers, Bart

AU - van Ginneken, Bram

AU - Pluim, Josien P W

AU - Veta, Mitko

AU - Sánchez, Clara I

AU - de Bruijne, Marleen

N1 - Copyright © 2021 The Author(s). Published by Elsevier B.V. All rights reserved.

PY - 2021

Y1 - 2021

N2 - Adversarial attacks are considered a potentially serious security threat for machine learning systems. Medical image analysis (MedIA) systems have recently been argued to be vulnerable to adversarial attacks due to strong financial incentives and the associated technological infrastructure. In this paper, we study previously unexplored factors affecting adversarial attack vulnerability of deep learning MedIA systems in three medical domains: ophthalmology, radiology, and pathology. We focus on adversarial black-box settings, in which the attacker does not have full access to the target model and usually uses another model, commonly referred to as surrogate model, to craft adversarial examples that are then transferred to the target model. We consider this to be the most realistic scenario for MedIA systems. Firstly, we study the effect of weight initialization (pre-training on ImageNet or random initialization) on the transferability of adversarial attacks from the surrogate model to the target model, i.e., how effective attacks crafted using the surrogate model are on the target model. Secondly, we study the influence of differences in development (training and validation) data between target and surrogate models. We further study the interaction of weight initialization and data differences with differences in model architecture. All experiments were done with a perturbation degree tuned to ensure maximal transferability at minimal visual perceptibility of the attacks. Our experiments show that pre-training may dramatically increase the transferability of adversarial examples, even when the target and surrogate's architectures are different: the larger the performance gain using pre-training, the larger the transferability. Differences in the development data between target and surrogate models considerably decrease the performance of the attack; this decrease is further amplified by difference in the model architecture. We believe these factors should be considered when developing security-critical MedIA systems planned to be deployed in clinical practice. We recommend avoiding using only standard components, such as pre-trained architectures and publicly available datasets, as well as disclosure of design specifications, in addition to using adversarial defense methods. When evaluating the vulnerability of MedIA systems to adversarial attacks, various attack scenarios and target-surrogate differences should be simulated to achieve realistic robustness estimates. The code and all trained models used in our experiments are publicly available.3.

AB - Adversarial attacks are considered a potentially serious security threat for machine learning systems. Medical image analysis (MedIA) systems have recently been argued to be vulnerable to adversarial attacks due to strong financial incentives and the associated technological infrastructure. In this paper, we study previously unexplored factors affecting adversarial attack vulnerability of deep learning MedIA systems in three medical domains: ophthalmology, radiology, and pathology. We focus on adversarial black-box settings, in which the attacker does not have full access to the target model and usually uses another model, commonly referred to as surrogate model, to craft adversarial examples that are then transferred to the target model. We consider this to be the most realistic scenario for MedIA systems. Firstly, we study the effect of weight initialization (pre-training on ImageNet or random initialization) on the transferability of adversarial attacks from the surrogate model to the target model, i.e., how effective attacks crafted using the surrogate model are on the target model. Secondly, we study the influence of differences in development (training and validation) data between target and surrogate models. We further study the interaction of weight initialization and data differences with differences in model architecture. All experiments were done with a perturbation degree tuned to ensure maximal transferability at minimal visual perceptibility of the attacks. Our experiments show that pre-training may dramatically increase the transferability of adversarial examples, even when the target and surrogate's architectures are different: the larger the performance gain using pre-training, the larger the transferability. Differences in the development data between target and surrogate models considerably decrease the performance of the attack; this decrease is further amplified by difference in the model architecture. We believe these factors should be considered when developing security-critical MedIA systems planned to be deployed in clinical practice. We recommend avoiding using only standard components, such as pre-trained architectures and publicly available datasets, as well as disclosure of design specifications, in addition to using adversarial defense methods. When evaluating the vulnerability of MedIA systems to adversarial attacks, various attack scenarios and target-surrogate differences should be simulated to achieve realistic robustness estimates. The code and all trained models used in our experiments are publicly available.3.

U2 - 10.1016/j.media.2021.102141

DO - 10.1016/j.media.2021.102141

M3 - Journal article

C2 - 34246850

VL - 73

JO - Medical Image Analysis

JF - Medical Image Analysis

SN - 1361-8415

M1 - 102141

ER -

ID: 274431709