Adversarial Reconstruction Loss for Domain Generalization

Publikation: Bidrag til tidsskriftTidsskriftartikelForskningfagfællebedømt

Dokumenter

  • Imad Eddine Ibrahim Bekkouch
  • Dragos Constantin Nicolae
  • Adil Khan
  • S. M.Ahsan Kazmi
  • Asad Masood Khattak
  • Ibragimov, Bulat

The biggest fear when deploying machine learning models to the real world is their ability to handle the new data. This problem is significant especially in medicine, where models trained on rich high-quality data extracted from large hospitals do not scale to small regional hospitals. One of the clinical challenges addressed in this work is magnetic resonance image generalization for improved visualization and diagnosis of hip abnormalities such as femoroacetabular impingement and dysplasia. Domain Generalization (DG) is a field in machine learning that tries to solve the model's dependency on the training data by leveraging many related but different data sources. We present a new method for DG that is both efficient and fast, unlike the most current state of art methods, which add a substantial computational burden making it hard to fine-tune. Our model trains an autoencoder setting on top of the classifier, but the encoder is trained on the adversarial reconstruction loss forcing it to forget style information while extracting features useful for classification. Our approach aims to force the encoder to generate domain-invariant representations that are still category informative by pushing it in both directions. Our method has proven universal and was validated on four different benchmarks for domain generalization, outperforming state of the art on RMNIST, VLCS and IXMAS with a 0.70% increase in accuracy and providing comparable results on PACS with a 0.02% difference. Our method was also evaluated for unsupervised domain adaptation and has shown to be quite an effective method against over-fitting.

OriginalsprogEngelsk
Artikelnummer9378518
TidsskriftIEEE Access
Vol/bind9
Sider (fra-til)42424-42437
Antal sider14
ISSN2169-3536
DOI
StatusUdgivet - 2021

Bibliografisk note

Funding Information:
This work was supported in part by the Foundation for Basic Research (RFBR) under Project 19-37-51034, and in part by the Zayed University Research Incentive Fund under Grant R19096.

Publisher Copyright:
© 2013 IEEE.

ID: 285249942