You are required to read and agree to the below before accessing a full-text version of an article in the IDE article repository.

The full-text document you are about to access is subject to national and international copyright laws. In most cases (but not necessarily all) the consequence is that personal use is allowed given that the copyright owner is duly acknowledged and respected. All other use (typically) require an explicit permission (often in writing) by the copyright owner.

For the reports in this repository we specifically note that

  • the use of articles under IEEE copyright is governed by the IEEE copyright policy (available at http://www.ieee.org/web/publications/rights/copyrightpolicy.html)
  • the use of articles under ACM copyright is governed by the ACM copyright policy (available at http://www.acm.org/pubs/copyright_policy/)
  • technical reports and other articles issued by M‰lardalen University is free for personal use. For other use, the explicit consent of the authors is required
  • in other cases, please contact the copyright owner for detailed information

By accepting I agree to acknowledge and respect the rights of the copyright owner of the document I am about to access.

If you are in doubt, feel free to contact webmaster@ide.mdh.se

Evaluating the robustness of ML models to out-of-distribution data through similarity analysis

Fulltext:


Authors:


Publication Type:

Conference/Workshop Paper

Venue:

1st Workshop on Advanced AI Techniques for Data Management and Analytics

Publisher:

Springer


Abstract

In Machine Learning systems, several factors impact the performance of a trained model. The most important ones include model architecture, the amount of training time, the dataset size and diversity. We present a method for analyzing datasets from a use-case scenario perspective, detecting and quantifying out-of-distribution (OOD) data on dataset level.Our main contribution is the novel use of similarity metrics for the evaluation of the robustness of a model by introducing relative Fréchet Inception Distance (FID) and relative Kernel Inception Distance (KID) measures. These relative measures are relative to a baseline in-distribution dataset and are used to estimate how the model will perform on OOD data (i.e. estimate the model accuracy drop). We find a correlation between our proposed relative FID/relative KID measure and the drop in Average Precision (AP) accuracy on unseen data.

Bibtex

@inproceedings{Linden6712,
author = {Joakim Lind{\'e}n and H{\aa}kan Forsberg and Ingemar S{\"o}derquist and Masoud Daneshtalab},
title = {Evaluating the robustness of ML models to out-of-distribution data through similarity analysis},
number = {1},
month = {September},
year = {2023},
booktitle = {1st Workshop on Advanced AI Techniques for Data Management and Analytics},
publisher = {Springer},
url = {http://www.es.mdu.se/publications/6712-}
}