Problem: In the medical domain, obtaining training images on a large scale is difficult due to privacy and cost issues, and as further disease incidence varies widely, class imbalance has been a critical problem. Importantly, securing samples for rare diseases will always be difficult, but a diagnosis model should be able to predict the diseased samples accurately. When a model is trained on imbalanced data, it is expected that (1) the model only exhibits biased results for the majority classes, and (2) the recognition performance for minority classes will decline rapidly, resulting in low specificity. Objective: We aim to improve the prediction performance of such few sample classes using our auxiliary data from an efficient generation model and style knowledge distillation. In this paper, we introduce a style knowledge distillation method to address the class imbalance problem. Methods: The proposed method generates auxiliary images in which the style of a training sample is transferred. We train our model with balanced training samples by distilling the styles of auxiliary images. Conclusion: We demonstrate that it considerably increases the accuracy of classes with few samples, notably without adversely impacting classes with large samples, using the APTOS2019, and ISIC2018. This approach is straightforward but very effective for medical image datasets with skewed class distributions.
This paper was supported in part by IITP-2019-0-01906 (AI Graduate Program at POSTECH) and the National Research Foundation of Korea (NRF) from the Korea Government (MSIT) under Grant NRF-2021R1F1A1062807 , Institute of Information and Communications Technology Planning and Evaluation (IITP) grant funded by the Korea Government (MSIT) (Artificial Intelligence Innovation Hub) under Grant 2021-0-02068 , and and under the Artificial Intelligence Convergence Innovation Human Resources Development RS-2023-00255968 Grant.