Resident Vision Transformer: Lightweight Deep Learning Model for Disease Diagnosis on Edge Devices
No Thumbnail Available
Date
2024
Journal Title
Journal ISSN
Volume Title
Publisher
Institute of Electrical and Electronics Engineers Inc.
Abstract
The deployment of AI-based medical image diagnosis on mobile edge devices face dual challenges of aliasing due to resizing images to lower resolutions, and huge total trainable parameters associated with deep learning architectures. This paper aims to address these challenges by proposing Resident Vision Transformer architecture with residual and dense connections enabling enhanced performance with lower number of parameters. The input to the proposed architecture is processed with adaptive padding, which not only maintains a constant input size, but also, preserves the spatial information. The proposed architecture has been trained and tested on three publicly available datasets, namely, breast cancer, skin cancer and brain tumor datasets. On the mini-MIAS database of mammograms, the proposed architecture achieved an accuracy of 92.94%, outperforming several related works. Similarly, on the PH2 database for Skin Cancer and the brain tumor dataset, the proposed architecture achieved an accuracy of 94.73% and 98.79%, respectively with fewer parameters. The proposed architecture paves way for feasibility of AI-driven medical image diagnosis on resource-constrained mobile edge devices. © 2024 IEEE.
Description
Keywords
Medical Image Diagnosis, Mobile Edge Devices, Vision Transformer
Citation
2024 10th International Conference on Smart Computing and Communication, ICSCC 2024, 2024, Vol., , p. 349-355
