An approach for multimodal medical image retrieval using latent dirichlet allocation

dc.contributor.authorVikram, M.
dc.contributor.authorSuhas, B.S.
dc.contributor.authorAnantharaman, A.
dc.contributor.authorKamath S․, S.S.
dc.date.accessioned2026-02-06T06:37:36Z
dc.date.issued2019
dc.description.abstractModern medical practices are increasingly dependent on Medical Imaging for clinical analysis and diagnoses of patient illnesses. A significant challenge when dealing with the extensively available medical data is that it often consists of heterogeneous modalities. Existing works in the field of Content based medical image retrieval (CBMIR) have several limitations as they focus mainly on visual or textual features for retrieval. Given the unique manifold of medical data, we seek to leverage both the visual and textual modalities to improve the image retrieval. We propose a Latent Dirichlet Allocation (LDA) based technique for encoding the visual features and show that these features effectively model the medical images. We explore early fusion and late fusion techniques to combine these visual features with the textual features. The proposed late fusion technique achieved a higher mAP than the state-of-the-art on the ImageCLEF 2009 dataset, underscoring its suitability for effective multimodal medical image retrieval. © 2019 Association for Computing Machinery.
dc.identifier.citationACM International Conference Proceeding Series, 2019, Vol., , p. 44-51
dc.identifier.issn21531633
dc.identifier.urihttps://doi.org/10.1145/3297001.3297007
dc.identifier.urihttps://idr.nitk.ac.in/handle/123456789/31119
dc.publisherAssociation for Computing Machinery
dc.subjectContent-based Image Retrieval (CBIR)
dc.subjectMedical Informatics
dc.subjectTopic modeling
dc.titleAn approach for multimodal medical image retrieval using latent dirichlet allocation

Files