Conference Papers

Permanent URI for this collectionhttps://idr.nitk.ac.in/handle/123456789/28506

Browse

Search Results

Now showing 1 - 2 of 2
  • Item
    Effect of Different Color Spaces on Deep Image Segmentation
    (Institute of Electrical and Electronics Engineers Inc., 2021) Sushma, B.; Pulikala, P.
    Image segmentation is an important application in computer vision, proposed to partition an image into meaningful regions on a specific criterion. In recent days, image segmentation tasks have achieved state of the art performance using deep neural and fully connected networks. The datasets used for the segmentation task mainly consist of image data in RGB color space and the deep segmentation architectures are trained without modifying the color space. In this study, the importance of color space is investigated and the obtained results show that the color space can affect the segmentation performance remarkably. Certain regions of interest in images belonging to a particular domain can be segmented better when represented in a certain form of color space. To explore on this two datasets from medical and satellite imagery are considered. The UNET model is modified to accept images as a combination of color spaces and is trained to segment the colonoscopy images for polyps and satellite images for roads under individual and combination of color spaces. Experiments show that the performance of polyp segmentation is better when a combination of HSV+YCbCr color space is considered. Road segmentation in satellite imagery is better in LAB+HSV color space. © 2021 IEEE.
  • Item
    Automated Summarization of Gastrointestinal Endoscopy Video
    (Springer Science and Business Media Deutschland GmbH, 2023) Sushma, B.; Aparna., P.
    Gastrointestinal (GI) endoscopy enables many minimally invasive procedures for diagnosing diseases such as esophagitis, ulcer, polyps and cancers. Guided by the endoscope’s video sequence, a physician can diagnose the diseases and administer the treatment. Unfortunately, due to the huge amount of data generated, physicians are currently discarding procedural video and rely on a small number of carefully chosen images to record a procedure. In addition, when a patient seeks a second opinion, the assessment of lesions in a huge video stream necessitates a thorough examination, which is a time-consuming process that demands much attention. To reduce the length of the video stream, an automated method to generate the summary of endoscopy video recordings consisting only of abnormal frames by using deep convolutional neural networks trained to classify normal, abnormal and uninformative frames is proposed. Results show that our method can efficiently detect abnormal frames and is robust to the variations in the frames. The proposed CNN architecture outperforms the other classification models with an accuracy of 0.9698 with less number of parameters. © IFIP International Federation for Information Processing 2023.