Faculty Publications
Permanent URI for this communityhttps://idr.nitk.ac.in/handle/123456789/18736
Publications by NITK Faculty
Browse
9 results
Search Results
Item Texture Classification based Efficient Image Compression Algorithm for Wireless Capsule Endoscopy(Institute of Electrical and Electronics Engineers Inc., 2019) Sushma, B.; Aparna., P.This paper presents a novel method for classification of blocks into smooth and edge blocks in transform domain and a compression scheme for Wireless Capsule Endoscopy (WCE) with block classifier. WCE involves capturing, transmission and processing of gastrointestinal images. Power consumption is a critical issue in WCE, as it uses a button battery driven capsule endoscope to capture and transmit images. The captured image needs to be compressed to save the transmission power and low complexity compressor should be used to avoid more power consumption from the compressor itself. JPEG based compression techniques which consists Discrete Cosine Transform(DCT), quantizer and entropy encoder provides the best compression performance with less complexity compared to other various techniques. Pixel distribution in smooth blocks is uniform and energy is compacted only into low frequency bands in spectral domain. Because high frequency bands are almost having zero energy, only low frequency bands are quantized and entropy coded which saves power in processing high bands. Most of the endoscopic image has smooth region, this method is more suitable to WCE. Proposed algorithm improves compression rate by 9% without sacrificing quality compared to JPEG based compression algorithm. © 2019 IEEE.Item Effect of Different Color Spaces on Deep Image Segmentation(Institute of Electrical and Electronics Engineers Inc., 2021) Sushma, B.; Pulikala, P.Image segmentation is an important application in computer vision, proposed to partition an image into meaningful regions on a specific criterion. In recent days, image segmentation tasks have achieved state of the art performance using deep neural and fully connected networks. The datasets used for the segmentation task mainly consist of image data in RGB color space and the deep segmentation architectures are trained without modifying the color space. In this study, the importance of color space is investigated and the obtained results show that the color space can affect the segmentation performance remarkably. Certain regions of interest in images belonging to a particular domain can be segmented better when represented in a certain form of color space. To explore on this two datasets from medical and satellite imagery are considered. The UNET model is modified to accept images as a combination of color spaces and is trained to segment the colonoscopy images for polyps and satellite images for roads under individual and combination of color spaces. Experiments show that the performance of polyp segmentation is better when a combination of HSV+YCbCr color space is considered. Road segmentation in satellite imagery is better in LAB+HSV color space. © 2021 IEEE.Item Automated Summarization of Gastrointestinal Endoscopy Video(Springer Science and Business Media Deutschland GmbH, 2023) Sushma, B.; Aparna., P.Gastrointestinal (GI) endoscopy enables many minimally invasive procedures for diagnosing diseases such as esophagitis, ulcer, polyps and cancers. Guided by the endoscope’s video sequence, a physician can diagnose the diseases and administer the treatment. Unfortunately, due to the huge amount of data generated, physicians are currently discarding procedural video and rely on a small number of carefully chosen images to record a procedure. In addition, when a patient seeks a second opinion, the assessment of lesions in a huge video stream necessitates a thorough examination, which is a time-consuming process that demands much attention. To reduce the length of the video stream, an automated method to generate the summary of endoscopy video recordings consisting only of abnormal frames by using deep convolutional neural networks trained to classify normal, abnormal and uninformative frames is proposed. Results show that our method can efficiently detect abnormal frames and is robust to the variations in the frames. The proposed CNN architecture outperforms the other classification models with an accuracy of 0.9698 with less number of parameters. © IFIP International Federation for Information Processing 2023.Item Speech Intelligibility Enhancement for Cochlear Implant using Multi-Objective Deep Denoising Autoencoder(Institute of Electrical and Electronics Engineers Inc., 2023) Vishnu, B.U.P.; Poluboina, V.; Sushma, B.; Pulikala, A.This study introduces a novel technique for enhancing the performance of deep denoising autoencoders (DDAE) in speech processing for cochlear implants (CIs). For individuals with hearing loss, cochlear implants are electronic devices that help to restore their ability to hear. However, the performance of CIs speech intelligibility in the noisy environment is limited. One of the most commonly used methods for reducing noise in CIs is through a preprocessing technique called deep denoising autoencoder. DDAE models have shown potential in learning various noise patterns, but their performance in enhancing speech intelligibility is relatively low due to a ineffective objective function. To address this limitation, this study proposes a multi-objective technique to fine-tune the DDAE model. When multiple objectives are optimized simultaneously, the model becomes more robust and better at handling real-time noise. Based on the experimental findings, it has been confirmed that the proposed multi-objective learning technique performs better than other models when it comes to speech intelligibility. Furthermore, the enhanced signal is presented to the acoustic cochlear implant simulator to evaluate the improvement of speech intelligibility in CIs. © 2023 IEEE.Item Recent developments in wireless capsule endoscopy imaging: Compression and summarization techniques(Elsevier Ltd, 2022) Sushma, B.; Aparna., P.Wireless capsule endoscopy (WCE) can be viewed as an innovative technology introduced in the medical domain to directly visualize the digestive system using a battery-powered electronic capsule. It is considered a desirable substitute for conventional digestive tract diagnostic methods for a comfortable and painless inspection. Despite many benefits, WCE results in poor video quality due to low frame resolution and diagnostic accuracy. Many research groups have presented diversified, low-complexity compression techniques to economize battery power consumed in the radio-frequency transmission of the captured video, which allows for capturing the images at high resolution. Many vision-based computational methods have been developed to improve the diagnostic yield. These methods include approaches for automatically detecting abnormalities and reducing the amount of time needed for video analysis. Though various research works have been put forth in the WCE imaging field, there is still a wide gap between the existing techniques and the current needs. Hence, this article systematically reviews recent WCE video compression and summarization techniques. The review's objectives are as follows: First, to provide the details of the requirement, challenges and design percepts for the low complexity WCE video compressor. Second, to discuss the most recent compression methods, emphasizing simple distributed video coding methods. Next, to review the most recent summarization techniques and the significance of using deep neural networks. Further, this review aims to provide a quantitative analysis of the state-of-the-art methods along with their advantages and drawbacks. At last, to discuss existing problems and possible future directions for building a robust WCE imaging framework. © 2022 Elsevier LtdItem Distributed video coding based on classification of frequency bands with block texture conditioned key frame encoder for wireless capsule endoscopy(Elsevier Ltd, 2020) Sushma, B.; Aparna., P.Wireless capsule endoscopy (WCE) has provided remarkable improvement in diagnosing gastrointestinal disorders by scanning the entire digestive tract. The system still need refinement, to upgrade the quality of images, frame rate and battery life. The principal component of the system that can address these issues is low complexity video compressor. Motivated by low computational complexity requirements of WCE video encoding, this paper presents a distributed video coding framework based on frequency bands classification. The lower frequency bands are used to generate good quality side information (SI) as they exhibit high temporal correlation. This reduces the complexity of hash generation at the encoder, thus eliminating the latency in SI creation. Apart from this, SI creation involves only a simple block search and doesn't depend on Wyner–Ziv (WZ) bands. Also different approach for distributed coding of sub-sampled chroma components of WZ frame is proposed. Low complexity JPEG based key frame encoding is proposed that take advantage of WCE image textural properties to reduce the complexity of encoding smooth blocks by 81% at the quantization and encoding stage. Other novel features include use of discrete Tchebichef transform (DTT), Golomb–Rice code for entropy coding. Performance evaluation shows that the proposed method achieves 60% improvement in compression over Motion JPEG with low computational complexity. © 2020 Elsevier LtdItem Summarization of Wireless Capsule Endoscopy Video Using Deep Feature Matching and Motion Analysis(Institute of Electrical and Electronics Engineers Inc., 2021) Sushma, B.; Aparna., P.Conventional Wireless capsule endoscopy (WCE) video summary generation techniques apprehend an image by extracting hand crafted features, which are not essentially sufficient to encapsulate the semantic similarity of endoscopic images. Use of supervised methods for extraction of deep features from an image need an enormous amount of accurate labelled data for training process. To solve this, we use an unsupervised learning method to extract features using convolutional auto encoder. Furthermore, WCE images are classified into similar and dissimilar pairs using fixed threshold derived through large number of experiments. Finally, keyframe extraction method based on motion analysis is used to derive a structured summary of WCE video. Proposed method achieves an average F-measure of 91.1% with compression ratio of 83.12%. The results indicate that the proposed method is more efficient compared to existing WCE video summarization techniques. © 2013 IEEE.Item Deep chroma prediction of Wyner–Ziv frames in distributed video coding of wireless capsule endoscopy video(Academic Press Inc., 2022) Sushma, B.; Aparna., P.Compression of captured video frames is crucial for saving the power in wireless capsule endoscopy (WCE). A low complexity encoder is desired to limit the power consumption required for compressing the WCE video. Distributed video coding (DVC) technique is best suitable for designing a low complexity encoder. In this technique, frames captured in RGB colour space are converted into YCbCr colour space. Both Y and CbCr representing luma and chroma components of the Wyner–Ziv (WZ) frames are processed and encoded in existing DVC techniques proposed for WCE video compression. In the WCE video, consecutive frames exhibit more similarity in texture and colour properties. The proposed work uses these properties to present a method for processing and encoding only the luma component of a WZ frame. The chroma components of the WZ frame are predicted by an encoder–decoder based deep chroma prediction model at the decoder by matching luma and texture information of the keyframe and WZ frame. The proposed method reduces the computations required for encoding and transmitting of WZ chroma component. The results show that the proposed DVC with a deep chroma prediction model performs better when compared to motion JPEG and existing DVC systems for WCE at the reduced encoder complexity. © 2022 Elsevier Inc.Item AAPFC-BUSnet: Hierarchical encoder–decoder based CNN with attention aggregation pyramid feature clustering for breast ultrasound image lesion segmentation(Elsevier Ltd, 2024) Sushma, B.; Pulikala, A.Breast cancer causes a serious menace to women's health and lives, underscoring the urgency of accurate tumor detection. Detecting both cancerous and non-cancerous breast tumors has become increasingly crucial, with ultrasound imaging emerging as a widely adopted modality for this purpose. However, identifying breast lesions in ultrasound images is a challenging task due to various tumor morphologies, geometry, similar color intensity distributions, and fuzzy boundaries, particularly irregularly shaped malignant tumors. This work proposes an encoder–decoder based U-shaped convolutional neural network (CNN) variant with an attention aggregation-based pyramid feature clustering module (AAPFC) to detect breast lesion regions. The network consists of the U-Net variant as a base network and AAPFC to fuse features extracted at the various levels of the base U-Net using a suitable feature fusion technique. Furthermore, the deformable convolution with adaptive self-attention mechanism is introduced to decode the pyramid features parallel to capture the various geometric features at multi-stages. Two public breast lesion ultrasound datasets consisting 263 malignant, 547 benign and 133 normal images are considered to evaluate the performance of the proposed model and state-of-the-art deep CNN-based segmentation models. The proposed model provides 96% accuracy, 68% Mean-IoU, 97% specificity, 82% sensitivity and 0.747 kappa score respectively. The conducted qualitative and quantitative performance analysis experiments show that the proposed model performs better in breast lesion segmentation on ultrasound images. © 2024 Elsevier Ltd
