Conference Papers

Permanent URI for this collectionhttps://idr.nitk.ac.in/handle/123456789/28506

Browse

Search Results

Now showing 1 - 4 of 4
  • Item
    Dynamic Checkpointing: Fault Tolerance in High-Performance Computing
    (Institute of Electrical and Electronics Engineers Inc., 2024) Bhowmik, B.; Verma, T.; Dineshbhai, N.D.; Reddy, M.R.V.; Girish, K.K.
    Parallel computing has become a cornerstone of modern computational systems, enabling the rapid processing of complex tasks by utilizing multiple processors simultaneously. However, the efficiency and reliability of these systems can be significantly compromised by inherent challenges such as hardware failures, communication delays, and uneven workload distribution. These issues not only slow down computations but also threaten the dependability of applications reliant on parallel processing. To address these challenges, researchers have developed strategies like dynamic checkpointing and load balancing, which are crucial for enhancing fault tolerance and optimizing performance. Dynamic checkpointing periodically saves the computational state, allowing for recovery from failures without significant data loss, while load balancing ensures that tasks are evenly distributed across processors, preventing bottlenecks and underutilization of resources. By integrating these mechanisms, this paper proposes a robust framework that improves the reliability and efficiency of parallel systems, particularly in high-performance computing environments where the ability to handle large-scale data processing with minimal downtime is critical. © 2024 IEEE.
  • Item
    Enhancing MPI Communication Efficiency for Grid-Based Stencil Computations
    (Institute of Electrical and Electronics Engineers Inc., 2024) Goudar, S.I.; Nayaka, P.S.J.; Girish, K.K.; Bhowmik, B.
    In parallel computing, where efficiency and speed are crucial, the Message Passing Interface (MPI) is a fundamental paradigm for managing large-scale distributed memory systems. MPI is critical to complex computational tasks, particularly in grid-based computations that solve intricate numerical problems by discretizing spatial domains into structured grids. However, MPI Cartesian communicators exhibit limitations in handling these computations effectively, especially when managing large-scale data exchanges and complex stencil patterns. This paper addresses these challenges by presenting an integrated approach that combines MPI collective and Cartesian communication methods. The proposed solution simplifies data distribution, eliminates redundant interfaces, and enhances communication efficiency. Experimental results show a 43% reduction in execution time and a 40% decrease in communication overhead, with scalability improvements achieving 12.5x speedup using 64 processes. These quantitative outcomes demonstrate the advan-tages of the proposed method over conventional MPI Cartesian approaches, establishing it as a reliable framework for advancing High-Performance Computing (HPC) capabilities in grid-based applications. © 2024 IEEE.
  • Item
    Performance Analysis and Predictive Modeling of MPI Collective Algorithms in Multi-Core Clusters: A Comparative Study
    (Institute of Electrical and Electronics Engineers Inc., 2025) Reddy, M.R.V.S.R.S.; Raju, S.R.; Girish, K.K.; Bhowmik, B.
    Efficient communication is the foundation of parallel computing systems, enabling seamless coordination across multiple processors for optimal performance. At the core of this communication lies the Message Passing Interface, a crucial framework designed to facilitate data exchange between processors through collective operations. However, these MPI operations often face challenges, including fluctuating process counts, varying message sizes, and increased communication overhead. These issues can significantly impact execution times and scalability, leading to potential bottlenecks in large-scale systems. To address these concerns, this paper provides an in-depth evaluation of key MPI collective algorithms - Flat Tree, Chain, and Binary Tree - by examining their performance under varying configurations. By analyzing execution times and communication overhead, the study reveals the trade-offs inherent in each algorithm, offering insights into strategies for reducing communication costs. Through this analysis, we aim to provide valuable guidance to improve the efficiency and scalability of parallel computing, particularly in high-performance systems where communication efficiency is critical. © 2025 IEEE.
  • Item
    Efficient Parallel Algorithm for Detecting Longest Flow Paths in Flow Direction Grids
    (Institute of Electrical and Electronics Engineers Inc., 2025) Jayarukshi, K.; Agarwal, S.; Girish, K.K.; Goudar, S.; Bhowmik, B.
    High-performance computing (HPC) has transformed the capacity to address complex computational tasks across various scientific fields by enabling the efficient processing of large datasets and intricate simulations. In hydrological modeling, a critical task is identifying the longest flow channel within a catchment, which is essential for understanding water flow patterns and managing resources. However, existing geographic information system (GIS) algorithms for flow path identification often suffer from inefficiencies and inaccuracies. To address these challenges, this paper introduces innovative parallel methods utilizing Open Multi-Processing (OpenMP), a widely-used API that supports multi-platform shared-memory parallel programming. This approach optimizes the analysis of flow direction data, resulting in faster and more accurate identification of flow channels. The results demonstrate that the proposed method outperforms current approaches, offering substantial improvements in both performance and precision. These advancements have the potential to significantly enhance hydrological modeling practices and water resource management. © 2025 IEEE.