Journal Articles
Permanent URI for this collectionhttps://idr.nitk.ac.in/handle/123456789/19884
Browse
16 results
Search Results
Item Containerized deployment of micro-services in fog devices: a reinforcement learning-based approach(Springer, 2022) Nath, S.B.; Chattopadhyay, S.; Karmakar, R.; Addya, S.K.; Chakraborty, S.; Ghosh, S.K.The real power of fog computing comes when deployed under a smart environment, where the raw data sensed by the Internet of Things (IoT) devices should not cross the data boundary to preserve the privacy of the environment, yet a fast computation and the processing of the data is required. Devices like home network gateway, WiFi access points or core network switches can work as a fog device in such scenarios as its computing resources can be leveraged by the applications for data processing. However, these devices have their primary workload (like packet forwarding in a router/switch) that is time-varying and often generates spikes in the resource demand when bandwidth-hungry end-user applications, are started. In this paper, we propose pick–test–choose, a dynamic micro-service deployment and execution model that considers such time-varying primary workloads and workload spikes in the fog nodes. The proposed mechanism uses a reinforcement learning mechanism, Bayesian optimization, to decide the target fog node for an application micro-service based on its prior observation of the system’s states. We implement PTC in a testbed setup and evaluate its performance. We observe that PTC performs better than four other baseline models for micro-service offloading in a fog computing framework. In the experiment with an optical character recognition service, the proposed PTC gives average response time in the range of 9.71 sec–50 sec, which is better than Foglets (24.21 sec–80.35 sec), first-fit (16.74 sec–88 sec), best-fit (11.48 sec–57.39 sec) and mobility-based method (12 sec–53 sec). A further scalability study with an emulated setup over Amazon EC2 further confirms the superiority of PTC over other baselines. © 2021, The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature.Item CoMCLOUD: Virtual Machine Coalition for Multi-Tier Applications over Multi-Cloud Environments(Institute of Electrical and Electronics Engineers Inc., 2023) Addya, S.K.; Satpathy, A.; Ghosh, B.C.; Chakraborty, S.; Ghosh, S.K.; Das, S.K.Applications hosted in commercial clouds are typically multi-tier and comprise multiple tightly coupled virtual machines (VMs). Service providers (SPs) cater to the users using VM instances with different configurations and pricing depending on the location of the data center (DC) hosting the VMs. However, selecting VMs to host multi-tier applications is challenging due to the trade-off between cost and quality of service (QoS) depending on the placement of VMs. This paper proposes a multi-cloud broker model called CoMCLOUD to select a sub-optimal VM coalition for multi-tier applications from an SP with minimum coalition pricing and maximum QoS. To strike a trade-off between the cost and QoS, we use an ant-colony-based optimization technique. The overall service selection game is modeled as a first-price sealed-bid auction aimed at maximizing the overall revenue of SPs. Further, as the hosted VMs often face demand spikes, we present a parallel migration strategy to migrate VMs with minimum disruption time. Detailed experiments show that our approach can improve the federation profit up to 23% at the expense of increased latency of approximately 15%, compared to the baselines. © 2013 IEEE.Item NORD: NOde Ranking-based efficient virtual network embedding over single Domain substrate networks(Elsevier B.V., 2023) Keerthan Kumar, T.G.; Addya, S.K.; Satpathy, A.; Koolagudi, S.G.Network virtualization (NV) allows the service providers (SPs) to partition the substrate resources in the form of isolated virtual networks (VNs) comprising multiple correlated virtual machines (VMs) and virtual links (VLs), capturing the dependencies. Though NV brought about multiple benefits, such as service isolation, improved quality-of-service (QoS), secure communication, and better utilization of substrate resources, it also introduced numerous research challenges. In this regard, one of the predominant challenges is assigning resources to the virtual components, i.e., VMs and VLs, also termed virtual network embedding (VNE). VNE comprises two closely related sub-problems, (i.) VM embedding and (ii.) VL embedding, and both the problems have been demonstrated to be NP-Hard. In the context of VNE, maximizing the revenue to cost ratio remains the focal point for the SPs as it not only boosts acceptance of VNRs but also effectively utilizes the substrate resources. However, the existing literature on VNE suffers from the following pitfalls: (i.) They only consider system resources or (ii.) limited topological attributes. However, both attributes are quintessential in accurately capturing the VNRs and the substrate network dependencies, thereby augmenting the revenue to cost ratio. This paper proposes an efficient VNE strategy called, NOde Ranking-based efficient virtual network embedding over single Domain substrate networks (NORD), to maximize the revenue to cost ratio. To address the problem of VM embedding, NORD utilizes a hybrid entropy and the technique for order of preference by similarity to ideal solution (TOPSIS) based ranking strategy for VMs and servers considering both system and topological attributes that effectively capture the dependencies. Once the ranking is generated, A greedy VM embedding followed by shortest path VL embedding completes the assignment. Simulation results confirm that NORD attains a 40% and 61% increment in average acceptance and revenue-to-cost ratios compared to the baselines. © 2023 Elsevier B.V.Item Shipping code towards data in an inter-region serverless environment to leverage latency(Springer, 2023) Sethi, B.; Addya, S.K.; Bhutada, J.; Ghosh, S.K.Serverless computing emerges as a new standard to build cloud applications, where developers write compact functions that respond to events in the cloud infrastructure. Several cloud service industries started adopting serverless for deploying their applications. But one key limitation in serverless computing is that it disregards the significance of data. In the age of big data, when applications run around a huge volume, to transfer data from the data side to the computation side to co-allocate the data and code, leads to high latency. All existing serverless architectures are based on the data shipping architecture. In this paper, we present an inter-region code shipping architecture for serverless, that enables the code to flow from computation side to the data side where the size of the code is negligible compared to the data size. We tested our proposed architecture over a real-time cloud platform Amazon Web Services with the integration of the Fission serverless tool. The evaluation of the proposed code shipping architecture shows for a data file size of 64 MB, the latency in the proposed code shipping architecture is 8.36 ms and in existing data shipped architecture is found to be 16.8 ms. Hence, the proposed architecture achieves a speedup of 2x on the round latency for high data sizes in a serverless environment. We define round latency to be the duration to read and write back the data in the storage. © 2023, The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature.Item Geo-Distributed Multi-Tier Workload Migration Over Multi-Timescale Electricity Markets(Institute of Electrical and Electronics Engineers Inc., 2023) Addya, S.K.; Satpathy, A.; Ghosh, B.C.; Chakraborty, S.; Ghosh, S.K.; Das, S.K.Virtual machine (VM) migration enables cloud service providers (CSPs) to balance workload, perform zero-downtime maintenance, and reduce applications' power consumption and response time. Migrating a VM consumes energy at the source, destination, and backbone networks, i.e., intermediate routers and switches, especially in a Geo-distributed setting. In this context, we propose a VM migration model called Low Energy Application Workload Migration (LEAWM) aimed at reducing the per-bit migration cost in migrating VMs over Geo-distributed clouds. With a Geo-distributed cloud connected through multiple Internet Service Providers (ISPs), we develop an approach to find out the migration path across ISPs leading to the most feasible destination. For this, we use the variation in the electricity price at the ISPs to decide the migration paths. However, reduced power consumption at the expense of higher migration time is intolerable for real-time applications. As finding an optimal relocation is $\mathcal {NP}$NP-Hard, we propose an Ant Colony Optimization (ACO) based bi-objective optimization technique to strike a balance between migration delay and migration power. A thorough simulation analysis of the proposed approach shows that the proposed model can reduce the migration time by 25%-30% and electricity cost by approximately 25% compared to the baseline. © 2008-2012 IEEE.Item ESMA: Towards elevating system happiness in a decentralized serverless edge computing framework(Academic Press Inc., 2024) Datta, S.; Addya, S.K.; Ghosh, S.K.Due to the rapid growth in the adoption of numerous technologies, such as smartphones and the Internet of Things (IoT), edge and serverless computing have started gaining momentum in today's computing infrastructure. It has led to the production of huge amounts of data and has also resulted in increased network traffic, which if not managed well can cause network congestion. To address this and maintain the quality of service (QoS), in this work, a novel dispatch (destination selection) algorithm called Egalitarian Stable Matching Algorithm (ESMA) for faster data processing has been developed while also considering the best use of server resources in a decentralized Serverless-Edge environment. This will allow us to effectively utilize the enormous volumes of data that are generated. The proposed algorithm has been able to achieve lower overall dissatisfaction scores for the entire system. Individually, the client's happiness as well as the server's happiness have improved over the baseline. Moreover, there has been a drop of 25.7% in the total execution time and the total network resources consumed are lower as compared to the baseline algorithm as well as random-allocation algorithm. © 2023 Elsevier Inc.Item InDS: Intelligent DRL Strategy for Effective Virtual Network Embedding of an Online Virtual Network Requests(Institute of Electrical and Electronics Engineers Inc., 2024) Keerthan Kumar, T.G.K.; Addya, S.K.; Koolagudi, S.G.Network virtualization is a demanding feature in the evolution of future Internet architectures. It enables on-demand virtualized resource provision for heterogeneous Virtual Network Requests (VNRs) from diverse end users over the underlying substrate network. However, network virtualization provides various benefits such as service separation, improved Quality of Service, security, and more prominent resource usage. It also introduces significant research challenges. One of the major such issues is allocating substrate network resources to VNR components such as virtual machines and virtual links, also named as the virtual network embedding, and it is proven to be mathbb {N}mathbb {P} -hard. To address the virtual network embedding problem, most of the existing works are 1) Single-objective, 2) They failed to address dynamic and time-varying network states 3) They neglected network-specific features. All these limitations hinder the performance of existing approaches. This work introduces an embedding framework called Intelligent Deep Reinforcement Learning (DRL) Strategy for effective virtual network embedding of an online VNRs (InDS). The proposed InDS uses an actor-critic model based on DRL architecture and Graph Convolutional Networks (GCNs). The GCN effectively captures dependencies between the VNRs and substrate network environment nodes by extracting both network and system-specific features. In DRL, the asynchronous advantage actor-critic agents can learn policies from these features during the training to decide which virtual machines to embed on which servers over time. The actor-critic helps in efficiently learning optimal policies in complex environments. The suggested reward function considers multiple objectives and guides the learning process effectively. Evaluation of simulation results shows the effectiveness of InDS in achieving optimal resource allocation and addressing diverse objectives, including minimizing congestion, maximizing acceptance, and revenue-to-cost ratios. The performance of InDS exhibits superiority in achieving 28% of the acceptance ratio and 45% of the revenue-to-cost ratio by effectively managing the network congestion compared to other existing baseline works. © 2013 IEEE.Item Optimizing Completion Time of Requests in Serverless Computing(Springer, 2024) Sherawat, A.; Nath, S.B.; Addya, S.K.Serverless computing offers people with the liberty of not thinking about the backend side of the things in an application development. They are scalable and cost efficient as they provide pay-for-use service. Providing acceptable performance while having no knowledge about the kind of application is the main challenge the cloud providers have. Many applications may have the need to be completed before the deadline. In that case, the request has to be completed before the deadline or else it will lead to service level agreement violation. If the cloud provider completes the requests faster, there would be less SLA violations. This will also reduce cost for the user as the functions will be completed sooner. Therefore, improving the completion time of the requests will benefit the user as well as the provider. In this paper, we present a method to improve the completion time of requests using genetic algorithm for allocation of requests to virtual machines that could provide optimal completion time for them. © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2024.Item FASE: fast deployment for dependent applications in serverless environments(Springer, 2024) Saha, R.; Satpathy, A.; Addya, S.K.Function-as-a-service has reduced the user burden by allowing cloud service providers to overtake operational activities such as resource allocation, service deployment, auto-scaling, and load-balancing, to name a few. The users are only responsible for developing the business logic through event-triggered functions catering to an application. Although FaaS brings about multiple user benefits, a typical challenge in this context is the time incurred in the environmental setup of the containers on which the functions execute, often referred to as the cold-start time leading to delayed execution and quality-of-service violations. This paper presents an efficient scheduling strategy FASE that uses a finite-sized warm pool to facilitate the instantaneous execution of functions on pre-warmed containers. Test-bed evaluations over AWS Lambda confirm that FASE achieves a 40% reduction in the average cold-start time and 1.29× speedup compared to the baselines. © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2023.Item EFraS: Emulated framework to develop and analyze dynamic Virtual Network Embedding strategies over SDN infrastructure(Elsevier B.V., 2024) Keerthan Kumar, K.K.; Tomar, S.; Addya, S.K.; Satpathy, A.; Koolagudi, S.G.The integration of Software-Defined Networking (SDN) into Network Virtualization (NV) significantly enhances network management, isolation, and troubleshooting capabilities. However, it brings forth the intricate challenge of allocating Substrate Network (SN) resources for various Virtual Network Requests (VNRs), a process known as Virtual Network Embedding (VNE). It encompasses solving two intractable sub-problems: embedding Virtual Machines (VMs) and embedding Virtual Links (VLs). While the research community has focused on formulating embedding strategies, there has been less emphasis on practical implementation at a laboratory scale, which is crucial for comprehensive design, development, testing, and validation policies for large-scale systems. However, conducting tests using commercial providers presents challenges due to the scale of the problem and associated costs. Moreover, current simulators lack accuracy in representing the complexities of communication patterns, resource allocation, and support for SDN-specific features. These limitations result in inefficient implementations and reduced adaptability, hindering seamless integration with commercial cloud providers. To address this gap, this work introduces EFraS (Emulated Framework for Dynamic VNE Strategies over SDN). The goal is to aid developers and researchers in iterating, testing, and evaluating VNE solutions seamlessly, leveraging a modular design and customized reconfigurability. EFraS offers various functionalities, including generating real-world SN topologies and VNRs. Additionally, it integrates with a diverse set of evaluation metrics to streamline the testing and validation process. EFraS leverages Mininet, Ryu controller, and OpenFlow switches to closely emulate real-time setups. Moreover, we integrate EFraS with various state-of-the-art VNE schemes, ensuring the effective validation of embedding algorithms. © 2024 Elsevier B.V.
