Faculty Publications
Permanent URI for this communityhttps://idr.nitk.ac.in/handle/123456789/18736
Publications by NITK Faculty
Browse
Search Results
Item LCS : Alleviating Total Cold Start Latency in Serverless Applications with LRU Warm Container Approach(Association for Computing Machinery, 2023) Sethi, B.; Addya, S.K.; Ghosh, S.K.Serverless computing offers "Function-as-a-Service"(FaaS), which promotes an application in the form of independent granular components called functions. FaaS goes well as a widespread standard that facilitates the development of applications in cloud-based environments. Clients can solely focus on developing applications in a serverless ecosystem, passing the overburden of resource governance to the service providers. However, FaaS platforms have to bear the degradation in performance originating from the cold starts of executables i.e. serverless functions. The cold start reflects the delay in provisioning a runtime container that processes the functions. Each serverless platform is handling the problem of cold start with its own solution. In recent times, approaches to deal with cold starts have received the attention of many researchers. This paper comes up with an extensive solution to handle the cold start problem. We propose a scheduling approach to reduce the cold start occurrences by keeping the containers alive for a longer period of time using the Least Recently Used warm Container Selection (LCS ) approach on Affinity-based scheduling. Further, we carried out an evaluation and compared the obtained results with the MRU container selection approach. The proposed LCS approach outperforms by approximately 48% compared to the MRU approach. © 2023 ACM.Item Shipping code towards data in an inter-region serverless environment to leverage latency(Springer, 2023) Sethi, B.; Addya, S.K.; Bhutada, J.; Ghosh, S.K.Serverless computing emerges as a new standard to build cloud applications, where developers write compact functions that respond to events in the cloud infrastructure. Several cloud service industries started adopting serverless for deploying their applications. But one key limitation in serverless computing is that it disregards the significance of data. In the age of big data, when applications run around a huge volume, to transfer data from the data side to the computation side to co-allocate the data and code, leads to high latency. All existing serverless architectures are based on the data shipping architecture. In this paper, we present an inter-region code shipping architecture for serverless, that enables the code to flow from computation side to the data side where the size of the code is negligible compared to the data size. We tested our proposed architecture over a real-time cloud platform Amazon Web Services with the integration of the Fission serverless tool. The evaluation of the proposed code shipping architecture shows for a data file size of 64 MB, the latency in the proposed code shipping architecture is 8.36 ms and in existing data shipped architecture is found to be 16.8 ms. Hence, the proposed architecture achieves a speedup of 2x on the round latency for high data sizes in a serverless environment. We define round latency to be the duration to read and write back the data in the storage. © 2023, The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature.
