Conference Papers

Permanent URI for this collectionhttps://idr.nitk.ac.in/handle/123456789/28506

Browse

Search Results

Now showing 1 - 2 of 2
  • Item
    Natural Language Inference: Detecting Contradiction and Entailment in Multilingual Text
    (Springer Science and Business Media Deutschland GmbH, 2021) Sree Harsha, S.; Krishna Swaroop, K.; Chandavarkar, B.R.
    Natural Language Inference (NLI) is the task of characterising the inferential relationship between a natural language premise and a natural language hypothesis. The premise and the hypothesis could be related in three distinct ways. The hypothesis could be a logical conclusion that follows from the given premise (entailment), the hypothesis could be false (contradiction), or the hypothesis and the premise could be unrelated (neutral). A robust and reliable system for NLI serves as a suitable evaluation measure for true natural language understanding and enables the use of such systems in several modern day application scenarios. We propose a novel technique for the NLI task by leveraging the recently proposed Bidirectional Encoder Representations from Transformers (BERT). We utilize a robustly optimized variant of BERT, integrate a contextualized definition embedding mechanism, and incorporate the use of global average pooling into our proposed NLI system. We use several different benchmark datasets, including a dataset containing premise-hypothesis pairs from 15 different languages to systematically evaluate the performance of our model and show that it yields superior results. © 2021, Springer Nature Switzerland AG.
  • Item
    NeuralDoc-Automating Code Translation Using Machine Learning
    (Springer Science and Business Media Deutschland GmbH, 2022) Sree Harsha, S.; Sohoni, A.C.; Chandrasekaran, K.
    Source code documentation is the process of writing concise, natural language descriptions of how the source code behaves during run time. In this work, we propose a novel approach called NeuralDoc, for automating source code documentation using machine learning techniques. We model automatic code documentation as a language translation task, where the source code serves as the input sequence, which is translated by the machine learning model to natural language sentences depicting the functionality of the program. The machine learning model that we use is the Transformer, which leverages the self-attention and multi-headed attention features to effectively capture long-range dependencies and has been shown to perform well on a range of natural language processing tasks. We integrate the copy attention mechanism and incorporate the use of BERT, which is a pre-training technique into the basic Transformer architecture to create a novel approach for automating code documentation. We build an intuitive interface for users to interact with our models and deploy our system as a web application. We carry out experiments on two datasets consisting of Java and Python source programs and their documentation, to demonstrate the effectiveness of our proposed method. © 2022, The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.