Improving convergence in Irgan with PPO

dc.contributor.authorJain, M.
dc.contributor.authorSowmya, Kamath S.
dc.date.accessioned2020-03-30T10:18:19Z
dc.date.available2020-03-30T10:18:19Z
dc.date.issued2020
dc.description.abstractInformation retrieval modeling aims to optimise generative and discriminative retrieval strategies, where, generative retrieval focuses on predicting query-specific relevant documents and discriminative retrieval tries to predict relevancy given a query-document pair. IRGAN unifies the generative and discriminative retrieval approaches through a minimax game. However, training IRGAN is unstable and varies largely with the random initialization of parameters. In this work, we propose improvements to IRGAN training through a novel optimization objective based on proximal policy optimisation and gumbel-softmax based sampling for the generator, along with a modified training algorithm which performs the gradient update on both the models simultaneously for each training iteration. We benchmark our proposed approach against IRGAN on three different information retrieval tasks and present empirical evidence of improved convergence. � 2020 Copyright held by the owner/author(s). Publication rights licensed to ACM.en_US
dc.identifier.citationACM International Conference Proceeding Series, 2020, Vol., , pp.328-329en_US
dc.identifier.urihttps://idr.nitk.ac.in/jspui/handle/123456789/8274
dc.titleImproving convergence in Irgan with PPOen_US
dc.typeBook chapteren_US

Files

Original bundle

Now showing 1 - 1 of 1
Thumbnail Image
Name:
13 Improving Convergence in IRGAN with PPO.pdf
Size:
622.69 KB
Format:
Adobe Portable Document Format