Education
PhD, Machine and Language Learning Lab (MALL Lab), 2016 – 2022 (Defended April 2023)
Dept. of Computer Science & Automation (CSA), Indian Institute of Science, Bangalore
Masters in Computer Science, 2014 – 2016
Chennai Mathematical Institute (CMI), Chennai
B.Tech, 2008 – 2012
Dept. of Mechanical Engineering, Indian Institute of Technology (IIT) , Kharagpur
Work Experience
Amazon Search Science Team, Amazon, Bangalore, August 2022 – Present
Applied Scientist II
Amazon Lex Team, AWS AI, Seattle (Remote), July 2020 – Oct 2020
Applied Scientist Intern
Linear Algebra and its applications, Indian Institute of Science, Bangalore, Aug 2017 – Dec 2017
Teaching Assistant
Amazon Core Machine Learning Team, Amazon, Bangalore, June 2017 – Aug 2017
Applied Science Intern
Machine Learning Team, Real Image Media Tech., Chennai, March 2016 – June 2016
Intern
Product Engineering, eGain Communications Pvt. Ltd., Pune, 2012 – 2014
Software Engineer
PhD Thesis
Inducing Constraints in Paraphrase Generation and Consistency in Paraphrase Detection
Thesis
slides
Publications
- Striking a Balance: Alleviating Inconsistency in Pre-trained Models for Symmetric Classification Tasks
Ashutosh Kumar, Aditya Joshi
Accepted as a Short Paper in the Findings of Association for Computational Linguistics (ACL) 2022
abstract paper codeWhile fine-tuning pre-trained models for downstream classification is the conventional paradigm in NLP, often task-specific nuances may not get captured in the resultant models. Specifically, for tasks that take two inputs and require the output to be invariant of the order of the inputs, inconsistency is often observed in the predicted labels or confidence scores. We highlight this model shortcoming and apply a consistency loss function to alleviate inconsistency in symmetric classification. Our results show an improved consistency in predictions for three paraphrase detection datasets without a significant drop in the accuracy scores. We examine the classification performance of six datasets (both symmetric and non-symmetric) to showcase the strengths and limitations of our approach. - Syntax-guided Controlled Generation of Paraphrases
Ashutosh Kumar, Kabir Ahuja, Raghuram Vadapalli, Partha Talukdar
Transactions of the Association for Computational Linguistics (TACL) 2020
Presented at ACL 2020
abstract paper code slidesGiven a sentence (e.g., "I like mangoes") and a constraint (e.g., sentiment flip), the goal of controlled text generation is to produce a sentence that adapts the input sentence to meet the requirements of the constraint (e.g., "I hate mangoes"). Going beyond such simple constraints, recent works have started exploring the incorporation of complex syntactic-guidance as constraints in the task of controlled paraphrase generation. In these methods, syntactic-guidance is sourced from a separate exemplar sentence. However, these prior works have only utilized limited syntactic information available in the parse tree of the exemplar sentence. We address this limitation in the paper and propose Syntax Guided Controlled Paraphraser (SGCP), an end-to-end framework for syntactic paraphrase generation. We find that SGCP can generate syntax conforming sentences while not compromising on relevance. We perform extensive automated and human evaluations over multiple real-world English language datasets to demonstrate the efficacy of SGCP over state-of-the-art baselines. To drive future research, we have made SGCP’s source code available. - Submodular Optimization-based Diverse Paraphrasing and its Effectiveness in Data Augmentation
Ashutosh Kumar*, Satwik Bhattamishra*, Manik Bhandari, Partha Talukdar
Oral Presentation : Long paper in NAACL 2019, Minneapolis, USA
*Equal Contributions
abstract paper code slidesInducing diversity in the task of paraphrasing is an important problem in NLP with applications in data augmentation and conversational agents. Previous paraphrasing approaches have mainly focused on the issue of generating semantically similar paraphrases, while paying little attention towards diversity. In fact, most of the methods rely solely on top-k beam search sequences to obtain a set of paraphrases. The resulting set, however, contains many structurally similar sentences. In this work, we focus on the task of obtaining highly diverse paraphrases while not compromising on paraphrasing quality. We provide a novel formulation of the problem in terms of monotone submodular function maximization, specifically targeted towards the task of paraphrasing. Additionally, we demonstrate the effectiveness of our method for data augmentation on multiple tasks such as intent classification and paraphrase recognition. In order to drive further research, we have made the source code available. - eCommerceGAN : A Generative Adversarial Network for E-commerce
Ashutosh Kumar, Arijit Biswas, Subhajit Sanyal
Accepted as a Workshop paper in ICLR 2018, Vancouver, Canada
abstract paper mediaE-commerce companies such as Amazon, Alibaba and Flipkart process billions of orders every year. However, these orders represent only a small fraction of all plausible orders. Exploring the space of all plausible orders could help us better understand the relationships between the various entities in an e-commerce ecosystem, namely the customers and the products they purchase. In this paper, we propose a Generative Adversarial Network (GAN) for orders made in e-commerce websites. Once trained, the generator in the GAN could generate any number of plausible orders. Our contributions include: (a) creating a dense and low-dimensional representation of e-commerce orders, (b) train an ecommerceGAN (ecGAN) with real orders to show the feasibility of the proposed paradigm, and (c) train an ecommerce-conditional-GAN (ec2GAN) to generate the plausible orders involving a particular product. We propose several qualitative methods to evaluate ecGAN and demonstrate its effectiveness. The ec2GAN is used for various kinds of characterization of possible orders involving a product that has just been introduced into the e-commerce system. The proposed approach ec2GAN performs significantly better than the baseline in most of the scenarios. - NL-Augmenter: A Framework for Task-Sensitive Natural Language Augmentation
Kaustubh D. Dhole et. al, (includes: Ashutosh Kumar )
Arxiv pre-print 2021
abstract paper codeData augmentation is an important component in the robustness evaluation of models in natural language processing (NLP) and in enhancing the diversity of the data they are trained on. In this paper, we present NL-Augmenter, a new participatory Python-based natural language augmentation framework which supports the creation of both transformations (modifications to the data) and filters (data splits according to specific features). We describe the framework and an initial set of 117 transformations and 23 filters for a variety of natural language tasks. We demonstrate the efficacy of NL-Augmenter by using several of its transformations to analyze the robustness of popular natural language models. The infrastructure, datacards and robustness analysis results are available publicly on the NL-Augmenter repository. - Discovering Non-Monotonic Autoregressive Ordering for Text Generation Models using Sinkhorn Distributions
Ashutosh Kumar
ICLR Blog Post 2022
abstract blogIn this post, we will be focusing on one of the lesser explored area - Non-Monotonic Autoregressive Order (NMAO) in Decoding. In particular, we will be discussing the ICLR 2021 paper “Discovering Non-monotonic Autoregressive Orderings with Variational Inference” [Li 2021], the background needed for understanding the work as well as the way-forward.
Service
Reviewer
2020: ICLR, ACL, EMNLP [Outstanding Reviewer]
2021: ICLR, EACL, NAACL, ACL-IJCNLP, EMNLP, NeurIPS
2022: ICLR, ACL ARR, NeurIPS
2023: ICLR, ACL ARR, EMNLP, NeurIPS
Judge
Initiative for Research & Innovation in STEM (IRIS) National Fair 2020
Contact
Address
Amazon Development Center India
Bagmane Constellation Business Park,
Bangalore - 560048
firstname"k2401"[at]gmail[dot]com