Bodhisattwa P. Majumder
Office @ 4146, CSE (EBU3B)
UC San Diego

I am a 3rd year Ph.D. student at the Artificial Intelligence Group, Computer Science, UC San Diego, advised by Prof. Julian McAuley. I am interested in Natural Language Generation, Conversational AI, and ML Interpretability to build empathetic, commonsensical and personalized interactive systems.

I am a Qualcomm Innovation Fellow in North America for 2020-21. In 2019-20, I led Team Bernard from UC San Diego in Amazon Alexa Prize.

So far, I spent wonderful summers at:
['20] Microsoft Research, NLP Group working with Sudha Rao and Michel Galley on clarifying question generation estimating missing context;
['19] Google AI Research with Sandeep Tata and Marc Najork on information extraction using representation learning from templatic documents.

Previously, I graduated (2017) summa cum laude from IIT Kharagpur, MS in Machine Learning. I wrote a book on Practical NLP, published by O'Reilly Media.

By Note-to-Self
Museum of Photographic Arts (MOPA)
A fusion with piano notes and digital image

CV  |  Google Scholar  |  Github
LinkedIn  |  Twitter

Publications·Experiences·Awards·Education·Book·Invited Talks

  • [Feb] Excited to be featured by Jacobs School of Engineering, UC San Diego, for our QIF Fellowship!
  • [Jan] Launched GEM Benchmark (shared task in ACL 2021) for evaluation in Natural Language Generation tasks!
  • [Jan] We are organizing SoCal ML & NLP symposium 2021 virtually! Please consider submitting by Feb 16, 2021.
  • [Jan] Joining Facebook AI Research for Summer 2021 to work with Y-Lan Boureau on Language Generation.
  • [Oct] Invited talk on Achieving Commonsense in Text Generation at NC State. See slides here.
  • [Sep] Two long papers (#1, #2) w/ Harsh, Taylor, Shuyang, Jianmo, and Julian got accepted in EMNLP (main), 2020.
  • [Aug] Received Qualcomm Innovation Fellowship 2020 for our proposal on Conversational Recommender Systems.
  • [July] Our book Practical Natural Language Processing has become #1 best seller in Amazon! Know more here.
  • [June] Excited that my internship work at Google got featured in the Google AI blog! Check out for more.
  • [April] Work at Google AI got accepted in ACL 2020 as a long paper w/ Navneet, Sandeep, James, Qi and Marc.
  • [Mar] New work on making deeper networks faster (ReZero) w/ Thomas, Henry, Gary and Julian.
  • [Feb] Organizing SoCal Machine Learning Symposium, 2020 w/ Julian, Jingbo and Hao at UC San Diego.
  • [Jan] Invited talk on Personalized NLG in the AI/ML track at CSE Research Open House, UC San Diego.
2019 2018
  • [Sept] Joined the NLP group at CSE, UC San Diego in Fall 2018.
  • [July] Paper w/ Amrith Krishna, Rajesh Bhat and Pawan Goyal got published in CoNLL, 2018.

Here in xkcd.


My research aims to achieve empathy, subjectivity and personalization while interpreting machine predictions. While current ML interpretability techniques fails to be accessible to non-expert users, I envision my research will successfully bridge the gap by marrying ideas from (a) interpretable ML; (b) language generation; and (c) dialog and interactive systems. In this regard, I'm interested in developing generative models that are personalized, capable of reasoning commonsense and world events, and providing subjective knowledge -- broadly connecting to an interactive system.

My previous research on NLP includes information extraction, sequence labeling, sequence generation, and natural language parsers. I also worked on statistical modeling, game theory, and machine learning applications. Eternally, I am inspired by the question that Richard Feynman once asked, can machines think?

My selected publications are listed here. The complete list of publications can be seen from my Google Scholar page.

Selected Publications
(* denotes equal contribution)

[13] The GEM Benchmark: Natural Language Generation, its Evaluation and Metrics
GEM team
Shared task at GEM workshop, ACL 2021
pdf | website

GEM is a community-driven effort with the goal to improve how progress in natural language generation is measured. As a shared task in ACL 2021, we invite for challenge set submissions for 11 datasets and 7 languages in various NLG challenges.

[12] Like hiking? You probably enjoy nature: Persona-grounded Dialog with Commonsense Expansions
Bodhisattwa P. Majumder, Harsh Jhamtani, Taylor Berg-Kirkpatrick, Julian McAuley
Empirical Methods in Natural Language Processing (EMNLP), 2020
pdf | code | slides

A variational learning framework to capture commonsense implications of input persona in a persona-grounded dialog agent using richer expansions obtained from existing commonsense knowledge bases.

[11] Interview: Large-scale Modeling of Media Dialog with Discourse Patterns and Knowledge Grounding
Bodhisattwa P. Majumder*, Shuyang Li*, Jianmo Ni, Julian McAuley
Empirical Methods in Natural Language Processing (EMNLP), 2020
pdf | code | data

The first large-scale analysis of discourse in media dialog ("Interview" - 105K conversations) and its impact on generative modeling of dialog turns, with a focus on interrogative patterns and use of external knowledge.

[10] Bernard: A Stateful Neural Open-domain Socialbot
Bodhisattwa P. Majumder, Shuyang Li, Jianmo Ni, Henry Mao, Sophia Sun, Julian McAuley
3rd Proceedings of Alexa Prize, Amazon Alexa Prize 2019-2020

A framework for an engaging open-domain socialbot with a stateful autonomous dialog manager using non-deterministic finite automata to control multi-turn conversations. This work was done for Alexa Prize 2019.

[9] Representation Learning for Information Extraction from Form-like Documents
Bodhisattwa P. Majumder, Navneet Potti, Sandeep Tata, James Wendt, Qi Zhao, Marc Najork
Annual Conference of the Association for Computational Linguistics (ACL), 2020
pdf | blog | slides

A novel approach to learn interpretable representations for target fields using spatial and contextual knowledge for extracting structured information from form-like document images, even with unseen templates. This work was done at Google AI as a part of 2019 summer internship.

[8] ReZero is All You Need: Fast Convergence at Large Depth
Thomas Bachlechner*, Bodhisattwa P. Majumder*, Henry Mao*, Gary Cottrell, Julian McAuley
Preprint. Work In Progress. arXiv, 2020
pdf | code

A novel deep neural network architecture that initializes an arbitrary layer as the identity map (ReZero), using a single additional learned parameter per layer to facilitate very deep signal propagation.

[7] Generating Personalized Recipes from Historical User Preferences
Bodhisattwa P. Majumder*, Shuyang Li*, Jianmo Ni, Julian McAuley
Conference on Empirical Methods in Natural Language Processing (EMNLP), 2019
pdf | code | data | poster

Media coverage: Science Node, UCSD CSE News, UCSD JSOE News

A new task of personalized recipe generation to help these users: expanding a name and incomplete ingredient details into complete natural-text instructions aligned with the user's historical preferences.

[6] Improving Neural Story Generation by Targeted Common Sense Grounding
Henry Mao, Bodhisattwa P. Majumder, Julian McAuley, Gary Cottrell
Conference on Empirical Methods in Natural Language Processing (EMNLP), 2019
pdf | code

A multi-task learning scheme to achieve quantitatively better common sense reasoning in language models by leveraging auxiliary training signals from datasets designed to provide common sense grounding.

[5] Upcycle Your OCR: Reusing OCRs for Post-OCR Text Correction in Romanised Sanskrit
Amrith Krishna, Bodhisattwa P. Majumder, Rajesh S. Bhat, Pawan Goyal
Conference on Computational Natural Language Learning (CoNLL), co-located with EMNLP, 2018
pdf | code+data | supplementary

A state-of-the-art approach towards post-OCR text correction for digitising texts in Romanised Sanskrit. This work was done in a collaboration with CNeRG.

[4] An 'Eklavya' approach to learning Context Free Grammar rules for Sanskrit using Adaptor Grammar
Amrith Krishna, Bodhisattwa P. Majumder, Anil K. Boga, Pawan Goyal
17th World Sanskrit Conference , 2018

A non-parametric Bayesian approach for learning (Probabilistic) Context Free Grammar productions for Sanskrit language at word-level supervised tasks such as compound type identification, identification of source and derived words from the corpora for derivational nouns and sentence-level structured prediction. This work was done at CNeRG.

[3] Deep Recurrent Neural Networks for Product Attribute Extraction in eCommerce
Bodhisattwa P. Majumder*, Aditya Subramanian*, Abhinandan Krishnan, Shreyansh Gandhi, Ajinkya More
ArXiv , 2017
pdf | system description | video

We demonstrate the potential of neural recurrent structures in product attribute extraction by improving overall F1 scores, as compared to the previous benchmarks. This has made Walmart e-commerce achieve a significant coverage of important facets or attributes of products. This work at Walmart Labs was followed by a US patent.

[2] Distributed Semantic Representations of Retail Products based on Large-scale Transaction Logs
Bodhisattwa P. Majumder*, Sumanth S Prabhu*, Julian McAuley

We processed 18 million transactions consisting of unique 325,548 products from 1,551 categories to obtain vector representations which preserve product analogy. These representations were effective in identifying substitutes and complements. This work was done at Walmart Labs.

[1] When lolcats meet philosoraptors! - What's in a 'meme'?
Bodhisattwa P. Majumder, Amrith Krishna, Unni Krishnan, Anil K. Boga, Animesh Mukherjee
Arxiv , 2018
pdf | presentation

How similar are the dynamics of meme based communities to that of text based communities? We try to explain the community dynamics by categorising each day based on temporal variations in the user engagement. This work was done in a collaboration with CNeRG.

  • REDCLAN - RElative Density based CLustering and Anomaly Detection, Wal-mart, 2018
  • Automated Extraction of Product Attributes from Images, Wal-mart, 2018
  • System and Method for Product Attribute Extraction Using a Deep Recurrent System, Wal-mart, 2017
  • Analytical Determination of Competitive Interrelationship between Item Pairs, Wal-mart, 2017

Microsoft Research, Redmond
Summer, 2020
Research Intern with Sudha Rao and Michel Galley at Natural Language Processing Group.

Developing a novel framework that estimates missing 'local' information from the knowledge of a closed-world to generate a follow-up clarification question.


Amazon Alexa Prize
Team Leader of Bernard, UC San Diego.

Media Coverage: cnet
Building free-form social conversational agent as a finalist in the Amazon Alexa Prize Challenge 2019-2020 along with 9 other finalist universities. We have been awarded $250,000 to support our research on dialog systems.


Google AI, Mountain View
Summer, 2019
Research Intern with Sandeep Tata and Navneet Potti from Team Juicer.

Media Coverage: Google AI blog, Google Engineering Newsletter (Intern Spotlight)
Developed an Information Extraction Framework for form-like documents using representation learning. The work was published as an Intern spotlight article in the Google-wide Newsletter and is being integrated with Google Cloud's Document AI. Our work got accepted as a long paper in ACL '20.


Walmart Labs
Research Engineer

Developed a neural multimodal attribute tagging framework to improve faceted product using both product description and product images. The work produced 2 US patents and one technical report published in arXiv. Other works on user modeling and product embeddings also have been patented.

  • Shivam Lakhotia, MS, CSE @ UCSD
  • Maximilian Halvax and Tatum Maston, Undergraduate, HDSI @ UCSD, as a part of HDSI scholar program
  • [2020] Winner of Qualcomm Innovation Fellowship, 2020 for North America
  • [2020] Nominated by UC San Diego (one of three from Dept. of CSE) for Microsoft Research PhD Fellowhip 2021
  • [2019] Nominated by UC San Diego (one of two from Dept. of CSE) for Google PhD Fellowhip 2020
  • [2019] Intern Spotlight in Google-wide Engineering Newsletter for summer internship project with the Juicer Team
  • [2019] Team Leader for Team Bernard represeting UC San Diego, a finalist in Alexa Prize 2019; awarded $250,000
  • [2018] Department Fellowship, 1st-year of PhD, Dept. of CSE, UC San Diego
  • [2017] Gold medal and Endowment for the highest academic performance (Rank-1) in Masters, IIT Kharagpur
  • [2016] Finalist, Data Science Game '16, Paris; Represented India (1 out of 3 teams), International Rank 14
  • [2015] Scholarship for academic excellence (obtaining CGPA > 9.5), Indian Statistical Institute
  • [2014] Officially entitled as contributor in NSF-CPS project (CNS -1136040) by PIs, Kansas State University
  • [2011] 4-year scholarship for academic excellence, Ministry of Human Resource & Development, India

PhD, Computer Science and Engineering
University of California, San Diego

Advised by Prof. Julian McAuley on Achieving Conversational Interpretability for Machine Predictions.


MS, Computer Science and Engineering
University of California, San Diego

CGPA: 4.0; Courses: Intro to NLP, Data Mining, Program Synthesis, Deep Learning for Sequences, Probabilistic Reasoning, Intro to Computer Vision, Convex Optimization, Human-centered Programming


MS, Data Science and Machine Learning
Indian Institute of Technology, Kharagpur

Summa cum laude (Gold Medalist); Advised by Prof. Animesh Mukherjee as a part of CNeRG lab. Courses: Algorithms, Intro to ML, Multivariate Analysis, Complex Networks, Information Retrieval

Book: Practical NLP by O'Reilly

Practical Natural Language Processing
O'Reilly Media, 2020
Sowmya Vajjala, Bodhisattwa P. Majumder, Anuj Gupta, Harshit Surana
amazon | safari online | website

Practical Natural Language Processing distills our collective wisdom on building real world applications such as data collection, working with noisy data and signals, incremental development of solutions, and issues involved in deploying the solutions as a part of a larger application - bridging a gap between current textbooks and online offerings.


  • Endorsed by Zach Lipton, Sebastian Ruder, Marc Najork et al.
  • #1 Best seller in in Data Mining category
  • #1 New release in in Natural Language Processing category
  • Read and adapted by 20+ AI companies and 6 academic courses internationally

Invited Talks
  • [2020] at NC State, AI club on Achieving Commonsense in Text Generation (slides)
  • [2020] at INFORMS 2020, Mining and Learning on Graphs session in Washington, DC (slides)
  • [2020] at UC San Diego, CSE Research Open House, on Personalization in Natural Language Generation
  • [2018] at Indian Inst of Management Calcutta, Industry Conclave & Graduate Orientation, on NLP - a primer
  • [2017] at Walmart Labs, on Information Extraction from Images - Application in e-Commerce
  • [2017] at Indian Statistical Institute, on Deep Neural Network: in light of Optimization and Regularization

Thanks to Jon Barron for this nice template! Art by Bekin M ~