COL873: Advanced Topics in NLP
Tuesday, Friday 2-3:20 pm in SIT 006


Instructor: Mausam
(mausam at cse dot iitd dot ac dot in)
Office hours: by appointment, SIT Building Room 402
TA: Prachi Jain
(prachi.jain at cse dot iitd dot ac dot in)
Office hours: by appointment

Course Contents

Information Extraction: NELL, Open IE, joint inference, distant supervision using graphical models, reinforcement learning for IE.

Pre-Trained Language models: Transformer, BERT, GPT2, ERNIE, T5, GPT3.

ML Techniques: Reinforcement Learning, GANs, Graph Neural Nets, Neural Module Networks, Tensor Factorization, Constrained Deep Learning.

Question Answering: open question answering, semantic parsing, reasoning, compositionality, knowledge-based reasoning.

Schedule

  1. Administrivia.
  2. Information Extraction

  3. Topic: Open IE
    Reading: Open Information Extraction from Conjunctive Sentences.
    Additional Reading: Open Information Extraction: the Second Generation.
    Additional Reading (section 2): Open Information Extraction Systems and Downstream Applications.
    [Slides]

  4. Topic: NELL
    Reading: Never-Ending Learning.
    Additional Reading: Toward an Architecture for Never-Ending Language Learning.
    [Slides]
  5. Topic: Distant supervision using probabilistic graphical models
    Reading: Multi-instance Multi-label Learning for Relation Extraction.
    Background Reading: Distant supervision for relation extraction without labeled data.
    Background Reading: Knowledge-Based Weak Supervision for Information Extraction of Overlapping Relations.
    Additional Reading: Modeling Missing Data in Distant Supervision for Information Extraction.
    [Slides]
  6. Topic: Neural distant supervision
    Reading: Robust Distant Supervision Relation Extraction via Deep Reinforcement Learning.
    Background Reading: Distant Supervision for Relation Extraction via Piecewise Convolutional Neural Networks.
    Additional Reading: Reinforcement Learning for Relation Classification from Noisy Data.
    [Slides]
  7. Explainability in NLP

  8. Topic: Attention vs. Explanation
    Reading: Attention and its (mis)interpretation.
    Additional Reading: Attention is not Explanation.
    Additional Reading: Attention is not not Explanation.
    Additional Reading: Is Attention Interpretable?.
    [Slides]
  9. Pre-Trained Language Models

  10. Topic: Transformer Architecture
    Reading: Attention is All You Need.
    [Slides]
  11. Topic: BERT
    Reading: BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding.
    Additional Reading: RoBERTa: A Robustly Optimized BERT Pretraining Approach.
    [Slides]
  12. Topic: T5
    Reading: Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer.
    [Slides]
  13. Topic: GPT3
    Reading: Language Models are Few-Shot Learners.
    [Slides]
  14. Topic: World Knowledge in Pre-trained Language Models
    Reading: Language Models as Knowledge Bases?.
    Additional Reading: BERT is Not a Knowledge Base (Yet): Factual Knowledge vs. Name-Based Reasoning in Unsupervised QA.
    [Slides]
  15. Topic: Probing Pre-trained Language Models
    Reading: Visualizing and Measuring the Geometry of BERT.
    Additional Reading: BERT rediscovers the classical NLP pipeline.
    [Slides]
  16. Machine Learning for NLP

  17. Topic: Generative Adversarial Networks
    Reading: SeqGAN: Sequence Generative Adversarial Nets with Policy Gradient.
    Background Reading: Adversarial Learning for Neural Dialogue Generation.
    [Slides]
  18. Topic: Constrained Deep Learning
    Reading: A Primal-Dual Formulation for Deep Learning with Constraints.
    [Slides]
  19. Topic: Tensor Factorization
    Reading: RotatE: Knowledge Graph Embedding by Relational Rotation in Complex Space.
    Reading: Knowledge Base Completion: Baseline Strikes Back (Again).
    [Slides]
  20. Question Answering

  21. Topic: Open Domain QA
    Reading: REALM: Retrieval-Augmented Language Model Pre-Training.
    Background Reading: Reading Wikipedia to Answer Open-Domain Questions.
    Background Reading: R3: Reinforced Ranker-Reader for Open-Domain Question Answering.
    [Slides]
  22. Topic: Reasoning for QA
    Reading: Neural Module Networks for Reasoning over Text.
    Background Reading: DROP: A reading comprehension benchmark requiring discrete reasoning over paragraphs.
    Background Reading: Neural Module Networks.
    [Slides]
  23. Topic: Multi-hop Reading Comprehension
    Reading: Select, Answer and Explain: Interpretable Multi-hop Reading Comprehension over Multiple Documents.
    Background Reading: HOTPOTQA: A Dataset for Diverse, Explainable Multi-hop Question Answering.
    Additional Reading: Hierarchical Graph Network for Multi-hop Question Answering.
    [Slides]
  24. Topic: Semantic Parsing for QA
    Reading: Memory Augmented Policy Optimization for Program Synthesis and Semantic Parsing.
    Background Reading: Compositional semantic parsing on semi-structured tables.
    [Slides]
  25. Topic: Knowledge-based QA
    Reading: Differentiable Reasoning over a Virtual Knowledge Base.
    Background Reading: PullNet: Open Domain Question Answering with Iterative Retrieval on Knowledge Bases and Text .
    [Slides]

Grading

Project: 35%; Midterm Survey: 15%; Final: 20%; Reviews: 20%, Presentation: 10%.

(Revised Grading Scheme due to Covid-19 Pandemic) Project: 22%; Midterm Survey: 18%; Final: 24%; Reviews: 24%, Presentation: 12%.