CIS 607    

    Seminar on Deep Learning for Natural Language Processing    

Course Description

Deep learning has recently proved itself as a powerful branch of machine learning that has applications in a variety of domains (i.e., computer vision, robotics, etc.). In natural language processing (NLP), deep learning has been transformative and led to a new generation of methods with better performance, portability and robustness (e.g., machine translation, text generation/dialog systems). Especially, with the recent breakthrough in pretrained language models (e.g., BERT, GPT), building state-of-the-art NLP models can be done efficiently for different domains and languages. All of those advances are very recent and the demand for data scientists with deep learning expertise is growing very quickly. At the beginning of this seminar, we will cover the basic concepts of deep learning and NLP, and possibly provide some hand-on experience to implement the models. Afterward, we will review and discuss a collection of research papers on NLP with deep learning, including but not limited to the typical tasks of language modeling, question answering, information extraction, machine translation, natural language inference, dialog, summarization, domain adaptation, transfer learning, and multilingual learning. NLP is growing fast these days and we expect to read many exciting recent papers in this field.

Instructor

Thien Huu Nguyen, [email protected]

Lectures

One 80-minute presentation and discussion session is delivered each week.
Zoom: Lecture
For June 1, please use this Zoom link: Link

Prerequisites

Textbooks and Readings

Major Topics

Expected Learning Outcomes

In this seminar, we will review and discuss a collection of research papers on NLP with deep learning, including but not limited to the typical tasks of language modeling, question answering, information extraction, natural language inference, dialog and summarization.

Upon successful completion of the course, students will be able to:

Acquired Skills

Upon successful completion of the course, students will have acquired the following skills:

Class Organization

Each student in the class will need to choose some paper(s) on particular topics, present them in one of the classes and lead the discussion on the topic.

Each presentation will be given 30 minutes along with 5-10 minutes for discussions. More discussion is encouraged on Piazza.

After the presentation, the presenter should submit a summary about the presented paper/topic. The summary should follow the NAACL format (a.k.a. ACL style for 8.5x11 paper). The required length of the summary is between 2 and 3 pages.

For each presentation, we will have one student (other than the presenter) to serve as the reviewer. The role of the reviewer is to provide judgement/comments/suggestion or ask questions about the paper/topic in the discussion time after the presentation. Although all students need to read the papers being presented before each class to be able to actively contribute to the discussions, the reviewer would help to provide deeper judgement by reading and thinking critically about the papers/topics ahead of time.

IMPORTANT:

Please select a paper you want to present in this list. Write your name next to the papers you select (for presentations and reviews). You are welcome to choose another paper that is not in the list. Please talk with the instructor if you want to do this. All the paper assignments should be done before April 4 (no later than that) so we can schedule the presentations.

Tentative Schedule

Please sign up on Piazza for discussions.
Week Topics Presenter Slides Reviewer
03/30 Introduction to Modern Natural Language Processing, Data Augmentation for NLP Thien Link, Link
04/06 Event Extraction by Answering (Almost) Natural Questions (EMNLP 2020) Minh Link -
CLIP-Event: Connecting Text and Images with Event Structures (2022) Amir Link -
04/13 mT6: Multilingual Pretrained Text-to-Text Transformer with Translation Pairs (EMNLP 2021) Viet Shravan
Focus on what matters: Applying Discourse Coherence Theory to Cross Document Coreference (EMNLP 2021) Masaki Link Ali
04/20 Want To Reduce Labeling Cost? GPT-3 Can Help (EMNLP 2021) Manish Ayushi
Revisiting Self-Training for Few-Shot Learning of Language Model (EMNLP 2021) Jared Masaki
04/27 Deduplicating Training Data Mitigates Privacy Risks in Language Models (2022) Zayd -
Generation-Augmented Retrieval for Open-Domain Question Answering (ACL 2021) Qiuhao Amir
05/04 Theoretical Analysis of Self-Training with Deep Networks on Unlabeled Data (ICLR 2021) Nghia Zayd
Why Do Pretrained Language Models Help in Downstream Tasks? An Analysis of Head and Prompt Tuning (NeurIPS 2021) Nghia Jared
05/11 Contrastive Explanations for Model Interpretability (EMNLP 2021) Steven Nghia
SentencePiece: A simple and language independent subword tokenizer and detokenizer for Neural Text Processing (EMNLP 2018)
Vocabulary Learning via Optimal Transport for Neural Machine Translation (ACL 2021)
Minh Viet
05/18 Exploring Example Selection for Few-shot Text-to-SQL Semantic Parsing (2022) Ayushi Manish
Self-Supervised Multimodal Opinion Summarization (ACL 2021) Amir Minh
05/25 Learning Transferable Visual Models From Natural Language Supervision (2021) Ali Steven
Efficient Mind-Map Generation via Sequence-to-Graph and Reinforced Graph Refinement (EMNLP 2021) Shravan Brad
06/01 ReGen: Reinforcement Learning for Text and Knowledge Base Generation using Pretrained Language Models (EMNLP 2021) Brad -
Discovering the Unknown Knowns: Turning Implicit Knowledge in the Dataset into Explicit Training Examples for Visual Question Answering (EMNLP 2021) Najmul -

Course Requirements and Grading

This course will be taught in-person. We might also stream and record the lectures over Zoom upon request. We use Canvas and Piazza for communication and discussion.

Grading will be based on P/NP.