CMPT 839 (Spring 2025): Advanced Natural Language Processing and Understanding

Overview

This course is a graduate-level research course covering advanced topics in NLP, introducing the state-of-the-art methods for computational understanding, analysis, and generation of natural language text. In recent years, advances in deep learning models for NLP has transformed the ability of computers to converse with human using language, giving us multi-lingual, multi-models that are capable of answering questions, composing message, translating and summarizing documents. The development of large language models (LLMs) are built on top of neural models such as transformers, that allows for the scaling up of models and training with large amounts of data. In this course, we will focus on current state-of-the-art methods in NLP including how to do parameter efficient fine-tuning, techniques for scaling models to long sequences, etc. We will also go beyond transformers to learn alternative architectures such as state-space models.

Students are expected to have prior experience with deep learning concepts and framework (Pytorch, Tensorflow, etc), and should also be familiar with basic natural language processing.

Each week, students will read papers in a particular area of natural language processing, and discuss the contributions, limitations and interconnections between the papers. Students will also work on a research project during the course, culminating in a final presentation and written report. The course aims to provide practical experience in comprehending, analyzing and synthesizing research in natural language processing and understanding.

Note: This course is NOT an introductory course to natural language processing. If you are interested in taking a introductory course about natural language processing, please take CMPT 413/713.

Background

There are no formal prerequisites for this class. However, you are expected to be familiar with the following:

Topics

Quick info

Syllabus

Below is a tentative outline for the course.

R: Readings, BG: (Optional) Background material / reading for deeper understanding. Provided for reference.

Date Topic Notes
Jan 6 Introduction, NLP review & logistics BG: Stanford CS324 intro to LLMs
Jan 8 Transformers and LLMs BG: Attention is all you need
BG: LLaMA
Jan 13 Training LLMs (pre-training, instruction tuning) BG: Instruction tuning survey
Jan 15 Paper discussion 0  
Jan 20 Training LLMs (PEFT, efficient training)  
Jan 22 Ensembling and mixture-of-experts BG: Review of mixture of experts
Jan 27 Using LLMs (Prompting, generation and retrieval)
Potential project topics
BG: Prompting survey
BG: ACL 2023 Text generation tutorial
BG: ACL 2023 RAG tutorial
Jan 29 Paper discussion 1  
Feb 3 Efficient LLMs: Pruning, quantization, distillation BG: Inference optimization (Lillian Weng)
Feb 5 Handling long context R: UniMem
Feb 10 State space models R: Mamba
Feb 12 Paper discussion 2  
Feb 17 No class - Reading break  
Feb 19 No class - Reading break  
Feb 24 Reasoning  
Feb 26 Project proposals Due: Project proposal
Mar 3 Multimodal models  
Mar 5 Paper discussion 3  
Mar 10 LLM agents  
Mar 12 Paper discussion 4  
Mar 17 LLM attacks and social impact  
Mar 19 Project milestone presentations Due: Project milestone
Mar 24 Guest lecturer  
Mar 26 Guest lecturer  
Mar 31 Interpreting LLMs: what is learned in these models?  
Apr 2 Paper discussion 5  
Apr 7 Project presentations  
Apr 9 Project presentations + Conclusion  
Apr 10   Due: Project final report

Grading

General policies

Academic integrity

SFU’s Academic Integrity web site is filled with information on what is meant by academic dishonesty, where you can find resources to help with your studies and the consequences of cheating. Check out the site for more information and videos that help explain the issues in plain English.

Each student is responsible for his or her conduct as it affects the University community. Academic dishonesty, in whatever form, is ultimately destructive of the values of the University. Furthermore, it is unfair and discouraging to the majority of students who pursue their studies honestly. Scholarly integrity is required of all members of the University. Please refer to this web site.