Link Search Menu Expand Document

CS662

Advanced Natural Language Processing

Staff

Instructor

Jonathan May

Office Hours: Mondays 12:00-1:00 pm TBD and Wednesdays 9:00–9:50 am TBD, or by appointment (at ISI on other days)

Teaching Assistant

Alexander Spangher

spangher@usc.edu

Office Hours: 12-2pm Monday, RTH 314

Lectures

  • Monday and Wednesday 10:00–11:50 am, DMC 261
  • See schedule for select days where class is canceled

Textbook

Grading

PercentageAssessment Component
10%In class participation
10%Posted questions before each in-class selected paper presentation and possible quizzes
10%In-class selected paper presentation
30%Three Homeworks (10% each)
40%Project, done in small groups, comprising:
 - Proposal (5%)
 - First version of report (5%)
 - In-class presentation (10%)
 - Final report (20%).
  • Written homeworks and project components except for final project report must be submitted on the date listed in the schedule, by 23:59:59 AoE.
  • Final project report is due Monday, December 16, 2024, 10:00 AM PST
  • A deduction of 1/5 of the total possible score will be assessed for each late day. After four late days (i.e. on the fifth), you get a 0 on the assignment (and you should come talk to us because your grade will likely suffer!)
  • You have four extension days, to be applied as you wish, throughout the entire class, for homeworks and project proposal / first report (NOT final report). No deduction will be assessed if an extension day is used. As an example, if an assignment is due November 10, you have two extension days remaining, you submit the assignment on November 12, and your score is 90/100. In this case you lose the extension days but your grade is not reduced; it remains 90/100. If you have one extension day, you lose it, and your grade is 70/100. If you have no extension days, your grade is 50/100.

Contact us

On Slack, or in class/office hours. Please do not email (unless notified otherwise).

Topics

(subject to change per instructor/class whim) (will not necessarily be presented in this order):
Fundamentals
Linguistic Stack (graphemes/phones - words - syntax - semantics - pragmatics - discourse
Corpora, Corpus statistics, Data cleaning, munging, and annotation
Evaluation
Linear and Nonlinear Models
Dense Representations and neural architectures (feed-forward, RNN, Transformer)
Language Models
Pre-training, Fine-tuning, Prompting, Reward Alignment
Ethics
Effective written and oral communication
Applications
Multilingualism and Translation
Syntax
Information Retrieval/Question Answering
Dialogue
Information Extraction
Multimodality
Speech Recognition and Generation
Agent Interaction
Discourse

Week 1

Week 2

Week 3

Sep 9
NO CLASS
Sep 11
Non-linear Classifiers, Backprop, Gradient Descent
E 3. JM 7.2–7.5 :
Sep 13
Early Drop (no W, refund)

Week 4

Week 5

Week 6

Week 7

Week 8

Week 9

Week 10

Week 11

Nov 4
Multimodal NLP (Guest Lecture by Xuezhe Ma)
Huang, Run - Assisting in Writing Wikipedia-like Articles From Scratch with Large Language Models
Questions by: Eric Boxer
Robinson, Joshua - I am a Strange Dataset: Metalinguistic Tests for Language Models
Questions by: Zeyu Liu
Nov 6
Spoken Language Processing (SLP) (Guest Lecture by Sudarsana Reddy Kadiri)
JM 16
Ye, Wen - CaMML: Context-Aware Multimodal Learner for Large Models
Questions by: Lei Liu
(MOVED TO 11/25) Xing, Hanwen - Visual Grounding Helps Learn Word Meanings in Low-Data Regimes
Questions by: Ziyi Liu
Nov 8
Project Report Version 1 due

Week 12

Nov 11
VETERANS DAY NO CLASS
Nov 13
EMNLP NO CLASS
-
Nov 15
Late Drop (W, No refund)

Week 13

Nov 18
Syntax (Not actually presenting this year, in lieu of IE, which is moved. Legacy notes for background interest) POS/HMM, Constituencies, Dependencies
Liu, Ziyi - Evaluating the Deductive Competence of Large Language Models
Questions by: Lucine Oganesian
Lin, Fandel - VariErr NLI: Separating Annotation Error from Human Label Variation
Questions by: Ting-Rui Chiang
Nov 20
Discourse (Guest Lecture by Alexander Spangher) Slides
Wu, Cheng-Han - Grounding Gaps in Language Model Generations
Questions by: Zhang, Tianyi
Wilber Blas Urrutia - Rethinking the Bounds of LLM Reasoning: Are Multi-Agent Discussions the Key?
Questions by: Chumeng Liang
Nov 22
HW 3 due

Week 14

Nov 25
Guest Lecture by Jonathan Choi (USC Gould School of Law) Paper
(MOVED FROM 11/6) Xing, Hanwen - Visual Grounding Helps Learn Word Meanings in Low-Data Regimes
Questions by: Ziyi Liu
(MOVED FROM 10/9) Mia Sultan - LLMRefine: Pinpointing and Refining Large Language Models via Fine-Grained Actionable Feedback
Questions by: Wilber Blas Urrutia
Nov 27
THANKSGIVING BREAK; NO CLASS

Week 15

Dec 2
Project Presentations
(10:00) Ziyi Liu, Ayush Goyal - Editing Common Sense in Transformers
Questions by: Xinyan Velocity Yu, Joshua Robinson, Chumeng Liang
(10:22) Ting-Rui Chiang, Fandel Lin - Anchoring Fine-tuning of Sentence Transformer with Semantic Label Information for Efficient Truly Few-shot Classification
Questions by: Wilber Leonardo Blas Urrutia, Skyler Hallinan
(10:44) Anirudh Ravi Kumar, Maksim Siniukov, Tianyi Zhang - Learning Retrieval Augmentation for Personalized Dialogue Generation
Questions by: Eric Boxer, Yifan Jiang, Lei Liu
(11:06) Wen Ye, Cheng-Han Wu, Zeyu Liu - SMoP: Towards efficient and effective prompt tuning with sparse mixture-of-prompts
Questions by: Run Huang, Mia Sultan
(11:28) Lucine Oganesian, Enes Burak Bilgin, Debaditya Pal - Adapting Language Models to Compress Contexts
Questions by: Ryan Lee, Emily Weiss, Hanwen Xing
Dec 4
Project presentations
(10:00) Xinyan Velocity Yu, Run Huang - Visual sketchpad: Sketching as a visual chain of thought for multimodal language models.
Questions by: Enes Burak Bilgin, Ziyi Liu, Lucine Oganesian
(10:22) Joshua Robinson, Ryan Lee, Hanwen Xing - Context Compression for Auto-regressive Transformers with Sentinel Tokens
Questions by: Wen Ye, Fandel Lin
(10:44) Emily Weiss, Skyler Hallinan, Wilber Leonardo Blas Urrutia - Text embeddings reveal (almost) as much as text
Questions by: Ting-Rui Chiang, Anirudh Ravi Kumar, Cheng-Han Wu
(11:06) Eric Boxer, Mia Sultan - Fighting Fire with Fire: The Dual Role of LLMs in Crafting and Detecting Elusive Disinformation
Questions by: Tianyi Zhang, Maksim Siniukov
(11:28) Yifan Jiang, Chumeng Liang, Lei Liu - Precedent-Enhanced Legal Judgment Prediction with LLM and Domain-Model Collaboration
Questions by: Zeyu Liu, Ayush Goyal, Debaditya Pal