CS291A - Special Topics in Foundation Models | home | schedule |
|
Announcements
March 31: The course time has changed to Monday/Friday 1:00-2:50pm,
Abstract: This is a graduate-level research course on Transformer and Large Language Models. Over the duration of this course, we will delve into the latest publications within the expansive domain of Large Language Models (LLMs), intelligent agents, and multimodal learning. Each student is entrusted with the following responsibilities: Conducting critical analyses and authoring paper reviews, programming with LLMs, delivering comprehensive paper presentations, and undertaking a substantial, research-quality course project. The intent behind this course is to foster a deep understanding of LLMs and their pivotal role in today's technological landscape.
Each student is expected to read papers before lecture, write paper reviews or program with LLMs, present papers, and complete a research-quality course project (e.g., implement an existing algorithm or solve a new problem creatively using deep learning. One team could have two students. ). Projects that simply apply Transformers/LLMs are not encouraged.
Prerequisites: Neural network building experience or successfully finished an introductory deep learning course.
Time: Monday/Friday 1:00- 2:50pm, Location: PHELP 3526 Office Hour: Wed 10:00-11:00pm, Henley Hall 2017
TA: N/A.
Text Books (not required)