Goals of the course:
Syllabus from SIS:
The course is part of the inter-university programme prg.ai Minor.
SIS code: NPFL140
Semester: summer
E-credits: 3
Examination: 0/2 C
Guarantors: Jindřich Helcl, Jindřich Libovický
The course is held on Thursdays at 12:20 in S5.
1. Introductory notes and discussion on large language models Slides
2. The Transformer architecture Slides Notes Recording
Unless otherwise stated, teaching materials for this course are available under CC BY-SA 4.0.
Feb 19 Slides
Instructor: Zdeněk Kasner
Covered topics: aims of the course, passing requirements. We informally discussed what are (large) language models, what are they for, what are their benefits and downsides. We also gathered ideas on how to train the models, how to use them, and how to evaluate them.
Instructor: Jindřich Libovický
Learning objectives. After the lecture you should be able to...
Explain the building blocks of the Transformer architecture to a non-technical person.
Describe the Transformer architecture using equations, especially the self-attention block.
Implement the Transformer architecture (in PyTorch or another framework that does automated differentiation).
Additional materials.
Transformers explained by AI Coffee Break with Letitia (20 min)
Let's build GPT: from scratch, in code, spelled out by Andrej Karpathy (2 hours)
The Illustrated Transformer by Jay Alammar
MicroGPT Project: Complete Transformer training and inference code in 200 lines of Python
You will work on a team project during the semester. The project will be presented during the final class. Each team will submit a report, consisting of:
The length of the report should be maximum 4 pages plus references and contributions. You might want to use the ACL paper template.
You will be asked at least once to read a paper before the class.
You need to take part in a final written test that will not be graded.