Глубинное обучение 2 2025

Материал из Wiki - Факультет компьютерных наук
Перейти к: навигация, поиск

Общая информация

Курс предназначен для студентов 4 курса ФКН ПМИ (МОП и КНАД).

Занятия проходят по понедельникам 14:40-17:40 (переносы будут сообщаться в чате).


Полезные ссылки:


Формула итоговой оценки (округление арифметическое):

  1. МОП: Оитог = 0.25 * Осоревнование + 0.75 * ОДЗ,
  2. КНАД: Оитог = ОДЗ,

где ОДЗ - средняя оценка за практические домашние задания.

Преподаватели и ассистенты

Кому писать, если кажется, что все пропало: Мишан Алиев

Группа Семинарист Ассистенты Чаты групп Инвайт в anytask
221 (МОП) Федя Великонивцев Динар Саберов, Анна Василева МОП 221 yvZZTIS
222 (МОП) Ева Неудачина Александр Матосян, Полина Кадейшвили МОП 222 W38CfZf
223 (МОП) Иван Ершов Андрей Уткин, Георгий Фатахов МОП 223 ph91Jlz
224 (МОП) Степан Беляков Анна Пономарчук, Татьяна Яковлева МОП 224 iPwd342
КНАД Даня Бураков Анастасия Лапшина, Иван Галий КНАД zTn4sRP

Лекции и семинары

Лекция / Семинар 1 (08.09).

Тема: Essentials of GPU, Deep Learning Bottlenecks, and Benchmarking Basics

Аннотация: In this session, we will explore the reasons behind the dominance of GPUs in Deep Learning and examine the common sources of performance bottlenecks in DL code. You will learn how to identify these bottlenecks using profiling tools and apply techniques to optimize and accelerate your code.

Лектор и семинарист: Fedor Velikonivtsev

Локация: лекция - R208, семинар - D102.

Материалы: запись лекции, запись семинара, материалы.


Лекция / Семинар 2 (15.09).

Тема: On Transformers and Bitter Lesson

Аннотация: In this talk we’ll dive into the landscape of model architectures in deep learning, with a focus on the world around transformers. We’ll briefly recall what a transformer is, trace the evolution from encoder–decoder to encoder-only and decoder-only models, and touch on the rise of “efficient mixers” such as state space models, linear attention, and beyond. We’ll conclude by reflecting on the role of data and compute. The lecture is inspired in part by The Bitter Lesson and blends a brain dump with some entertaining insights from recent years of architectural exploration.

Лектор и семинарист: Ivan Rubachev

Локация: лекция - R206, семинар - R208.

Материалы: запись лекции и семинара, материалы.


Лекция / Семинар 3 (22.09).

Тема: Modern LLMs essentials

Аннотация лекции: This week, we will discuss LLMs. We will discuss why they are so effective for text generation, how they can be applied to different NLP problems, and the risks they pose. You will learn the details of RLHF, PEFT, and RAG, which make LLMs robust in various cases.

Лектор: Alexander Shabalin

Локация: онлайн.


Аннотация семинара: In this seminar, we will explore the concept of LLM-based agents and how they extend the capabilities of modern language models. We will discuss function calling as a way to integrate external tools, chain-of-thought reasoning for structured problem solving, and reinforcement learning techniques for training agents.

Семинарист: Ivan Ershov

Локация: R208.


Материалы: запись лекции, запись семинара, материалы.


Лекция / Семинар 4 (29.09).

Тема: Basics of Efficient LLM Training Infrastructure

Аннотация: In this lecture, we will study the basic rules that underpin the infrastructure for efficient large language model training. We will also examine common problems that may arise in this process and explore practical ways to address them.

Лектор и семинарист: Michael Khrushchev

Локация: лекция - G002, семинар - R208.

Материалы: запись лекции и семинара, материалы.


Лекция / Семинар 5 (06.10).

Тема: Segmentation and Detection

Аннотация лекции: This week, we'll explore the evolution of object detection and segmentation — from R-CNN to real-time methods like YOLO-World and CLIP-based approaches. We'll examine how U-Net architectures have transcended computer vision to power neural operators and how to use diffusion models for segmentation tasks.

Аннотация семинара: We will implement architectures and train a semantic segmentation model. We will discuss regularization methods for convolutional layers. In addition, we will learn how to use pre-trained models and apply them.

Лектор и семинарист: Alexander Oganov

Локация: лекция - G002, семинар - R208.

Материалы: запись лекции, запись семинара, материалы.


Лекция / Семинар 6 (13.10).

Тема: Segmentation and Detection 2

Аннотация лекции и семинара: In this lecture, we will continue in more detail about segmentation and detection.

Лектор: Sergey Zagoruyko

Семинарист: Eva Neudachina

Локация: лекция - G002, семинар - R208.

Материалы: запись лекции, запись семинара, материалы.


Лекция / Семинар 7 (20.10).

Тема: Diffusion models 1

Аннотация лекции и семинара: Tomorrow's lecture and seminar will be devoted to an introduction to diffusion models. Diffusion models are currently the most popular approach to generative modeling due to their high-quality generation and diversity (mode coverage) of the learned distribution. The idea behind diffusion models is to consider the process of gradually transforming data into pure noise and construct its inverse in time, which will transform noise into data. In the lecture and seminar, we will work with noise processes and derive the classic DDPM model, which proposes to minimize the KL-divergence between the “true” reverse process that converts noise into data and the denoising process specified by the neural network. In the process, we will see that this procedure is equivalent to training a denoiser neural network that predicts a clean object from a noisy one. In addition, we will interpret the resulting denoising process: in it, each step corresponds to replacing part of the current noisy image with an (increasingly high-quality) prediction of the denoiser.

Лектор и семинарист: Denis Rakitin

Локация: лекция - G002, семинар - R208.

Материалы: запись лекции, запись семинара, материалы.


Лекция / Семинар 8 (03.11).

Тема: Diffusion models 2

Аннотация лекции и семинара: In this lecture and seminar, we will continue our exploration of diffusion models. We will introduce the score function and score identity, present classifier and classifier-free guidance, and derive DDIM model.

Лектор и семинарист: Denis Rakitin

Локация: лекция - R205, семинар - R208.

Материалы: запись лекции, запись семинара, материалы.


Лекция / Семинар (11.11).

Тема: Diffusion models 3

Аннотация лекции: This lecture presents applied aspects of diffusion models, a class of generative methods that have demonstrated strong performance across images, text, and audio. We will review modern architectures, with a focus on FLUX, and unpack the key principles of their design and training.

Лектор: Nikita Starodubcev

Аннотация семинара: The seminar will provide a hands-on, in-depth inspection of PixArt model. Then ee will attempt further fine-tuning via DreamBooth, review common evaluation metrics, and compare different models and configurations.

Семинарист: Eva Neudachina

Локация: лекция - R205, семинар - R208.

Материалы: tba.

Домашние задания

Домашнее задание Ссылка Дедлайн (жёсткий)
1 Tensor and DL Libraries https://github.com/thecrazymage/DL2_HSE/tree/main/homeworks/homework_01 30 сентября, 23:59
2 Transformers for Named Entity Recognition https://github.com/thecrazymage/DL2_HSE/tree/main/homeworks/homework_02 14 октября, 23:59
3 Image Segmentation https://github.com/thecrazymage/DL2_HSE/tree/main/homeworks/homework_03 2 ноября, 23:59
4 Diffusion Models https://github.com/thecrazymage/DL2_HSE/tree/main/homeworks/homework_04 18 ноября, 23:59

Соревнование

TBA