You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This repository contains materials for the Efficient Deep Learning Systems course taught at the [Faculty of Computer Science](https://cs.hse.ru/en/) of [HSE University](https://www.hse.ru/en/) and [Yandex School of Data Analysis](https://academy.yandex.com/dataschool/).
3
3
4
-
__This branch corresponds to the ongoing 2024 course. If you want to see full materials of past years, see the ["Past versions"](#past-versions) section.__
4
+
__This branch corresponds to the ongoing 2025 course. If you want to see full materials of past years, see the ["Past versions"](#past-versions) section.__
5
5
6
6
# Syllabus
7
7
-[__Week 1:__](./week01_intro)__Introduction__
@@ -10,20 +10,18 @@ __This branch corresponds to the ongoing 2024 course. If you want to see full ma
10
10
-[__Week 2:__](./week02_management_and_testing)__Experiment tracking, model and data versioning, testing DL code in Python__
11
11
- Lecture: Experiment management basics and pipeline versioning. Configuring Python applications. Intro to regular and property-based testing.
12
12
- Seminar: Example DVC+Weights & Biases project walkthrough. Intro to testing with pytest.
- Lecture: Mixed-precision training. Data storage and loading optimizations. Tools for profiling deep learning workloads.
15
-
- Seminar: Automatic Mixed Precision in PyTorch. Dynamic padding for sequence data and JPEG decoding benchmarks. Basics of profiling with py-spy, PyTorch Profiler, PyTorch TensorBoard Profiler, nvprof and Nsight Systems.
16
-
-[__Week 4:__](./week04_distributed)__Basics of distributed ML__
17
-
- Lecture: Introduction to distributed training. Process-based communication. Parameter Server architecture.
-[__Week 5:__](./week05_data_parallel)__Data-parallel training and All-Reduce__
20
-
- Lecture: Data-parallel training of neural networks. All-Reduce and its efficient implementations.
13
+
-[__Week 3:__](./week03_fast_pipelines)__Training optimizations, FP16/BF16/FP8 formats, profiling deep learning code__
14
+
- Lecture: Measuring performance of GPU-accelerated software. Mixed-precision training. Data storage and loading optimizations. Tools for profiling deep learning workloads.
15
+
- Seminar: Automatic Mixed Precision in PyTorch. Dynamic padding for sequence data and JPEG decoding benchmarks. Basics of profiling with py-spy, PyTorch Profiler, Memory Snapshot and Nsight Systems.
16
+
-[__Week 4:__](./week04_data_parallel)__Data-parallel training and All-Reduce__
17
+
- Lecture: Introduction to distributed training. Data-parallel training of neural networks. All-Reduce and its efficient implementations.
21
18
- Seminar: Introduction to PyTorch Distributed. Data-parallel training primitives.
19
+
-__Week 5:____Sharded data-parallel training, distributed training optimizations__
22
20
-__Week 6:____Training large models__
23
21
-__Week 7:____Python web application deployment__
24
-
-__Week 8:____Software for serving neural networks__
22
+
-__Week 8:____LLM inference optimizations and software__
25
23
-__Week 9:____Efficient model inference__
26
-
-__Week 10:____Guest lecture__
24
+
-__Week 10:__Guest lecture
27
25
28
26
## Grading
29
27
There will be several home assignments (spread over multiple weeks) on the following topics:
@@ -37,11 +35,16 @@ Please refer to the course page of your institution for details.
Copy file name to clipboardExpand all lines: week01_intro/README.md
+3-1Lines changed: 3 additions & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,12 +1,14 @@
1
1
# Week 1: Introduction
2
2
3
3
* Lecture: [link](./lecture.pdf)
4
-
* Seminar + bonus home assignment: [link](./seminar.ipynb)
4
+
* Seminar: [link](./seminar.ipynb)
5
5
6
6
## Further reading
7
7
*[CUDA MODE reading group Resource Stream](https://github.com/cuda-mode/resource-stream)
8
8
*[CUDA Programming Guide](https://docs.nvidia.com/cuda/cuda-c-programming-guide/index.html) and [CUDA C++ Best Practices Guide](https://docs.nvidia.com/cuda/cuda-c-best-practices-guide/index.html)
0 commit comments