C. Xu
Distributed-memory Parallel Programming and its Applications
S1 3 ECTS 24h OPT EN C. Xu
This course mainly focus on distributed-memory parallel framework, where each processor (core, CPU, GPU, IoT device) has its own memory space which can not be shared with others. This parallel framework is popular as it corresponds to many real application scenarios such as the cluster, the sensor network and IoT.
The goal of this course is to explore the most popular strategies for parallelizing machine learning (ML) tasks without data-sharing across various real-world application scenarios. Additionally, the students will learn to maintain the privacy and security of such distributed systems. For the lab, the students will parallelize their ML tasks in a real GPU-equipped cluster (NEF) using PyTorch and Flower. This hand-on experience will enable students to accelerate heavy computations involved in training large neural networks.
Content:
- Distributed Learning Algorithms
- Server-Clients Framework
- Decentralized Framework
- Privacy
- Vulnerabilities in Distributed Systems
- Differential Privacy Mechanisms and its theoretical guarantees
- Security
- Robust Algorithms, Secure aggregation, TEE
- Theoretical Guarantees