Picture of Marvin Xhemrishi

M.Sc. Marvin Xhemrishi

Technical University of Munich

Associate Professorship of Coding and Cryptography (Prof. Wachter-Zeh)

Postal address

Postal:
Theresienstr. 90
80333 München

Theses

Available Theses

Secure Federated Learning with Differential Privacy

Description

Federated learning is a machine learning paradigm that aims to learn collaboratively from decentralized private data owned by entities referred to as clients. However, due to its decentralized nature, federated learning is susceptible to model poisoning attacks, where malicious clients try to corrupt the learning process by modifying local model updates. Moreover, the updates sent by the clients might leak information about the private data involved in the learning. The goal of this work is to investigate and combine existing robust aggregation techniques in FL with differential privacy techniques.

References:

[1] - https://arxiv.org/pdf/2304.09762.pdf

[2] - https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=9757841

[3] - https://dl.acm.org/doi/abs/10.1145/3465084.3467919

Prerequisites

- Basic knowledge about machine learning and gradient descent optimization

- First experience with machine learning in python

- Undergraduate statistics courses

- Prior knowledge about differential privacy is a plus

Supervisor:

Theses in Progress

Contribution scoring in Federated Learning

Description

Federated learning (FL) is a machine learning paradigm that aims to learn collaboratively from decentralized private data, owned by entities referred to as clients. In real-world applications of FL it is important to score the contribution of each client. The goal of this seminar is to provide a high-level overview of existing contribution-scoring techniques in federated learning using [1-2] and other references. 

Prerequisites

References: 

[1] - https://ieeexplore.ieee.org/document/10138056

[2] - https://arxiv.org/pdf/2403.07151.pdf

Supervisor:

Implementation of model poisoning attacks in federated learning

Description

Federated learning is a machine learning paradigm where decentralized entities (clients) collaboratively learn using their private data. A central server acts as a coordinator of the learning process. Due to the sensitivity of the private data involved,  the data cannot be transferred. A salient problem of federated learning is the presence of malicious clients, which are clients that try to destroy the learning process. Malicious clients can do this by corrupting their data and/or by modifying their local model updates. The goal of this project is to understand how model poisoning attacks and defense strategies perform under different scenarios of federated learning using experiments. 

References: 

[1]- https://www.ndss-symposium.org/wp-content/uploads/ndss2021_6C-3_24498_paper.pdf

[2]- https://arxiv.org/pdf/1903.03936.pdf

[3]- https://arxiv.org/pdf/2304.00160.pdf

 

Prerequisites

  • Basic knowledge of machine learning
  • Python programming skills, knowledge of PyTorch is an advantage 

Contact

marvin.xhemrishi@tum.de

Supervisor:

Publications

2022

  • Bitar, R.; Xhemrishi M.; Wachter-Zeh A.: Adaptive private distributed matrix multiplication. IEEE Transactions on Information Theory, 2022 more…
  • Garb K.; Xhemrishi M.; Kürzinger L.; Frisch C.: The Wiretap Channel for Capacitive PUF-Based Security Enclosures. IACR Transactions on Cryptographic Hardware and Embedded Systems, 2022 more…
  • Maringer G.; Xhemrishi M.; Puchinger S.; Garb K.; Liu H.; Jerkovits T.; Kürzinger L.; Hiller M.; Wachter-Zeh A.: Analysis of Communication Channels Related to Physical Unclonable Functions. Workshop on Coding and Cryptography (WCC), 2022 more…
  • Xhemrishi M.: Distributed Linear Computations over Private Sparse Matrices. IEEE European School of Information Theory, 2022 more…
  • Xhemrishi M.: Distributed Linear Computations over Private Sparse Matrices. Munich Workshop on Coding and Cryptography, 2022 more…
  • Xhemrishi M.: Computational Code-Based Privacy for Coded Federated Learning. TUM ICE Workshop Raitenhaslach, 2022 more…
  • Xhemrishi M.; Bitar R.; Wachter-Zeh A.: Distributed Matrix-Vector Multiplication with Sparsity and Privacy Guarantees. IEEE International Symposium on Information Theory, 2022 more…
  • Xhemrishi M.; Egger M. Bitar R.: Efficient Private Storage of Sparse Machine Learning Data. IEEE Information Theory Workshop, 2022 more…
  • Xhemrishi M.; Graell i Amat A.; Rosnes E.; Wachter-Zeh A.: Computational Code-Based Privacy in Coded Federated Learning. IEEE International Symposium on Information Theory, 2022 more…

2021

  • Maringer G.; Xhemrishi M.; Puchinger S.; Garb K.; Liu H.; Jerkovits T.; Hiller M.; Wachter-Zeh A.Kürzinger L.;: Analysis of Communication Channels related to Physically Unclonable Functions. arXiv, 2021 more…
  • Xhemrishi M.: Trade-off between privacy and sparsity in coded computing. 36th meeting of ITG Professional Group "Applied Information Theory", 2021 more…

2020

  • Bitar R.; Xhemrishi M.; Wachter-Zeh A.: Rateless Private Matrix-Matrix Multiplication. European School of Information Theory, 2020 more…
  • Bitar R.; Xhemrishi M.; Wachter-Zeh A.: Fountain codes for Private Distributed Matrix-Matrix Multiplication. International Symposium in Information Theory and its Applications, 2020 more…

2019

  • Xhemrishi, M.; Coşkun, M. C.; Liva, G.; Östman, J.; Durisi, G. : List Decoding of Short Codes for Communication over Unknown Fading Channels. , Workshop on Coding, Cooperation, and Security in Modern Communication Networks (COCO 2019) . , 2019 more…
  • Xhemrishi, M.; Coşkun, M. C.; Liva, G.: List Decoding for Fading Channels. Oberpfaffenhofen Workshop on High Throughput Coding (OWHTC) , 2019 more…
  • Xhemrishi, M.; Coşkun, M. C.; Liva, G.; Östman, J.; Durisi, G.: List Decoding of Short Codes for Communication over Unknown Fading Channels. Asilomar Conference on Signals, Systems, and Computers, 2019 more…