Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
Next revision Both sides next revision
education [2020/11/24 12:02]
fablpd
education [2020/12/09 17:59]
fablpd
Line 38: Line 38:
   * **Probabilistic Byzantine Resilience**: ​ Development of high-performance,​ Byzantine-resilient distributed systems with provable probabilistic guarantees. Two options are currently available, both building on previous work on probabilistic Byzantine broadcast: (i) a theoretical project, focused the correctness of probabilistic Byzantine-tolerant distributed algorithms; (ii) a practical project, focused on numerically evaluating of our theoretical results. Please contact [[matteo.monti@epfl.ch|Matteo Monti]] to get more information.   * **Probabilistic Byzantine Resilience**: ​ Development of high-performance,​ Byzantine-resilient distributed systems with provable probabilistic guarantees. Two options are currently available, both building on previous work on probabilistic Byzantine broadcast: (i) a theoretical project, focused the correctness of probabilistic Byzantine-tolerant distributed algorithms; (ii) a practical project, focused on numerically evaluating of our theoretical results. Please contact [[matteo.monti@epfl.ch|Matteo Monti]] to get more information.
  
- +  ​* **Distributed ​coordination ​using RDMA.** RDMA (Remote Direct Memory Access) allows accessing a remote machine'​s memory without interrupting its CPU. This technology ​is gaining traction over the last couple ​of years, ​as it allows ​for the creation ​of real-time ​distributed ​systemsRDMA allows for communication to take place close to the μsec scalewhich enables ​the design and implementation ​of systems that process requests in only tens of μsecCurrent research focuses on achieving real-time failure detection through a combination ​of novel algorithm design, latest hardware and linux kernel customizationFast failure detection over RDMA brings ​the notion of availability ​to a new level, essentially allowing modern systems to enter the era of 7 nines of availabilityContact [[https://people.epfl.ch/athanasios.xygkis|Athanasios Xygkis]] ​and [[https://​people.epfl.ch/​antoine.murat|Antoine Murat]] for more information.
-  ​* **Distributed ​computing ​using RDMA and/or NVRAM.** RDMA (Remote Direct Memory Access) allows accessing a remote machine'​s memory without interrupting its CPU. NVRAM is byte-addressable persistent (non-volatile) memory with access times on the same order of magnitude ​as traditional (volatile) RAM. These two recent technologies pose novel challenges and raise new opportunities in distributed system design and implementation. Contact [[https://​people.epfl.ch/​igor.zablotchi|Igor Zablotchi]] ​for more information. +
- +
-  * **Robust Distributed Machine Learning**: With the proliferation ​of big datasets and models, Machine Learning is becoming ​distributed. ​Following ​the standard parameter server model, the learning phase is taken by two categories ​of machines: parameter servers and workersAny of these machines could behave arbitrarily (i.e., said Byzantine) affecting ​the model convergence in the learning phase. Our goal in this project is to build system that is robust against Byzantine behavior ​of both parameter server and workersOur first prototype, AggregaThor(https://mlsys.org/Conferences/​2019/​doc/​2019/​54.pdf), describes the first scalable robust Machine Learning framework. It fixed a severe vulnerability in TensorFlow ​and it showed how to make TensorFlow even faster, while robust. Contact ​[[https://​people.epfl.ch/​arsany.guirguis|Arsany Guirguis]] for more information.+
  
  
 + 
   * **Consistency in global-scale storage systems**: We offer several projects in the context of storage systems, ranging from implementation of social applications (similar to [[http://​retwis.redis.io/​|Retwis]],​ or [[https://​github.com/​share/​sharejs|ShareJS]]) to recommender systems, static content storage services (à la [[https://​www.usenix.org/​legacy/​event/​osdi10/​tech/​full_papers/​Beaver.pdf|Facebook'​s Haystack]]),​ or experimenting with well-known cloud serving benchmarks (such as [[https://​github.com/​brianfrankcooper/​YCSB|YCSB]]);​ please contact [[http://​people.epfl.ch/​dragos-adrian.seredinschi|Adi Seredinschi]] or [[https://​people.epfl.ch/​karolos.antoniadis|Karolos Antoniadis]] ​ for further information.   * **Consistency in global-scale storage systems**: We offer several projects in the context of storage systems, ranging from implementation of social applications (similar to [[http://​retwis.redis.io/​|Retwis]],​ or [[https://​github.com/​share/​sharejs|ShareJS]]) to recommender systems, static content storage services (à la [[https://​www.usenix.org/​legacy/​event/​osdi10/​tech/​full_papers/​Beaver.pdf|Facebook'​s Haystack]]),​ or experimenting with well-known cloud serving benchmarks (such as [[https://​github.com/​brianfrankcooper/​YCSB|YCSB]]);​ please contact [[http://​people.epfl.ch/​dragos-adrian.seredinschi|Adi Seredinschi]] or [[https://​people.epfl.ch/​karolos.antoniadis|Karolos Antoniadis]] ​ for further information.
  
  
 \\ \\
-  * **Consistency in global-scale storage systems**: Generative adversarial networks (GANs) have achieved some spectacular results in the six years since they have been introduced. However, their training process is still fraught with issues such as mode collapse, non-convegence or gradients collapse. Those issues have still not been completely resolved, making multiple restarts and selection of well-performing Generator-Discriminator pairs part of GAN training process. The process of adversarial generator-discriminator training is not dissimilar from co-evolution of two adversarial species - such as for instance hosts and pathogens, except that rounds of mutation/​recombination/​selection in search of fitness optimum are replaced by gradient descent. Our goal is to investigate - both experimentally and theoretically - if we can further stabilize and improve GAN training with evolutionary mechanisms, such as speciation, aneuploidization,​ neutral variability buffering or meta-evolutionary mechanisms. This work wold have implications on developing efficient solutions to detecting GAN products (aka deepfakes). You will need to have experience with scientific computing in Python, ideally with PyTorch experience, and ideally you should have some knowledge of population genetics. ​ Contact [[https://​people.epfl.ch/​andrei.kucharavy|Andrei Kucharavy]] for more information.+  * **Theory of evolution to improve GAN training**: Generative adversarial networks (GANs) have achieved some spectacular results in the six years since they have been introduced. However, their training process is still fraught with issues such as mode collapse, non-convegence or gradients collapse. Those issues have still not been completely resolved, making multiple restarts and selection of well-performing Generator-Discriminator pairs part of GAN training process. The process of adversarial generator-discriminator training is not dissimilar from co-evolution of two adversarial species - such as for instance hosts and pathogens, except that rounds of mutation/​recombination/​selection in search of fitness optimum are replaced by gradient descent. Our goal is to investigate - both experimentally and theoretically - if we can further stabilize and improve GAN training with evolutionary mechanisms, such as speciation, aneuploidization,​ neutral variability buffering or meta-evolutionary mechanisms. This work wold have implications on developing efficient solutions to detecting GAN products (aka deepfakes). You will need to have experience with scientific computing in Python, ideally with PyTorch experience, and ideally you should have some knowledge of population genetics. ​ Contact [[https://​people.epfl.ch/​andrei.kucharavy|Andrei Kucharavy]] for more information.
  
 +  * **Text GANs**: For a long time the Generative adversarial networks (GANs) were limited to differentiable data generation - notably images and videos. Recent advances allow text-generating GANs to be build. Lighter and easier to train than other solution capable of text generation, their capabilities remain unexpored as of now. The goal of this project would be implementing possible GAN architectures and evaluating their ability to generate texts with pre-set features - such as style or topics. You will need to have experience with scientific computing in Python, ideally with PyTorch experience, and ideally some experience with NLP. Contact [[https://​people.epfl.ch/​andrei.kucharavy|Andrei Kucharavy]] for more information.
 +
 +  * **Byzantine-resilient heterogeneous GANs**: Byzantine-resilient federated learning has emerged as a major theme over the last couple of years, in grand part due to the need to distribute machine learning across many nodes, due to performance and privacy concerns. Until now it has focused on training a single model across many workers and many parameter serves. While this approach has brought on formidable results - including in GAN training, the topic of efficient, distributed and byzantine-resilient training of heterogeneous architectures remain relatively unexplored. In the context of Generative adversarial networks (GANs), such learning is critical to training light discriminators that can specialize in detecting specific featuers of generator-generated images. The goal of this project will be to investigate the potential for GAN training process poisonning by malicious discriminators and generators and investigate efficient protocols to ensure the training process robustness. You will need to have experience with scientific computing in Python, ideally with PyTorch experience, and notions of distributed computing. Contact [[https://​people.epfl.ch/​andrei.kucharavy|Andrei Kucharavy]] for more information.
 +
 +\\
  
 ===== Semester Projects ===== ===== Semester Projects =====