Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
Next revision Both sides next revision
education [2019/11/01 13:27]
seredins
education [2022/10/31 10:25]
fablpd
Line 8: Line 8:
 \\ \\
  
-  * [[education/​ca_2019|Concurrent Algorithms]] (theory & practice)+  * [[education/​ca_2021|Concurrent Algorithms]] (theory & practice)
   * [[education/​da|Distributed Algorithms]] (theory & practice)   * [[education/​da|Distributed Algorithms]] (theory & practice)
 \\ \\
Line 28: Line 28:
   * **[[cryptocurrencies|Cryptocurrencies]]**:​ We have several project openings as part of our ongoing research on designing new cryptocurrency systems. Please contact [[rachid.guerraoui@epfl.ch|Prof. Rachid Guerraoui]].   * **[[cryptocurrencies|Cryptocurrencies]]**:​ We have several project openings as part of our ongoing research on designing new cryptocurrency systems. Please contact [[rachid.guerraoui@epfl.ch|Prof. Rachid Guerraoui]].
  
-  * **Probabilistic Byzantine Resilience**: ​ Development of high-performance,​ Byzantine-resilient distributed systems with provable probabilistic guarantees. Two options are currently available, both building on previous work on probabilistic Byzantine broadcast: (i) a theoretical project, focused the correctness of probabilistic Byzantine-tolerant distributed algorithms; (ii) a practical project, focused on numerically evaluating of our theoretical results. Please contact [[matteo.monti@epfl.ch|Matteo Monti]] to get more information. 
  
 +  * **Proof systems for Byzantine systems**: Cryptographic proof systems enable the rapid verification of computation between mutually distrustful parties. Recent advances in proof systems include (1) recursive proofs, transition proofs and accumulators which are of prime interest to shrink long chains of computation and/or their associated storage, and (2) zero-knowledge scalable proofs useful for privacy-preserving systems. Motivated by cryptocurrencies,​ the goal of this project is to devise and implement Byzantine-resilient systems that incorporate new cryptographic proof systems for efficiency and/or privacy. Contact Pierre-Louis Roman <​pierre-louis.roman@epfl.ch>​ for more information.
  
-  * **Distributed computing using RDMA and/or NVRAM.** RDMA (Remote Direct Memory Access) allows accessing ​remote machine'​s memory without interrupting its CPUNVRAM is byte-addressable persistent (non-volatile) memory with access times on the same order of magnitude ​as traditional (volatile) RAMThese two recent technologies pose novel challenges and raise new opportunities ​in distributed ​system design and implementation. Contact ​[[https://​people.epfl.ch/​igor.zablotchi|Igor Zablotchi]] ​for more information.+  * **Topology-aware mempool for cryptocurrencies**: The mempool is core component of cryptocurrency systemsIt disseminates user transactions to the miner nodes before they reach consensus.Current mempools assume an homogeneous network topology where all machines have the same bandwidth and latency.This unrealitic assumption forces the system to progress at the same speed as the slowest node in the systemThis project aims at implementing a mempool which exploits the heterogeneity of the network to speed up data dissemination for cryptocurrency systems. This is a practical project which requires good knowledge ​in network programming,​ either Go or C++, distributed ​algorithms. Contact ​Gauthier Voron <​gauthier.voron@epfl.chfor more information.
  
-  * **[[Distributed ML|Distributed Machine Learning]]**: contact ​[[http://​people.epfl.ch/​georgios.damaskinos|Georgios Damaskinos]] for more information.+  * **Robust mean estimation**:​ In recent years, many algorithms have been proposed to perform robust mean estimation, which has been shown to be equivalent to robust gradient-based machine learning. A new concept has been proposed to define the performance of a robust mean estimator, called the [[https://​arxiv.org/​abs/​2008.00742|averaging constant]] (along with the Byzantine resilience). This research project consists of computing the theoretical averaging constant of different proposed robust mean estimators, and to study their empirical performances on randomly generated vectors. Contact ​[[https://​people.epfl.ch/​sadegh.farhadkhani?​lang=en|Sadegh Farhadkhani]] for more information.
  
-  * **Robust Distributed Machine Learning**: With the proliferation of big datasets and models, Machine Learning is becoming distributed. Following the standard parameter server model, the learning phase is taken by two categories of machines: parameter servers and workers. Any of these machines could behave arbitrarily (i.e., said Byzantine) affecting the model convergence in the learning phase. Our goal in this project is to build a system that is robust against Byzantine behavior of both parameter server and workers. Our first prototype, AggregaThor(https://​www.sysml.cc/​doc/​2019/​54.pdf),​ describes the first scalable robust Machine Learning framework. It fixed a severe vulnerability in TensorFlow and it showed how to make TensorFlow even faster, while robust. Contact [[https://​people.epfl.ch/​arsany.guirguis|Arsany Guirguis]] for more information. 
  
-  * **Stochastic gradient: (artificial) reduction of the ratio variance/​norm for adversarial distributed SGD**: One computationally-efficient and non-intrusive line of defense for adversarial distributed SGD (e.g1 parameter server distributing ​the gradient estimation to severalpossibly adversarial workers) relies on the honest workers ​to send back gradient estimations ​with sufficiently low variance; assumption which is sometimes hard to satisfy in practiceOne solution could be to (drastically) increase the batch-size at the workersbut doing so may as well defeat the very purpose ​of distributing the computation. \\ In this project, we propose two approaches that you can choose to explore (also you may propose ​different approach) ​to (artificially) reduce the ratio variance/​norm of the stochastic gradientswhile keeping ​the benefits ​of the distribution. The first proposed approachspeculative,​ boils down to "​intelligent"​ coordinate selection. The second makes use of some kind of "​momentum"​ at the workers. \\ [1] [[https://​papers.nips.cc/​paper/​6617-machine-learning-with-adversaries-byzantine-tolerant-gradient-descent|"​Machine Learning with Adversaries:​ Byzantine Tolerant Gradient Descent"​ ]]  \\ [2] [[https://​arxiv.org/​abs/​1610.05492|"​Federated Learning: Strategies for Improving Communication Efficiency"​]] \\ Contact ​ [[https://​people.epfl.ch/​sebastien.rouault|Sébastien Rouault]] for more information.+  * **Accelerate Byzantine collaborative learning**: [[https://​arxiv.org/​abs/​2008.00742|Our recent NeurIPS paper]] proposed algorithms for collaborative machine learning in the presence of Byzantine nodeswhich have been proved ​to be near optimal ​with respect ​to optimality at convergenceHowever, these algorithms require all-to-all communication ​at every roundwhich is suboptimal. This research consists ​of designing ​practical solution ​to Byzantine collaborative learningbased on the idea of a random communication network at each round, with both theoretical guarantees and practical implementation. Contact [[https://​people.epfl.ch/​sadegh.farhadkhani?​lang=en|Sadegh Farhadkhani]] for more information.
  
  
-  * **Consistency in global-scale storage systems**: We offer several projects in the context of storage systems, ranging from implementation of social applications (similar to [[http://​retwis.redis.io/​|Retwis]],​ or [[https://​github.com/​share/​sharejs|ShareJS]]) to recommender systems, static content storage services (à la [[https://​www.usenix.org/​legacy/​event/​osdi10/​tech/​full_papers/​Beaver.pdf|Facebook'​s Haystack]]),​ or experimenting with well-known cloud serving benchmarks (such as [[https://​github.com/​brianfrankcooper/​YCSB|YCSB]]);​ please contact [[http://​people.epfl.ch/​dragos-adrian.seredinschi|Adi Seredinschi]] or [[https://​people.epfl.ch/​karolos.antoniadis|Karolos Antoniadis]] ​ for further information. 
  
 +  * **Probabilistic Byzantine Resilience**: ​ Development of high-performance,​ Byzantine-resilient distributed systems with provable probabilistic guarantees. Two options are currently available, both building on previous work on probabilistic Byzantine broadcast: (i) a theoretical project, focused the correctness of probabilistic Byzantine-tolerant distributed algorithms; (ii) a practical project, focused on numerically evaluating of our theoretical results. Please contact [[matteo.monti@epfl.ch|Matteo Monti]] to get more information.
  
-\\+  * **Microsecond-scale dependable systems.** Modern networking technologies such as RDMA (Remote Direct Memory Access) allow for sub-microsecond communication latency. Combined with emerging data center architectures,​ such as disaggregated resources pools, they open the door to novel blazing-fast and resource-efficient systems. Our research focuses on designing such microsecond-scale systems that can also tolerate faults. Our vision is that tolerating network asynchrony as well as faults (crash and/or Byzantine) is a must, but that it shouldn'​t affect the overall performance of a system. We achieve this goal by devising and implementing novel algorithms tailored for new hardware and revisiting theoretical models to better reflect modern data centers. Previous work encompasses microsecond-scale (BFT) State Machine Replication,​ Group Membership Services and Key-Value Stores (OSDI'​20,​ ATC'22 and ASPLOS'​23). Overall, if you are interested in making data centers faster and safer, contact [[https://​people.epfl.ch/​athanasios.xygkis|Athanasios Xygkis]] and [[https://​people.epfl.ch/​antoine.murat|Antoine Murat]] for more information.
  
  
-===== Semester Projects =====+ 
  
-If the subject of a Master Project interests you as a Semester Project, please contact the supervisor of the Master Project to see if it can be considered for a Semester Project. 
  
-EPFL I&C duration, credits and workload information are available [[https://​www.epfl.ch/​schools/​ic/​education/​|here]]. Don't hesitate to contact the project supervisor if you want to complete your Semester Project outside the regular semester period.+\\
  
-===== Collaborative Projects =====+  * **Byzantine-resilient heterogeneous GANs**: Byzantine-resilient federated learning has emerged as a major theme over the last couple of years, in grand part due to the need to distribute machine learning across many nodes, due to performance and privacy concerns. Until now it has focused on training a single model across many workers and many parameter serves. While this approach has brought on formidable results - including in GAN training, the topic of efficient, distributed and byzantine-resilient training of heterogeneous architectures remain relatively unexplored. In the context of Generative adversarial networks (GANs), such learning is critical to training light discriminators that can specialize in detecting specific featuers of generator-generated images. The goal of this project will be to investigate the potential for GAN training process poisonning by malicious discriminators and generators and investigate efficient protocols to ensure the training process robustness. You will need to have experience with scientific computing in Python, ideally with PyTorch experience, and notions of distributed computing. Contact [[https://​people.epfl.ch/​andrei.kucharavy|Andrei Kucharavy]] for more information. 
 +\\
  
-The lab is also collaborating ​with the industry ​and other labs at EPFL to offer interesting student projects motivated from real-world problems. ​With [[LARA||http://lara.epfl.ch]] and [[interchain.io|Interchain Foundation]] we have several projects:+  * **Hijacking proof-of-work to make it useful: distributed gradient-free learning approach**: Proof-of-work blockchains - notably Bitcoin and Ethereum - reach a probabilistic consensus about the contents of the blockchain by a mechanism of probabilistic leader election. Every contributor to the consensus tries to solve a puzzle, and the first one to succeed ​is elected a leader, allowed to create the next block and publicly add information to it. The puzzle needs to be hard to solve and easy to verify, solvable only by random guessing and not allowing for any shortcuts and allow for its difficulty to be tuned so that nodes don't find answers to it simultaneously and take different leaderships forking the chain in two. Partial cryptographic hash reversal has traditionally been a perfect candidate for such puzzle, but it has no interest outside being a challenge for blockchain. And with 100-300 PetaFLOP/s (drawing 100 TWh/y) of general purpose computational power being tied into Ethereum blockchain alone as of early 2022, the waste of computational resources ​and energy is colossal. While the interest of blockchains and the suitability of proof-of-work as a mechanism to run them is widely debated, it'​s ​at this day the mechanism for the two largest ones. We try to at least use some of that challenge useful by injecting a "​try"​ step of a (1,λ)-ES evolutionary search algorithm into the hash computation loop, slowing it down and making it do something useful in during the slowdown period. This class of evolutionary search algorithm achieves a good performance on black-bock optimization tasks (sometimes exceeding RL approaches in traditionally RL problems), is embarrassingly parallel, fits well the requirements for a proof-of-work function and can be empirically optimized to minimize the waste of computational resources during a training run. However, in its current state the (1,​λ)-ES-based useful proof-of-work has been proven to work in cases where the data used for the training tasks can be fully replicated among the nodes. For numerous applications,​ it is not an option. Finding ways to solve that problem, both from a theoretical and an experimental perspective will be the goal of this project. You will need solid skills in Python (Rust and WebAssembly are a plus), basic understanding of distributed algorithms and of machine learning concepts. Some familiarity with blockchains and black box optimization is a plus, but is not a requirementContact ​[[https://people.epfl.ch/andrei.kucharavy|Andrei Kucharavy]] for more information. 
 +\\
  
-  - **[[https://​dcl.epfl.ch/​site/​cryptocurrencies|AT2]]:​** Integration of an asynchronous (consensus-less) payment system in the Cosmos Hub.+\\
  
-  - **[[https://github.com/cosmos/ics/tree/master/ibc|Interblockchain Communication (IBC)]]:** Protocols description (and optional implementation) for enabling ​the inter-operation of independent blockchain applications.+===== Semester Projects ===== 
 + 
 +If the subject of a Master Project interests you as a Semester Project, please contact the supervisor of the Master Project to see if it can be considered for a Semester Project. 
 + 
 +EPFL I&C duration, credits and workload information are available ​[[https://www.epfl.ch/schools/ic/education/|here]]. Don't hesitate to contact ​the project supervisor if you want to complete your Semester Project outside the regular semester period.
  
-- **[[http://​stainless.epfl.ch|Stainless]]**:​ Implementation of Tendermint modules (consensus, mempool, fast sync) using Stainless and Scala. 
-- **[[|Prusti]]** implementation of modules in Tendermint (consensus, mempool, fast sync). 
-- Mempool performance analysis and algorithm improvement 
-- Experimental evaluation of Tendermint in adversarial settings (Jepsen++) 
-- Test generation out of spec (TLA+ or Stainless) for consensus module 
-- Using HotStuff trick to improve Tendermint 
  
-[[education#​collaborative projects|test link]]