Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
Next revision Both sides next revision
education [2019/05/24 16:59]
fablpd
education [2019/05/24 17:12]
fablpd
Line 31: Line 31:
   * **Distributed computing using RDMA and/or NVRAM**: contact [[https://​people.epfl.ch/​igor.zablotchi|Igor Zablotchi]] for more information.   * **Distributed computing using RDMA and/or NVRAM**: contact [[https://​people.epfl.ch/​igor.zablotchi|Igor Zablotchi]] for more information.
  
-  * **[[Distributed ML|Distributed Machine Learning]]**+  * **[[Distributed ML|Distributed Machine Learning]]**: contact [[http://​people.epfl.ch/​georgios.damaskinos|Georgios Damaskinos]] for more information.
  
-  * **Distributed ​and Fault-tolerant algorithms**: projects here would consist in designing failure detection mechanisms suited for large-scale systems, real-time systems, ​and systems ​with unreliable communication ​or partial synchronyThis task also involves implementingevaluatingand simulating ​the performance of the developed mechanisms ​to verify the achievable guarantees; please contact ​[[http://​people.epfl.ch/​david.kozhaya|David Kozhaya]] to get more information.+  * **Robust ​Distributed ​Machine Learning**: The goal of such project is to work on the design and implementation of algorithms ​and systems ​to improve the robustness of distributed ML schemes that would tolerate poisoned data, software bugs as well as hardware failures. The practical work will be done on TensorFlow ​or PytorchOur first prototypeAggregaThor(https://​www.sysml.cc/​doc/​2019/​54.pdf)describes ​the first scalable robust Machine Learning framework. It fixed a severe vulnerability in TensorFlow and it showed how to make TensorFlow even faster, while robust. Contact ​[[https://​people.epfl.ch/​arsany.guirguis|Arsany Guirguis]] or [[https://​people.epfl.ch/​sebastien.rouault|Sébastien Rouault]] for more information.
  
-  * **Consistency in global-scale storage systems**: We offer several projects in the context of storage systems, ranging from implementation of social applications (similar to [[http://​retwis.redis.io/​|Retwis]],​ or [[https://​github.com/​share/​sharejs|ShareJS]]) to recommender systems, static content storage services (à la [[https://​www.usenix.org/​legacy/​event/​osdi10/​tech/​full_papers/​Beaver.pdf|Facebook'​s Haystack]]),​ or experimenting with well-known cloud serving benchmarks (such as [[https://​github.com/​brianfrankcooper/​YCSB|YCSB]]);​ please contact [[http://​people.epfl.ch/​dragos-adrian.seredinschi|Adrian Seredinschi]] ​for further information. +  * **Consistency in global-scale storage systems**: We offer several projects in the context of storage systems, ranging from implementation of social applications (similar to [[http://​retwis.redis.io/​|Retwis]],​ or [[https://​github.com/​share/​sharejs|ShareJS]]) to recommender systems, static content storage services (à la [[https://​www.usenix.org/​legacy/​event/​osdi10/​tech/​full_papers/​Beaver.pdf|Facebook'​s Haystack]]),​ or experimenting with well-known cloud serving benchmarks (such as [[https://​github.com/​brianfrankcooper/​YCSB|YCSB]]);​ please contact [[http://​people.epfl.ch/​dragos-adrian.seredinschi|Adrian Seredinschi]] ​or [[https://​people.epfl.ch/​karolos.antoniadis|Karolos Antoniadis]]  for further ​information.
- +
-  * **Distributed database algorithms**:​ a project here would consist in implementing and evaluating protocols that are running in today'​s database systems, e.g., [[https://​en.wikipedia.org/​wiki/​Two-phase_commit_protocol|2PC]],​ and comparing them with those protocols that can  potentially be used in future database systems; please contact [[http://​people.epfl.ch/​jingjing.wang|Jingjing Wang]] to get more information.+