Alistarh Group

Distributed Algorithms and Systems

Distribution has been one of the key trends in computing over the last decade: processor architectures are multi-core, while large-scale systems for machine learning and data processing can be distributed across several machines or even data centers. The Alistarh group works to enable these applications by creating algorithms that scale—that is, they improve their performance when more computational units are available.

This fundamental shift to distributed computing performed puts forward exciting open questions: How do we design algorithms to extract every last bit of performance from the current generation of architectures? How do we design future architectures to support more scalable algorithms? Are there clean abstractions to render high-performance distribution accessible to programmers? The group’s research is focused on answering these questions. In particular, they are interested in designing efficient, practical algorithms for fundamental problems in distributed computing, in understanding the inherent limitations of distributed systems, and in developing new ways to overcome these limitations. One particular area of focus over the past few years has been distributed machine learning.

On this site:


Current Projects

Distributed machine learning | Concurrent data structures and applications | Molecular computation


Brandt S, Keller B, Rybicki J, Suomela J, Uitto J. 2021. Efficient load-balancing through distributed token dropping. Annual ACM Symposium on Parallelism in Algorithms and Architectures. SPAA: Symposium on Parallelism in Algorithms and Architectures , 129–139. View

Alistarh D-A, Davies P. 2021. Collecting coupons is faster with friends. Structural Information and Communication Complexity. SIROCCO: International Colloquium on Structural Information and Communication Complexity, LNCS, vol. 12810, 3–12. View

Czumaj A, Davies P, Parter M. 2021. Graph sparsification for derandomizing massively parallel computation with low space. ACM Transactions on Algorithms. 17(2), 16. View

Davies P, Gurunanthan V, Moshrefi N, Ashkboos S, Alistarh D-A. 2021. New bounds for distributed mean estimation and variance reduction. 9th International Conference on Learning Representations. ICLR: International Conference on Learning Representations. View

Ramezani-Kebrya A, Faghri F, Markov I, Aksenov V, Alistarh D-A, Roy DM. 2021. NUQSGD: Provably communication-efficient data-parallel SGD via nonuniform quantization. Journal of Machine Learning Research. 22(114), 1−43. View

View All Publications


since 2017 Assistant Professor, IST Austria
2016 – 2017 “Ambizione Fellow”, Computer Science Department, ETH Zurich
2014 – 2016 Researcher, Microsoft Research, Cambridge, UK
2014 – 2016 Morgan Fellow, Downing College, University of Cambridge, UK
2012 – 2013 Postdoc, Massachusetts Institute of Technology, Cambridge, USA
2012 PhD, EPFL, Lausanne, Switzerland

Selected Distinctions

2018 ERC Starting Grant
2015 Awarded Swiss National Foundation “Ambizione” Fellowship
2014 Elected Morgan Fellow at Downing College, University of Cambridge
2012 Postdoctoral Fellowship of the Swiss National Foundation
2011 Best Paper Award at the International Conference on Distributed Computing and Networking

Additional Information

Dan Alistarh’s website

Back to Top