Arto Maranjyan

prof_pic.jpg

I’m a PhD student at KAUST, working under the guidance of Prof. Peter Richtárik. My research focuses on optimization for machine learning and federated learning, where I contribute to the development of distributed and randomized optimization algorithms.

Before starting my PhD, I earned my MSc and BSc from Yerevan State University. During my bachelor’s studies, I coauthored several papers in Harmonic Analysis under the supervision of Prof. Martin Grigoryan.

Outside of academics, I enjoy dancing bachata, playing board games, ultimate frisbee and foosball.

News

Nov 23, 2024 I had the pleasure of giving a talk at the Apple MLR seminar, thanks to an invitation from Samy Bengio. It was an amazing opportunity to share our work on MindFlayer. Feel free to check out the talk slides here.
Oct 17, 2024 I will be giving a talk at the International Conference on Algebra, Logic, and their Applications on October 18, on our paper, MindFlayer: Efficient Asynchronous Parallel SGD in the Presence of Heterogeneous and Random Worker Compute Times.
Oct 17, 2024 I am reviewing for SIAM Journal on Mathematics of Data Science (SIMODS).
Oct 10, 2024 Three papers are accepted to Optimization for Machine Learning Workshop (NeurIPS 2024):
  • MindFlayer: Efficient Asynchronous Parallel SGD in the Presence of Heterogeneous and Random Worker Compute Times
  • Differentially Private Random Block Coordinate Descent
  • LoCoDL: Communication-Efficient Distributed Learning with Local Training and Compression
Oct 08, 2024 New paper out; we have MindFlayer, Vecna, and other Stranger Things: MindFlayer: Efficient Asynchronous Parallel SGD in the Presence of Heterogeneous and Random Worker Compute Times

Selected Publications

  1. Differentially Private Random Block Coordinate Descent
    Artavazd Maranjyan, Abdurakhmon Sadiev, and Peter Richtárik
    arXiv preprint arXiv:2412.17054, 2024
  2. MindFlayer: Efficient Asynchronous Parallel SGD in the Presence of Heterogeneous and Random Worker Compute Times
    Artavazd Maranjyan, Omar Shaikh Omar, and Peter Richtárik
    arXiv preprint arXiv:2410.04285, 2024
  3. LoCoDL: Communication-Efficient Distributed Learning with Local Training and Compression
    Laurent Condat, Artavazd Maranjyan, and Peter Richtárik
    arXiv preprint arXiv:2403.04348, 2024
  4. Gradskip: Communication-accelerated local gradient methods with better computational complexity
    Artavazd Maranjyan, Mher Safaryan, and Peter Richtárik
    arXiv preprint arXiv:2210.16402, 2022