Follow
Konstantin Mishchenko
Konstantin Mishchenko
CNRS & École Normale Supérieure & Inria
Verified email at inria.fr - Homepage
Title
Cited by
Cited by
Year
Tighter Theory for Local SGD on Identical and Heterogeneous Data
A Khaled, K Mishchenko, P Richtarik
International Conference on Artificial Intelligence and Statistics, 4519-4529, 2020
1762020
Distributed Learning with Compressed Gradient Differences
K Mishchenko, E Gorbunov, M Takáč, P Richtárik
arXiv preprint arXiv:1901.09269, 2019
1112019
Stochastic Distributed Learning with Gradient Quantization and Variance Reduction
S Horváth, D Kovalev, K Mishchenko, S Stich, P Richtárik
arXiv preprint arXiv:1904.05115, 2019
1002019
First Analysis of Local GD on Heterogeneous Data
A Khaled, K Mishchenko, P Richtárik
NeurIPS FL Workshop, arXiv preprint arXiv:1909.04715, 2019
812019
SEGA: Variance Reduction via Gradient Sketching
F Hanzely, K Mishchenko, P Richtárik
Advances in Neural Information Processing Systems, 2082-2093, 2018
512018
Random Reshuffling: Simple Analysis with Vast Improvements
K Mishchenko, A Khaled, P Richtárik
Advances in Neural Information Processing Systems 33, 2020
372020
Revisiting Stochastic Extragradient
K Mishchenko, D Kovalev, E Shulgin, P Richtárik, Y Malitsky
AISTATS 2020, 2019
332019
A Delay-tolerant Proximal-Gradient Algorithm for Distributed Learning
K Mishchenko, F Iutzeler, J Malick, MR Amini
International Conference on Machine Learning, 3584-3592, 2018
332018
Adaptive Gradient Descent without Descent
K Mishchenko, Y Malitsky
37th International Conference on Machine Learning (ICML 2020), 2020
31*2020
Stochastic Newton and cubic Newton methods with simple local linear-quadratic rates
D Kovalev, K Mishchenko, P Richtárik
NeurIPS Workshop Beyond First Order Methods in ML, arXiv preprint arXiv:1912 …, 2019
242019
99% of worker-master communication in distributed optimization is not needed
K Mishchenko, F Hanzely, P Richtárik
Conference on Uncertainty in Artificial Intelligence, 979-988, 2020
23*2020
A distributed flexible delay-tolerant proximal gradient algorithm
K Mishchenko, F Iutzeler, J Malick
SIAM Journal on Optimization 30 (1), 933-959, 2020
232020
MISO is Making a Comeback With Better Proofs and Rates
X Qian, A Sailanbayev, K Mishchenko, P Richtárik
arXiv preprint arXiv:1906.01474, 2019
152019
DAve-QN: A Distributed Averaged Quasi-Newton Method with Local Superlinear Convergence Rate
S Soori, K Mischenko, A Mokhtari, MM Dehnavi, M Gurbuzbalaban
AISTATS 2020, 2019
132019
A Stochastic Decoupling Method for Minimizing the Sum of Smooth and Non-Smooth Functions
K Mishchenko, P Richtárik
arXiv preprint arXiv:1905.11535, 2019
122019
Dualize, split, randomize: Fast nonsmooth optimization algorithms
A Salim, L Condat, K Mishchenko, P Richtárik
Journal of Optimization Theory and Applications, 2020
102020
Proximal and federated random reshuffling
K Mishchenko, A Khaled, P Richtárik
International Conference on Machine Learning, 2022
92022
Sinkhorn Algorithm as a Special Case of Stochastic Mirror Descent
K Mishchenko
NeurIPS OTML Workshop, arXiv preprint arXiv:1909.06918, 2019
92019
A Stochastic Penalty Model for Convex and Nonconvex Optimization with Big Constraints
K Mishchenko, P Richtárik
arXiv preprint arXiv:1810.13387, 2018
62018
IntSGD: Adaptive Floatless Compression of Stochastic Gradients
K Mishchenko, B Wang, D Kovalev, P Richtárik
ICLR 2022 - International Conference on Learning Representations, 2022
5*2022
The system can't perform the operation now. Try again later.
Articles 1–20