Follow
Sashank J. Reddi
Sashank J. Reddi
Other namesSashank Reddi, Sashank Jakkam Reddi
Research Scientist, Google Research
Verified email at cs.cmu.edu - Homepage
Title
Cited by
Year
Connections are Expressive Enough: Universal Approximability of Sparse Transformers
AS Rawat, C Yun, S Kumar, S Reddi, S Bhojanapalli, YW Chang
2020
2017 Theses by Author
K EARLY, Y MA, GD MONTAÑEZ, SJ REDDI, YX WANG, J XIANG
A field guide to federated optimization
J Wang, Z Charles, Z Xu, G Joshi, HB McMahan, M Al-Shedivat, G Andrew, ...
arXiv preprint arXiv:2107.06917, 2021
3372021
A generic approach for escaping saddle points
S Reddi, M Zaheer, S Sra, B Poczos, F Bach, R Salakhutdinov, A Smola
International conference on artificial intelligence and statistics, 1233-1242, 2018
752018
A maximum likelihood approach for selecting sets of alternatives
AD Procaccia, SJ Reddi, N Shah
arXiv preprint arXiv:1210.4882, 2012
892012
A statistical perspective on distillation
AK Menon, AS Rawat, S Reddi, S Kim, S Kumar
International Conference on Machine Learning, 7632-7642, 2021
782021
Adaclip: Adaptive clipping for private sgd
V Pichapati, AT Suresh, FX Yu, SJ Reddi, S Kumar
arXiv preprint arXiv:1908.07643, 2019
1222019
Adaptive federated optimization
S Reddi, Z Charles, M Zaheer, Z Garrett, K Rush, J Konečný, S Kumar, ...
arXiv preprint arXiv:2003.00295, 2020
13102020
Adaptive methods for nonconvex optimization
M Zaheer, S Reddi, D Sachan, S Kale, S Kumar
Advances in neural information processing systems 31, 2018
4642018
Adaptive optimization with improved convergence
SJ Reddi, S Kumar, SC Kale
US Patent 11,586,904, 2023
12023
Adaptive Optimization with Improved Convergence
SJ Reddi, S Kumar, SC Kale
US Patent App. 18/081,403, 2023
2023
Adaptive sampling distributed stochastic variance reduced gradient for heterogeneous distributed datasets
I Ramazanli, H Nguyen, H Pham, SJ Reddi, B Poczos
arXiv preprint arXiv:2002.08528, 2020
162020
Adaptivity and computation-statistics tradeoffs for kernel and distance based high dimensional two sample testing
A Ramdas, SJ Reddi, B Poczos, A Singh, L Wasserman
arXiv preprint arXiv:1508.00655, 2015
342015
Aide: Fast and communication efficient distributed optimization
SJ Reddi, J Konečný, P Richtárik, B Póczós, A Smola
arXiv preprint arXiv:1608.06879, 2016
1752016
Are transformers universal approximators of sequence-to-sequence functions?
C Yun, S Bhojanapalli, AS Rawat, SJ Reddi, S Kumar
arXiv preprint arXiv:1912.10077, 2019
3142019
Breaking the centralized barrier for cross-device federated learning
SP Karimireddy, M Jaggi, S Kale, M Mohri, S Reddi, SU Stich, AT Suresh
Advances in Neural Information Processing Systems 34, 28663-28676, 2021
732021
Breaking the glass ceiling for embedding-based classifiers for large output spaces
C Guo, A Mousavi, X Wu, DN Holtmann-Rice, S Kale, S Reddi, S Kumar
Advances in Neural Information Processing Systems 32, 2019
672019
Can gradient clipping mitigate label noise?
AK Menon, AS Rawat, SJ Reddi, S Kumar
International Conference on Learning Representations, 2020
1612020
Can Looped Transformers Learn to Implement Multi-step Gradient Descent for In-context Learning?
K Gatmiry, N Saunshi, SJ Reddi, S Jegelka, S Kumar
Forty-first International Conference on Machine Learning, 0
1
Communication Efficient Coresets for Empirical Loss Minimization.
SJ Reddi, B Póczos, AJ Smola
UAI, 752-761, 2015
372015
The system can't perform the operation now. Try again later.
Articles 1–20