Summer Graduate Internship Opportunity
2025 Summer Internship Opportunity
There is a growing interest in designing optimization algorithm using adaptive learning rates (a.k.a. stepsize) to solve (stochastic) optimization problems. This has become particular relevant to train large language model systems and more general neural networks. This project aims to extend these methods to decentralized settings where data are distributed among different machines (referred to as agents). This is challenging because agents do not have access locally to the optimization functions and/or data at the other nodes of the network, and they may not know the exact topology of the network. This prevent the use of existing approaches as developed in the centralized setting. The project will focus on understanding the existing methods in the centralized setting first, and then extend such techniques to the networking setting. Potential outcomes of this project will be a report and possibly a paper for submission to a leading machine learning conference.