作者
Martin Rapp, Ramin Khalili, Kilian Pfeiffer, Jörg Henkel
发表日期
2022/6/28
期刊
Proceedings of the AAAI Conference on Artificial Intelligence
卷号
36
期号
7
页码范围
8062-8071
简介
We study the problem of distributed training of neural networks (NNs) on devices with heterogeneous, limited, and time-varying availability of computational resources. We present an adaptive, resource-aware, on-device learning mechanism, DISTREAL, which is able to fully and efficiently utilize the available resources on devices in a distributed manner, increasing the convergence speed. This is achieved with a dropout mechanism that dynamically adjusts the computational complexity of training an NN by randomly dropping filters of convolutional layers of the model. Our main contribution is the introduction of a design space exploration (DSE) technique, which finds Pareto-optimal per-layer dropout vectors with respect to resource requirements and convergence speed of the training. Applying this technique, each device is able to dynamically select the dropout vector that fits its available resource without requiring any assistance from the server. We implement our solution in a federated learning (FL) system, where the availability of computational resources varies both between devices and over time, and show through extensive evaluation that we are able to significantly increase the convergence speed over the state of the art without compromising on the final accuracy.
引用总数
学术搜索中的文章
M Rapp, R Khalili, K Pfeiffer, J Henkel - Proceedings of the AAAI Conference on Artificial …, 2022