Subarchitecture ensemble pruning in neural architecture search

Y Bian, Q Song, M Du, J Yao… - IEEE Transactions on …, 2021 - ieeexplore.ieee.org
Y Bian, Q Song, M Du, J Yao, H Chen, X Hu
IEEE Transactions on Neural Networks and Learning Systems, 2021ieeexplore.ieee.org
Neural architecture search (NAS) is gaining more and more attention in recent years
because of its flexibility and remarkable capability to reduce the burden of neural network
design. To achieve better performance, however, the searching process usually costs
massive computations that might not be affordable for researchers and practitioners.
Although recent attempts have employed ensemble learning methods to mitigate the
enormous computational cost, however, they neglect a key property of ensemble methods …
Neural architecture search (NAS) is gaining more and more attention in recent years because of its flexibility and remarkable capability to reduce the burden of neural network design. To achieve better performance, however, the searching process usually costs massive computations that might not be affordable for researchers and practitioners. Although recent attempts have employed ensemble learning methods to mitigate the enormous computational cost, however, they neglect a key property of ensemble methods, namely diversity, which leads to collecting more similar subarchitectures with potential redundancy in the final design. To tackle this problem, we propose a pruning method for NAS ensembles called “ subarchitecture ensemble pruning in neural architecture search (SAEP).” It targets to leverage diversity and to achieve subensemble architectures at a smaller size with comparable performance to ensemble architectures that are not pruned. Three possible solutions are proposed to decide which subarchitectures to prune during the searching process. Experimental results exhibit the effectiveness of the proposed method by largely reducing the number of subarchitectures without degrading the performance.
ieeexplore.ieee.org
以上显示的是最相近的搜索结果。 查看全部搜索结果