Why deep learning works: A manifold disentanglement perspective

PP Brahma, D Wu, Y She - IEEE transactions on neural …, 2015 - ieeexplore.ieee.org
IEEE transactions on neural networks and learning systems, 2015ieeexplore.ieee.org
Deep hierarchical representations of the data have been found out to provide better
informative features for several machine learning applications. In addition, multilayer neural
networks surprisingly tend to achieve better performance when they are subject to an
unsupervised pretraining. The booming of deep learning motivates researchers to identify
the factors that contribute to its success. One possible reason identified is the flattening of
manifold-shaped data in higher layers of neural networks. However, it is not clear how to …
Deep hierarchical representations of the data have been found out to provide better informative features for several machine learning applications. In addition, multilayer neural networks surprisingly tend to achieve better performance when they are subject to an unsupervised pretraining. The booming of deep learning motivates researchers to identify the factors that contribute to its success. One possible reason identified is the flattening of manifold-shaped data in higher layers of neural networks. However, it is not clear how to measure the flattening of such manifold-shaped data and what amount of flattening a deep neural network can achieve. For the first time, this paper provides quantitative evidence to validate the flattening hypothesis. To achieve this, we propose a few quantities for measuring manifold entanglement under certain assumptions and conduct experiments with both synthetic and real-world data. Our experimental results validate the proposition and lead to new insights on deep learning.
ieeexplore.ieee.org
以上显示的是最相近的搜索结果。 查看全部搜索结果