A new method is given for speeding up learning in a deep neural network with many hidden layers, by partially partitioning the network rather than fully interconnecting the layers. Empirical results are shown both for learning a simple Boolean function on a standard back-prop network, and for learning two different, complex, real-world vision tasks on a more sophisticated convolutional network. In all cases, the performance of the proposed system was better than traditional systems. The partially-partitioned network outperformed both the fully-partitioned and fully-unpartitioned networks.
name of conference
2009 International Joint Conference on Neural Networks