Alert button
Picture for Yimin Yang

Yimin Yang

Alert button

Lakehead University

Deep Networks with Fast Retraining

Aug 13, 2020
Wandong Zhang, Yimin Yang, Jonathan Wu

Figure 1 for Deep Networks with Fast Retraining
Figure 2 for Deep Networks with Fast Retraining
Figure 3 for Deep Networks with Fast Retraining
Figure 4 for Deep Networks with Fast Retraining

Recent wor [1] has utilized Moore-Penrose (MP) inverse in deep convolutional neural network (DCNN) training, which achieves better generalization performance over the DCNN with a stochastic gradient descent (SGD) pipeline. However, the MP technique cannot be processed in the GPU environment due to its high demands of computational resources. This paper proposes a fast DCNN learning strategy with MP inverse to achieve better testing performance without introducing a large calculation burden. We achieve this goal through an SGD and MP inverse-based two-stage training procedure. In each training epoch, a random learning strategy that controls the number of convolutional layers trained in backward pass is utilized, and an MP inverse-based batch-by-batch learning strategy is developed that enables the network to be implemented with GPU acceleration and to refine the parameters in dense layer. Through experiments on image classification datasets with various training images ranging in amount from 3,060 (Caltech101) to 1,803,460 (Place365), we empirically demonstrate that the fast retraining is a unified strategy that can be utilized in all DCNNs. Our method obtains up to 1% Top-1 testing accuracy boosts over the state-of-the-art DCNN learning pipeline, yielding a savings in training time of 15% to 25% over the work in [1]. [1] Y. Yang, J. Wu, X. Feng, and A. Thangarajah, "Recomputation of dense layers for the perfor-238mance improvement of dcnn," IEEE Trans. Pattern Anal. Mach. Intell., 2019.

* 10 pages, 3 figures 
Viaarxiv icon

Non-iterative recomputation of dense layers for performance improvement of DCNN

Sep 14, 2018
Yimin Yang, Q. M. Jonathan Wu, Xiexing Feng, Thangarajah Akilan

Figure 1 for Non-iterative recomputation of dense layers for performance improvement of DCNN
Figure 2 for Non-iterative recomputation of dense layers for performance improvement of DCNN
Figure 3 for Non-iterative recomputation of dense layers for performance improvement of DCNN
Figure 4 for Non-iterative recomputation of dense layers for performance improvement of DCNN

An iterative method of learning has become a paradigm for training deep convolutional neural networks (DCNN). However, utilizing a non-iterative learning strategy can accelerate the training process of the DCNN and surprisingly such approach has been rarely explored by the deep learning (DL) community. It motivates this paper to introduce a non-iterative learning strategy that eliminates the backpropagation (BP) at the top dense or fully connected (FC) layers of DCNN, resulting in, lower training time and higher performance. The proposed method exploits the Moore-Penrose Inverse to pull back the current residual error to each FC layer, generating well-generalized features. Then using the recomputed features, i.e., the new generalized features the weights of each FC layer is computed according to the Moore-Penrose Inverse. We evaluate the proposed approach on six widely accepted object recognition benchmark datasets: Scene-15, CIFAR-10, CIFAR-100, SUN-397, Places365, and ImageNet. The experimental results show that the proposed method obtains significant improvements over 30 state-of-the-art methods. Interestingly, it also indicates that any DCNN with the proposed method can provide better performance than the same network with its original training based on BP.

* 11 
Viaarxiv icon

Pulling back error to the hidden-node parameter technology: Single-hidden-layer feedforward network without output weight

May 06, 2014
Yimin Yang, Q. M. Jonathan Wu, Guangbin Huang, Yaonan Wang

Figure 1 for Pulling back error to the hidden-node parameter technology: Single-hidden-layer feedforward network without output weight
Figure 2 for Pulling back error to the hidden-node parameter technology: Single-hidden-layer feedforward network without output weight
Figure 3 for Pulling back error to the hidden-node parameter technology: Single-hidden-layer feedforward network without output weight
Figure 4 for Pulling back error to the hidden-node parameter technology: Single-hidden-layer feedforward network without output weight

According to conventional neural network theories, the feature of single-hidden-layer feedforward neural networks(SLFNs) resorts to parameters of the weighted connections and hidden nodes. SLFNs are universal approximators when at least the parameters of the networks including hidden-node parameter and output weight are exist. Unlike above neural network theories, this paper indicates that in order to let SLFNs work as universal approximators, one may simply calculate the hidden node parameter only and the output weight is not needed at all. In other words, this proposed neural network architecture can be considered as a standard SLFNs with fixing output weight equal to an unit vector. Further more, this paper presents experiments which show that the proposed learning method tends to extremely reduce network output error to a very small number with only 1 hidden node. Simulation results demonstrate that the proposed method can provide several to thousands of times faster than other learning algorithm including BP, SVM/SVR and other ELM methods.

* 7 pages 
Viaarxiv icon