计算机科学与应用

Vol.5 No.12 (December 2015)

深度学习模型各层参数数目对于性能的影响
The Influence of the Amount of Parameters in Different Layers on the Performance of Deep Learning Models

 

作者:

岳喜斌 , 唐亮 :北京林业大学工学院,北京

胡晓林 :清华大学计算机科学与技术系,清华信息科学与技术国家实验室,北京

 

关键词:

卷积神经网络递归卷积神经网络深度学习Convolutional Neural Network Recurrent Convolutional Neural Network Deep Learning

 

摘要:

近年来深度学习在图像识别、语音识别等领域得到了广泛的应用,取得了优异的效果,但深度学习网络的结构设计没有一般规律可循。本文基于卷积神经网络和递归卷积神经网络模型探究了深度学习网络不同层级间参数分布对网络性能的影响,在CIFAR-10、CIFAR-100和SVHN数据集上进行了大量的实验。结果表明:在保证网络总参数大致相等并稳定在饱和的临界值附近的条件下,增加高层参数数量的能够提升网络性能,而增加低层参数数量的会降低网络性能。通过这一简单的规则,我们设计的递归卷积神经网络模型结构在CIFAR-100和SVHN两个数据集上达到了目前单模型最低的识别错误率。

In recent years, deep learning has been widely used in many pattern recognition tasks including image classification and speech recognition due to its excellent performance. But a general rule for the structure design is lacked. We explored the influence of the amount of parameters in different layers of two deep learning models, convolutional neural network (CNN) and recurrent convolutional neural network (RCNN). Experiments on three benchmark datasets, CIFAR-10, CIFAR-100 and SVHN showed that when the total number of parameters was fixed, increasing the number of parameters in higher layers could boost the performance of the models while increasing the number of parameters in lower layers could be harmful to the performance of the models. Based on this simple rule, we obtained the state-of-the-art classification accuracy on CIFAR-100 and SVHN with single models.

文章引用:

岳喜斌 , 胡晓林 , 唐亮 (2015) 深度学习模型各层参数数目对于性能的影响。 计算机科学与应用, 5, 445-453. doi: 10.12677/CSA.2015.512056

 

参考文献

分享
Top