论文标题

我们是否需要卷积网络中的完全连接的输出层?

Do We Need Fully Connected Output Layers in Convolutional Networks?

论文作者

Qian, Zhongchao, Hayes, Tyler L., Kafle, Kushal, Kanan, Christopher

论文摘要

传统上,深度卷积神经网络由一系列卷积和合并层组成,然后是一个或多个完全连接的(FC)层以执行最终分类。尽管此设计已经成功,但对于具有大量类别的数据集,完全连接的层通常是网络参数的很大一部分。对于具有内存约束的应用程序,例如移动设备和嵌入式平台,这不是理想的。最近,已经提出了一个涉及用固定层代替完全连接的输出层的体系结构,以实现提高效率。在本文中,我们进一步研究了这一想法,并证明固定分类器与简单地删除输出层及其参数相比没有其他好处。我们进一步证明,在参数计数方面,具有完全连接的最终输出层的典型方法效率低下。我们能够与传统学到的完全连接的分类输出层在Imagenet-1k,Cifar-100,Stanford Cars-196和Oxford Flowers-102数据集上实现可比的性能,同时根本没有完全连接的输出层。

Traditionally, deep convolutional neural networks consist of a series of convolutional and pooling layers followed by one or more fully connected (FC) layers to perform the final classification. While this design has been successful, for datasets with a large number of categories, the fully connected layers often account for a large percentage of the network's parameters. For applications with memory constraints, such as mobile devices and embedded platforms, this is not ideal. Recently, a family of architectures that involve replacing the learned fully connected output layer with a fixed layer has been proposed as a way to achieve better efficiency. In this paper we examine this idea further and demonstrate that fixed classifiers offer no additional benefit compared to simply removing the output layer along with its parameters. We further demonstrate that the typical approach of having a fully connected final output layer is inefficient in terms of parameter count. We are able to achieve comparable performance to a traditionally learned fully connected classification output layer on the ImageNet-1K, CIFAR-100, Stanford Cars-196, and Oxford Flowers-102 datasets, while not having a fully connected output layer at all.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源