Learning discriminative features via weights-biased softmax loss

作者:

Highlights:

• We determine the minimum number of units in FC layer by rigorous theoretical analysis and extensive experiments for various classes tasks, which reduces CNNs’ parameter memory and training time.

• We present a new W-Softmax loss to make CNNs learn more discriminative features, and it can effectively improve the classification performance by avoiding premature convergence.

• The size of decision margins can be optionally adjusted by a positive real-value paremeter α. By increasing the value of α, CNNs can maximize inter-class variance and minimize intra-class variance. Extensive experiments on benchmark datasets show the effectiveness of W-Softmax loss.

摘要

•We determine the minimum number of units in FC layer by rigorous theoretical analysis and extensive experiments for various classes tasks, which reduces CNNs’ parameter memory and training time.•We present a new W-Softmax loss to make CNNs learn more discriminative features, and it can effectively improve the classification performance by avoiding premature convergence.•The size of decision margins can be optionally adjusted by a positive real-value paremeter α. By increasing the value of α, CNNs can maximize inter-class variance and minimize intra-class variance. Extensive experiments on benchmark datasets show the effectiveness of W-Softmax loss.

论文关键词:Classification,Softmax,CNNs,Fully connected layer units,Classifier weights

论文评审过程:Received 22 April 2019, Revised 1 January 2020, Accepted 28 April 2020, Available online 31 May 2020, Version of Record 8 June 2020.

论文官网地址:https://doi.org/10.1016/j.patcog.2020.107405