|
This article is cited in 1 scientific paper (total in 1 paper)
Optimal recurrent neural network model in paraphrase detection
A. N. Smerdova, O. Yu. Bakhteeva, V. V. Strijovab a Moscow Institute of Physics and Technology, 9 Institutskiy Per., Dolgoprudny, Moscow Region 141700, Russian Federation
b A. A. Dorodnicyn Computing Center, Federal Research Center “Computer Science and Control” of the Russian
Academy of Sciences, 40 Vavilov Str., Moscow 119333, Russian Federation
Abstract:
This paper addresses the problem of optimal recurrent neural network selection. It asserts the neural network evidence lower bound as the optimal criterion for selection. It investigates variational inference methods to approximate the posterior distribution of the network parameters. As a particular case, the normal distribution of the parameters with different types of the covariance matrix is investigated. The authors propose a method of pruning parameters with the highest probability density in zero to increase the model marginal likelihood. As an illustrative example, a computational experiment of multiclass classification on the SemEval 2015 dataset was carried out.
Keywords:
deep learning, recurrent neural network, neural network pruning, variational approach.
Received: 05.05.2018
Citation:
A. N. Smerdov, O. Yu. Bakhteev, V. V. Strijov, “Optimal recurrent neural network model in paraphrase detection”, Inform. Primen., 12:4 (2018), 63–69
Linking options:
https://www.mathnet.ru/eng/ia564 https://www.mathnet.ru/eng/ia/v12/i4/p63
|
|