|
This article is cited in 1 scientific paper (total in 1 paper)
SPECIAL ISSUE: ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING TECHNOLOGIES
A new computationally simple approach for implementing neural networks with output hard constraints
A. V. Konstantinov, L. V. Utkin Higher School of Artificial Intelligence Technologies, Peter the Great St.Petersburg Polytechnic University, St. Petersburg, Russia
Abstract:
A new computationally simple method of imposing hard convex constraints on the neural network output values is proposed. The key idea behind the method is to map a vector of hidden parameters of the network to a point that is guaranteed to be inside the feasible set defined by a set of constraints. The mapping is implemented by the additional neural network layer with constraints for output. The proposed method is simply extended to the case when constraints are imposed not only on the output vectors, but also on joint constraints depending on inputs. The projection approach to imposing constraints on outputs can simply be implemented in the framework of the proposed method. It is shown how to incorporate different types of constraints into the proposed method, including linear and quadratic constraints, equality constraints, and dynamic constraints, constraints in the form of boundaries. An important feature of the method is its computational simplicity. Complexities of the forward pass of the proposed neural network layer by linear and quadratic constraints are $O(nm)$ and $O(n^2m)$, respectively, where $n$ is the number of variables, $m$ is the number of constraints. Numerical experiments illustrate the method by solving optimization and classification problems. The code implementing the method is publicly available.
Keywords:
neural network, hard constraints, convex set, projection model, optimization problem, classification.
Citation:
A. V. Konstantinov, L. V. Utkin, “A new computationally simple approach for implementing neural networks with output hard constraints”, Dokl. RAN. Math. Inf. Proc. Upr., 514:2 (2023), 80–90; Dokl. Math., 108:suppl. 2 (2023), S233–S241
Linking options:
https://www.mathnet.ru/eng/danma453 https://www.mathnet.ru/eng/danma/v514/i2/p80
|
Statistics & downloads: |
Abstract page: | 61 | References: | 12 |
|